-
Notifications
You must be signed in to change notification settings - Fork 50
Move scaling logic to input generation #338
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
3484c61 to
a5bc2fa
Compare
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
a5bc2fa to
edfdb2a
Compare
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
edfdb2a to
a609415
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
a609415 to
b024198
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
b024198 to
4efc17d
Compare
c817c15 to
e3adb1b
Compare
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
e3adb1b to
d9e0627
Compare
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
d9e0627 to
1acc603
Compare
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
1acc603 to
ad791c4
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
ad791c4 to
da77dd8
Compare
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Pull Request resolved: meta-pytorch#338 Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Reviewed By: xuzhao9 Differential Revision: D80571223 Pulled By: jananisriram
|
This pull request was exported from Phabricator. Differential Revision: D80571223 |
da77dd8 to
7d67063
Compare
Summary:
Move scaling logic for FP8 benchmarks to
get_input_iter().This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (
bfloat16,float16), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (float8_e4m3fn).This diff also circumvents performing unsupported operations, like
torch.maxandtorch.abs, on low-precision data types.Reviewed By: NikhilAPatel
Differential Revision: D80571223