-
Notifications
You must be signed in to change notification settings - Fork 45
Move scaling logic to input generation #338
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
This pull request was exported from Phabricator. Differential Revision: D80571223 |
This pull request was exported from Phabricator. Differential Revision: D80571223 |
3484c61
to
a5bc2fa
Compare
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
a5bc2fa
to
edfdb2a
Compare
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
edfdb2a
to
a609415
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
a609415
to
b024198
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Pull Request resolved: meta-pytorch#338 Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Reviewed By: NikhilAPatel Differential Revision: D80571223
b024198
to
4efc17d
Compare
c817c15
to
e3adb1b
Compare
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
e3adb1b
to
d9e0627
Compare
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
d9e0627
to
1acc603
Compare
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
1acc603
to
ad791c4
Compare
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Differential Revision: D80571223 Pulled By: jananisriram
ad791c4
to
da77dd8
Compare
This pull request was exported from Phabricator. Differential Revision: D80571223 |
Summary: Move scaling logic for FP8 benchmarks to `get_input_iter()`. This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (`bfloat16`, `float16`), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (`float8_e4m3fn`). This diff also circumvents performing unsupported operations, like `torch.max` and `torch.abs`, on low-precision data types. Pull Request resolved: meta-pytorch#338 Test Plan: Imported from GitHub, without a `Test Plan:` line. Rollback Plan: Reviewed By: xuzhao9 Differential Revision: D80571223 Pulled By: jananisriram
This pull request was exported from Phabricator. Differential Revision: D80571223 |
da77dd8
to
7d67063
Compare
Summary:
Move scaling logic for FP8 benchmarks to
get_input_iter()
.This diff aligns our fp8_gemm benchmarking suite with real-world practices: input tensors are of high precision types (
bfloat16
,float16
), scales are computed on the high-precision input tensors, and input tensors are then casted to a lower precision (float8_e4m3fn
).This diff also circumvents performing unsupported operations, like
torch.max
andtorch.abs
, on low-precision data types.Reviewed By: NikhilAPatel
Differential Revision: D80571223