-
Notifications
You must be signed in to change notification settings - Fork 257
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support fp8 cast WOQ #1746
support fp8 cast WOQ #1746
Conversation
Signed-off-by: xin3he <xin3.he@intel.com>
⚡ Required checks status: All passing 🟢Groups summary🟢 Code Scan Tests workflow
These checks are required after the changes to 🟢 Model Tests 3x workflow
These checks are required after the changes to 🟢 Unit Tests 3x-PyTorch workflow
These checks are required after the changes to Thank you for your contribution! 💜
|
Signed-off-by: xin3he <xin3.he@intel.com>
Type of Change
feature
Description
https://jira.devtools.intel.com/browse/ILITV-3505
support fp8 cast WOQ using official PyTorch >= 2.2
Expected Behavior & Potential Risk
Torch supports 4 fp8 dtypes as shown below. In INC, we use short strings for mapping.
An introduction to the differences between FP8 data types:
Float8E4M3FN, Float8E4M3FNUZ, Float8E5M2, Float8E5M2FNUZ
stablehlo/rfcs/20230321-fp8_fnuz.md at main · openxla/stablehlo (github.com)
How has this PR been tested?
local test