-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Inference] update fakequant support #9047
[Inference] update fakequant support #9047
Conversation
2. add llama3.1 and qwen2 ptq config 3. update quantization.md
…nto add_new_fakequant_type
Thanks for your contribution! |
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## develop #9047 +/- ##
===========================================
- Coverage 53.76% 53.65% -0.12%
===========================================
Files 652 652
Lines 104507 104867 +360
===========================================
+ Hits 56190 56264 +74
- Misses 48317 48603 +286 ☔ View full report in Codecov by Sentry. |
@@ -1,4 +1,4 @@ | |||
# 大模型量化教程 | |||
p# 大模型量化教程 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这里多了一个‘p’
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* 1. add a8w8(fp8) a8w8c8(int8) quant_type support 2. add llama3.1 and qwen2 ptq config 3. update quantization.md * fix load_quant_model bug * fix load quant bug * update ll/README.md * remove useless code * update quant observer config * resolve wrong modify * fix prepare_qconfig * remove unuse files
PR types
Bug fixes
PR changes
Others
Description