-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[LLM Inference] support llama3.1 #8929
[LLM Inference] support llama3.1 #8929
Conversation
Thanks for your contribution! |
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## develop #8929 +/- ##
===========================================
+ Coverage 55.05% 55.11% +0.06%
===========================================
Files 635 635
Lines 99410 99548 +138
===========================================
+ Hits 54729 54870 +141
+ Misses 44681 44678 -3 ☔ View full report in Codecov by Sentry. |
self.vocab_size, | ||
self.hidden_size, | ||
) | ||
self.embed_tokens = nn.Embedding(self.vocab_size, self.hidden_size) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
多卡推理遇到两个问题,1. block_attn为false时动转静会报错,2. block_attn为true时运行会报错
建议后续针对 Embedding TP切分的推理报错进行修复,而不是直接改成不切分。 @yuanlehome |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
收到,已记todo |
PR types
New features
PR changes
Others
Description
support llama3.1 --block_attn 支持多卡推理