Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

xpu devices support llama-7b basic mode inference (turn on BlockAtten… #8588

Merged
merged 3 commits into from
Jun 13, 2024

Conversation

zhink
Copy link
Contributor

@zhink zhink commented Jun 12, 2024

…tion)

PR types

New features

PR changes

Others

Description

xpu devices support llama-7b basic mode inference (turn on BlockAttention)

Copy link

paddle-bot bot commented Jun 12, 2024

Thanks for your contribution!

Copy link

codecov bot commented Jun 12, 2024

Codecov Report

Attention: Patch coverage is 0% with 31 lines in your changes missing coverage. Please review.

Project coverage is 54.41%. Comparing base (525eef7) to head (8717c5e).
Report is 239 commits behind head on develop.

Files with missing lines Patch % Lines
...erimental/transformers/fused_transformer_layers.py 0.00% 11 Missing ⚠️
...enlp/experimental/transformers/generation_utils.py 0.00% 9 Missing ⚠️
...dlenlp/experimental/transformers/llama/modeling.py 0.00% 3 Missing ⚠️
...dlenlp/experimental/transformers/bloom/modeling.py 0.00% 2 Missing ⚠️
...ddlenlp/experimental/transformers/qwen/modeling.py 0.00% 2 Missing ⚠️
...enlp/experimental/transformers/chatglm/modeling.py 0.00% 1 Missing ⚠️
...p/experimental/transformers/chatglm_v2/modeling.py 0.00% 1 Missing ⚠️
...addlenlp/experimental/transformers/gpt/modeling.py 0.00% 1 Missing ⚠️
...addlenlp/experimental/transformers/opt/modeling.py 0.00% 1 Missing ⚠️
Additional details and impacted files
@@             Coverage Diff             @@
##           develop    #8588      +/-   ##
===========================================
- Coverage    54.42%   54.41%   -0.02%     
===========================================
  Files          632      632              
  Lines        99451    99470      +19     
===========================================
  Hits         54129    54129              
- Misses       45322    45341      +19     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@@ -25,27 +25,27 @@
from paddle.nn import Layer
from paddle.nn.initializer import Constant
from paddle.nn.quant import weight_only_linear
from paddlenlp_ops import rebuild_padding_v2
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

额,这力直接 import了,后面才判断 is_paddlenlp_ops_available ? 已经晚了吧

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

@zhink zhink requested a review from ZHUI June 13, 2024 03:04
@ZHUI ZHUI merged commit 3d777c1 into PaddlePaddle:develop Jun 13, 2024
7 of 12 checks passed
@zhink zhink deleted the xpullama2 branch June 13, 2024 03:11
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants