-
Notifications
You must be signed in to change notification settings - Fork 74
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flash-Attention #85
Comments
I run into NaNs if I enable flash attention. unicore.nan_detector | NaN detected in output of model.evoformer.blocks.47.tri_att_end.mha.linear_v, shape: torch.Size([1, 292, 292, 128]), backward WARNING | unicore.nan_detector | NaN detected in output of model.evoformer.blocks.21.msa_att_row.mha.linear_v, shape: torch.Size([1, 256, 184, 256]), backward I get also lots of new warnings: Is it working for you @Xreki? A100 with bfloat16 enabled |
Can you provide some details for the installation of flash attention? It seems that the backward did not work correctly. |
@lhatsk It seems OK for me. I use the docker image |
I installed flash attention from source according to the README. torch 1.12.1 + CUDA 11.2 |
Can you write a single test for the |
Just running |
you now can use this branch: https://github.com/dptech-corp/Uni-Fold/tree/flash-attn , to try the flash-attention. |
No description provided.
The text was updated successfully, but these errors were encountered: