Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[AMD backend] fix test_dot_without_load #3357

Merged
merged 1 commit into from
Mar 13, 2024

Conversation

scxiao
Copy link
Contributor

@scxiao scxiao commented Mar 12, 2024

This PR is actually fix the regression in the reverted PR: #3338, which caused a regression for the test test_masked_load_shared_memory. The reason is for type used in packing dot_op for bfloat16. We should use the type i16 for bf16 when packing dot_op for mfma.

This time I ran all the tests in test_core.py locally, and all work fine.

@scxiao scxiao requested a review from ptillet as a code owner March 12, 2024 19:45
@scxiao scxiao changed the title fix test_dot_without_load [AMD backend] fix test_dot_without_load Mar 12, 2024
@scxiao scxiao force-pushed the fix_dot_without_load1 branch from cf6614b to dcb0690 Compare March 13, 2024 13:57
@scxiao
Copy link
Contributor Author

scxiao commented Mar 13, 2024

It seems like there is a problem the get the H100 device in the CI build

@zhanglx13 zhanglx13 merged commit ce9618a into triton-lang:main Mar 13, 2024
4 checks passed
joviliast added a commit to joviliast/triton that referenced this pull request Mar 13, 2024
TypeConverter:
-Support WMMA dot operand type;
DecomposeUnsupportedAMDConversions:
-Replace `wmma -> dot_op` with `wmma -> blocked -> dot_op`;
-Add lit test;
RmoveLayoutConversion:
-Do not propagate WMMA layout if it is not a chain dot case;
WMMA operation convertion:
-Fix related to dot operand element size,
See also triton-lang#3357

Signed-off-by: joviliast <iveselov.nn@gmail.com>
htyu pushed a commit to htyu/triton that referenced this pull request Mar 20, 2024
This PR is actually fix the regression in the reverted PR:
triton-lang#3338, which caused a regression
for the test `test_masked_load_shared_memory`. The reason is for type
used in packing dot_op for bfloat16. We should use the type `i16` for
`bf16` when packing dot_op for mfma.

This time I ran all the tests in `test_core.py` locally, and all work
fine.
karupayun pushed a commit to openxla/triton that referenced this pull request Apr 3, 2024
This PR is actually fix the regression in the reverted PR:
triton-lang#3338, which caused a regression
for the test `test_masked_load_shared_memory`. The reason is for type
used in packing dot_op for bfloat16. We should use the type `i16` for
`bf16` when packing dot_op for mfma.

This time I ran all the tests in `test_core.py` locally, and all work
fine.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants