Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

all tests got segmentation fault after launch #17

Closed
arom4github opened this issue Jan 25, 2017 · 4 comments
Closed

all tests got segmentation fault after launch #17

arom4github opened this issue Jan 25, 2017 · 4 comments

Comments

@arom4github
Copy link

arom4github commented Jan 25, 2017

Building process looks correct except some warnings. But after that all tests fail.

$ ./bench/bench-blas
Devices available:
------------------
[x] - gfx803 on AMD Accelerated Parallel Processing
------------------
BENCH   M       N       K       AT      BT      ISAAC
Segmentation fault (core dumped)

$ gdb $_
GNU gdb (Ubuntu 7.7.1-0ubuntu5~14.04.2) 7.7.1
......
Reading symbols from ./bench/bench-blas...(no debugging symbols found)...done.
(gdb) run
Starting program: /tmp/isaac/build/bench/bench-blas
Devices available:
------------------
Warning: couldn't activate thread debugging using libthread_db: Cannot find new threads: generic error
warning: File "/lib/x86_64-linux-gnu/libthread_db-1.0.so" auto-loading has been declined by your `auto-load safe-path' set to "$debugdir:$datadir/auto-load".
To enable execution of this file add
        add-auto-load-safe-path /lib/x86_64-linux-gnu/libthread_db-1.0.so
line to your configuration file "/var/data0/sandbox/user1/.gdbinit".
To completely disable this security protection add
        set auto-load safe-path /
line to your configuration file "/var/data0/sandbox/user1/.gdbinit".
For more information about this security protection see the
"Auto-loading safe path" section in the GDB manual.  E.g., run from the shell:
        info "(gdb)Auto-loading safe path"
warning: Unable to find libthread_db matching inferior's thread library, thread debugging will not be available.
[x] - gfx803 on AMD Accelerated Parallel Processing
------------------
BENCH   M       N       K       AT      BT      ISAAC

Program received signal SIGSEGV, Segmentation fault.
0x00007ffff7750a1d in isaac::runtime::profiles::value_type::init(isaac::runtime::execution_handler const&) () from /tmp/isaac/build/lib/libisaac.so
(gdb) where
#0  0x00007ffff7750a1d in isaac::runtime::profiles::value_type::init(isaac::runtime::execution_handler const&) () from /tmp/isaac/build/lib/libisaac.so
#1  0x00007ffff7751440 in isaac::runtime::profiles::value_type::execute(isaac::runtime::execution_handler const&) () from /tmp/isaac/build/lib/libisaac.so
#2  0x00007ffff775b71e in isaac::runtime::execute(isaac::runtime::execution_handler const&, std::map<std::pair<isaac::expression_type, isaac::numeric_type>, std::shared_ptr<isaac::runtime::profiles::value_type>, std::less<std::pair<isaac::expression_type, isaac::numeric_type> >, std::allocator<std::pair<std::pair<isaac::expression_type, isaac::numeric_type> const, std::shared_ptr<isaac::runtime::profiles::value_type> > > >&) () from /tmp/isaac/build/lib/libisaac.so
#3  0x00007ffff7741200 in isaac::array_base::operator=(isaac::runtime::execution_handler const&) () from /tmp/isaac/build/lib/libisaac.so
#4  0x00007ffff77419a7 in isaac::array_base::operator=(isaac::expression_tree const&) () from /tmp/isaac/build/lib/libisaac.so
#5  0x0000000000409c7e in double bench<void bench<float>(isaac::numeric_type, std::string)::{lambda()#8}, void bench<float>(isaac::numeric_type, std::string)::{lambda()#1}>(void bench<float>(isaac::numeric_type, std::string)::{lambda()#8} const&, void bench<float>(isaac::numeric_type, std::string)::{lambda()#1} const&) ()
#6  0x000000000040d4f8 in void bench<float>(isaac::numeric_type, std::string) ()
#7  0x0000000000405caf in main ()
(gdb)
@ptillet
Copy link
Collaborator

ptillet commented Jan 25, 2017

Hi,

Interesting, the OpenCL driver reports gfx803 instead of Fiji. What is the GPU and driver version?

@ptillet
Copy link
Collaborator

ptillet commented Jan 25, 2017

Also, can you retry now with the latest version of origin/master?

@arom4github
Copy link
Author

arom4github commented Jan 26, 2017

Hi,

Updated version works much better now

$ ./bench-blas
Devices available:
------------------
[x] - gfx803 on AMD Accelerated Parallel Processing
------------------
BENCH   M       N       K       AT      BT      ISAAC
Deep    1760    16      1760    N       N       0.30
Deep    1760    32      1760    N       N       0.59
Deep    1760    64      1760    N       N       0.70
Deep    1760    128     1760    N       N       0.74
Deep    1760    7000    1760    N       N       1.27
Deep    2048    16      2048    N       N       0.34
Deep    2048    32      2048    N       N       0.66
Deep    2048    64      2048    N       N       0.80
Deep    2048    128     2048    N       N       0.81
Deep    2048    7000    2048    N       N       1.02
Deep    2560    16      2560    N       N       0.54
Deep    2560    32      2560    N       N       0.56
Deep    2560    64      2560    N       N       0.62
Deep    2560    128     2560    N       N       0.81
Deep    2560    7000    2560    N       N       1.09
Deep    1760    16      1760    T       N       0.28
Deep    1760    32      1760    T       N       0.55
Deep    1760    64      1760    T       N       1.01
Deep    1760    128     1760    T       N       1.13
Deep    1760    7000    1760    T       N       1.58
Deep    2048    16      2048    T       N       0.07
Deep    2048    32      2048    T       N       0.11
Deep    2048    64      2048    T       N       0.38
Deep    2048    128     2048    T       N       0.28
Deep    2048    7000    2048    T       N       0.50
Deep    2560    16      2560    T       N       0.25
Deep    2560    32      2560    T       N       0.48
Deep    2560    64      2560    T       N       0.53
Deep    2560    128     2560    T       N       0.98
Deep    2560    7000    2560    T       N       0.83
Deep    1760    7133    1760    N       T       0.98
Memory access fault by GPU node-1 on address 0x916cf2000. Reason: Page not present or supervisor privilege.
Aborted (core dumped)

The error I see is not benchmark error I guess. clBLAS tests throw the same error. Probably it's on driver side.

I have RX 480 installed

01:00.0 VGA compatible controller: Advanced Micro Devices, Inc. [AMD/ATI] Device 67df (rev c7)

I'm also confused by this "gfx803". I use driver from ROCm.

$ uname -r
4.6.0-kfd-compute-rocm-rel-1.4-16
$ /opt/rocm/opencl/bin/x86_64/clinfo
Number of platforms:                             1
  Platform Profile:                              FULL_PROFILE
  Platform Version:                              OpenCL 2.0 AMD-APP (2300.5)
  Platform Name:                                 AMD Accelerated Parallel Processing
  Platform Vendor:                               Advanced Micro Devices, Inc.
  Platform Extensions:                           cl_khr_icd cl_amd_event_callback cl_amd_offline_devices


  Platform Name:                                 AMD Accelerated Parallel Processing
Number of devices:                               1
  Device Type:                                   CL_DEVICE_TYPE_GPU
  Vendor ID:                                     1002h
  Board name:                                    Device 67df
  Device Topology:                               PCI[ B#1, D#0, F#0 ]
  Max compute units:                             36
...
  Platform ID:                                   0x7f32f2e9f198
  Name:                                          gfx803
  Vendor:                                        Advanced Micro Devices, Inc.
  Device OpenCL C version:                       OpenCL C 2.0
  Driver version:                                1.1 (HSA,LC)
  Profile:                                       FULL_PROFILE
  Version:                                       OpenCL 1.2
...

@ptillet
Copy link
Collaborator

ptillet commented Jan 26, 2017

Glad that it solved your problem.

So gfx803 is actually Polaris? The performance is very poor (1.6TFLOPS vs 5.0TFLOPS peak). Is it an R480m or the full desktop version? I get 70-75% of the peak on R9 Fury so that's a little odd.

@ptillet ptillet closed this as completed Jan 30, 2017
goostavz pushed a commit to goostavz/triton that referenced this issue Aug 4, 2023
jlebar pushed a commit that referenced this issue Jun 21, 2024
When running
[convert_blocked1d_to_slice0](https://github.com/triton-lang/triton/blob/0ba5f0c3cd029d5c3d1f01b9bf29dac32c27345e/test/Conversion/tritongpu_to_llvm.mlir#L924)
Triton ends up computing a rank of a matrix with 0 columns during linear
layout lowering, which trips up f2reduce, and causes undefined behavior,
detectable through
[UBSAN](https://clang.llvm.org/docs/UndefinedBehaviorSanitizer.html).

Fix this by returning the rank (0) early in these cases, without calling
f2reduce.

<details><summary>Stack trace</summary>
<p>

```
third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30: runtime error: shift exponent 18446744073709551615 is too large for 64-bit type 'unsigned long long'
    #0 0x556ee2fea3be in inplace_rref_small third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30
    #1 0x556ee2fea3be in f2reduce::inplace_rref_strided(unsigned long*, unsigned long, unsigned long, unsigned long) third_party/triton/third_party/f2reduce/f2reduce.cpp:470:9
    #2 0x556ee2ea70da in getMatrixRank third_party/triton/lib/Tools/LinearLayout.cpp:125:3
    #3 0x556ee2ea70da in mlir::triton::LinearLayout::checkInvariants(bool) third_party/triton/lib/Tools/LinearLayout.cpp:299:7
    #4 0x556ee2ea656d in mlir::triton::LinearLayout::tryCreate(llvm::MapVector<mlir::StringAttr, std::__u::vector<std::__u::vector<int, std::__u::allocator<int>>, std::__u::allocator<std::__u::vector<int, std::__u::allocator<int>>>>, llvm::DenseMap<mlir::StringAttr, unsigned int, llvm::DenseMapInfo<mlir::StringAttr, void>, llvm::detail::DenseMapPair<mlir::StringAttr, unsigned int>>, llvm::SmallVector<std::__u::pair<mlir::StringAttr, std::__u::vector<std::__u::vector<int, std::__u::allocator<int>>, std::__u::allocator<std::__u::vector<int, std::__u::allocator<int>>>>>, 0u>>, llvm::ArrayRef<std::__u::pair<mlir::StringAttr, int>>, bool) third_party/triton/lib/Tools/LinearLayout.cpp:190:41
    #5 0x556ee2eb2150 in mlir::triton::LinearLayout::divideRight(mlir::triton::LinearLayout const&) third_party/triton/lib/Tools/LinearLayout.cpp:654:51
    #6 0x556ee2ee1c39 in mlir::cvtNeedsSharedMemory(mlir::RankedTensorType, mlir::RankedTensorType) third_party/triton/lib/Analysis/Utility.cpp:652:14
    #7 0x556ee2cf38fd in mlir::triton::getRepShapeForCvtLayout(mlir::triton::gpu::ConvertLayoutOp) third_party/triton/lib/Analysis/Allocation.cpp:66:8
    #8 0x556ee2cf3efa in mlir::triton::getScratchConfigForCvtLayout(mlir::triton::gpu::ConvertLayoutOp, unsigned int&, unsigned int&) third_party/triton/lib/Analysis/Allocation.cpp:95:19
    #9 0x556ee2cf6057 in mlir::triton::AllocationAnalysis::getScratchValueSize(mlir::Operation*) third_party/triton/lib/Analysis/Allocation.cpp:272:24
    #10 0x556ee2cf5499 in operator() third_party/triton/lib/Analysis/Allocation.cpp:343:7
    #11 0x556ee2cf5499 in void llvm::function_ref<void (mlir::Operation*)>::callback_fn<mlir::triton::AllocationAnalysis::getValuesAndSizes()::'lambda'(mlir::Operation*)>(long, mlir::Operation*) third_party/llvm/llvm-project/llvm/include/llvm/ADT/STLFunctionalExtras.h:45:12
    #12 0x556edeeee7a9 in operator() third_party/llvm/llvm-project/llvm/include/llvm/ADT/STLFunctionalExtras.h:68:12
    #13 0x556edeeee7a9 in void mlir::detail::walk<mlir::ForwardIterator>(mlir::Operation*, llvm::function_ref<void (mlir::Operation*)>, mlir::WalkOrder) third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:174:5
    #14 0x556edeeee87c in void mlir::detail::walk<mlir::ForwardIterator>(mlir::Operation*, llvm::function_ref<void (mlir::Operation*)>, mlir::WalkOrder) third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:182:9
    #15 0x556ee2cf49e7 in walk<(mlir::WalkOrder)0, mlir::ForwardIterator, (lambda at third_party/triton/lib/Analysis/Allocation.cpp:341:42), mlir::Operation *, void> third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:313:10
    #16 0x556ee2cf49e7 in walk<(mlir::WalkOrder)0, mlir::ForwardIterator, (lambda at third_party/triton/lib/Analysis/Allocation.cpp:341:42), void> third_party/llvm/llvm-project/mlir/include/mlir/IR/Operation.h:794:12
    #17 0x556ee2cf49e7 in mlir::triton::AllocationAnalysis::getValuesAndSizes() third_party/triton/lib/Analysis/Allocation.cpp:341:16
    #18 0x556ee2cf4852 in run third_party/triton/lib/Analysis/Allocation.cpp:182:5
    #19 0x556ee2cf4852 in AllocationAnalysis third_party/triton/lib/Analysis/Allocation.cpp:169:5
    #20 0x556ee2cf4852 in mlir::Allocation::run(llvm::DenseMap<mlir::FunctionOpInterface, mlir::Allocation, llvm::DenseMapInfo<mlir::FunctionOpInterface, void>, llvm::detail::DenseMapPair<mlir::FunctionOpInterface, mlir::Allocation>>&) third_party/triton/lib/Analysis/Allocation.cpp:627:3
    #21 0x556ee1677402 in operator() third_party/triton/include/triton/Analysis/Allocation.h:227:26
    #22 0x556ee1677402 in void mlir::CallGraph<mlir::Allocation>::doWalk<(mlir::WalkOrder)0, (mlir::WalkOrder)1, mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::CallOpInterface, mlir::FunctionOpInterface), mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::FunctionOpInterface)>(mlir::FunctionOpInterface, llvm::DenseSet<mlir::FunctionOpInterface, llvm::DenseMapInfo<mlir::FunctionOpInterface, void>>&, mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::CallOpInterface, mlir::FunctionOpInterface), mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::FunctionOpInterface)) third_party/triton/include/triton/Analysis/Utility.h:350:7
    #23 0x556ee16756b3 in walk<(mlir::WalkOrder)0, (mlir::WalkOrder)1, (lambda at third_party/triton/include/triton/Analysis/Allocation.h:222:9), (lambda at third_party/triton/include/triton/Analysis/Allocation.h:224:9)> third_party/triton/include/triton/Analysis/Utility.h:242:7
    #24 0x556ee16756b3 in mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp) third_party/triton/include/triton/Analysis/Allocation.h:220:5
    #25 0x556ee2c2bf18 in (anonymous namespace)::AllocateSharedMemory::runOnOperation() third_party/triton/lib/Conversion/TritonGPUToLLVM/AllocateSharedMemory.cpp:26:22
...
UndefinedBehaviorSanitizer: invalid-shift-exponent third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30 
```
</p>
</details>
oraluben pushed a commit to oraluben/triton that referenced this issue Sep 11, 2024
…ng#17)

* Fixed yaml syntax

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Removed cpu label from run-on

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing zlib-dev

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing apt-get update

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Remove pip cache because on self-hosted runner it slows things down

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Corrected path to tests

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added installation of torch==2.1.2

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

---------

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>
gglin001 pushed a commit to gglin001/triton that referenced this issue Nov 13, 2024
…ng#17)

* Fixed yaml syntax

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Removed cpu label from run-on

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing zlib-dev

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing apt-get update

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Remove pip cache because on self-hosted runner it slows things down

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Corrected path to tests

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added installation of torch==2.1.2

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

---------

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>
bertmaher pushed a commit to bertmaher/triton that referenced this issue Dec 10, 2024
When running
[convert_blocked1d_to_slice0](https://github.com/triton-lang/triton/blob/0ba5f0c3cd029d5c3d1f01b9bf29dac32c27345e/test/Conversion/tritongpu_to_llvm.mlir#L924)
Triton ends up computing a rank of a matrix with 0 columns during linear
layout lowering, which trips up f2reduce, and causes undefined behavior,
detectable through
[UBSAN](https://clang.llvm.org/docs/UndefinedBehaviorSanitizer.html).

Fix this by returning the rank (0) early in these cases, without calling
f2reduce.

<details><summary>Stack trace</summary>
<p>

```
third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30: runtime error: shift exponent 18446744073709551615 is too large for 64-bit type 'unsigned long long'
    #0 0x556ee2fea3be in inplace_rref_small third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30
    triton-lang#1 0x556ee2fea3be in f2reduce::inplace_rref_strided(unsigned long*, unsigned long, unsigned long, unsigned long) third_party/triton/third_party/f2reduce/f2reduce.cpp:470:9
    triton-lang#2 0x556ee2ea70da in getMatrixRank third_party/triton/lib/Tools/LinearLayout.cpp:125:3
    triton-lang#3 0x556ee2ea70da in mlir::triton::LinearLayout::checkInvariants(bool) third_party/triton/lib/Tools/LinearLayout.cpp:299:7
    triton-lang#4 0x556ee2ea656d in mlir::triton::LinearLayout::tryCreate(llvm::MapVector<mlir::StringAttr, std::__u::vector<std::__u::vector<int, std::__u::allocator<int>>, std::__u::allocator<std::__u::vector<int, std::__u::allocator<int>>>>, llvm::DenseMap<mlir::StringAttr, unsigned int, llvm::DenseMapInfo<mlir::StringAttr, void>, llvm::detail::DenseMapPair<mlir::StringAttr, unsigned int>>, llvm::SmallVector<std::__u::pair<mlir::StringAttr, std::__u::vector<std::__u::vector<int, std::__u::allocator<int>>, std::__u::allocator<std::__u::vector<int, std::__u::allocator<int>>>>>, 0u>>, llvm::ArrayRef<std::__u::pair<mlir::StringAttr, int>>, bool) third_party/triton/lib/Tools/LinearLayout.cpp:190:41
    triton-lang#5 0x556ee2eb2150 in mlir::triton::LinearLayout::divideRight(mlir::triton::LinearLayout const&) third_party/triton/lib/Tools/LinearLayout.cpp:654:51
    triton-lang#6 0x556ee2ee1c39 in mlir::cvtNeedsSharedMemory(mlir::RankedTensorType, mlir::RankedTensorType) third_party/triton/lib/Analysis/Utility.cpp:652:14
    triton-lang#7 0x556ee2cf38fd in mlir::triton::getRepShapeForCvtLayout(mlir::triton::gpu::ConvertLayoutOp) third_party/triton/lib/Analysis/Allocation.cpp:66:8
    triton-lang#8 0x556ee2cf3efa in mlir::triton::getScratchConfigForCvtLayout(mlir::triton::gpu::ConvertLayoutOp, unsigned int&, unsigned int&) third_party/triton/lib/Analysis/Allocation.cpp:95:19
    triton-lang#9 0x556ee2cf6057 in mlir::triton::AllocationAnalysis::getScratchValueSize(mlir::Operation*) third_party/triton/lib/Analysis/Allocation.cpp:272:24
    triton-lang#10 0x556ee2cf5499 in operator() third_party/triton/lib/Analysis/Allocation.cpp:343:7
    triton-lang#11 0x556ee2cf5499 in void llvm::function_ref<void (mlir::Operation*)>::callback_fn<mlir::triton::AllocationAnalysis::getValuesAndSizes()::'lambda'(mlir::Operation*)>(long, mlir::Operation*) third_party/llvm/llvm-project/llvm/include/llvm/ADT/STLFunctionalExtras.h:45:12
    triton-lang#12 0x556edeeee7a9 in operator() third_party/llvm/llvm-project/llvm/include/llvm/ADT/STLFunctionalExtras.h:68:12
    triton-lang#13 0x556edeeee7a9 in void mlir::detail::walk<mlir::ForwardIterator>(mlir::Operation*, llvm::function_ref<void (mlir::Operation*)>, mlir::WalkOrder) third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:174:5
    triton-lang#14 0x556edeeee87c in void mlir::detail::walk<mlir::ForwardIterator>(mlir::Operation*, llvm::function_ref<void (mlir::Operation*)>, mlir::WalkOrder) third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:182:9
    triton-lang#15 0x556ee2cf49e7 in walk<(mlir::WalkOrder)0, mlir::ForwardIterator, (lambda at third_party/triton/lib/Analysis/Allocation.cpp:341:42), mlir::Operation *, void> third_party/llvm/llvm-project/mlir/include/mlir/IR/Visitors.h:313:10
    triton-lang#16 0x556ee2cf49e7 in walk<(mlir::WalkOrder)0, mlir::ForwardIterator, (lambda at third_party/triton/lib/Analysis/Allocation.cpp:341:42), void> third_party/llvm/llvm-project/mlir/include/mlir/IR/Operation.h:794:12
    triton-lang#17 0x556ee2cf49e7 in mlir::triton::AllocationAnalysis::getValuesAndSizes() third_party/triton/lib/Analysis/Allocation.cpp:341:16
    triton-lang#18 0x556ee2cf4852 in run third_party/triton/lib/Analysis/Allocation.cpp:182:5
    triton-lang#19 0x556ee2cf4852 in AllocationAnalysis third_party/triton/lib/Analysis/Allocation.cpp:169:5
    triton-lang#20 0x556ee2cf4852 in mlir::Allocation::run(llvm::DenseMap<mlir::FunctionOpInterface, mlir::Allocation, llvm::DenseMapInfo<mlir::FunctionOpInterface, void>, llvm::detail::DenseMapPair<mlir::FunctionOpInterface, mlir::Allocation>>&) third_party/triton/lib/Analysis/Allocation.cpp:627:3
    triton-lang#21 0x556ee1677402 in operator() third_party/triton/include/triton/Analysis/Allocation.h:227:26
    triton-lang#22 0x556ee1677402 in void mlir::CallGraph<mlir::Allocation>::doWalk<(mlir::WalkOrder)0, (mlir::WalkOrder)1, mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::CallOpInterface, mlir::FunctionOpInterface), mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::FunctionOpInterface)>(mlir::FunctionOpInterface, llvm::DenseSet<mlir::FunctionOpInterface, llvm::DenseMapInfo<mlir::FunctionOpInterface, void>>&, mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::CallOpInterface, mlir::FunctionOpInterface), mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp)::'lambda'(mlir::FunctionOpInterface)) third_party/triton/include/triton/Analysis/Utility.h:350:7
    triton-lang#23 0x556ee16756b3 in walk<(mlir::WalkOrder)0, (mlir::WalkOrder)1, (lambda at third_party/triton/include/triton/Analysis/Allocation.h:222:9), (lambda at third_party/triton/include/triton/Analysis/Allocation.h:224:9)> third_party/triton/include/triton/Analysis/Utility.h:242:7
    triton-lang#24 0x556ee16756b3 in mlir::ModuleAllocation::ModuleAllocation(mlir::ModuleOp) third_party/triton/include/triton/Analysis/Allocation.h:220:5
    triton-lang#25 0x556ee2c2bf18 in (anonymous namespace)::AllocateSharedMemory::runOnOperation() third_party/triton/lib/Conversion/TritonGPUToLLVM/AllocateSharedMemory.cpp:26:22
...
UndefinedBehaviorSanitizer: invalid-shift-exponent third_party/triton/third_party/f2reduce/f2reduce.cpp:421:30 
```
</p>
</details>
stephen-huan pushed a commit to stephen-huan/triton that referenced this issue Dec 24, 2024
…ng#17)

* Fixed yaml syntax

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Removed cpu label from run-on

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing zlib-dev

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added missing apt-get update

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Remove pip cache because on self-hosted runner it slows things down

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Corrected path to tests

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

* Added installation of torch==2.1.2

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>

---------

Signed-off-by: Gregory Shimansky <gshimansky@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants