Use existing helper instead of handling only a subset of indices lowering arithmetic. Also relax the restriction on the memref rank for the GPU mma ops as we can now support any rank.
Details
Details
Diff Detail
Diff Detail
- Repository
- rG LLVM Github Monorepo
Event Timeline
mlir/lib/Conversion/GPUToNVVM/WmmaOpsToNvvm.cpp | ||
---|---|---|
206 | Why is this cast no longer needed? |
mlir/lib/Conversion/GPUToNVVM/WmmaOpsToNvvm.cpp | ||
---|---|---|
206 | I don't think this comment was correct, the intrinsic is overloaded and can take any pointer type. The pointer type itself doesn't matter since the intrinsic is converted one to one to ptx op wmma.load.a.sync.row.m*n*k*.* where pointers are not typed. |
Why is this cast no longer needed?