This is an archive of the discontinued LLVM Phabricator instance.

[mlir][sparse][GPU] add 2:4 integration test
ClosedPublic

Authored by K-Wu on Jun 6 2023, 9:48 AM.

Diff Detail

Event Timeline

K-Wu created this revision.Jun 6 2023, 9:48 AM
Herald added a project: Restricted Project. · View Herald Transcript
K-Wu requested review of this revision.Jun 6 2023, 9:48 AM
K-Wu edited the summary of this revision. (Show Details)Jun 6 2023, 9:51 AM
K-Wu planned changes to this revision.Jun 6 2023, 9:51 AM
K-Wu updated this revision to Diff 529392.Jun 7 2023, 11:49 AM
K-Wu edited the summary of this revision. (Show Details)

test untracking file

K-Wu updated this revision to Diff 529393.Jun 7 2023, 11:51 AM

not working

K-Wu planned changes to this revision.Jun 7 2023, 11:51 AM
K-Wu updated this revision to Diff 529394.Jun 7 2023, 11:57 AM

try again

K-Wu updated this revision to Diff 529395.Jun 7 2023, 11:57 AM

recover

K-Wu updated this revision to Diff 529401.Jun 7 2023, 12:29 PM

try again

K-Wu updated this revision to Diff 529402.Jun 7 2023, 12:29 PM

recover

K-Wu updated this revision to Diff 529740.Jun 8 2023, 3:10 PM

initial

K-Wu updated this revision to Diff 529774.Jun 8 2023, 4:30 PM

rebase origin/main

K-Wu updated this revision to Diff 529791.Jun 8 2023, 5:30 PM

upd test

K-Wu updated this revision to Diff 529809.Jun 8 2023, 7:46 PM

update

K-Wu updated this revision to Diff 530112.Jun 9 2023, 4:20 PM

working now

aartbik added inline comments.Jun 12 2023, 1:00 PM
mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-matmul-2-4-lib.mlir
17

This seems a bit copy-and-past from the sparse-mma-2-4-f16.mlir test (which really uses device code for this method by means of e.g. nvgpu.mma.sp.sync). Here, however ,the library calls are still made from the host. So I would remove the while device/host comments here at L17 and at L62).

Also, the gpu.container_module is not needed, since no method is defined as gpu.module

34

commented out code?

207

avoid commented out code

K-Wu updated this revision to Diff 530651.Jun 12 2023, 1:21 PM

address comments

K-Wu marked 3 inline comments as done.Jun 12 2023, 1:22 PM
K-Wu added inline comments.
mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-matmul-2-4-lib.mlir
17

Thanks for all these comments! They are all addressed now

aartbik added inline comments.Jun 12 2023, 4:30 PM
mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-matmul-2-4-lib.mlir
5

it looks like this pipeline can be simplified quite a bit, all the gpu.module(....) can go, right?
Also, the vector to llvm and probably more. Perhaps you can actually get rid of the first mlir-opt call and just start at L7 (bit hard to tell just by looking, but run it by hand and see how far you can strip it)

15

remove gpu.container_module

25

add comment to magic constant here

42

does it work without? in any case, let the TODO jump out a bi t more

119

Copy and paste comment, this is no longer the compressed matrix, but the full 2:4 matrix A

145

empty // line after this comment to seperate it from the CHECK

197

there are no warps in this code, so simply Call the kernel
this all still runs on host

K-Wu edited the summary of this revision. (Show Details)Jun 12 2023, 4:41 PM
K-Wu marked an inline comment as done.
K-Wu marked 6 inline comments as done.Jun 12 2023, 4:59 PM
K-Wu marked an inline comment as done.Jun 12 2023, 5:17 PM
K-Wu updated this revision to Diff 530727.Jun 12 2023, 5:18 PM

rebase origin/main; addressing comments

K-Wu updated this revision to Diff 530728.Jun 12 2023, 5:22 PM

fix test error

aartbik accepted this revision.Jun 12 2023, 6:49 PM
This revision is now accepted and ready to land.Jun 12 2023, 6:49 PM
This revision was automatically updated to reflect the committed changes.