matmul with transposed RHS operand allows better memory access
patterns on several architectures including common GPUs. Having a named
op for it allows to handle this kind of matmul in a more explicit way.
Details
Details
Diff Detail
Diff Detail
- Repository
- rG LLVM Github Monorepo
Event Timeline
Comment Actions
Hello, it appears that the changes made only include the generated yaml file and additional tests. However, it seems that core_named_ops.py was not included. Could you please confirm this?