MatMul and FullyConnected have transposed dimensions for the weights.
Also, removed uneeded tensor reshape for bias.
Details
Details
Diff Detail
Diff Detail
- Repository
- rG LLVM Github Monorepo
Event Timeline
mlir/lib/Conversion/TosaToLinalg/TosaToLinalg.cpp | ||
---|---|---|
751 | I understand that you want to do this, but just to throw this out there, might want to actually have the transpose variants of matmul. You could write compiler passes to combine transpose with matmuls and present different variants to code-generation. |
mlir/lib/Conversion/TosaToLinalg/TosaToLinalg.cpp | ||
---|---|---|
751 | In progress of discussing named op variants of matmul. To avoid having this be a block issue I'll submit and update to whatever solution we find. |
I understand that you want to do this, but just to throw this out there, might want to actually have the transpose variants of matmul. You could write compiler passes to combine transpose with matmuls and present different variants to code-generation.