Some necessary yak shaving before lowering *_EXTEND_VECTOR_INREG 256-bit vectors on AVX1 targets as suggested by D52964.
Details
Details
Diff Detail
Diff Detail
- Repository
- rL LLVM
Event Timeline
test/CodeGen/X86/pr35443.ll | ||
---|---|---|
18 ↗ | (On Diff #168587) | @craig.topper Please can you confirm if the pr35443.ll change is acceptable? An alternative is to set the passthrough value zeroinitializer, which instead adds a vpmovzxbq op after the vmovd (some kind of demanded bits failure that could be fixed in a future patch). |
test/CodeGen/X86/pr35443.ll | ||
---|---|---|
18 ↗ | (On Diff #168587) | What if you just change the alignment of @ac to 1? That should prevent the single byte load from the masked.load from promoting to a wider size I think. |
Comment Actions
Tweaked load alignment of test - the additional vpmovzxbq /should/ be removable with a suitable demandedelts+demandedbits combine (probably D52935 in reverse).