Index: lib/Target/X86/X86InstrInfo.cpp =================================================================== --- lib/Target/X86/X86InstrInfo.cpp +++ lib/Target/X86/X86InstrInfo.cpp @@ -3275,6 +3275,20 @@ MI->getOperand(3).setImm(Imm); return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2); } + case X86::VPERM2F128rr: + case X86::VPERM2I128rr: { + // Flip permute source immediate. + // Imm & 0x02: If set, select Op1.lo else Op0.lo. + // Imm & 0x20: If set, select Op1.hi else Op0.hi. + unsigned Imm = MI->getOperand(3).getImm() & 0xFF; + if (NewMI) { + MachineFunction &MF = *MI->getParent()->getParent(); + MI = MF.CloneMachineInstr(MI); + NewMI = false; + } + MI->getOperand(3).setImm(Imm ^ 0x22); + return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2); + } case X86::CMOVB16rr: case X86::CMOVB32rr: case X86::CMOVB64rr: case X86::CMOVAE16rr: case X86::CMOVAE32rr: case X86::CMOVAE64rr: case X86::CMOVE16rr: case X86::CMOVE32rr: case X86::CMOVE64rr: Index: lib/Target/X86/X86InstrSSE.td =================================================================== --- lib/Target/X86/X86InstrSSE.td +++ lib/Target/X86/X86InstrSSE.td @@ -8158,6 +8158,7 @@ // VPERM2F128 - Permute Floating-Point Values in 128-bit chunks // let ExeDomain = SSEPackedSingle in { +let isCommutable = 1 in def VPERM2F128rr : AVXAIi8<0x06, MRMSrcReg, (outs VR256:$dst), (ins VR256:$src1, VR256:$src2, u8imm:$src3), "vperm2f128\t{$src3, $src2, $src1, $dst|$dst, $src1, $src2, $src3}", @@ -8531,6 +8532,7 @@ //===----------------------------------------------------------------------===// // VPERM2I128 - Permute Floating-Point Values in 128-bit chunks // +let isCommutable = 1 in def VPERM2I128rr : AVX2AIi8<0x46, MRMSrcReg, (outs VR256:$dst), (ins VR256:$src1, VR256:$src2, u8imm:$src3), "vperm2i128\t{$src3, $src2, $src1, $dst|$dst, $src1, $src2, $src3}", Index: test/CodeGen/X86/avx-vperm2x128.ll =================================================================== --- test/CodeGen/X86/avx-vperm2x128.ll +++ test/CodeGen/X86/avx-vperm2x128.ll @@ -366,3 +366,174 @@ ret <4 x i64> %c } +;;; Memory folding cases + +define <4 x double> @ld0_hi0_lo1_4f64(<4 x double> * %pa, <4 x double> %b) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld0_hi0_lo1_4f64: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX1-NEXT: vaddpd {{.*}}(%rip), %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld0_hi0_lo1_4f64: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX2-NEXT: vbroadcastsd {{.*}}(%rip), %ymm1 +; AVX2-NEXT: vaddpd %ymm1, %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %a = load <4 x double>, <4 x double> * %pa + %shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> + %res = fadd <4 x double> %shuffle, + ret <4 x double> %res +} + +define <4 x double> @ld1_hi0_hi1_4f64(<4 x double> %a, <4 x double> * %pb) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld1_hi0_hi1_4f64: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX1-NEXT: vaddpd {{.*}}(%rip), %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld1_hi0_hi1_4f64: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX2-NEXT: vbroadcastsd {{.*}}(%rip), %ymm1 +; AVX2-NEXT: vaddpd %ymm1, %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %b = load <4 x double>, <4 x double> * %pb + %shuffle = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> + %res = fadd <4 x double> %shuffle, + ret <4 x double> %res +} + +define <8 x float> @ld0_hi0_lo1_8f32(<8 x float> * %pa, <8 x float> %b) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld0_hi0_lo1_8f32: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX1-NEXT: vaddps {{.*}}(%rip), %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld0_hi0_lo1_8f32: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX2-NEXT: vbroadcastss {{.*}}(%rip), %ymm1 +; AVX2-NEXT: vaddps %ymm1, %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %a = load <8 x float>, <8 x float> * %pa + %shuffle = shufflevector <8 x float> %a, <8 x float> %b, <8 x i32> + %res = fadd <8 x float> %shuffle, + ret <8 x float> %res +} + +define <8 x float> @ld1_hi0_hi1_8f32(<8 x float> %a, <8 x float> * %pb) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld1_hi0_hi1_8f32: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX1-NEXT: vaddps {{.*}}(%rip), %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld1_hi0_hi1_8f32: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX2-NEXT: vbroadcastss {{.*}}(%rip), %ymm1 +; AVX2-NEXT: vaddps %ymm1, %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %b = load <8 x float>, <8 x float> * %pb + %shuffle = shufflevector <8 x float> %a, <8 x float> %b, <8 x i32> + %res = fadd <8 x float> %shuffle, + ret <8 x float> %res +} + +define <4 x i64> @ld0_hi0_lo1_4i64(<4 x i64> * %pa, <4 x i64> %b) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld0_hi0_lo1_4i64: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX1-NEXT: vpaddq {{.*}}(%rip), %xmm0, %xmm1 +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0 +; AVX1-NEXT: vpaddq {{.*}}(%rip), %xmm0, %xmm0 +; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld0_hi0_lo1_4i64: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX2-NEXT: vpaddq {{.*}}(%rip), %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %a = load <4 x i64>, <4 x i64> * %pa + %shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> + %res = add <4 x i64> %shuffle, + ret <4 x i64> %res +} + +define <4 x i64> @ld1_hi0_hi1_4i64(<4 x i64> %a, <4 x i64> * %pb) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld1_hi0_hi1_4i64: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX1-NEXT: vpaddq {{.*}}(%rip), %xmm0, %xmm1 +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0 +; AVX1-NEXT: vpaddq {{.*}}(%rip), %xmm0, %xmm0 +; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld1_hi0_hi1_4i64: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX2-NEXT: vpaddq {{.*}}(%rip), %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %b = load <4 x i64>, <4 x i64> * %pb + %shuffle = shufflevector <4 x i64> %a, <4 x i64> %b, <4 x i32> + %res = add <4 x i64> %shuffle, + ret <4 x i64> %res +} + +define <8 x i32> @ld0_hi0_lo1_8i32(<8 x i32> * %pa, <8 x i32> %b) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld0_hi0_lo1_8i32: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1 +; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [1,2,3,4] +; AVX1-NEXT: vpaddd %xmm2, %xmm1, %xmm1 +; AVX1-NEXT: vpaddd %xmm2, %xmm0, %xmm0 +; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld0_hi0_lo1_8i32: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = mem[2,3],ymm0[0,1] +; AVX2-NEXT: vpaddd {{.*}}(%rip), %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %a = load <8 x i32>, <8 x i32> * %pa + %shuffle = shufflevector <8 x i32> %a, <8 x i32> %b, <8 x i32> + %res = add <8 x i32> %shuffle, + ret <8 x i32> %res +} + +define <8 x i32> @ld1_hi0_hi1_8i32(<8 x i32> %a, <8 x i32> * %pb) nounwind uwtable readnone ssp { +; AVX1-LABEL: ld1_hi0_hi1_8i32: +; AVX1: ## BB#0: ## %entry +; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1 +; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [1,2,3,4] +; AVX1-NEXT: vpaddd %xmm2, %xmm1, %xmm1 +; AVX1-NEXT: vpaddd %xmm2, %xmm0, %xmm0 +; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0 +; AVX1-NEXT: retq +; +; AVX2-LABEL: ld1_hi0_hi1_8i32: +; AVX2: ## BB#0: ## %entry +; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],mem[2,3] +; AVX2-NEXT: vpaddd {{.*}}(%rip), %ymm0, %ymm0 +; AVX2-NEXT: retq +entry: + %b = load <8 x i32>, <8 x i32> * %pb + %shuffle = shufflevector <8 x i32> %a, <8 x i32> %b, <8 x i32> + %res = add <8 x i32> %shuffle, + ret <8 x i32> %res +}