Index: llvm/trunk/lib/Target/Mips/MipsInstrInfo.h =================================================================== --- llvm/trunk/lib/Target/Mips/MipsInstrInfo.h +++ llvm/trunk/lib/Target/Mips/MipsInstrInfo.h @@ -135,6 +135,9 @@ MachineInstrBuilder genInstrWithNewOpc(unsigned NewOpc, MachineBasicBlock::iterator I) const; + bool findCommutedOpIndices(MachineInstr &MI, unsigned &SrcOpIdx1, + unsigned &SrcOpIdx2) const override; + protected: bool isZeroImm(const MachineOperand &op) const; Index: llvm/trunk/lib/Target/Mips/MipsInstrInfo.cpp =================================================================== --- llvm/trunk/lib/Target/Mips/MipsInstrInfo.cpp +++ llvm/trunk/lib/Target/Mips/MipsInstrInfo.cpp @@ -501,3 +501,31 @@ MIB.setMemRefs(I->memoperands_begin(), I->memoperands_end()); return MIB; } + +bool MipsInstrInfo::findCommutedOpIndices(MachineInstr &MI, unsigned &SrcOpIdx1, + unsigned &SrcOpIdx2) const { + assert(!MI.isBundle() && + "TargetInstrInfo::findCommutedOpIndices() can't handle bundles"); + + const MCInstrDesc &MCID = MI.getDesc(); + if (!MCID.isCommutable()) + return false; + + switch (MI.getOpcode()) { + case Mips::DPADD_U_H: + case Mips::DPADD_U_W: + case Mips::DPADD_U_D: + case Mips::DPADD_S_H: + case Mips::DPADD_S_W: + case Mips::DPADD_S_D: { + // The first operand is both input and output, so it should not commute + if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3)) + return false; + + if (!MI.getOperand(SrcOpIdx1).isReg() || !MI.getOperand(SrcOpIdx2).isReg()) + return false; + return true; + } + } + return TargetInstrInfo::findCommutedOpIndices(MI, SrcOpIdx1, SrcOpIdx2); +} Index: llvm/trunk/test/CodeGen/Mips/msa/3r_4r_widen.ll =================================================================== --- llvm/trunk/test/CodeGen/Mips/msa/3r_4r_widen.ll +++ llvm/trunk/test/CodeGen/Mips/msa/3r_4r_widen.ll @@ -5,18 +5,16 @@ ; RUN: llc -march=mips -mattr=+msa,+fp64 < %s | FileCheck %s ; RUN: llc -march=mipsel -mattr=+msa,+fp64 < %s | FileCheck %s -@llvm_mips_dpadd_s_h_ARG1 = global <8 x i16> , align 16 @llvm_mips_dpadd_s_h_ARG2 = global <16 x i8> , align 16 @llvm_mips_dpadd_s_h_ARG3 = global <16 x i8> , align 16 @llvm_mips_dpadd_s_h_RES = global <8 x i16> , align 16 define void @llvm_mips_dpadd_s_h_test() nounwind { entry: - %0 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_s_h_ARG1 - %1 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_s_h_ARG2 - %2 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_s_h_ARG3 - %3 = tail call <8 x i16> @llvm.mips.dpadd.s.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) - store <8 x i16> %3, <8 x i16>* @llvm_mips_dpadd_s_h_RES + %0 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_s_h_ARG2 + %1 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_s_h_ARG3 + %2 = tail call <8 x i16> @llvm.mips.dpadd.s.h(<8 x i16> , <16 x i8> %0, <16 x i8> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_dpadd_s_h_RES ret void } @@ -25,23 +23,21 @@ ; CHECK: llvm_mips_dpadd_s_h_test: ; CHECK: ld.b ; CHECK: ld.b -; CHECK: ld.h -; CHECK: dpadd_s.h +; CHECK: ldi.h [[R1:\$w[0-9]+]], +; CHECK: dpadd_s.h [[R1]], ; CHECK: st.h ; CHECK: .size llvm_mips_dpadd_s_h_test ; -@llvm_mips_dpadd_s_w_ARG1 = global <4 x i32> , align 16 @llvm_mips_dpadd_s_w_ARG2 = global <8 x i16> , align 16 @llvm_mips_dpadd_s_w_ARG3 = global <8 x i16> , align 16 @llvm_mips_dpadd_s_w_RES = global <4 x i32> , align 16 define void @llvm_mips_dpadd_s_w_test() nounwind { entry: - %0 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_s_w_ARG1 - %1 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_s_w_ARG2 - %2 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_s_w_ARG3 - %3 = tail call <4 x i32> @llvm.mips.dpadd.s.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) - store <4 x i32> %3, <4 x i32>* @llvm_mips_dpadd_s_w_RES + %0 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_s_w_ARG2 + %1 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_s_w_ARG3 + %2 = tail call <4 x i32> @llvm.mips.dpadd.s.w(<4 x i32> , <8 x i16> %0, <8 x i16> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_dpadd_s_w_RES ret void } @@ -50,48 +46,44 @@ ; CHECK: llvm_mips_dpadd_s_w_test: ; CHECK: ld.h ; CHECK: ld.h -; CHECK: ld.w -; CHECK: dpadd_s.w +; CHECK: ldi.w [[R1:\$w[0-9]+]], +; CHECK: dpadd_s.w [[R1]], ; CHECK: st.w ; CHECK: .size llvm_mips_dpadd_s_w_test ; -@llvm_mips_dpadd_s_d_ARG1 = global <2 x i64> , align 16 @llvm_mips_dpadd_s_d_ARG2 = global <4 x i32> , align 16 @llvm_mips_dpadd_s_d_ARG3 = global <4 x i32> , align 16 @llvm_mips_dpadd_s_d_RES = global <2 x i64> , align 16 define void @llvm_mips_dpadd_s_d_test() nounwind { entry: - %0 = load <2 x i64>, <2 x i64>* @llvm_mips_dpadd_s_d_ARG1 - %1 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_s_d_ARG2 - %2 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_s_d_ARG3 - %3 = tail call <2 x i64> @llvm.mips.dpadd.s.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) - store <2 x i64> %3, <2 x i64>* @llvm_mips_dpadd_s_d_RES + %0 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_s_d_ARG2 + %1 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_s_d_ARG3 + %2 = tail call <2 x i64> @llvm.mips.dpadd.s.d(<2 x i64> , <4 x i32> %0, <4 x i32> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_dpadd_s_d_RES ret void } declare <2 x i64> @llvm.mips.dpadd.s.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind ; CHECK: llvm_mips_dpadd_s_d_test: +; CHECK: ldi.d [[R1:\$w[0-9]+]], ; CHECK: ld.w ; CHECK: ld.w -; CHECK: ld.d -; CHECK: dpadd_s.d +; CHECK: dpadd_s.d [[R1]], ; CHECK: st.d ; CHECK: .size llvm_mips_dpadd_s_d_test ; -@llvm_mips_dpadd_u_h_ARG1 = global <8 x i16> , align 16 @llvm_mips_dpadd_u_h_ARG2 = global <16 x i8> , align 16 @llvm_mips_dpadd_u_h_ARG3 = global <16 x i8> , align 16 @llvm_mips_dpadd_u_h_RES = global <8 x i16> , align 16 define void @llvm_mips_dpadd_u_h_test() nounwind { entry: - %0 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_u_h_ARG1 - %1 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_u_h_ARG2 - %2 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_u_h_ARG3 - %3 = tail call <8 x i16> @llvm.mips.dpadd.u.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) - store <8 x i16> %3, <8 x i16>* @llvm_mips_dpadd_u_h_RES + %0 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_u_h_ARG2 + %1 = load <16 x i8>, <16 x i8>* @llvm_mips_dpadd_u_h_ARG3 + %2 = tail call <8 x i16> @llvm.mips.dpadd.u.h(<8 x i16> , <16 x i8> %0, <16 x i8> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_dpadd_u_h_RES ret void } @@ -100,23 +92,21 @@ ; CHECK: llvm_mips_dpadd_u_h_test: ; CHECK: ld.b ; CHECK: ld.b -; CHECK: ld.h -; CHECK: dpadd_u.h +; CHECK: ldi.h [[R1:\$w[0-9]+]], +; CHECK: dpadd_u.h [[R1]], ; CHECK: st.h ; CHECK: .size llvm_mips_dpadd_u_h_test ; -@llvm_mips_dpadd_u_w_ARG1 = global <4 x i32> , align 16 @llvm_mips_dpadd_u_w_ARG2 = global <8 x i16> , align 16 @llvm_mips_dpadd_u_w_ARG3 = global <8 x i16> , align 16 @llvm_mips_dpadd_u_w_RES = global <4 x i32> , align 16 define void @llvm_mips_dpadd_u_w_test() nounwind { entry: - %0 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_u_w_ARG1 - %1 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_u_w_ARG2 - %2 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_u_w_ARG3 - %3 = tail call <4 x i32> @llvm.mips.dpadd.u.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) - store <4 x i32> %3, <4 x i32>* @llvm_mips_dpadd_u_w_RES + %0 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_u_w_ARG2 + %1 = load <8 x i16>, <8 x i16>* @llvm_mips_dpadd_u_w_ARG3 + %2 = tail call <4 x i32> @llvm.mips.dpadd.u.w(<4 x i32> , <8 x i16> %0, <8 x i16> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_dpadd_u_w_RES ret void } @@ -125,33 +115,31 @@ ; CHECK: llvm_mips_dpadd_u_w_test: ; CHECK: ld.h ; CHECK: ld.h -; CHECK: ld.w -; CHECK: dpadd_u.w +; CHECK: ldi.w [[R1:\$w[0-9]+]], +; CHECK: dpadd_u.w [[R1]], ; CHECK: st.w ; CHECK: .size llvm_mips_dpadd_u_w_test ; -@llvm_mips_dpadd_u_d_ARG1 = global <2 x i64> , align 16 @llvm_mips_dpadd_u_d_ARG2 = global <4 x i32> , align 16 @llvm_mips_dpadd_u_d_ARG3 = global <4 x i32> , align 16 @llvm_mips_dpadd_u_d_RES = global <2 x i64> , align 16 define void @llvm_mips_dpadd_u_d_test() nounwind { entry: - %0 = load <2 x i64>, <2 x i64>* @llvm_mips_dpadd_u_d_ARG1 - %1 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_u_d_ARG2 - %2 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_u_d_ARG3 - %3 = tail call <2 x i64> @llvm.mips.dpadd.u.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) - store <2 x i64> %3, <2 x i64>* @llvm_mips_dpadd_u_d_RES + %0 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_u_d_ARG2 + %1 = load <4 x i32>, <4 x i32>* @llvm_mips_dpadd_u_d_ARG3 + %2 = tail call <2 x i64> @llvm.mips.dpadd.u.d(<2 x i64> , <4 x i32> %0, <4 x i32> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_dpadd_u_d_RES ret void } declare <2 x i64> @llvm.mips.dpadd.u.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind ; CHECK: llvm_mips_dpadd_u_d_test: +; CHECK: ldi.d [[R1:\$w[0-9]+]], ; CHECK: ld.w ; CHECK: ld.w -; CHECK: ld.d -; CHECK: dpadd_u.d +; CHECK: dpadd_u.d [[R1]], ; CHECK: st.d ; CHECK: .size llvm_mips_dpadd_u_d_test ;