diff --git a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp --- a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp +++ b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp @@ -12963,6 +12963,24 @@ return performTRUNCATECombine(N, DAG, Subtarget); case ISD::SELECT: return performSELECTCombine(N, DAG, Subtarget); + case RISCVISD::CZERO_EQZ: + case RISCVISD::CZERO_NEZ: + // czero_eq X, (xor Y, 1) -> czero_ne X, Y if Y is 0 or 1. + // czero_ne X, (xor Y, 1) -> czero_eq X, Y if Y is 0 or 1. + if (N->getOperand(1).getOpcode() == ISD::XOR && + isOneConstant(N->getOperand(1).getOperand(1))) { + SDValue Cond = N->getOperand(1).getOperand(0); + APInt Mask = APInt::getBitsSetFrom(Cond.getValueSizeInBits(), 1); + if (DAG.MaskedValueIsZero(Cond, Mask)) { + unsigned NewOpc = N->getOpcode() == RISCVISD::CZERO_EQZ + ? RISCVISD::CZERO_NEZ + : RISCVISD::CZERO_EQZ; + return DAG.getNode(NewOpc, SDLoc(N), N->getValueType(0), + N->getOperand(0), Cond); + } + } + return SDValue(); + case RISCVISD::SELECT_CC: { // Transform SDValue LHS = N->getOperand(0); diff --git a/llvm/test/CodeGen/RISCV/condops.ll b/llvm/test/CodeGen/RISCV/condops.ll --- a/llvm/test/CodeGen/RISCV/condops.ll +++ b/llvm/test/CodeGen/RISCV/condops.ll @@ -1291,9 +1291,8 @@ ; RV64XVENTANACONDOPS-LABEL: setge: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a0, a1 -; RV64XVENTANACONDOPS-NEXT: xori a0, a0, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a3, a0 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a3, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -1325,9 +1324,8 @@ ; RV64ZICOND-LABEL: setge: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: slt a0, a0, a1 -; RV64ZICOND-NEXT: xori a0, a0, 1 -; RV64ZICOND-NEXT: czero.nez a1, a3, a0 -; RV64ZICOND-NEXT: czero.eqz a0, a2, a0 +; RV64ZICOND-NEXT: czero.eqz a1, a3, a0 +; RV64ZICOND-NEXT: czero.nez a0, a2, a0 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = icmp sge i64 %a, %b @@ -1439,9 +1437,8 @@ ; RV64XVENTANACONDOPS-LABEL: setle: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: slt a0, a1, a0 -; RV64XVENTANACONDOPS-NEXT: xori a0, a0, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a3, a0 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a3, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -1473,9 +1470,8 @@ ; RV64ZICOND-LABEL: setle: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: slt a0, a1, a0 -; RV64ZICOND-NEXT: xori a0, a0, 1 -; RV64ZICOND-NEXT: czero.nez a1, a3, a0 -; RV64ZICOND-NEXT: czero.eqz a0, a2, a0 +; RV64ZICOND-NEXT: czero.eqz a1, a3, a0 +; RV64ZICOND-NEXT: czero.nez a0, a2, a0 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = icmp sle i64 %a, %b @@ -1587,9 +1583,8 @@ ; RV64XVENTANACONDOPS-LABEL: setuge: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a0, a1 -; RV64XVENTANACONDOPS-NEXT: xori a0, a0, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a3, a0 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a3, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -1621,9 +1616,8 @@ ; RV64ZICOND-LABEL: setuge: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: sltu a0, a0, a1 -; RV64ZICOND-NEXT: xori a0, a0, 1 -; RV64ZICOND-NEXT: czero.nez a1, a3, a0 -; RV64ZICOND-NEXT: czero.eqz a0, a2, a0 +; RV64ZICOND-NEXT: czero.eqz a1, a3, a0 +; RV64ZICOND-NEXT: czero.nez a0, a2, a0 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = icmp uge i64 %a, %b @@ -1735,9 +1729,8 @@ ; RV64XVENTANACONDOPS-LABEL: setule: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: sltu a0, a1, a0 -; RV64XVENTANACONDOPS-NEXT: xori a0, a0, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a3, a0 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a2, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a3, a0 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a2, a0 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -1769,9 +1762,8 @@ ; RV64ZICOND-LABEL: setule: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: sltu a0, a1, a0 -; RV64ZICOND-NEXT: xori a0, a0, 1 -; RV64ZICOND-NEXT: czero.nez a1, a3, a0 -; RV64ZICOND-NEXT: czero.eqz a0, a2, a0 +; RV64ZICOND-NEXT: czero.eqz a1, a3, a0 +; RV64ZICOND-NEXT: czero.nez a0, a2, a0 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = icmp ule i64 %a, %b @@ -3336,9 +3328,8 @@ ; RV64XVENTANACONDOPS-LABEL: setune_32: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: feq.s a2, fa0, fa1 -; RV64XVENTANACONDOPS-NEXT: xori a2, a2, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a1, a2 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a0, a2 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a1, a2 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a0, a2 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -3351,18 +3342,16 @@ ; RV32ZICOND-LABEL: setune_32: ; RV32ZICOND: # %bb.0: ; RV32ZICOND-NEXT: feq.s a2, fa0, fa1 -; RV32ZICOND-NEXT: xori a2, a2, 1 -; RV32ZICOND-NEXT: czero.nez a1, a1, a2 -; RV32ZICOND-NEXT: czero.eqz a0, a0, a2 +; RV32ZICOND-NEXT: czero.eqz a1, a1, a2 +; RV32ZICOND-NEXT: czero.nez a0, a0, a2 ; RV32ZICOND-NEXT: or a0, a0, a1 ; RV32ZICOND-NEXT: ret ; ; RV64ZICOND-LABEL: setune_32: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: feq.s a2, fa0, fa1 -; RV64ZICOND-NEXT: xori a2, a2, 1 -; RV64ZICOND-NEXT: czero.nez a1, a1, a2 -; RV64ZICOND-NEXT: czero.eqz a0, a0, a2 +; RV64ZICOND-NEXT: czero.eqz a1, a1, a2 +; RV64ZICOND-NEXT: czero.nez a0, a0, a2 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = fcmp une float %a, %b @@ -3393,9 +3382,8 @@ ; RV64XVENTANACONDOPS-LABEL: setune_64: ; RV64XVENTANACONDOPS: # %bb.0: ; RV64XVENTANACONDOPS-NEXT: feq.s a2, fa0, fa1 -; RV64XVENTANACONDOPS-NEXT: xori a2, a2, 1 -; RV64XVENTANACONDOPS-NEXT: vt.maskcn a1, a1, a2 -; RV64XVENTANACONDOPS-NEXT: vt.maskc a0, a0, a2 +; RV64XVENTANACONDOPS-NEXT: vt.maskc a1, a1, a2 +; RV64XVENTANACONDOPS-NEXT: vt.maskcn a0, a0, a2 ; RV64XVENTANACONDOPS-NEXT: or a0, a0, a1 ; RV64XVENTANACONDOPS-NEXT: ret ; @@ -3408,21 +3396,19 @@ ; RV32ZICOND-LABEL: setune_64: ; RV32ZICOND: # %bb.0: ; RV32ZICOND-NEXT: feq.s a4, fa0, fa1 -; RV32ZICOND-NEXT: xori a4, a4, 1 -; RV32ZICOND-NEXT: czero.nez a2, a2, a4 -; RV32ZICOND-NEXT: czero.eqz a0, a0, a4 +; RV32ZICOND-NEXT: czero.eqz a2, a2, a4 +; RV32ZICOND-NEXT: czero.nez a0, a0, a4 ; RV32ZICOND-NEXT: or a0, a0, a2 -; RV32ZICOND-NEXT: czero.nez a2, a3, a4 -; RV32ZICOND-NEXT: czero.eqz a1, a1, a4 +; RV32ZICOND-NEXT: czero.eqz a2, a3, a4 +; RV32ZICOND-NEXT: czero.nez a1, a1, a4 ; RV32ZICOND-NEXT: or a1, a1, a2 ; RV32ZICOND-NEXT: ret ; ; RV64ZICOND-LABEL: setune_64: ; RV64ZICOND: # %bb.0: ; RV64ZICOND-NEXT: feq.s a2, fa0, fa1 -; RV64ZICOND-NEXT: xori a2, a2, 1 -; RV64ZICOND-NEXT: czero.nez a1, a1, a2 -; RV64ZICOND-NEXT: czero.eqz a0, a0, a2 +; RV64ZICOND-NEXT: czero.eqz a1, a1, a2 +; RV64ZICOND-NEXT: czero.nez a0, a0, a2 ; RV64ZICOND-NEXT: or a0, a0, a1 ; RV64ZICOND-NEXT: ret %rc = fcmp une float %a, %b