Index: llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp =================================================================== --- llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp +++ llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp @@ -7240,6 +7240,12 @@ if (N1 == N2) return N1; + // fold (select, C, X, undef) -> X + if (N2.isUndef()) + return N1; + if (N1.isUndef()) + return N2; + if (const ConstantSDNode *N0C = dyn_cast(N0)) { // fold (select true, X, Y) -> X // fold (select false, X, Y) -> Y Index: llvm/trunk/test/CodeGen/AArch64/half.ll =================================================================== --- llvm/trunk/test/CodeGen/AArch64/half.ll +++ llvm/trunk/test/CodeGen/AArch64/half.ll @@ -87,9 +87,9 @@ ;CHECK: fcmp %cmp0 = fcmp ogt half 0xH3333, undef %cmp1 = fcmp ogt half 0xH2222, undef - %x = select i1 %cmp0, i16 0, i16 undef + %x = select i1 %cmp0, i16 0, i16 1 %or = or i1 %cmp1, %cmp0 - %y = select i1 %or, i16 4, i16 undef + %y = select i1 %or, i16 4, i16 1 %r = add i16 %x, %y ret i16 %r } Index: llvm/trunk/test/CodeGen/AMDGPU/select-undef.ll =================================================================== --- llvm/trunk/test/CodeGen/AMDGPU/select-undef.ll +++ llvm/trunk/test/CodeGen/AMDGPU/select-undef.ll @@ -0,0 +1,25 @@ +; RUN: llc -mtriple=amdgcn-amd-amdhsa < %s | FileCheck -check-prefix=GCN %s + +; GCN-LABEL: {{^}}select_undef_lhs: +; GCN: s_waitcnt +; GCN-NOT: v_cmp +; GCN-NOT: v_cndmask +; GCN-NEXT: s_setpc_b64 +define float @select_undef_lhs(float %val, i1 %cond) { + %undef = call float @llvm.amdgcn.rcp.f32(float undef) + %sel = select i1 %cond, float %undef, float %val + ret float %sel +} + +; GCN-LABEL: {{^}}select_undef_rhs: +; GCN: s_waitcnt +; GCN-NOT: v_cmp +; GCN-NOT: v_cndmask +; GCN-NEXT: s_setpc_b64 +define float @select_undef_rhs(float %val, i1 %cond) { + %undef = call float @llvm.amdgcn.rcp.f32(float undef) + %sel = select i1 %cond, float %val, float %undef + ret float %sel +} + +declare float @llvm.amdgcn.rcp.f32(float) Index: llvm/trunk/test/CodeGen/ARM/sub-cmp-peephole.ll =================================================================== --- llvm/trunk/test/CodeGen/ARM/sub-cmp-peephole.ll +++ llvm/trunk/test/CodeGen/ARM/sub-cmp-peephole.ll @@ -75,7 +75,7 @@ ; CHECK: cmp define i32 @bc_raise(i1 %cond) nounwind ssp { entry: - %val.2.i = select i1 %cond, i32 0, i32 undef + %val.2.i = select i1 %cond, i32 0, i32 1 %sub.i = sub nsw i32 0, %val.2.i %retval.0.i = select i1 %cond, i32 %val.2.i, i32 %sub.i %cmp1 = icmp eq i32 %retval.0.i, 0 Index: llvm/trunk/test/CodeGen/Generic/dag-combine-select-undef.ll =================================================================== --- llvm/trunk/test/CodeGen/Generic/dag-combine-select-undef.ll +++ llvm/trunk/test/CodeGen/Generic/dag-combine-select-undef.ll @@ -0,0 +1,24 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s | FileCheck %s + +define void @select_undef_n1(float addrspace(1)* %a, i32 %c) { +; CHECK-LABEL: select_undef_n1: +; CHECK: # %bb.0: +; CHECK-NEXT: movl $1065353216, (%rdi) # imm = 0x3F800000 +; CHECK-NEXT: retq + %cc = icmp eq i32 %c, 0 + %sel = select i1 %cc, float 1.000000e+00, float undef + store float %sel, float addrspace(1)* %a + ret void +} + +define void @select_undef_n2(float addrspace(1)* %a, i32 %c) { +; CHECK-LABEL: select_undef_n2: +; CHECK: # %bb.0: +; CHECK-NEXT: movl $1065353216, (%rdi) # imm = 0x3F800000 +; CHECK-NEXT: retq + %cc = icmp eq i32 %c, 0 + %sel = select i1 %cc, float undef, float 1.000000e+00 + store float %sel, float addrspace(1)* %a + ret void +} Index: llvm/trunk/test/CodeGen/Hexagon/autohvx/isel-extractelt-illegal-type.ll =================================================================== --- llvm/trunk/test/CodeGen/Hexagon/autohvx/isel-extractelt-illegal-type.ll +++ llvm/trunk/test/CodeGen/Hexagon/autohvx/isel-extractelt-illegal-type.ll @@ -21,7 +21,7 @@ %v7 = tail call i32 @llvm.hexagon.A2.subh.l16.sat.ll(i32 %v6, i32 16) %v8 = trunc i32 %v7 to i16 %v9 = icmp sgt i16 %v8, -1 - %v10 = select i1 %v9, i16 0, i16 undef + %v10 = select i1 %v9, i16 0, i16 1 ret i16 %v10 } Index: llvm/trunk/test/CodeGen/Hexagon/swp-const-tc1.ll =================================================================== --- llvm/trunk/test/CodeGen/Hexagon/swp-const-tc1.ll +++ llvm/trunk/test/CodeGen/Hexagon/swp-const-tc1.ll @@ -34,7 +34,7 @@ %v7 = add i32 %v6, undef %v8 = icmp slt i32 undef, %v7 %v9 = add nsw i32 %v7, 1 - %v10 = select i1 undef, i32 undef, i32 %v9 + %v10 = select i1 undef, i32 1, i32 %v9 %v11 = add i32 %v10, 0 %v12 = getelementptr inbounds i8, i8* null, i32 %v11 %v13 = load i8, i8* %v12, align 1, !tbaa !4 Index: llvm/trunk/test/CodeGen/SystemZ/subregliveness-04.ll =================================================================== --- llvm/trunk/test/CodeGen/SystemZ/subregliveness-04.ll +++ llvm/trunk/test/CodeGen/SystemZ/subregliveness-04.ll @@ -20,7 +20,7 @@ bb5: ; preds = %bb3 %tmp6 = or i1 %tmp2, false - %tmp7 = select i1 %tmp6, i32 0, i32 undef + %tmp7 = select i1 %tmp6, i32 0, i32 100 %tmp8 = ashr i32 %tmp1, %tmp7 %tmp9 = zext i32 %tmp8 to i64 %tmp10 = shl i64 %tmp9, 48 Index: llvm/trunk/test/CodeGen/X86/2012-08-07-CmpISelBug.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/2012-08-07-CmpISelBug.ll +++ llvm/trunk/test/CodeGen/X86/2012-08-07-CmpISelBug.ll @@ -26,7 +26,7 @@ %tmp113 = sub i8 %tmp106, 0 %tmp114 = add i8 %tmp113, -72 %tmp141 = icmp ne i32 %tmp67, -1263900958 - %tmp142 = select i1 %tmp141, i8 %tmp114, i8 undef + %tmp142 = select i1 %tmp141, i8 %tmp114, i8 1 %tmp143 = xor i8 %tmp142, 81 %tmp144 = zext i8 %tmp143 to i32 %tmp145 = add i32 %tmp144, 2062143348 Index: llvm/trunk/test/CodeGen/X86/hoist-spill.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/hoist-spill.ll +++ llvm/trunk/test/CodeGen/X86/hoist-spill.ll @@ -27,10 +27,10 @@ %c.0 = phi i32 [ %inc15, %for.inc14 ], [ 1, %entry ] %k.0 = phi i32 [ %k.1.lcssa, %for.inc14 ], [ undef, %entry ] %tmp3 = icmp sgt i32 undef, 0 - %smax52 = select i1 %tmp3, i32 undef, i32 0 + %smax52 = select i1 %tmp3, i32 %c.0, i32 0 %tmp4 = zext i32 %smax52 to i64 %tmp5 = icmp sgt i64 undef, %tmp4 - %smax53 = select i1 %tmp5, i64 undef, i64 %tmp4 + %smax53 = select i1 %tmp5, i64 %tmp2, i64 %tmp4 %tmp6 = add nsw i64 %smax53, 1 %tmp7 = sub nsw i64 %tmp6, %tmp4 %tmp8 = add nsw i64 %tmp7, -8 Index: llvm/trunk/test/CodeGen/X86/pr31045.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/pr31045.ll +++ llvm/trunk/test/CodeGen/X86/pr31045.ll @@ -19,28 +19,16 @@ ; CHECK-LABEL: _Z1av: ; CHECK: # %bb.0: # %entry ; CHECK-NEXT: movl struct_obj_3+{{.*}}(%rip), %eax -; CHECK-NEXT: movsbl {{.*}}(%rip), %ecx +; CHECK-NEXT: movzbl {{.*}}(%rip), %ecx ; CHECK-NEXT: movzbl {{.*}}(%rip), %edx ; CHECK-NEXT: andl $1, %eax -; CHECK-NEXT: leal (%rax,%rax), %esi -; CHECK-NEXT: subl %ecx, %esi -; CHECK-NEXT: subl %edx, %esi -; CHECK-NEXT: xorl %ecx, %ecx -; CHECK-NEXT: testb %al, %al -; CHECK-NEXT: cmovel %eax, %ecx -; CHECK-NEXT: movzbl {{.*}}(%rip), %edx -; CHECK-NEXT: andl struct_obj_8+{{.*}}(%rip), %ecx -; CHECK-NEXT: andl $1, %ecx -; CHECK-NEXT: negl %ecx -; CHECK-NEXT: andl %edx, %ecx -; CHECK-NEXT: negl %ecx -; CHECK-NEXT: andl %eax, %ecx -; CHECK-NEXT: negl %ecx -; CHECK-NEXT: testl %ecx, %esi -; CHECK-NEXT: notl %esi -; CHECK-NEXT: movzbl %sil, %eax +; CHECK-NEXT: addl %eax, %eax +; CHECK-NEXT: subl %ecx, %eax +; CHECK-NEXT: subl %edx, %eax +; CHECK-NEXT: notl %eax +; CHECK-NEXT: movzbl %al, %eax ; CHECK-NEXT: movw %ax, struct_obj_12+{{.*}}(%rip) -; CHECK-NEXT: setne {{.*}}(%rip) +; CHECK-NEXT: movb $0, {{.*}}(%rip) ; CHECK-NEXT: retq entry: %bf.load = load i32, i32* bitcast (i24* getelementptr inbounds (%struct.d.3.7.11.15.39.71.75.91.95.99.107.123.363, %struct.d.3.7.11.15.39.71.75.91.95.99.107.123.363* @struct_obj_3, i64 0, i32 0, i32 2) to i32*), align 2 Index: llvm/trunk/test/CodeGen/X86/pr32610.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/pr32610.ll +++ llvm/trunk/test/CodeGen/X86/pr32610.ll @@ -20,11 +20,11 @@ %conv = zext i1 %cmp to i32 %tobool1.i = icmp ne i32 undef, 0 %or.cond.i = and i1 %cmp, %tobool1.i - %cond.i = select i1 %or.cond.i, i32 %conv, i32 undef + %cond.i = select i1 %or.cond.i, i32 %conv, i32 1 store i32 %cond.i, i32* @c, align 4, !tbaa !1 %1 = load i32, i32* getelementptr inbounds ([1 x i32], [1 x i32]* @b, i32 0, i32 0), align 4 %tobool = icmp ne i32 %1, 0 - %2 = select i1 %tobool, i32 %1, i32 undef + %2 = select i1 %tobool, i32 %1, i32 2 store i32 %2, i32* @d, align 4, !tbaa !1 ret void } Index: llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll +++ llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll @@ -13,449 +13,245 @@ ; ILP-LABEL: test1: ; ILP: # %bb.0: ; ILP-NEXT: pushq %rbp -; ILP-NEXT: pushq %r15 -; ILP-NEXT: pushq %r14 -; ILP-NEXT: pushq %r13 -; ILP-NEXT: pushq %r12 ; ILP-NEXT: pushq %rbx -; ILP-NEXT: movq %rcx, %r9 -; ILP-NEXT: movq %rdi, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill -; ILP-NEXT: xorl %eax, %eax -; ILP-NEXT: addq $1, %rsi -; ILP-NEXT: adcq $0, %rdx -; ILP-NEXT: adcq $0, %r9 -; ILP-NEXT: adcq $0, %r8 -; ILP-NEXT: leal 1(%rsi,%rsi), %edi -; ILP-NEXT: movl $1, %ebp -; ILP-NEXT: xorl %r14d, %r14d -; ILP-NEXT: movl %edi, %ecx -; ILP-NEXT: shldq %cl, %rbp, %r14 -; ILP-NEXT: movl $1, %r11d -; ILP-NEXT: shlq %cl, %r11 -; ILP-NEXT: movb $-128, %r10b -; ILP-NEXT: subb %dil, %r10b -; ILP-NEXT: movq %r9, %r13 -; ILP-NEXT: movl %r10d, %ecx -; ILP-NEXT: shlq %cl, %r13 -; ILP-NEXT: movl $1, %r12d -; ILP-NEXT: shrdq %cl, %rax, %r12 -; ILP-NEXT: xorl %r15d, %r15d -; ILP-NEXT: movl %edi, %ecx -; ILP-NEXT: shldq %cl, %r15, %r15 -; ILP-NEXT: movq %rsi, %rbx -; ILP-NEXT: shrdq %cl, %rdx, %rbx -; ILP-NEXT: shrq %cl, %rdx -; ILP-NEXT: addb $-128, %cl -; ILP-NEXT: shrdq %cl, %r8, %r9 -; ILP-NEXT: testb $64, %dil -; ILP-NEXT: cmovneq %r11, %r14 -; ILP-NEXT: cmoveq %rbx, %rdx -; ILP-NEXT: cmovneq %rax, %r15 -; ILP-NEXT: cmovneq %rax, %r11 -; ILP-NEXT: testb $64, %r10b -; ILP-NEXT: cmovneq %rax, %r12 -; ILP-NEXT: cmovneq %rax, %r13 +; ILP-NEXT: movq %rdi, %rax +; ILP-NEXT: leal 3(%rsi,%rsi), %ebp +; ILP-NEXT: movl %ebp, %r11d +; ILP-NEXT: addb $-128, %r11b +; ILP-NEXT: xorl %r8d, %r8d +; ILP-NEXT: movl $1, %r10d +; ILP-NEXT: xorl %edi, %edi +; ILP-NEXT: movl %ebp, %ecx +; ILP-NEXT: shldq %cl, %r10, %rdi +; ILP-NEXT: xorl %r9d, %r9d +; ILP-NEXT: movl %r11d, %ecx +; ILP-NEXT: shldq %cl, %r10, %r9 +; ILP-NEXT: xorl %esi, %esi +; ILP-NEXT: movl %ebp, %ecx +; ILP-NEXT: shldq %cl, %rsi, %rsi +; ILP-NEXT: movl $1, %edx +; ILP-NEXT: shlq %cl, %rdx ; ILP-NEXT: movl $1, %ebx +; ILP-NEXT: movl %r11d, %ecx ; ILP-NEXT: shlq %cl, %rbx -; ILP-NEXT: orl %edx, %r13d -; ILP-NEXT: xorl %edx, %edx -; ILP-NEXT: movl $1, %ebp -; ILP-NEXT: shldq %cl, %rbp, %rdx -; ILP-NEXT: shrq %cl, %r8 +; ILP-NEXT: movb $-128, %cl +; ILP-NEXT: subb %bpl, %cl +; ILP-NEXT: shrdq %cl, %r8, %r10 ; ILP-NEXT: testb $64, %cl -; ILP-NEXT: cmoveq %r9, %r8 -; ILP-NEXT: cmovneq %rbx, %rdx -; ILP-NEXT: cmovneq %rax, %rbx -; ILP-NEXT: testb %dil, %dil -; ILP-NEXT: cmovsq %rax, %r14 -; ILP-NEXT: cmovsq %rax, %r11 -; ILP-NEXT: jns .LBB0_2 -; ILP-NEXT: # %bb.1: -; ILP-NEXT: movl %r8d, %r13d -; ILP-NEXT: .LBB0_2: -; ILP-NEXT: je .LBB0_4 -; ILP-NEXT: # %bb.3: -; ILP-NEXT: movl %r13d, %esi -; ILP-NEXT: .LBB0_4: -; ILP-NEXT: cmovnsq %r12, %rbx -; ILP-NEXT: cmoveq %rax, %rbx -; ILP-NEXT: cmovnsq %r15, %rdx -; ILP-NEXT: cmoveq %rax, %rdx -; ILP-NEXT: testb $1, %sil -; ILP-NEXT: cmovneq %rax, %rdx -; ILP-NEXT: movq {{[-0-9]+}}(%r{{[sb]}}p), %rax # 8-byte Reload -; ILP-NEXT: movq %rdx, 24(%rax) -; ILP-NEXT: cmovneq %rax, %rbx +; ILP-NEXT: cmovneq %r8, %r10 +; ILP-NEXT: testb $64, %r11b +; ILP-NEXT: cmovneq %rbx, %r9 +; ILP-NEXT: cmovneq %r8, %rbx +; ILP-NEXT: testb $64, %bpl +; ILP-NEXT: cmovneq %rdx, %rdi +; ILP-NEXT: cmovneq %r8, %rdx +; ILP-NEXT: cmovneq %r8, %rsi +; ILP-NEXT: testb %bpl, %bpl +; ILP-NEXT: cmovsq %r8, %rdi +; ILP-NEXT: cmovsq %r8, %rdx +; ILP-NEXT: movq %rdi, 8(%rax) +; ILP-NEXT: movq %rdx, (%rax) +; ILP-NEXT: cmovsq %r9, %rsi +; ILP-NEXT: cmoveq %r8, %rsi +; ILP-NEXT: movq %rsi, 24(%rax) +; ILP-NEXT: cmovnsq %r10, %rbx +; ILP-NEXT: cmoveq %r8, %rbx ; ILP-NEXT: movq %rbx, 16(%rax) -; ILP-NEXT: cmovneq %rax, %r14 -; ILP-NEXT: movq %r14, 8(%rax) -; ILP-NEXT: cmovneq %rax, %r11 -; ILP-NEXT: movq %r11, (%rax) ; ILP-NEXT: popq %rbx -; ILP-NEXT: popq %r12 -; ILP-NEXT: popq %r13 -; ILP-NEXT: popq %r14 -; ILP-NEXT: popq %r15 ; ILP-NEXT: popq %rbp ; ILP-NEXT: retq ; ; HYBRID-LABEL: test1: ; HYBRID: # %bb.0: -; HYBRID-NEXT: pushq %rbp -; HYBRID-NEXT: pushq %r15 -; HYBRID-NEXT: pushq %r14 -; HYBRID-NEXT: pushq %r13 -; HYBRID-NEXT: pushq %r12 ; HYBRID-NEXT: pushq %rbx -; HYBRID-NEXT: movq %rcx, %r9 ; HYBRID-NEXT: movq %rdi, %rax -; HYBRID-NEXT: addq $1, %rsi -; HYBRID-NEXT: adcq $0, %rdx -; HYBRID-NEXT: adcq $0, %r9 -; HYBRID-NEXT: adcq $0, %r8 -; HYBRID-NEXT: xorl %r10d, %r10d -; HYBRID-NEXT: leal 1(%rsi,%rsi), %edi -; HYBRID-NEXT: xorl %r14d, %r14d -; HYBRID-NEXT: movl %edi, %ecx -; HYBRID-NEXT: shldq %cl, %r14, %r14 -; HYBRID-NEXT: testb $64, %dil -; HYBRID-NEXT: cmovneq %r10, %r14 -; HYBRID-NEXT: movl $1, %ebp -; HYBRID-NEXT: movl $1, %r12d -; HYBRID-NEXT: shlq %cl, %r12 -; HYBRID-NEXT: testb $64, %dil -; HYBRID-NEXT: movq %r12, %r11 -; HYBRID-NEXT: cmovneq %r10, %r11 -; HYBRID-NEXT: movq %rsi, %rbx -; HYBRID-NEXT: shrdq %cl, %rdx, %rbx -; HYBRID-NEXT: shrq %cl, %rdx -; HYBRID-NEXT: testb $64, %dil -; HYBRID-NEXT: cmoveq %rbx, %rdx -; HYBRID-NEXT: xorl %r15d, %r15d -; HYBRID-NEXT: shldq %cl, %rbp, %r15 -; HYBRID-NEXT: testb $64, %dil -; HYBRID-NEXT: cmovneq %r12, %r15 +; HYBRID-NEXT: leal 3(%rsi,%rsi), %r10d ; HYBRID-NEXT: movb $-128, %cl -; HYBRID-NEXT: subb %dil, %cl -; HYBRID-NEXT: movq %r9, %r13 -; HYBRID-NEXT: shlq %cl, %r13 -; HYBRID-NEXT: movl $1, %r12d -; HYBRID-NEXT: shrdq %cl, %r10, %r12 -; HYBRID-NEXT: testb $64, %cl -; HYBRID-NEXT: cmovneq %r10, %r12 -; HYBRID-NEXT: cmovneq %r10, %r13 -; HYBRID-NEXT: orl %edx, %r13d -; HYBRID-NEXT: movl %edi, %ecx -; HYBRID-NEXT: addb $-128, %cl +; HYBRID-NEXT: subb %r10b, %cl +; HYBRID-NEXT: xorl %r8d, %r8d +; HYBRID-NEXT: movl $1, %esi +; HYBRID-NEXT: movl $1, %r9d ; HYBRID-NEXT: shrdq %cl, %r8, %r9 -; HYBRID-NEXT: shrq %cl, %r8 +; HYBRID-NEXT: testb $64, %cl +; HYBRID-NEXT: cmovneq %r8, %r9 ; HYBRID-NEXT: xorl %edx, %edx -; HYBRID-NEXT: shldq %cl, %rbp, %rdx -; HYBRID-NEXT: shlq %cl, %rbp +; HYBRID-NEXT: movl %r10d, %ecx +; HYBRID-NEXT: shldq %cl, %rsi, %rdx +; HYBRID-NEXT: addb $-128, %cl +; HYBRID-NEXT: xorl %r11d, %r11d +; HYBRID-NEXT: shldq %cl, %rsi, %r11 +; HYBRID-NEXT: movl $1, %edi +; HYBRID-NEXT: shlq %cl, %rdi ; HYBRID-NEXT: testb $64, %cl -; HYBRID-NEXT: cmovneq %rbp, %rdx -; HYBRID-NEXT: cmoveq %r9, %r8 -; HYBRID-NEXT: cmovneq %r10, %rbp -; HYBRID-NEXT: testb %dil, %dil -; HYBRID-NEXT: jns .LBB0_2 -; HYBRID-NEXT: # %bb.1: -; HYBRID-NEXT: movl %r8d, %r13d -; HYBRID-NEXT: .LBB0_2: -; HYBRID-NEXT: je .LBB0_4 -; HYBRID-NEXT: # %bb.3: -; HYBRID-NEXT: movl %r13d, %esi -; HYBRID-NEXT: .LBB0_4: -; HYBRID-NEXT: cmovsq %r10, %r15 -; HYBRID-NEXT: cmovnsq %r12, %rbp -; HYBRID-NEXT: cmoveq %r10, %rbp -; HYBRID-NEXT: cmovnsq %r14, %rdx -; HYBRID-NEXT: cmoveq %r10, %rdx -; HYBRID-NEXT: cmovsq %r10, %r11 -; HYBRID-NEXT: testb $1, %sil -; HYBRID-NEXT: cmovneq %rax, %rdx -; HYBRID-NEXT: movq %rdx, 24(%rax) -; HYBRID-NEXT: cmovneq %rax, %rbp -; HYBRID-NEXT: movq %rbp, 16(%rax) -; HYBRID-NEXT: cmovneq %rax, %r15 -; HYBRID-NEXT: movq %r15, 8(%rax) -; HYBRID-NEXT: cmovneq %rax, %r11 -; HYBRID-NEXT: movq %r11, (%rax) +; HYBRID-NEXT: cmovneq %rdi, %r11 +; HYBRID-NEXT: cmovneq %r8, %rdi +; HYBRID-NEXT: xorl %ebx, %ebx +; HYBRID-NEXT: movl %r10d, %ecx +; HYBRID-NEXT: shldq %cl, %rbx, %rbx +; HYBRID-NEXT: shlq %cl, %rsi +; HYBRID-NEXT: testb $64, %r10b +; HYBRID-NEXT: cmovneq %rsi, %rdx +; HYBRID-NEXT: cmovneq %r8, %rbx +; HYBRID-NEXT: cmovneq %r8, %rsi +; HYBRID-NEXT: testb %r10b, %r10b +; HYBRID-NEXT: cmovsq %r8, %rdx +; HYBRID-NEXT: movq %rdx, 8(%rax) +; HYBRID-NEXT: cmovsq %r8, %rsi +; HYBRID-NEXT: movq %rsi, (%rax) +; HYBRID-NEXT: cmovsq %r11, %rbx +; HYBRID-NEXT: cmoveq %r8, %rbx +; HYBRID-NEXT: movq %rbx, 24(%rax) +; HYBRID-NEXT: cmovnsq %r9, %rdi +; HYBRID-NEXT: cmoveq %r8, %rdi +; HYBRID-NEXT: movq %rdi, 16(%rax) ; HYBRID-NEXT: popq %rbx -; HYBRID-NEXT: popq %r12 -; HYBRID-NEXT: popq %r13 -; HYBRID-NEXT: popq %r14 -; HYBRID-NEXT: popq %r15 -; HYBRID-NEXT: popq %rbp ; HYBRID-NEXT: retq ; ; BURR-LABEL: test1: ; BURR: # %bb.0: -; BURR-NEXT: pushq %rbp -; BURR-NEXT: pushq %r15 -; BURR-NEXT: pushq %r14 -; BURR-NEXT: pushq %r13 -; BURR-NEXT: pushq %r12 ; BURR-NEXT: pushq %rbx -; BURR-NEXT: movq %rcx, %r9 ; BURR-NEXT: movq %rdi, %rax -; BURR-NEXT: addq $1, %rsi -; BURR-NEXT: adcq $0, %rdx -; BURR-NEXT: adcq $0, %r9 -; BURR-NEXT: adcq $0, %r8 -; BURR-NEXT: xorl %r10d, %r10d -; BURR-NEXT: leal 1(%rsi,%rsi), %edi -; BURR-NEXT: xorl %r14d, %r14d -; BURR-NEXT: movl %edi, %ecx -; BURR-NEXT: shldq %cl, %r14, %r14 -; BURR-NEXT: testb $64, %dil -; BURR-NEXT: cmovneq %r10, %r14 -; BURR-NEXT: movl $1, %ebp -; BURR-NEXT: movl $1, %r12d -; BURR-NEXT: shlq %cl, %r12 -; BURR-NEXT: testb $64, %dil -; BURR-NEXT: movq %r12, %r11 -; BURR-NEXT: cmovneq %r10, %r11 -; BURR-NEXT: movq %rsi, %rbx -; BURR-NEXT: shrdq %cl, %rdx, %rbx -; BURR-NEXT: shrq %cl, %rdx -; BURR-NEXT: testb $64, %dil -; BURR-NEXT: cmoveq %rbx, %rdx -; BURR-NEXT: xorl %r15d, %r15d -; BURR-NEXT: shldq %cl, %rbp, %r15 -; BURR-NEXT: testb $64, %dil -; BURR-NEXT: cmovneq %r12, %r15 +; BURR-NEXT: leal 3(%rsi,%rsi), %r10d ; BURR-NEXT: movb $-128, %cl -; BURR-NEXT: subb %dil, %cl -; BURR-NEXT: movq %r9, %r13 -; BURR-NEXT: shlq %cl, %r13 -; BURR-NEXT: movl $1, %r12d -; BURR-NEXT: shrdq %cl, %r10, %r12 -; BURR-NEXT: testb $64, %cl -; BURR-NEXT: cmovneq %r10, %r12 -; BURR-NEXT: cmovneq %r10, %r13 -; BURR-NEXT: orl %edx, %r13d -; BURR-NEXT: movl %edi, %ecx -; BURR-NEXT: addb $-128, %cl +; BURR-NEXT: subb %r10b, %cl +; BURR-NEXT: xorl %r8d, %r8d +; BURR-NEXT: movl $1, %esi +; BURR-NEXT: movl $1, %r9d ; BURR-NEXT: shrdq %cl, %r8, %r9 +; BURR-NEXT: testb $64, %cl +; BURR-NEXT: cmovneq %r8, %r9 ; BURR-NEXT: xorl %edx, %edx -; BURR-NEXT: shldq %cl, %rbp, %rdx -; BURR-NEXT: shrq %cl, %r8 -; BURR-NEXT: shlq %cl, %rbp +; BURR-NEXT: movl %r10d, %ecx +; BURR-NEXT: shldq %cl, %rsi, %rdx +; BURR-NEXT: addb $-128, %cl +; BURR-NEXT: xorl %r11d, %r11d +; BURR-NEXT: shldq %cl, %rsi, %r11 +; BURR-NEXT: movl $1, %edi +; BURR-NEXT: shlq %cl, %rdi ; BURR-NEXT: testb $64, %cl -; BURR-NEXT: cmovneq %rbp, %rdx -; BURR-NEXT: cmoveq %r9, %r8 -; BURR-NEXT: cmovneq %r10, %rbp -; BURR-NEXT: testb %dil, %dil -; BURR-NEXT: jns .LBB0_2 -; BURR-NEXT: # %bb.1: -; BURR-NEXT: movl %r8d, %r13d -; BURR-NEXT: .LBB0_2: -; BURR-NEXT: je .LBB0_4 -; BURR-NEXT: # %bb.3: -; BURR-NEXT: movl %r13d, %esi -; BURR-NEXT: .LBB0_4: -; BURR-NEXT: cmovsq %r10, %r15 -; BURR-NEXT: cmovnsq %r12, %rbp -; BURR-NEXT: cmoveq %r10, %rbp -; BURR-NEXT: cmovnsq %r14, %rdx -; BURR-NEXT: cmoveq %r10, %rdx -; BURR-NEXT: cmovsq %r10, %r11 -; BURR-NEXT: testb $1, %sil -; BURR-NEXT: cmovneq %rax, %rdx -; BURR-NEXT: movq %rdx, 24(%rax) -; BURR-NEXT: cmovneq %rax, %rbp -; BURR-NEXT: movq %rbp, 16(%rax) -; BURR-NEXT: cmovneq %rax, %r15 -; BURR-NEXT: movq %r15, 8(%rax) -; BURR-NEXT: cmovneq %rax, %r11 -; BURR-NEXT: movq %r11, (%rax) +; BURR-NEXT: cmovneq %rdi, %r11 +; BURR-NEXT: cmovneq %r8, %rdi +; BURR-NEXT: xorl %ebx, %ebx +; BURR-NEXT: movl %r10d, %ecx +; BURR-NEXT: shldq %cl, %rbx, %rbx +; BURR-NEXT: shlq %cl, %rsi +; BURR-NEXT: testb $64, %r10b +; BURR-NEXT: cmovneq %rsi, %rdx +; BURR-NEXT: cmovneq %r8, %rbx +; BURR-NEXT: cmovneq %r8, %rsi +; BURR-NEXT: testb %r10b, %r10b +; BURR-NEXT: cmovsq %r8, %rdx +; BURR-NEXT: movq %rdx, 8(%rax) +; BURR-NEXT: cmovsq %r8, %rsi +; BURR-NEXT: movq %rsi, (%rax) +; BURR-NEXT: cmovsq %r11, %rbx +; BURR-NEXT: cmoveq %r8, %rbx +; BURR-NEXT: movq %rbx, 24(%rax) +; BURR-NEXT: cmovnsq %r9, %rdi +; BURR-NEXT: cmoveq %r8, %rdi +; BURR-NEXT: movq %rdi, 16(%rax) ; BURR-NEXT: popq %rbx -; BURR-NEXT: popq %r12 -; BURR-NEXT: popq %r13 -; BURR-NEXT: popq %r14 -; BURR-NEXT: popq %r15 -; BURR-NEXT: popq %rbp ; BURR-NEXT: retq ; ; SRC-LABEL: test1: ; SRC: # %bb.0: -; SRC-NEXT: pushq %rbp -; SRC-NEXT: pushq %r15 ; SRC-NEXT: pushq %r14 -; SRC-NEXT: pushq %r13 -; SRC-NEXT: pushq %r12 ; SRC-NEXT: pushq %rbx -; SRC-NEXT: movq %rcx, %r9 ; SRC-NEXT: movq %rdi, %rax -; SRC-NEXT: addq $1, %rsi -; SRC-NEXT: adcq $0, %rdx -; SRC-NEXT: adcq $0, %r9 -; SRC-NEXT: adcq $0, %r8 -; SRC-NEXT: leal 1(%rsi,%rsi), %r11d -; SRC-NEXT: movb $-128, %r10b -; SRC-NEXT: subb %r11b, %r10b -; SRC-NEXT: movq %r9, %r12 -; SRC-NEXT: movl %r10d, %ecx -; SRC-NEXT: shlq %cl, %r12 -; SRC-NEXT: movq %rsi, %rbp -; SRC-NEXT: movl %r11d, %ecx -; SRC-NEXT: shrdq %cl, %rdx, %rbp -; SRC-NEXT: shrq %cl, %rdx -; SRC-NEXT: xorl %r15d, %r15d +; SRC-NEXT: leal 3(%rsi,%rsi), %r9d +; SRC-NEXT: movb $-128, %cl +; SRC-NEXT: subb %r9b, %cl +; SRC-NEXT: xorl %r8d, %r8d ; SRC-NEXT: movl $1, %edi +; SRC-NEXT: movl $1, %r10d +; SRC-NEXT: shrdq %cl, %r8, %r10 +; SRC-NEXT: testb $64, %cl +; SRC-NEXT: cmovneq %r8, %r10 +; SRC-NEXT: movl %r9d, %r11d +; SRC-NEXT: addb $-128, %r11b +; SRC-NEXT: xorl %esi, %esi +; SRC-NEXT: movl %r11d, %ecx +; SRC-NEXT: shldq %cl, %rdi, %rsi +; SRC-NEXT: xorl %edx, %edx +; SRC-NEXT: movl %r9d, %ecx +; SRC-NEXT: shldq %cl, %rdi, %rdx ; SRC-NEXT: xorl %r14d, %r14d -; SRC-NEXT: shldq %cl, %rdi, %r14 -; SRC-NEXT: xorl %r13d, %r13d -; SRC-NEXT: shldq %cl, %r13, %r13 +; SRC-NEXT: shldq %cl, %r14, %r14 ; SRC-NEXT: movl $1, %ebx ; SRC-NEXT: shlq %cl, %rbx -; SRC-NEXT: testb $64, %r11b -; SRC-NEXT: cmoveq %rbp, %rdx -; SRC-NEXT: cmovneq %rbx, %r14 -; SRC-NEXT: cmovneq %r15, %rbx -; SRC-NEXT: cmovneq %r15, %r13 -; SRC-NEXT: movl $1, %ebp -; SRC-NEXT: movl %r10d, %ecx -; SRC-NEXT: shrdq %cl, %r15, %rbp -; SRC-NEXT: testb $64, %r10b -; SRC-NEXT: cmovneq %r15, %r12 -; SRC-NEXT: cmovneq %r15, %rbp -; SRC-NEXT: orl %edx, %r12d +; SRC-NEXT: testb $64, %r9b +; SRC-NEXT: cmovneq %rbx, %rdx +; SRC-NEXT: cmovneq %r8, %rbx +; SRC-NEXT: cmovneq %r8, %r14 ; SRC-NEXT: movl %r11d, %ecx -; SRC-NEXT: addb $-128, %cl -; SRC-NEXT: shrdq %cl, %r8, %r9 -; SRC-NEXT: shrq %cl, %r8 -; SRC-NEXT: xorl %edx, %edx -; SRC-NEXT: shldq %cl, %rdi, %rdx ; SRC-NEXT: shlq %cl, %rdi -; SRC-NEXT: testb $64, %cl -; SRC-NEXT: cmoveq %r9, %r8 -; SRC-NEXT: cmovneq %rdi, %rdx -; SRC-NEXT: cmovneq %r15, %rdi -; SRC-NEXT: testb %r11b, %r11b -; SRC-NEXT: jns .LBB0_2 -; SRC-NEXT: # %bb.1: -; SRC-NEXT: movl %r8d, %r12d -; SRC-NEXT: .LBB0_2: -; SRC-NEXT: je .LBB0_4 -; SRC-NEXT: # %bb.3: -; SRC-NEXT: movl %r12d, %esi -; SRC-NEXT: .LBB0_4: -; SRC-NEXT: cmovnsq %r13, %rdx -; SRC-NEXT: cmoveq %r15, %rdx -; SRC-NEXT: cmovnsq %rbp, %rdi -; SRC-NEXT: cmoveq %r15, %rdi -; SRC-NEXT: cmovsq %r15, %r14 -; SRC-NEXT: cmovsq %r15, %rbx -; SRC-NEXT: testb $1, %sil -; SRC-NEXT: cmovneq %rax, %rbx -; SRC-NEXT: cmovneq %rax, %r14 -; SRC-NEXT: cmovneq %rax, %rdi -; SRC-NEXT: cmovneq %rax, %rdx -; SRC-NEXT: movq %rdx, 24(%rax) -; SRC-NEXT: movq %rdi, 16(%rax) -; SRC-NEXT: movq %r14, 8(%rax) +; SRC-NEXT: testb $64, %r11b +; SRC-NEXT: cmovneq %rdi, %rsi +; SRC-NEXT: cmovneq %r8, %rdi +; SRC-NEXT: testb %r9b, %r9b +; SRC-NEXT: cmovnsq %r10, %rdi +; SRC-NEXT: cmoveq %r8, %rdi +; SRC-NEXT: cmovnsq %r14, %rsi +; SRC-NEXT: cmoveq %r8, %rsi +; SRC-NEXT: cmovsq %r8, %rdx +; SRC-NEXT: cmovsq %r8, %rbx +; SRC-NEXT: movq %rdx, 8(%rax) ; SRC-NEXT: movq %rbx, (%rax) +; SRC-NEXT: movq %rsi, 24(%rax) +; SRC-NEXT: movq %rdi, 16(%rax) ; SRC-NEXT: popq %rbx -; SRC-NEXT: popq %r12 -; SRC-NEXT: popq %r13 ; SRC-NEXT: popq %r14 -; SRC-NEXT: popq %r15 -; SRC-NEXT: popq %rbp ; SRC-NEXT: retq ; ; LIN-LABEL: test1: ; LIN: # %bb.0: -; LIN-NEXT: pushq %rbp -; LIN-NEXT: pushq %r15 -; LIN-NEXT: pushq %r14 -; LIN-NEXT: pushq %r12 -; LIN-NEXT: pushq %rbx -; LIN-NEXT: movq %rcx, %r9 ; LIN-NEXT: movq %rdi, %rax -; LIN-NEXT: xorl %r15d, %r15d -; LIN-NEXT: movl $1, %r14d -; LIN-NEXT: addq $1, %rsi -; LIN-NEXT: leal 1(%rsi,%rsi), %ebp -; LIN-NEXT: movl $1, %r12d -; LIN-NEXT: movl %ebp, %ecx -; LIN-NEXT: shlq %cl, %r12 -; LIN-NEXT: testb $64, %bpl -; LIN-NEXT: movq %r12, %rbx -; LIN-NEXT: cmovneq %r15, %rbx -; LIN-NEXT: testb %bpl, %bpl -; LIN-NEXT: cmovsq %r15, %rbx -; LIN-NEXT: adcq $0, %rdx -; LIN-NEXT: adcq $0, %r9 -; LIN-NEXT: adcq $0, %r8 -; LIN-NEXT: movl %ebp, %r10d -; LIN-NEXT: addb $-128, %r10b -; LIN-NEXT: movq %r9, %rdi -; LIN-NEXT: movl %r10d, %ecx -; LIN-NEXT: shrdq %cl, %r8, %rdi -; LIN-NEXT: shrq %cl, %r8 -; LIN-NEXT: testb $64, %r10b -; LIN-NEXT: cmoveq %rdi, %r8 -; LIN-NEXT: movq %rsi, %rdi -; LIN-NEXT: movl %ebp, %ecx -; LIN-NEXT: shrdq %cl, %rdx, %rdi -; LIN-NEXT: shrq %cl, %rdx -; LIN-NEXT: cmoveq %rdi, %rdx -; LIN-NEXT: movb $-128, %r11b -; LIN-NEXT: subb %bpl, %r11b -; LIN-NEXT: movl %r11d, %ecx -; LIN-NEXT: shlq %cl, %r9 -; LIN-NEXT: testb $64, %r11b -; LIN-NEXT: cmovneq %r15, %r9 -; LIN-NEXT: orl %edx, %r9d -; LIN-NEXT: jns .LBB0_2 -; LIN-NEXT: # %bb.1: -; LIN-NEXT: movl %r8d, %r9d -; LIN-NEXT: .LBB0_2: -; LIN-NEXT: je .LBB0_4 -; LIN-NEXT: # %bb.3: -; LIN-NEXT: movl %r9d, %esi -; LIN-NEXT: .LBB0_4: -; LIN-NEXT: testb $1, %sil -; LIN-NEXT: cmovneq %rax, %rbx -; LIN-NEXT: movq %rbx, (%rax) -; LIN-NEXT: xorl %edx, %edx -; LIN-NEXT: movl %ebp, %ecx -; LIN-NEXT: shldq %cl, %r14, %rdx -; LIN-NEXT: cmovneq %r12, %rdx -; LIN-NEXT: cmovsq %r15, %rdx -; LIN-NEXT: cmovneq %rax, %rdx -; LIN-NEXT: movq %rdx, 8(%rax) +; LIN-NEXT: xorl %r9d, %r9d +; LIN-NEXT: movl $1, %r8d +; LIN-NEXT: leal 3(%rsi,%rsi), %r11d ; LIN-NEXT: movl $1, %edx -; LIN-NEXT: movl %r10d, %ecx +; LIN-NEXT: movl %r11d, %ecx ; LIN-NEXT: shlq %cl, %rdx -; LIN-NEXT: movq %rdx, %rsi -; LIN-NEXT: cmovneq %r15, %rsi -; LIN-NEXT: movl $1, %edi +; LIN-NEXT: testb $64, %r11b +; LIN-NEXT: movq %rdx, %rcx +; LIN-NEXT: cmovneq %r9, %rcx +; LIN-NEXT: testb %r11b, %r11b +; LIN-NEXT: cmovsq %r9, %rcx +; LIN-NEXT: movq %rcx, (%rdi) +; LIN-NEXT: xorl %edi, %edi ; LIN-NEXT: movl %r11d, %ecx -; LIN-NEXT: shrdq %cl, %r15, %rdi -; LIN-NEXT: cmovneq %r15, %rdi -; LIN-NEXT: cmovsq %rsi, %rdi -; LIN-NEXT: cmoveq %r15, %rdi -; LIN-NEXT: cmovneq %rax, %rdi -; LIN-NEXT: movq %rdi, 16(%rax) +; LIN-NEXT: shldq %cl, %r8, %rdi +; LIN-NEXT: cmovneq %rdx, %rdi +; LIN-NEXT: cmovsq %r9, %rdi +; LIN-NEXT: movq %rdi, 8(%rax) +; LIN-NEXT: movl %r11d, %edx +; LIN-NEXT: addb $-128, %dl +; LIN-NEXT: movl $1, %r10d +; LIN-NEXT: movl %edx, %ecx +; LIN-NEXT: shlq %cl, %r10 +; LIN-NEXT: testb $64, %dl +; LIN-NEXT: movq %r10, %rdi +; LIN-NEXT: cmovneq %r9, %rdi +; LIN-NEXT: movb $-128, %cl +; LIN-NEXT: subb %r11b, %cl +; LIN-NEXT: movl $1, %esi +; LIN-NEXT: shrdq %cl, %r9, %rsi +; LIN-NEXT: testb $64, %cl +; LIN-NEXT: cmovneq %r9, %rsi +; LIN-NEXT: cmovsq %rdi, %rsi +; LIN-NEXT: cmoveq %r9, %rsi +; LIN-NEXT: movq %rsi, 16(%rax) ; LIN-NEXT: xorl %esi, %esi -; LIN-NEXT: movl %r10d, %ecx -; LIN-NEXT: shldq %cl, %r14, %rsi -; LIN-NEXT: cmovneq %rdx, %rsi +; LIN-NEXT: movl %edx, %ecx +; LIN-NEXT: shldq %cl, %r8, %rsi +; LIN-NEXT: cmovneq %r10, %rsi ; LIN-NEXT: xorl %edx, %edx -; LIN-NEXT: movl %ebp, %ecx +; LIN-NEXT: movl %r11d, %ecx ; LIN-NEXT: shldq %cl, %rdx, %rdx -; LIN-NEXT: cmovneq %r15, %rdx +; LIN-NEXT: cmovneq %r9, %rdx ; LIN-NEXT: cmovsq %rsi, %rdx -; LIN-NEXT: cmoveq %r15, %rdx -; LIN-NEXT: cmovneq %rax, %rdx +; LIN-NEXT: cmoveq %r9, %rdx ; LIN-NEXT: movq %rdx, 24(%rax) -; LIN-NEXT: popq %rbx -; LIN-NEXT: popq %r12 -; LIN-NEXT: popq %r14 -; LIN-NEXT: popq %r15 -; LIN-NEXT: popq %rbp ; LIN-NEXT: retq %b = add i256 %a, 1 %m = shl i256 %b, 1