Index: lib/Target/X86/InstPrinter/X86ATTInstPrinter.cpp =================================================================== --- lib/Target/X86/InstPrinter/X86ATTInstPrinter.cpp +++ lib/Target/X86/InstPrinter/X86ATTInstPrinter.cpp @@ -172,9 +172,15 @@ // If there are no instruction-specific comments, add a comment clarifying // the hex value of the immediate operand when it isn't in the range // [-256,255]. - if (CommentStream && !HasCustomInstComment && (Imm > 255 || Imm < -256)) - *CommentStream << format("imm = 0x%" PRIX64 "\n", (uint64_t)Imm); - + if (CommentStream && !HasCustomInstComment && (Imm > 255 || Imm < -256)) { + // Don't print unnecessary hex sign bits. + if (Imm == (int16_t)(Imm)) + *CommentStream << format("imm = 0x%" PRIX16 "\n", (uint16_t)Imm); + else if (Imm == (int32_t)(Imm)) + *CommentStream << format("imm = 0x%" PRIX32 "\n", (uint32_t)Imm); + else + *CommentStream << format("imm = 0x%" PRIX64 "\n", (uint64_t)Imm); + } } else { assert(Op.isExpr() && "unknown operand kind in printOperand"); O << markup(" * nocapture %aFOO, <8 x float>* nocapture %RET) nounwind { ; CHECK-LABEL: mp_11193: ; CHECK: # BB#0: # %allocas -; CHECK-NEXT: movl $-1082130432, (%rsi) # imm = 0xFFFFFFFFBF800000 +; CHECK-NEXT: movl $-1082130432, (%rsi) # imm = 0xBF800000 ; CHECK-NEXT: retq allocas: %bincmp = fcmp olt <8 x float> , Index: test/CodeGen/X86/avx512-cmp.ll =================================================================== --- test/CodeGen/X86/avx512-cmp.ll +++ test/CodeGen/X86/avx512-cmp.ll @@ -122,7 +122,7 @@ ; ALL-NEXT: testl %edx, %edx ; ALL-NEXT: movl $1, %eax ; ALL-NEXT: cmovel %eax, %edx -; ALL-NEXT: cmpl $-2147483648, %esi ## imm = 0xFFFFFFFF80000000 +; ALL-NEXT: cmpl $-2147483648, %esi ## imm = 0x80000000 ; ALL-NEXT: cmovnel %edx, %eax ; ALL-NEXT: cmpl $-1, %edi ; ALL-NEXT: cmovnel %edx, %eax Index: test/CodeGen/X86/bitreverse.ll =================================================================== --- test/CodeGen/X86/bitreverse.ll +++ test/CodeGen/X86/bitreverse.ll @@ -26,7 +26,7 @@ define <2 x i16> @fold_v2i16() { ; CHECK-LABEL: fold_v2i16: ; CHECK: # BB#0: -; CHECK-NEXT: movw $-4096, %ax # imm = 0xFFFFFFFFFFFFF000 +; CHECK-NEXT: movw $-4096, %ax # imm = 0xF000 ; CHECK-NEXT: movw $240, %dx ; CHECK-NEXT: retl %b = call <2 x i16> @llvm.bitreverse.v2i16(<2 x i16> ) Index: test/CodeGen/X86/pr16360.ll =================================================================== --- test/CodeGen/X86/pr16360.ll +++ test/CodeGen/X86/pr16360.ll @@ -6,7 +6,7 @@ ; CHECK: # BB#0: # %entry ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax ; CHECK-NEXT: shrl $2, %eax -; CHECK-NEXT: orl $-67108864, %eax # imm = 0xFFFFFFFFFC000000 +; CHECK-NEXT: orl $-67108864, %eax # imm = 0xFC000000 ; CHECK-NEXT: movl $1073741823, %edx # imm = 0x3FFFFFFF ; CHECK-NEXT: retl entry: Index: test/CodeGen/X86/rem.ll =================================================================== --- test/CodeGen/X86/rem.ll +++ test/CodeGen/X86/rem.ll @@ -5,7 +5,7 @@ ; CHECK-LABEL: test1: ; CHECK: # BB#0: ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx -; CHECK-NEXT: movl $-2139062143, %edx # imm = 0xFFFFFFFF80808081 +; CHECK-NEXT: movl $-2139062143, %edx # imm = 0x80808081 ; CHECK-NEXT: movl %ecx, %eax ; CHECK-NEXT: imull %edx ; CHECK-NEXT: addl %ecx, %edx @@ -44,7 +44,7 @@ ; CHECK-LABEL: test3: ; CHECK: # BB#0: ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx -; CHECK-NEXT: movl $-2139062143, %edx # imm = 0xFFFFFFFF80808081 +; CHECK-NEXT: movl $-2139062143, %edx # imm = 0x80808081 ; CHECK-NEXT: movl %ecx, %eax ; CHECK-NEXT: mull %edx ; CHECK-NEXT: shrl $7, %edx Index: test/CodeGen/X86/vec_fneg.ll =================================================================== --- test/CodeGen/X86/vec_fneg.ll +++ test/CodeGen/X86/vec_fneg.ll @@ -58,7 +58,7 @@ ; X32-SSE1-NEXT: movl %esp, %ebp ; X32-SSE1-NEXT: andl $-16, %esp ; X32-SSE1-NEXT: subl $32, %esp -; X32-SSE1-NEXT: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000 +; X32-SSE1-NEXT: movl $-2147483648, %eax # imm = 0x80000000 ; X32-SSE1-NEXT: movl 12(%ebp), %ecx ; X32-SSE1-NEXT: xorl %eax, %ecx ; X32-SSE1-NEXT: movl %ecx, {{[0-9]+}}(%esp) @@ -71,7 +71,7 @@ ; ; X32-SSE2-LABEL: fneg_bitcast: ; X32-SSE2: # BB#0: -; X32-SSE2-NEXT: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000 +; X32-SSE2-NEXT: movl $-2147483648, %eax # imm = 0x80000000 ; X32-SSE2-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X32-SSE2-NEXT: xorl %eax, %ecx ; X32-SSE2-NEXT: xorl {{[0-9]+}}(%esp), %eax Index: test/CodeGen/X86/vector-bitreverse.ll =================================================================== --- test/CodeGen/X86/vector-bitreverse.ll +++ test/CodeGen/X86/vector-bitreverse.ll @@ -615,7 +615,7 @@ ; SSE-NEXT: orq %rcx, %rdx ; SSE-NEXT: movq %rdi, %rcx ; SSE-NEXT: shrq %rcx -; SSE-NEXT: andl $-2147483648, %ecx # imm = 0xFFFFFFFF80000000 +; SSE-NEXT: andl $-2147483648, %ecx # imm = 0x80000000 ; SSE-NEXT: orq %rdx, %rcx ; SSE-NEXT: movq %rdi, %rdx ; SSE-NEXT: shrq $3, %rdx @@ -871,7 +871,7 @@ ; AVX-NEXT: orq %rcx, %rdx ; AVX-NEXT: movq %rdi, %rcx ; AVX-NEXT: shrq %rcx -; AVX-NEXT: andl $-2147483648, %ecx # imm = 0xFFFFFFFF80000000 +; AVX-NEXT: andl $-2147483648, %ecx # imm = 0x80000000 ; AVX-NEXT: orq %rdx, %rcx ; AVX-NEXT: movq %rdi, %rdx ; AVX-NEXT: shrq $3, %rdx Index: test/CodeGen/X86/vector-idiv-sdiv-512.ll =================================================================== --- test/CodeGen/X86/vector-idiv-sdiv-512.ll +++ test/CodeGen/X86/vector-idiv-sdiv-512.ll @@ -87,7 +87,7 @@ ; AVX-NEXT: vextracti32x4 $3, %zmm0, %xmm1 ; AVX-NEXT: vpextrd $1, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -96,7 +96,7 @@ ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: vmovd %xmm1, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %edx, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -107,7 +107,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm2, %xmm2 ; AVX-NEXT: vpextrd $2, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -117,7 +117,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm2, %xmm2 ; AVX-NEXT: vpextrd $3, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -128,7 +128,7 @@ ; AVX-NEXT: vextracti32x4 $2, %zmm0, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -137,7 +137,7 @@ ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: vmovd %xmm2, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %edx, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -148,7 +148,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -158,7 +158,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -170,7 +170,7 @@ ; AVX-NEXT: vextracti32x4 $1, %zmm0, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -179,7 +179,7 @@ ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: vmovd %xmm2, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %edx, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -190,7 +190,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -200,7 +200,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -210,7 +210,7 @@ ; AVX-NEXT: vpinsrd $3, %eax, %xmm3, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -219,7 +219,7 @@ ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: vmovd %xmm0, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %edx, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -230,7 +230,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -240,7 +240,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %ecx, %eax ; AVX-NEXT: movl %eax, %ecx @@ -1162,7 +1162,7 @@ ; AVX-NEXT: vextracti32x4 $3, %zmm0, %xmm1 ; AVX-NEXT: vpextrd $1, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1174,7 +1174,7 @@ ; AVX-NEXT: subl %edx, %eax ; AVX-NEXT: vmovd %xmm1, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %ecx, %edx ; AVX-NEXT: movl %edx, %esi @@ -1188,7 +1188,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm2, %xmm2 ; AVX-NEXT: vpextrd $2, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1201,7 +1201,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm2, %xmm2 ; AVX-NEXT: vpextrd $3, %xmm1, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1215,7 +1215,7 @@ ; AVX-NEXT: vextracti32x4 $2, %zmm0, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1227,7 +1227,7 @@ ; AVX-NEXT: subl %edx, %eax ; AVX-NEXT: vmovd %xmm2, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %ecx, %edx ; AVX-NEXT: movl %edx, %esi @@ -1241,7 +1241,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1254,7 +1254,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1269,7 +1269,7 @@ ; AVX-NEXT: vextracti32x4 $1, %zmm0, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1281,7 +1281,7 @@ ; AVX-NEXT: subl %edx, %eax ; AVX-NEXT: vmovd %xmm2, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %ecx, %edx ; AVX-NEXT: movl %edx, %esi @@ -1295,7 +1295,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1308,7 +1308,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm2, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1321,7 +1321,7 @@ ; AVX-NEXT: vpinsrd $3, %eax, %xmm3, %xmm2 ; AVX-NEXT: vpextrd $1, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1333,7 +1333,7 @@ ; AVX-NEXT: subl %edx, %eax ; AVX-NEXT: vmovd %xmm0, %ecx ; AVX-NEXT: movslq %ecx, %rcx -; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rdx ; AVX-NEXT: addl %ecx, %edx ; AVX-NEXT: movl %edx, %esi @@ -1347,7 +1347,7 @@ ; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $2, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx @@ -1360,7 +1360,7 @@ ; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3 ; AVX-NEXT: vpextrd $3, %xmm0, %eax ; AVX-NEXT: cltq -; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493 +; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493 ; AVX-NEXT: shrq $32, %rcx ; AVX-NEXT: addl %eax, %ecx ; AVX-NEXT: movl %ecx, %edx Index: test/MC/X86/imm-comments.s =================================================================== --- test/MC/X86/imm-comments.s +++ test/MC/X86/imm-comments.s @@ -16,10 +16,10 @@ # CHECK: movb $-128, %al # CHECK: movw $32767, %ax # imm = 0x7FFF -# CHECK: movw $-32768, %ax # imm = 0xFFFFFFFFFFFF8000 +# CHECK: movw $-32768, %ax # imm = 0x8000 # CHECK: movl $2147483647, %eax # imm = 0x7FFFFFFF -# CHECK: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000 +# CHECK: movl $-2147483648, %eax # imm = 0x80000000 # CHECK: movabsq $9223372036854775807, %rax # imm = 0x7FFFFFFFFFFFFFFF # CHECK: movabsq $-9223372036854775808, %rax # imm = 0x8000000000000000