diff --git a/llvm/lib/Target/X86/X86ISelLowering.cpp b/llvm/lib/Target/X86/X86ISelLowering.cpp --- a/llvm/lib/Target/X86/X86ISelLowering.cpp +++ b/llvm/lib/Target/X86/X86ISelLowering.cpp @@ -33552,7 +33552,9 @@ // changing, all we need is a lowering for the *ordering* impacts of the // atomicrmw. As such, we can chose a different operation and memory // location to minimize impact on other code. - if (Opc == ISD::ATOMIC_LOAD_OR && isNullConstant(RHS)) { + // The above holds unless the node is marked volatile in which + // case it needs to be preserved according to the langref. + if (Opc == ISD::ATOMIC_LOAD_OR && isNullConstant(RHS) && !AN->isVolatile()) { // On X86, the only ordering which actually requires an instruction is // seq_cst which isn't SingleThread, everything just needs to be preserved // during codegen and then dropped. Note that we expect (but don't assume), diff --git a/llvm/test/CodeGen/X86/pr63692.ll b/llvm/test/CodeGen/X86/pr63692.ll --- a/llvm/test/CodeGen/X86/pr63692.ll +++ b/llvm/test/CodeGen/X86/pr63692.ll @@ -9,7 +9,7 @@ ; CHECK-NEXT: .p2align 4, 0x90 ; CHECK-NEXT: .LBB0_1: # %while.body ; CHECK-NEXT: # =>This Inner Loop Header: Depth=1 -; CHECK-NEXT: #MEMBARRIER +; CHECK-NEXT: lock orb $0, (%rdi) ; CHECK-NEXT: addq $4096, %rdi # imm = 0x1000 ; CHECK-NEXT: cmpq %rsi, %rdi ; CHECK-NEXT: jb .LBB0_1