diff --git a/llvm/test/CodeGen/X86/fma-signed-zero.ll b/llvm/test/CodeGen/X86/fma-signed-zero.ll
new file mode 100644
--- /dev/null
+++ b/llvm/test/CodeGen/X86/fma-signed-zero.ll
@@ -0,0 +1,57 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=+fma | FileCheck %s
+
+; This test checks that (fneg (fma (fneg x), y, (fneg z))) can't be folded to (fma x, y, z)
+; without no signed zeros flag (nsz).
+
+declare float @llvm.fma.f32(float, float, float)
+
+define float @fneg_fma32(float %x, float %y, float %z) {
+; CHECK-LABEL: fneg_fma32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vfmadd213ss {{.*#+}} xmm0 = (xmm1 * xmm0) + xmm2
+; CHECK-NEXT:    retq
+  %negx = fneg float %x
+  %negz = fneg float %z
+  %fma = call float @llvm.fma.f32(float %negx, float %y, float %negz)
+  %n = fneg float %fma
+  ret float %n
+}
+
+define float @fneg_fma32_nsz(float %x, float %y, float %z) {
+; CHECK-LABEL: fneg_fma32_nsz:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vfmadd213ss {{.*#+}} xmm0 = (xmm1 * xmm0) + xmm2
+; CHECK-NEXT:    retq
+  %negx = fneg float %x
+  %negz = fneg float %z
+  %fma = call nsz float @llvm.fma.f32(float %negx, float %y, float %negz)
+  %n = fneg float %fma
+  ret float %n
+}
+
+declare double @llvm.fma.f64(double, double, double)
+
+define double @fneg_fma64(double %x, double %y, double %z) {
+; CHECK-LABEL: fneg_fma64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vfmadd213sd {{.*#+}} xmm0 = (xmm1 * xmm0) + xmm2
+; CHECK-NEXT:    retq
+  %negx = fneg double %x
+  %negz = fneg double %z
+  %fma = call double @llvm.fma.f64(double %negx, double %y, double %negz)
+  %n = fneg double %fma
+  ret double %n
+}
+
+define double @fneg_fma64_nsz(double %x, double %y, double %z) {
+; CHECK-LABEL: fneg_fma64_nsz:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vfmadd213sd {{.*#+}} xmm0 = (xmm1 * xmm0) + xmm2
+; CHECK-NEXT:    retq
+  %negx = fneg double %x
+  %negz = fneg double %z
+  %fma = call nsz double @llvm.fma.f64(double %negx, double %y, double %negz)
+  %n = fneg double %fma
+  ret double %n
+}