@@ -4162,7 +4162,9 @@ define <16 x float>@test_int_x86_avx512_mask_shuf_f32x4(<16 x float> %x0, <16 x
4162
4162
; CHECK: ## BB#0:
4163
4163
; CHECK-NEXT: kmovw %edi, %k1
4164
4164
; CHECK-NEXT: vshuff32x4 $22, %zmm1, %zmm0, %zmm2 {%k1}
4165
+ ; CHECK-NEXT: ## zmm2 = zmm0[8,9,10,11,4,5,6,7],zmm1[4,5,6,7,0,1,2,3]
4165
4166
; CHECK-NEXT: vshuff32x4 $22, %zmm1, %zmm0, %zmm0
4167
+ ; CHECK-NEXT: ## zmm0 = zmm0[8,9,10,11,4,5,6,7],zmm1[4,5,6,7,0,1,2,3]
4166
4168
; CHECK-NEXT: vaddps %zmm0, %zmm2, %zmm0
4167
4169
; CHECK-NEXT: retq
4168
4170
%res = call <16 x float > @llvm.x86.avx512.mask.shuf.f32x4 (<16 x float > %x0 , <16 x float > %x1 , i32 22 , <16 x float > %x3 , i16 %x4 )
@@ -4179,8 +4181,11 @@ define <8 x double>@test_int_x86_avx512_mask_shuf_f64x2(<8 x double> %x0, <8 x d
4179
4181
; CHECK-NEXT: movzbl %dil, %eax
4180
4182
; CHECK-NEXT: kmovw %eax, %k1
4181
4183
; CHECK-NEXT: vshuff64x2 $22, %zmm1, %zmm0, %zmm2 {%k1}
4184
+ ; CHECK-NEXT: ## zmm2 = zmm0[4,5,2,3],zmm1[2,3,0,1]
4182
4185
; CHECK-NEXT: vshuff64x2 $22, %zmm1, %zmm0, %zmm3 {%k1} {z}
4186
+ ; CHECK-NEXT: ## zmm3 = zmm0[4,5,2,3],zmm1[2,3,0,1]
4183
4187
; CHECK-NEXT: vshuff64x2 $22, %zmm1, %zmm0, %zmm0
4188
+ ; CHECK-NEXT: ## zmm0 = zmm0[4,5,2,3],zmm1[2,3,0,1]
4184
4189
; CHECK-NEXT: vaddpd %zmm0, %zmm2, %zmm0
4185
4190
; CHECK-NEXT: vaddpd %zmm3, %zmm0, %zmm0
4186
4191
; CHECK-NEXT: retq
@@ -4200,7 +4205,9 @@ define <16 x i32>@test_int_x86_avx512_mask_shuf_i32x4(<16 x i32> %x0, <16 x i32>
4200
4205
; CHECK: ## BB#0:
4201
4206
; CHECK-NEXT: kmovw %edi, %k1
4202
4207
; CHECK-NEXT: vshufi32x4 $22, %zmm1, %zmm0, %zmm2 {%k1}
4208
+ ; CHECK-NEXT: ## zmm2 = zmm0[8,9,10,11,4,5,6,7],zmm1[4,5,6,7,0,1,2,3]
4203
4209
; CHECK-NEXT: vshufi32x4 $22, %zmm1, %zmm0, %zmm0
4210
+ ; CHECK-NEXT: ## zmm0 = zmm0[8,9,10,11,4,5,6,7],zmm1[4,5,6,7,0,1,2,3]
4204
4211
; CHECK-NEXT: vpaddd %zmm0, %zmm2, %zmm0
4205
4212
; CHECK-NEXT: retq
4206
4213
%res = call <16 x i32 > @llvm.x86.avx512.mask.shuf.i32x4 (<16 x i32 > %x0 , <16 x i32 > %x1 , i32 22 , <16 x i32 > %x3 , i16 %x4 )
@@ -4217,7 +4224,9 @@ define <8 x i64>@test_int_x86_avx512_mask_shuf_i64x2(<8 x i64> %x0, <8 x i64> %x
4217
4224
; CHECK-NEXT: movzbl %dil, %eax
4218
4225
; CHECK-NEXT: kmovw %eax, %k1
4219
4226
; CHECK-NEXT: vshufi64x2 $22, %zmm1, %zmm0, %zmm2 {%k1}
4227
+ ; CHECK-NEXT: ## zmm2 = zmm0[4,5,2,3],zmm1[2,3,0,1]
4220
4228
; CHECK-NEXT: vshufi64x2 $22, %zmm1, %zmm0, %zmm0
4229
+ ; CHECK-NEXT: ## zmm0 = zmm0[4,5,2,3],zmm1[2,3,0,1]
4221
4230
; CHECK-NEXT: vpaddq %zmm0, %zmm2, %zmm0
4222
4231
; CHECK-NEXT: retq
4223
4232
%res = call <8 x i64 > @llvm.x86.avx512.mask.shuf.i64x2 (<8 x i64 > %x0 , <8 x i64 > %x1 , i32 22 , <8 x i64 > %x3 , i8 %x4 )
0 commit comments