@@ -29,7 +29,7 @@ define <2 x double> @test_mm_mask_movddup_pd(<2 x double> %a0, i8 %a1, <2 x doub
29
29
; X32-NEXT: movb %al, {{[0-9]+}}(%esp)
30
30
; X32-NEXT: movzbl {{[0-9]+}}(%esp), %eax
31
31
; X32-NEXT: kmovw %eax, %k1
32
- ; X32-NEXT: vmovddup {{.*#+}} xmm0 = xmm1[0,0]
32
+ ; X32-NEXT: vmovddup {{.*#+}} xmm0 {%k1} = xmm1[0,0]
33
33
; X32-NEXT: popl %eax
34
34
; X32-NEXT: retl
35
35
;
@@ -39,7 +39,7 @@ define <2 x double> @test_mm_mask_movddup_pd(<2 x double> %a0, i8 %a1, <2 x doub
39
39
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
40
40
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
41
41
; X64-NEXT: kmovw %eax, %k1
42
- ; X64-NEXT: vmovddup {{.*#+}} xmm0 = xmm1[0,0]
42
+ ; X64-NEXT: vmovddup {{.*#+}} xmm0 {%k1} = xmm1[0,0]
43
43
; X64-NEXT: retq
44
44
%trn1 = trunc i8 %a1 to i2
45
45
%arg1 = bitcast i2 %trn1 to <2 x i1 >
@@ -59,7 +59,7 @@ define <2 x double> @test_mm_maskz_movddup_pd(i8 %a0, <2 x double> %a1) {
59
59
; X32-NEXT: movb %al, {{[0-9]+}}(%esp)
60
60
; X32-NEXT: movzbl {{[0-9]+}}(%esp), %eax
61
61
; X32-NEXT: kmovw %eax, %k1
62
- ; X32-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
62
+ ; X32-NEXT: vmovddup {{.*#+}} xmm0 {%k1} {z} = xmm0[0,0]
63
63
; X32-NEXT: popl %eax
64
64
; X32-NEXT: retl
65
65
;
@@ -69,7 +69,7 @@ define <2 x double> @test_mm_maskz_movddup_pd(i8 %a0, <2 x double> %a1) {
69
69
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
70
70
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
71
71
; X64-NEXT: kmovw %eax, %k1
72
- ; X64-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
72
+ ; X64-NEXT: vmovddup {{.*#+}} xmm0 {%k1} {z} = xmm0[0,0]
73
73
; X64-NEXT: retq
74
74
%trn1 = trunc i8 %a0 to i2
75
75
%arg0 = bitcast i2 %trn1 to <2 x i1 >
@@ -103,7 +103,7 @@ define <4 x double> @test_mm256_mask_movddup_pd(<4 x double> %a0, i8 %a1, <4 x d
103
103
; X32-NEXT: movb %al, (%esp)
104
104
; X32-NEXT: movzbl (%esp), %eax
105
105
; X32-NEXT: kmovw %eax, %k1
106
- ; X32-NEXT: vmovddup {{.*#+}} ymm0 = ymm1[0,0,2,2]
106
+ ; X32-NEXT: vmovddup {{.*#+}} ymm0 {%k1} = ymm1[0,0,2,2]
107
107
; X32-NEXT: popl %eax
108
108
; X32-NEXT: retl
109
109
;
@@ -113,7 +113,7 @@ define <4 x double> @test_mm256_mask_movddup_pd(<4 x double> %a0, i8 %a1, <4 x d
113
113
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
114
114
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
115
115
; X64-NEXT: kmovw %eax, %k1
116
- ; X64-NEXT: vmovddup {{.*#+}} ymm0 = ymm1[0,0,2,2]
116
+ ; X64-NEXT: vmovddup {{.*#+}} ymm0 {%k1} = ymm1[0,0,2,2]
117
117
; X64-NEXT: retq
118
118
%trn1 = trunc i8 %a1 to i4
119
119
%arg1 = bitcast i4 %trn1 to <4 x i1 >
@@ -133,7 +133,7 @@ define <4 x double> @test_mm256_maskz_movddup_pd(i8 %a0, <4 x double> %a1) {
133
133
; X32-NEXT: movb %al, (%esp)
134
134
; X32-NEXT: movzbl (%esp), %eax
135
135
; X32-NEXT: kmovw %eax, %k1
136
- ; X32-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
136
+ ; X32-NEXT: vmovddup {{.*#+}} ymm0 {%k1} {z} = ymm0[0,0,2,2]
137
137
; X32-NEXT: popl %eax
138
138
; X32-NEXT: retl
139
139
;
@@ -143,7 +143,7 @@ define <4 x double> @test_mm256_maskz_movddup_pd(i8 %a0, <4 x double> %a1) {
143
143
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
144
144
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
145
145
; X64-NEXT: kmovw %eax, %k1
146
- ; X64-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
146
+ ; X64-NEXT: vmovddup {{.*#+}} ymm0 {%k1} {z} = ymm0[0,0,2,2]
147
147
; X64-NEXT: retq
148
148
%trn1 = trunc i8 %a0 to i4
149
149
%arg0 = bitcast i4 %trn1 to <4 x i1 >
@@ -177,7 +177,7 @@ define <4 x float> @test_mm_mask_movehdup_ps(<4 x float> %a0, i8 %a1, <4 x float
177
177
; X32-NEXT: movb %al, (%esp)
178
178
; X32-NEXT: movzbl (%esp), %eax
179
179
; X32-NEXT: kmovw %eax, %k1
180
- ; X32-NEXT: vmovshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
180
+ ; X32-NEXT: vmovshdup {{.*#+}} xmm0 {%k1} = xmm1[1,1,3,3]
181
181
; X32-NEXT: popl %eax
182
182
; X32-NEXT: retl
183
183
;
@@ -187,7 +187,7 @@ define <4 x float> @test_mm_mask_movehdup_ps(<4 x float> %a0, i8 %a1, <4 x float
187
187
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
188
188
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
189
189
; X64-NEXT: kmovw %eax, %k1
190
- ; X64-NEXT: vmovshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
190
+ ; X64-NEXT: vmovshdup {{.*#+}} xmm0 {%k1} = xmm1[1,1,3,3]
191
191
; X64-NEXT: retq
192
192
%trn1 = trunc i8 %a1 to i4
193
193
%arg1 = bitcast i4 %trn1 to <4 x i1 >
@@ -207,7 +207,7 @@ define <4 x float> @test_mm_maskz_movehdup_ps(i8 %a0, <4 x float> %a1) {
207
207
; X32-NEXT: movb %al, (%esp)
208
208
; X32-NEXT: movzbl (%esp), %eax
209
209
; X32-NEXT: kmovw %eax, %k1
210
- ; X32-NEXT: vmovshdup {{.*#+}} xmm0 = xmm0[1,1,3,3]
210
+ ; X32-NEXT: vmovshdup {{.*#+}} xmm0 {%k1} {z} = xmm0[1,1,3,3]
211
211
; X32-NEXT: popl %eax
212
212
; X32-NEXT: retl
213
213
;
@@ -217,7 +217,7 @@ define <4 x float> @test_mm_maskz_movehdup_ps(i8 %a0, <4 x float> %a1) {
217
217
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
218
218
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
219
219
; X64-NEXT: kmovw %eax, %k1
220
- ; X64-NEXT: vmovshdup {{.*#+}} xmm0 = xmm0[1,1,3,3]
220
+ ; X64-NEXT: vmovshdup {{.*#+}} xmm0 {%k1} {z} = xmm0[1,1,3,3]
221
221
; X64-NEXT: retq
222
222
%trn0 = trunc i8 %a0 to i4
223
223
%arg0 = bitcast i4 %trn0 to <4 x i1 >
@@ -245,13 +245,13 @@ define <8 x float> @test_mm256_mask_movehdup_ps(<8 x float> %a0, i8 %a1, <8 x fl
245
245
; X32: # BB#0:
246
246
; X32-NEXT: movb {{[0-9]+}}(%esp), %al
247
247
; X32-NEXT: kmovw %eax, %k1
248
- ; X32-NEXT: vmovshdup {{.*#+}} ymm0 = ymm1[1,1,3,3,5,5,7,7]
248
+ ; X32-NEXT: vmovshdup {{.*#+}} ymm0 {%k1} = ymm1[1,1,3,3,5,5,7,7]
249
249
; X32-NEXT: retl
250
250
;
251
251
; X64-LABEL: test_mm256_mask_movehdup_ps:
252
252
; X64: # BB#0:
253
253
; X64-NEXT: kmovw %edi, %k1
254
- ; X64-NEXT: vmovshdup {{.*#+}} ymm0 = ymm1[1,1,3,3,5,5,7,7]
254
+ ; X64-NEXT: vmovshdup {{.*#+}} ymm0 {%k1} = ymm1[1,1,3,3,5,5,7,7]
255
255
; X64-NEXT: retq
256
256
%arg1 = bitcast i8 %a1 to <8 x i1 >
257
257
%res0 = shufflevector <8 x float > %a2 , <8 x float > undef , <8 x i32 > <i32 1 , i32 1 , i32 3 , i32 3 , i32 5 , i32 5 , i32 7 , i32 7 >
@@ -264,13 +264,13 @@ define <8 x float> @test_mm256_maskz_movehdup_ps(i8 %a0, <8 x float> %a1) {
264
264
; X32: # BB#0:
265
265
; X32-NEXT: movb {{[0-9]+}}(%esp), %al
266
266
; X32-NEXT: kmovw %eax, %k1
267
- ; X32-NEXT: vmovshdup {{.*#+}} ymm0 = ymm0[1,1,3,3,5,5,7,7]
267
+ ; X32-NEXT: vmovshdup {{.*#+}} ymm0 {%k1} {z} = ymm0[1,1,3,3,5,5,7,7]
268
268
; X32-NEXT: retl
269
269
;
270
270
; X64-LABEL: test_mm256_maskz_movehdup_ps:
271
271
; X64: # BB#0:
272
272
; X64-NEXT: kmovw %edi, %k1
273
- ; X64-NEXT: vmovshdup {{.*#+}} ymm0 = ymm0[1,1,3,3,5,5,7,7]
273
+ ; X64-NEXT: vmovshdup {{.*#+}} ymm0 {%k1} {z} = ymm0[1,1,3,3,5,5,7,7]
274
274
; X64-NEXT: retq
275
275
%arg0 = bitcast i8 %a0 to <8 x i1 >
276
276
%res0 = shufflevector <8 x float > %a1 , <8 x float > undef , <8 x i32 > <i32 1 , i32 1 , i32 3 , i32 3 , i32 5 , i32 5 , i32 7 , i32 7 >
@@ -303,7 +303,7 @@ define <4 x float> @test_mm_mask_moveldup_ps(<4 x float> %a0, i8 %a1, <4 x float
303
303
; X32-NEXT: movb %al, (%esp)
304
304
; X32-NEXT: movzbl (%esp), %eax
305
305
; X32-NEXT: kmovw %eax, %k1
306
- ; X32-NEXT: vmovsldup {{.*#+}} xmm0 = xmm1[0,0,2,2]
306
+ ; X32-NEXT: vmovsldup {{.*#+}} xmm0 {%k1} = xmm1[0,0,2,2]
307
307
; X32-NEXT: popl %eax
308
308
; X32-NEXT: retl
309
309
;
@@ -313,7 +313,7 @@ define <4 x float> @test_mm_mask_moveldup_ps(<4 x float> %a0, i8 %a1, <4 x float
313
313
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
314
314
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
315
315
; X64-NEXT: kmovw %eax, %k1
316
- ; X64-NEXT: vmovsldup {{.*#+}} xmm0 = xmm1[0,0,2,2]
316
+ ; X64-NEXT: vmovsldup {{.*#+}} xmm0 {%k1} = xmm1[0,0,2,2]
317
317
; X64-NEXT: retq
318
318
%trn1 = trunc i8 %a1 to i4
319
319
%arg1 = bitcast i4 %trn1 to <4 x i1 >
@@ -333,7 +333,7 @@ define <4 x float> @test_mm_maskz_moveldup_ps(i8 %a0, <4 x float> %a1) {
333
333
; X32-NEXT: movb %al, (%esp)
334
334
; X32-NEXT: movzbl (%esp), %eax
335
335
; X32-NEXT: kmovw %eax, %k1
336
- ; X32-NEXT: vmovsldup {{.*#+}} xmm0 = xmm0[0,0,2,2]
336
+ ; X32-NEXT: vmovsldup {{.*#+}} xmm0 {%k1} {z} = xmm0[0,0,2,2]
337
337
; X32-NEXT: popl %eax
338
338
; X32-NEXT: retl
339
339
;
@@ -343,7 +343,7 @@ define <4 x float> @test_mm_maskz_moveldup_ps(i8 %a0, <4 x float> %a1) {
343
343
; X64-NEXT: movb %dil, -{{[0-9]+}}(%rsp)
344
344
; X64-NEXT: movzbl -{{[0-9]+}}(%rsp), %eax
345
345
; X64-NEXT: kmovw %eax, %k1
346
- ; X64-NEXT: vmovsldup {{.*#+}} xmm0 = xmm0[0,0,2,2]
346
+ ; X64-NEXT: vmovsldup {{.*#+}} xmm0 {%k1} {z} = xmm0[0,0,2,2]
347
347
; X64-NEXT: retq
348
348
%trn0 = trunc i8 %a0 to i4
349
349
%arg0 = bitcast i4 %trn0 to <4 x i1 >
@@ -371,13 +371,13 @@ define <8 x float> @test_mm256_mask_moveldup_ps(<8 x float> %a0, i8 %a1, <8 x fl
371
371
; X32: # BB#0:
372
372
; X32-NEXT: movb {{[0-9]+}}(%esp), %al
373
373
; X32-NEXT: kmovw %eax, %k1
374
- ; X32-NEXT: vmovsldup {{.*#+}} ymm0 = ymm1[0,0,2,2,4,4,6,6]
374
+ ; X32-NEXT: vmovsldup {{.*#+}} ymm0 {%k1} = ymm1[0,0,2,2,4,4,6,6]
375
375
; X32-NEXT: retl
376
376
;
377
377
; X64-LABEL: test_mm256_mask_moveldup_ps:
378
378
; X64: # BB#0:
379
379
; X64-NEXT: kmovw %edi, %k1
380
- ; X64-NEXT: vmovsldup {{.*#+}} ymm0 = ymm1[0,0,2,2,4,4,6,6]
380
+ ; X64-NEXT: vmovsldup {{.*#+}} ymm0 {%k1} = ymm1[0,0,2,2,4,4,6,6]
381
381
; X64-NEXT: retq
382
382
%arg1 = bitcast i8 %a1 to <8 x i1 >
383
383
%res0 = shufflevector <8 x float > %a2 , <8 x float > undef , <8 x i32 > <i32 0 , i32 0 , i32 2 , i32 2 , i32 4 , i32 4 , i32 6 , i32 6 >
@@ -390,13 +390,13 @@ define <8 x float> @test_mm256_maskz_moveldup_ps(i8 %a0, <8 x float> %a1) {
390
390
; X32: # BB#0:
391
391
; X32-NEXT: movb {{[0-9]+}}(%esp), %al
392
392
; X32-NEXT: kmovw %eax, %k1
393
- ; X32-NEXT: vmovsldup {{.*#+}} ymm0 = ymm0[0,0,2,2,4,4,6,6]
393
+ ; X32-NEXT: vmovsldup {{.*#+}} ymm0 {%k1} {z} = ymm0[0,0,2,2,4,4,6,6]
394
394
; X32-NEXT: retl
395
395
;
396
396
; X64-LABEL: test_mm256_maskz_moveldup_ps:
397
397
; X64: # BB#0:
398
398
; X64-NEXT: kmovw %edi, %k1
399
- ; X64-NEXT: vmovsldup {{.*#+}} ymm0 = ymm0[0,0,2,2,4,4,6,6]
399
+ ; X64-NEXT: vmovsldup {{.*#+}} ymm0 {%k1} {z} = ymm0[0,0,2,2,4,4,6,6]
400
400
; X64-NEXT: retq
401
401
%arg0 = bitcast i8 %a0 to <8 x i1 >
402
402
%res0 = shufflevector <8 x float > %a1 , <8 x float > undef , <8 x i32 > <i32 0 , i32 0 , i32 2 , i32 2 , i32 4 , i32 4 , i32 6 , i32 6 >
0 commit comments