Changeset View
Changeset View
Standalone View
Standalone View
llvm/trunk/test/CodeGen/X86/vec_umulo.ll
Show First 20 Lines • Show All 837 Lines • ▼ Show 20 Lines | |||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm5 = xmm0[1,1,3,3] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm5 = xmm0[1,1,3,3] | ||||
; AVX1-NEXT: vpmuludq %xmm7, %xmm5, %xmm5 | ; AVX1-NEXT: vpmuludq %xmm7, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpmuludq %xmm1, %xmm0, %xmm7 | ; AVX1-NEXT: vpmuludq %xmm1, %xmm0, %xmm7 | ||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm7 = xmm7[1,1,3,3] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm7 = xmm7[1,1,3,3] | ||||
; AVX1-NEXT: vpblendw {{.*#+}} xmm5 = xmm7[0,1],xmm5[2,3],xmm7[4,5],xmm5[6,7] | ; AVX1-NEXT: vpblendw {{.*#+}} xmm5 = xmm7[0,1],xmm5[2,3],xmm7[4,5],xmm5[6,7] | ||||
; AVX1-NEXT: vpcmpeqd %xmm8, %xmm5, %xmm5 | ; AVX1-NEXT: vpcmpeqd %xmm8, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpxor %xmm6, %xmm5, %xmm5 | ; AVX1-NEXT: vpxor %xmm6, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm5, %ymm2 | ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm5, %ymm2 | ||||
; AVX1-NEXT: vpmulld %xmm3, %xmm4, %xmm3 | |||||
; AVX1-NEXT: vpmulld %xmm1, %xmm0, %xmm0 | ; AVX1-NEXT: vpmulld %xmm1, %xmm0, %xmm0 | ||||
; AVX1-NEXT: vinsertf128 $1, %xmm3, %ymm0, %ymm0 | ; AVX1-NEXT: vpmulld %xmm3, %xmm4, %xmm1 | ||||
; AVX1-NEXT: vmovaps %ymm0, (%rdi) | ; AVX1-NEXT: vmovdqa %xmm1, 16(%rdi) | ||||
; AVX1-NEXT: vmovdqa %xmm0, (%rdi) | |||||
; AVX1-NEXT: vmovaps %ymm2, %ymm0 | ; AVX1-NEXT: vmovaps %ymm2, %ymm0 | ||||
; AVX1-NEXT: retq | ; AVX1-NEXT: retq | ||||
; | ; | ||||
; AVX2-LABEL: umulo_v8i32: | ; AVX2-LABEL: umulo_v8i32: | ||||
; AVX2: # %bb.0: | ; AVX2: # %bb.0: | ||||
; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7] | ; AVX2-NEXT: vpshufd {{.*#+}} ymm2 = ymm1[1,1,3,3,5,5,7,7] | ||||
; AVX2-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[1,1,3,3,5,5,7,7] | ; AVX2-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[1,1,3,3,5,5,7,7] | ||||
; AVX2-NEXT: vpmuludq %ymm2, %ymm3, %ymm2 | ; AVX2-NEXT: vpmuludq %ymm2, %ymm3, %ymm2 | ||||
▲ Show 20 Lines • Show All 248 Lines • ▼ Show 20 Lines | |||||
; AVX1-NEXT: vpmuludq %xmm7, %xmm5, %xmm5 | ; AVX1-NEXT: vpmuludq %xmm7, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpmuludq %xmm2, %xmm0, %xmm7 | ; AVX1-NEXT: vpmuludq %xmm2, %xmm0, %xmm7 | ||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm7 = xmm7[1,1,3,3] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm7 = xmm7[1,1,3,3] | ||||
; AVX1-NEXT: vpblendw {{.*#+}} xmm5 = xmm7[0,1],xmm5[2,3],xmm7[4,5],xmm5[6,7] | ; AVX1-NEXT: vpblendw {{.*#+}} xmm5 = xmm7[0,1],xmm5[2,3],xmm7[4,5],xmm5[6,7] | ||||
; AVX1-NEXT: vpcmpeqd %xmm8, %xmm5, %xmm5 | ; AVX1-NEXT: vpcmpeqd %xmm8, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpxor %xmm9, %xmm5, %xmm5 | ; AVX1-NEXT: vpxor %xmm9, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpackssdw %xmm13, %xmm5, %xmm5 | ; AVX1-NEXT: vpackssdw %xmm13, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpacksswb %xmm11, %xmm5, %xmm5 | ; AVX1-NEXT: vpacksswb %xmm11, %xmm5, %xmm5 | ||||
; AVX1-NEXT: vpmulld %xmm2, %xmm0, %xmm2 | |||||
; AVX1-NEXT: vpmulld %xmm6, %xmm4, %xmm4 | ; AVX1-NEXT: vpmulld %xmm6, %xmm4, %xmm4 | ||||
; AVX1-NEXT: vpmulld %xmm2, %xmm0, %xmm0 | ; AVX1-NEXT: vpmulld %xmm3, %xmm1, %xmm3 | ||||
; AVX1-NEXT: vinsertf128 $1, %xmm4, %ymm0, %ymm2 | ; AVX1-NEXT: vpmulld %xmm10, %xmm12, %xmm6 | ||||
; AVX1-NEXT: vpmulld %xmm10, %xmm12, %xmm0 | |||||
; AVX1-NEXT: vpmulld %xmm3, %xmm1, %xmm1 | |||||
; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm3 | |||||
; AVX1-NEXT: vpmovsxbd %xmm5, %xmm0 | ; AVX1-NEXT: vpmovsxbd %xmm5, %xmm0 | ||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm5[1,1,2,3] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm5[1,1,2,3] | ||||
; AVX1-NEXT: vpmovsxbd %xmm1, %xmm1 | ; AVX1-NEXT: vpmovsxbd %xmm1, %xmm1 | ||||
; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0 | ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0 | ||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm5[2,3,0,1] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm5[2,3,0,1] | ||||
; AVX1-NEXT: vpmovsxbd %xmm1, %xmm1 | ; AVX1-NEXT: vpmovsxbd %xmm1, %xmm1 | ||||
; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm5[3,3,0,1] | ; AVX1-NEXT: vpshufd {{.*#+}} xmm5 = xmm5[3,3,0,1] | ||||
; AVX1-NEXT: vpmovsxbd %xmm4, %xmm4 | ; AVX1-NEXT: vpmovsxbd %xmm5, %xmm5 | ||||
; AVX1-NEXT: vinsertf128 $1, %xmm4, %ymm1, %ymm1 | ; AVX1-NEXT: vinsertf128 $1, %xmm5, %ymm1, %ymm1 | ||||
; AVX1-NEXT: vmovaps %ymm3, 32(%rdi) | ; AVX1-NEXT: vmovdqa %xmm6, 48(%rdi) | ||||
; AVX1-NEXT: vmovaps %ymm2, (%rdi) | ; AVX1-NEXT: vmovdqa %xmm3, 32(%rdi) | ||||
; AVX1-NEXT: vmovdqa %xmm4, 16(%rdi) | |||||
; AVX1-NEXT: vmovdqa %xmm2, (%rdi) | |||||
; AVX1-NEXT: retq | ; AVX1-NEXT: retq | ||||
; | ; | ||||
; AVX2-LABEL: umulo_v16i32: | ; AVX2-LABEL: umulo_v16i32: | ||||
; AVX2: # %bb.0: | ; AVX2: # %bb.0: | ||||
; AVX2-NEXT: vpshufd {{.*#+}} ymm4 = ymm3[1,1,3,3,5,5,7,7] | ; AVX2-NEXT: vpshufd {{.*#+}} ymm4 = ymm3[1,1,3,3,5,5,7,7] | ||||
; AVX2-NEXT: vpshufd {{.*#+}} ymm5 = ymm1[1,1,3,3,5,5,7,7] | ; AVX2-NEXT: vpshufd {{.*#+}} ymm5 = ymm1[1,1,3,3,5,5,7,7] | ||||
; AVX2-NEXT: vpmuludq %ymm4, %ymm5, %ymm4 | ; AVX2-NEXT: vpmuludq %ymm4, %ymm5, %ymm4 | ||||
; AVX2-NEXT: vpmuludq %ymm3, %ymm1, %ymm5 | ; AVX2-NEXT: vpmuludq %ymm3, %ymm1, %ymm5 | ||||
▲ Show 20 Lines • Show All 1,485 Lines • Show Last 20 Lines |