| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py | 
 | ; RUN: llc < %s -mtriple=i686-darwin -mattr=+mmx,+sse2 | FileCheck --check-prefix=X32 %s | 
 | ; RUN: llc < %s -mtriple=x86_64-darwin -mattr=+mmx,+sse2 | FileCheck --check-prefix=X64 %s | 
 |  | 
 | ; If there is no explicit MMX type usage, always promote to XMM. | 
 |  | 
 | define void @test0(ptr %x) { | 
 | ; X32-LABEL: test0: | 
 | ; X32:       ## %bb.0: ## %entry | 
 | ; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X32-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero | 
 | ; X32-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1,1,1] | 
 | ; X32-NEXT:    movlps %xmm0, (%eax) | 
 | ; X32-NEXT:    retl | 
 | ; | 
 | ; X64-LABEL: test0: | 
 | ; X64:       ## %bb.0: ## %entry | 
 | ; X64-NEXT:    movq {{.*#+}} xmm0 = mem[0],zero | 
 | ; X64-NEXT:    pshufd {{.*#+}} xmm0 = xmm0[1,1,1,1] | 
 | ; X64-NEXT:    movq %xmm0, (%rdi) | 
 | ; X64-NEXT:    retq | 
 | entry: | 
 |   %tmp2 = load <1 x i64>, ptr %x | 
 |   %tmp6 = bitcast <1 x i64> %tmp2 to <2 x i32> | 
 |   %tmp9 = shufflevector <2 x i32> %tmp6, <2 x i32> undef, <2 x i32> < i32 1, i32 1 > | 
 |   %tmp10 = bitcast <2 x i32> %tmp9 to <1 x i64> | 
 |   store <1 x i64> %tmp10, ptr %x | 
 |   ret void | 
 | } | 
 |  | 
 | define void @test1() { | 
 | ; X32-LABEL: test1: | 
 | ; X32:       ## %bb.0: ## %entry | 
 | ; X32-NEXT:    pushl %edi | 
 | ; X32-NEXT:    .cfi_def_cfa_offset 8 | 
 | ; X32-NEXT:    .cfi_offset %edi, -8 | 
 | ; X32-NEXT:    pxor %mm0, %mm0 | 
 | ; X32-NEXT:    movq {{\.?LCPI[0-9]+_[0-9]+}}, %mm1 ## mm1 = 0x7070606040400000 | 
 | ; X32-NEXT:    xorl %edi, %edi | 
 | ; X32-NEXT:    maskmovq %mm1, %mm0 | 
 | ; X32-NEXT:    popl %edi | 
 | ; X32-NEXT:    retl | 
 | ; | 
 | ; X64-LABEL: test1: | 
 | ; X64:       ## %bb.0: ## %entry | 
 | ; X64-NEXT:    pxor %mm0, %mm0 | 
 | ; X64-NEXT:    movq {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %mm1 ## mm1 = 0x7070606040400000 | 
 | ; X64-NEXT:    xorl %edi, %edi | 
 | ; X64-NEXT:    maskmovq %mm1, %mm0 | 
 | ; X64-NEXT:    retq | 
 | entry: | 
 |   %tmp528 = bitcast <8 x i8> zeroinitializer to <2 x i32> | 
 |   %tmp529 = and <2 x i32> %tmp528, bitcast (<4 x i16> < i16 -32640, i16 16448, i16 8224, i16 4112 > to <2 x i32>) | 
 |   %tmp542 = bitcast <2 x i32> %tmp529 to <4 x i16> | 
 |   %tmp543 = add <4 x i16> %tmp542, < i16 0, i16 16448, i16 24672, i16 28784 > | 
 |   %tmp555 = bitcast <4 x i16> %tmp543 to <8 x i8> | 
 |   %tmp556 = bitcast <8 x i8> %tmp555 to x86_mmx | 
 |   %tmp557 = bitcast <8 x i8> zeroinitializer to x86_mmx | 
 |   tail call void @llvm.x86.mmx.maskmovq( x86_mmx %tmp557, x86_mmx %tmp556, ptr null) | 
 |   ret void | 
 | } | 
 |  | 
 | @tmp_V2i = common global <2 x i32> zeroinitializer | 
 |  | 
 | define void @test2() nounwind { | 
 | ; X32-LABEL: test2: | 
 | ; X32:       ## %bb.0: ## %entry | 
 | ; X32-NEXT:    movl L_tmp_V2i$non_lazy_ptr, %eax | 
 | ; X32-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero | 
 | ; X32-NEXT:    unpcklps {{.*#+}} xmm0 = xmm0[0,0,1,1] | 
 | ; X32-NEXT:    movlps %xmm0, (%eax) | 
 | ; X32-NEXT:    retl | 
 | ; | 
 | ; X64-LABEL: test2: | 
 | ; X64:       ## %bb.0: ## %entry | 
 | ; X64-NEXT:    movq _tmp_V2i@GOTPCREL(%rip), %rax | 
 | ; X64-NEXT:    movd {{.*#+}} xmm0 = mem[0],zero,zero,zero | 
 | ; X64-NEXT:    pshufd {{.*#+}} xmm0 = xmm0[0,0,0,0] | 
 | ; X64-NEXT:    movq %xmm0, (%rax) | 
 | ; X64-NEXT:    retq | 
 | entry: | 
 |   %0 = load <2 x i32>, ptr @tmp_V2i, align 8 | 
 |   %1 = shufflevector <2 x i32> %0, <2 x i32> undef, <2 x i32> zeroinitializer | 
 |   store <2 x i32> %1, ptr @tmp_V2i, align 8 | 
 |   ret void | 
 | } | 
 |  | 
 | define <4 x float> @pr35869() nounwind { | 
 | ; X32-LABEL: pr35869: | 
 | ; X32:       ## %bb.0: | 
 | ; X32-NEXT:    movl $64, %eax | 
 | ; X32-NEXT:    movd %eax, %mm0 | 
 | ; X32-NEXT:    pxor %mm1, %mm1 | 
 | ; X32-NEXT:    punpcklbw %mm1, %mm0 ## mm0 = mm0[0],mm1[0],mm0[1],mm1[1],mm0[2],mm1[2],mm0[3],mm1[3] | 
 | ; X32-NEXT:    pcmpgtw %mm0, %mm1 | 
 | ; X32-NEXT:    movq %mm0, %mm2 | 
 | ; X32-NEXT:    punpckhwd %mm1, %mm2 ## mm2 = mm2[2],mm1[2],mm2[3],mm1[3] | 
 | ; X32-NEXT:    xorps %xmm0, %xmm0 | 
 | ; X32-NEXT:    cvtpi2ps %mm2, %xmm0 | 
 | ; X32-NEXT:    movlhps {{.*#+}} xmm0 = xmm0[0,0] | 
 | ; X32-NEXT:    punpcklwd %mm1, %mm0 ## mm0 = mm0[0],mm1[0],mm0[1],mm1[1] | 
 | ; X32-NEXT:    cvtpi2ps %mm0, %xmm0 | 
 | ; X32-NEXT:    retl | 
 | ; | 
 | ; X64-LABEL: pr35869: | 
 | ; X64:       ## %bb.0: | 
 | ; X64-NEXT:    movl $64, %eax | 
 | ; X64-NEXT:    movd %eax, %mm0 | 
 | ; X64-NEXT:    pxor %mm1, %mm1 | 
 | ; X64-NEXT:    punpcklbw %mm1, %mm0 ## mm0 = mm0[0],mm1[0],mm0[1],mm1[1],mm0[2],mm1[2],mm0[3],mm1[3] | 
 | ; X64-NEXT:    pcmpgtw %mm0, %mm1 | 
 | ; X64-NEXT:    movq %mm0, %mm2 | 
 | ; X64-NEXT:    punpckhwd %mm1, %mm2 ## mm2 = mm2[2],mm1[2],mm2[3],mm1[3] | 
 | ; X64-NEXT:    xorps %xmm0, %xmm0 | 
 | ; X64-NEXT:    cvtpi2ps %mm2, %xmm0 | 
 | ; X64-NEXT:    movlhps {{.*#+}} xmm0 = xmm0[0,0] | 
 | ; X64-NEXT:    punpcklwd %mm1, %mm0 ## mm0 = mm0[0],mm1[0],mm0[1],mm1[1] | 
 | ; X64-NEXT:    cvtpi2ps %mm0, %xmm0 | 
 | ; X64-NEXT:    retq | 
 |   %1 = tail call x86_mmx @llvm.x86.mmx.punpcklbw(x86_mmx bitcast (<8 x i8> <i8 64, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0> to x86_mmx), x86_mmx bitcast (<8 x i8> zeroinitializer to x86_mmx)) | 
 |   %2 = tail call x86_mmx @llvm.x86.mmx.pcmpgt.w(x86_mmx bitcast (<4 x i16> zeroinitializer to x86_mmx), x86_mmx %1) | 
 |   %3 = tail call x86_mmx @llvm.x86.mmx.punpckhwd(x86_mmx %1, x86_mmx %2) | 
 |   %4 = tail call <4 x float> @llvm.x86.sse.cvtpi2ps(<4 x float> zeroinitializer, x86_mmx %3) | 
 |   %5 = shufflevector <4 x float> %4, <4 x float> undef, <4 x i32> <i32 0, i32 1, i32 0, i32 1> | 
 |   %6 = tail call x86_mmx @llvm.x86.mmx.punpcklwd(x86_mmx %1, x86_mmx %2) | 
 |   %7 = tail call <4 x float> @llvm.x86.sse.cvtpi2ps(<4 x float> %5, x86_mmx %6) | 
 |   ret <4 x float> %7 | 
 | } | 
 |  | 
 | declare void @llvm.x86.mmx.maskmovq(x86_mmx, x86_mmx, ptr) | 
 | declare x86_mmx @llvm.x86.mmx.pcmpgt.w(x86_mmx, x86_mmx) | 
 | declare x86_mmx @llvm.x86.mmx.punpcklbw(x86_mmx, x86_mmx) | 
 | declare x86_mmx @llvm.x86.mmx.punpcklwd(x86_mmx, x86_mmx) | 
 | declare x86_mmx @llvm.x86.mmx.punpckhwd(x86_mmx, x86_mmx) | 
 | declare <4 x float> @llvm.x86.sse.cvtpi2ps(<4 x float>, x86_mmx) |