| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py | 
 | ; RUN: llc -mtriple=x86_64-unknown-linux-gnu -o - -mattr=-x87,+sse,-sse2 %s  | FileCheck --check-prefixes=X64SSE %s | 
 | ; RUN: llc -mtriple=i686-unknown-linux-gnu -o - -mattr=-x87,+sse,-sse2 %s    | FileCheck --check-prefixes=X86SSE %s | 
 | ; RUN: llc -mtriple=x86_64-unknown-linux-gnu -o - -mattr=-x87,+sse2,-sse3 %s | FileCheck --check-prefixes=X64SSE2 %s | 
 | ; RUN: llc -mtriple=i686-unknown-linux-gnu -o - -mattr=-x87,+sse2,-sse3 %s   | FileCheck --check-prefixes=X86SSE2 %s | 
 | ; RUN: llc -mtriple=x86_64-unknown-linux-gnu -o - -mattr=-x87,+avx,-avx2 %s  | FileCheck --check-prefixes=X64AVX %s | 
 | ; RUN: llc -mtriple=i686-unknown-linux-gnu -o - -mattr=-x87,+avx,-avx2 %s    | FileCheck --check-prefixes=X86AVX %s | 
 |  | 
 |  | 
 | %struct.params = type { double, double } | 
 |  | 
 | define i32 @pr38738() { | 
 | ; X64SSE-LABEL: pr38738: | 
 | ; X64SSE:       # %bb.0: # %entry | 
 | ; X64SSE-NEXT:    xorps %xmm0, %xmm0 | 
 | ; X64SSE-NEXT:    movaps %xmm0, -{{[0-9]+}}(%rsp) | 
 | ; X64SSE-NEXT:    movl $0, -{{[0-9]+}}(%rsp) | 
 | ; X64SSE-NEXT:    movl -{{[0-9]+}}(%rsp), %eax | 
 | ; X64SSE-NEXT:    retq | 
 | ; | 
 | ; X86SSE-LABEL: pr38738: | 
 | ; X86SSE:       # %bb.0: # %entry | 
 | ; X86SSE-NEXT:    subl $28, %esp | 
 | ; X86SSE-NEXT:    .cfi_def_cfa_offset 32 | 
 | ; X86SSE-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86SSE-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86SSE-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86SSE-NEXT:    movl $0, (%esp) | 
 | ; X86SSE-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE-NEXT:    addl $28, %esp | 
 | ; X86SSE-NEXT:    .cfi_def_cfa_offset 4 | 
 | ; X86SSE-NEXT:    retl | 
 | ; | 
 | ; X64SSE2-LABEL: pr38738: | 
 | ; X64SSE2:       # %bb.0: # %entry | 
 | ; X64SSE2-NEXT:    xorps %xmm0, %xmm0 | 
 | ; X64SSE2-NEXT:    movaps %xmm0, -{{[0-9]+}}(%rsp) | 
 | ; X64SSE2-NEXT:    movl $0, -{{[0-9]+}}(%rsp) | 
 | ; X64SSE2-NEXT:    movl -{{[0-9]+}}(%rsp), %eax | 
 | ; X64SSE2-NEXT:    retq | 
 | ; | 
 | ; X86SSE2-LABEL: pr38738: | 
 | ; X86SSE2:       # %bb.0: # %entry | 
 | ; X86SSE2-NEXT:    subl $44, %esp | 
 | ; X86SSE2-NEXT:    .cfi_def_cfa_offset 48 | 
 | ; X86SSE2-NEXT:    xorps %xmm0, %xmm0 | 
 | ; X86SSE2-NEXT:    movaps %xmm0, {{[0-9]+}}(%esp) | 
 | ; X86SSE2-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86SSE2-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE2-NEXT:    addl $44, %esp | 
 | ; X86SSE2-NEXT:    .cfi_def_cfa_offset 4 | 
 | ; X86SSE2-NEXT:    retl | 
 | ; | 
 | ; X64AVX-LABEL: pr38738: | 
 | ; X64AVX:       # %bb.0: # %entry | 
 | ; X64AVX-NEXT:    vxorps %xmm0, %xmm0, %xmm0 | 
 | ; X64AVX-NEXT:    vmovaps %xmm0, -{{[0-9]+}}(%rsp) | 
 | ; X64AVX-NEXT:    movl $0, -{{[0-9]+}}(%rsp) | 
 | ; X64AVX-NEXT:    movl -{{[0-9]+}}(%rsp), %eax | 
 | ; X64AVX-NEXT:    retq | 
 | ; | 
 | ; X86AVX-LABEL: pr38738: | 
 | ; X86AVX:       # %bb.0: # %entry | 
 | ; X86AVX-NEXT:    subl $44, %esp | 
 | ; X86AVX-NEXT:    .cfi_def_cfa_offset 48 | 
 | ; X86AVX-NEXT:    vxorps %xmm0, %xmm0, %xmm0 | 
 | ; X86AVX-NEXT:    vmovaps %xmm0, {{[0-9]+}}(%esp) | 
 | ; X86AVX-NEXT:    movl $0, {{[0-9]+}}(%esp) | 
 | ; X86AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86AVX-NEXT:    addl $44, %esp | 
 | ; X86AVX-NEXT:    .cfi_def_cfa_offset 4 | 
 | ; X86AVX-NEXT:    retl | 
 | entry: | 
 |   %retval = alloca i32, align 4 | 
 |   %dlg_sys_param = alloca %struct.params, align 8 | 
 |   %total_active_bw = alloca float, align 4 | 
 |   call void @llvm.memset.p0.i64(ptr align 8 %dlg_sys_param, i8 0, i64 16, i1 false) | 
 |   store float 0.000000e+00, ptr %total_active_bw, align 4 | 
 |   %0 = load i32, ptr %retval, align 4 | 
 |   ret i32 %0 | 
 | } | 
 |  | 
 | define void @tryset(ptr nocapture %x) { | 
 | ; X64SSE-LABEL: tryset: | 
 | ; X64SSE:       # %bb.0: | 
 | ; X64SSE-NEXT:    movq $0, 56(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 48(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 40(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 32(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 24(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 16(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, 8(%rdi) | 
 | ; X64SSE-NEXT:    movq $0, (%rdi) | 
 | ; X64SSE-NEXT:    retq | 
 | ; | 
 | ; X86SSE-LABEL: tryset: | 
 | ; X86SSE:       # %bb.0: | 
 | ; X86SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE-NEXT:    movl $0, 60(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 56(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 52(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 48(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 44(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 40(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 36(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 32(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 28(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 24(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 20(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 16(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 12(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 8(%eax) | 
 | ; X86SSE-NEXT:    movl $0, 4(%eax) | 
 | ; X86SSE-NEXT:    movl $0, (%eax) | 
 | ; X86SSE-NEXT:    retl | 
 | ; | 
 | ; X64SSE2-LABEL: tryset: | 
 | ; X64SSE2:       # %bb.0: | 
 | ; X64SSE2-NEXT:    movq $0, 56(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 48(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 40(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 32(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 24(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 16(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, 8(%rdi) | 
 | ; X64SSE2-NEXT:    movq $0, (%rdi) | 
 | ; X64SSE2-NEXT:    retq | 
 | ; | 
 | ; X86SSE2-LABEL: tryset: | 
 | ; X86SSE2:       # %bb.0: | 
 | ; X86SSE2-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE2-NEXT:    movl $0, 4(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, (%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 12(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 8(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 20(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 16(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 28(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 24(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 36(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 32(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 44(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 40(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 52(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 48(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 60(%eax) | 
 | ; X86SSE2-NEXT:    movl $0, 56(%eax) | 
 | ; X86SSE2-NEXT:    retl | 
 | ; | 
 | ; X64AVX-LABEL: tryset: | 
 | ; X64AVX:       # %bb.0: | 
 | ; X64AVX-NEXT:    vxorps %xmm0, %xmm0, %xmm0 | 
 | ; X64AVX-NEXT:    vmovups %ymm0, 32(%rdi) | 
 | ; X64AVX-NEXT:    vmovups %ymm0, (%rdi) | 
 | ; X64AVX-NEXT:    vzeroupper | 
 | ; X64AVX-NEXT:    retq | 
 | ; | 
 | ; X86AVX-LABEL: tryset: | 
 | ; X86AVX:       # %bb.0: | 
 | ; X86AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86AVX-NEXT:    vxorps %xmm0, %xmm0, %xmm0 | 
 | ; X86AVX-NEXT:    vmovups %ymm0, 32(%eax) | 
 | ; X86AVX-NEXT:    vmovups %ymm0, (%eax) | 
 | ; X86AVX-NEXT:    vzeroupper | 
 | ; X86AVX-NEXT:    retl | 
 |   tail call void @llvm.memset.p0.i64(ptr align 1 %x, i8 0, i64 64, i1 false) | 
 |   ret void | 
 | } | 
 |  | 
 | define void @trycpy(ptr nocapture %x, ptr nocapture readonly %y) { | 
 | ; X64SSE-LABEL: trycpy: | 
 | ; X64SSE:       # %bb.0: | 
 | ; X64SSE-NEXT:    movq 24(%rsi), %rax | 
 | ; X64SSE-NEXT:    movq %rax, 24(%rdi) | 
 | ; X64SSE-NEXT:    movq 16(%rsi), %rax | 
 | ; X64SSE-NEXT:    movq %rax, 16(%rdi) | 
 | ; X64SSE-NEXT:    movq (%rsi), %rax | 
 | ; X64SSE-NEXT:    movq 8(%rsi), %rcx | 
 | ; X64SSE-NEXT:    movq %rcx, 8(%rdi) | 
 | ; X64SSE-NEXT:    movq %rax, (%rdi) | 
 | ; X64SSE-NEXT:    retq | 
 | ; | 
 | ; X86SSE-LABEL: trycpy: | 
 | ; X86SSE:       # %bb.0: | 
 | ; X86SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE-NEXT:    movl {{[0-9]+}}(%esp), %ecx | 
 | ; X86SSE-NEXT:    movl 28(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 28(%eax) | 
 | ; X86SSE-NEXT:    movl 24(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 24(%eax) | 
 | ; X86SSE-NEXT:    movl 20(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 20(%eax) | 
 | ; X86SSE-NEXT:    movl 16(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 16(%eax) | 
 | ; X86SSE-NEXT:    movl 12(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 12(%eax) | 
 | ; X86SSE-NEXT:    movl 8(%ecx), %edx | 
 | ; X86SSE-NEXT:    movl %edx, 8(%eax) | 
 | ; X86SSE-NEXT:    movl (%ecx), %edx | 
 | ; X86SSE-NEXT:    movl 4(%ecx), %ecx | 
 | ; X86SSE-NEXT:    movl %ecx, 4(%eax) | 
 | ; X86SSE-NEXT:    movl %edx, (%eax) | 
 | ; X86SSE-NEXT:    retl | 
 | ; | 
 | ; X64SSE2-LABEL: trycpy: | 
 | ; X64SSE2:       # %bb.0: | 
 | ; X64SSE2-NEXT:    movq 24(%rsi), %rax | 
 | ; X64SSE2-NEXT:    movq %rax, 24(%rdi) | 
 | ; X64SSE2-NEXT:    movq 16(%rsi), %rax | 
 | ; X64SSE2-NEXT:    movq %rax, 16(%rdi) | 
 | ; X64SSE2-NEXT:    movq (%rsi), %rax | 
 | ; X64SSE2-NEXT:    movq 8(%rsi), %rcx | 
 | ; X64SSE2-NEXT:    movq %rcx, 8(%rdi) | 
 | ; X64SSE2-NEXT:    movq %rax, (%rdi) | 
 | ; X64SSE2-NEXT:    retq | 
 | ; | 
 | ; X86SSE2-LABEL: trycpy: | 
 | ; X86SSE2:       # %bb.0: | 
 | ; X86SSE2-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86SSE2-NEXT:    movl {{[0-9]+}}(%esp), %ecx | 
 | ; X86SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero | 
 | ; X86SSE2-NEXT:    movsd %xmm0, 24(%eax) | 
 | ; X86SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero | 
 | ; X86SSE2-NEXT:    movsd %xmm0, 16(%eax) | 
 | ; X86SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero | 
 | ; X86SSE2-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero | 
 | ; X86SSE2-NEXT:    movsd %xmm1, 8(%eax) | 
 | ; X86SSE2-NEXT:    movsd %xmm0, (%eax) | 
 | ; X86SSE2-NEXT:    retl | 
 | ; | 
 | ; X64AVX-LABEL: trycpy: | 
 | ; X64AVX:       # %bb.0: | 
 | ; X64AVX-NEXT:    vmovups (%rsi), %ymm0 | 
 | ; X64AVX-NEXT:    vmovups %ymm0, (%rdi) | 
 | ; X64AVX-NEXT:    vzeroupper | 
 | ; X64AVX-NEXT:    retq | 
 | ; | 
 | ; X86AVX-LABEL: trycpy: | 
 | ; X86AVX:       # %bb.0: | 
 | ; X86AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax | 
 | ; X86AVX-NEXT:    movl {{[0-9]+}}(%esp), %ecx | 
 | ; X86AVX-NEXT:    vmovups (%ecx), %ymm0 | 
 | ; X86AVX-NEXT:    vmovups %ymm0, (%eax) | 
 | ; X86AVX-NEXT:    vzeroupper | 
 | ; X86AVX-NEXT:    retl | 
 |   tail call void @llvm.memcpy.p0.p0.i64(ptr align 1 %x, ptr align 1 %y, i64 32, i1 false) | 
 |   ret void | 
 | } | 
 |  | 
 | declare void @llvm.memset.p0.i64(ptr nocapture writeonly, i8, i64, i1) argmemonly nounwind | 
 | declare void @llvm.memcpy.p0.p0.i64(ptr nocapture writeonly, ptr nocapture readonly, i64, i1) #2 | 
 |  |