| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc < %s -mtriple=x86_64-linux-generic -verify-machineinstrs -mattr=-sse | FileCheck %s --check-prefix=X64-NOSSE |
| ; RUN: llc < %s -mtriple=x86_64-linux-generic -verify-machineinstrs | FileCheck %s --check-prefix=X64-SSE |
| |
| ; Note: This test is testing that the lowering for atomics matches what we |
| ; currently emit for non-atomics + the atomic restriction. The presence of |
| ; particular lowering detail in these tests should not be read as requiring |
| ; that detail for correctness unless it's related to the atomicity itself. |
| ; (Specifically, there were reviewer questions about the lowering for halfs |
| ; and their calling convention which remain unresolved.) |
| |
| define void @store_fp128(fp128* %fptr, fp128 %v) { |
| ; X64-NOSSE-LABEL: store_fp128: |
| ; X64-NOSSE: # %bb.0: |
| ; X64-NOSSE-NEXT: pushq %rax |
| ; X64-NOSSE-NEXT: .cfi_def_cfa_offset 16 |
| ; X64-NOSSE-NEXT: callq __sync_lock_test_and_set_16@PLT |
| ; X64-NOSSE-NEXT: popq %rax |
| ; X64-NOSSE-NEXT: .cfi_def_cfa_offset 8 |
| ; X64-NOSSE-NEXT: retq |
| ; |
| ; X64-SSE-LABEL: store_fp128: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: subq $24, %rsp |
| ; X64-SSE-NEXT: .cfi_def_cfa_offset 32 |
| ; X64-SSE-NEXT: movaps %xmm0, (%rsp) |
| ; X64-SSE-NEXT: movq (%rsp), %rsi |
| ; X64-SSE-NEXT: movq {{[0-9]+}}(%rsp), %rdx |
| ; X64-SSE-NEXT: callq __sync_lock_test_and_set_16@PLT |
| ; X64-SSE-NEXT: addq $24, %rsp |
| ; X64-SSE-NEXT: .cfi_def_cfa_offset 8 |
| ; X64-SSE-NEXT: retq |
| store atomic fp128 %v, fp128* %fptr unordered, align 16 |
| ret void |
| } |