| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc -mcpu=pwr9 -verify-machineinstrs -ppc-asm-full-reg-names \ |
| ; RUN: -mtriple=powerpc64le-unknown-linux-gnu < %s | FileCheck %s --check-prefix=LE |
| ; RUN: llc -mcpu=pwr9 -verify-machineinstrs -ppc-asm-full-reg-names \ |
| ; RUN: -mtriple=powerpc64-unknown-linux-gnu < %s | FileCheck %s --check-prefix=P9BE |
| ; RUN: llc -mcpu=pwr8 -verify-machineinstrs -ppc-asm-full-reg-names \ |
| ; RUN: -mtriple=powerpc64le-unknown-linux-gnu < %s | FileCheck %s --check-prefix=LE |
| ; RUN: llc -mcpu=pwr8 -verify-machineinstrs -ppc-asm-full-reg-names \ |
| ; RUN: -mtriple=powerpc64-unknown-linux-gnu < %s | FileCheck %s --check-prefix=P8BE |
| |
| declare ppc_fp128 @llvm.fmuladd.ppcf128(ppc_fp128, ppc_fp128, ppc_fp128) #2 |
| |
| define ppc_fp128 @test_ctr0() { |
| ; LE-LABEL: test_ctr0: |
| ; LE: # %bb.0: # %bb |
| ; LE-NEXT: mflr r0 |
| ; LE-NEXT: .cfi_def_cfa_offset 48 |
| ; LE-NEXT: .cfi_offset lr, 16 |
| ; LE-NEXT: .cfi_offset r30, -16 |
| ; LE-NEXT: std r30, -16(r1) # 8-byte Folded Spill |
| ; LE-NEXT: std r0, 16(r1) |
| ; LE-NEXT: stdu r1, -48(r1) |
| ; LE-NEXT: xxlxor f1, f1, f1 |
| ; LE-NEXT: li r30, 0 |
| ; LE-NEXT: xxlxor f2, f2, f2 |
| ; LE-NEXT: .p2align 5 |
| ; LE-NEXT: .LBB0_1: # %bb6 |
| ; LE-NEXT: # |
| ; LE-NEXT: xxlxor f3, f3, f3 |
| ; LE-NEXT: xxlxor f4, f4, f4 |
| ; LE-NEXT: bl __gcc_qadd |
| ; LE-NEXT: nop |
| ; LE-NEXT: addi r30, r30, 4 |
| ; LE-NEXT: cmpldi r30, 0 |
| ; LE-NEXT: bne cr0, .LBB0_1 |
| ; LE-NEXT: # %bb.2: # %bb14 |
| ; LE-NEXT: addi r1, r1, 48 |
| ; LE-NEXT: ld r0, 16(r1) |
| ; LE-NEXT: ld r30, -16(r1) # 8-byte Folded Reload |
| ; LE-NEXT: mtlr r0 |
| ; LE-NEXT: blr |
| ; |
| ; P9BE-LABEL: test_ctr0: |
| ; P9BE: # %bb.0: # %bb |
| ; P9BE-NEXT: mflr r0 |
| ; P9BE-NEXT: std r0, 16(r1) |
| ; P9BE-NEXT: stdu r1, -128(r1) |
| ; P9BE-NEXT: .cfi_def_cfa_offset 128 |
| ; P9BE-NEXT: .cfi_offset lr, 16 |
| ; P9BE-NEXT: .cfi_offset r30, -16 |
| ; P9BE-NEXT: std r30, 112(r1) # 8-byte Folded Spill |
| ; P9BE-NEXT: xxlxor f1, f1, f1 |
| ; P9BE-NEXT: li r30, 0 |
| ; P9BE-NEXT: xxlxor f2, f2, f2 |
| ; P9BE-NEXT: .p2align 5 |
| ; P9BE-NEXT: .LBB0_1: # %bb6 |
| ; P9BE-NEXT: # |
| ; P9BE-NEXT: xxlxor f3, f3, f3 |
| ; P9BE-NEXT: xxlxor f4, f4, f4 |
| ; P9BE-NEXT: bl __gcc_qadd |
| ; P9BE-NEXT: nop |
| ; P9BE-NEXT: addi r30, r30, 4 |
| ; P9BE-NEXT: cmpldi r30, 0 |
| ; P9BE-NEXT: bne cr0, .LBB0_1 |
| ; P9BE-NEXT: # %bb.2: # %bb14 |
| ; P9BE-NEXT: ld r30, 112(r1) # 8-byte Folded Reload |
| ; P9BE-NEXT: addi r1, r1, 128 |
| ; P9BE-NEXT: ld r0, 16(r1) |
| ; P9BE-NEXT: mtlr r0 |
| ; P9BE-NEXT: blr |
| ; |
| ; P8BE-LABEL: test_ctr0: |
| ; P8BE: # %bb.0: # %bb |
| ; P8BE-NEXT: mflr r0 |
| ; P8BE-NEXT: std r0, 16(r1) |
| ; P8BE-NEXT: stdu r1, -128(r1) |
| ; P8BE-NEXT: .cfi_def_cfa_offset 128 |
| ; P8BE-NEXT: .cfi_offset lr, 16 |
| ; P8BE-NEXT: .cfi_offset r30, -16 |
| ; P8BE-NEXT: xxlxor f1, f1, f1 |
| ; P8BE-NEXT: std r30, 112(r1) # 8-byte Folded Spill |
| ; P8BE-NEXT: li r30, 0 |
| ; P8BE-NEXT: xxlxor f2, f2, f2 |
| ; P8BE-NEXT: .p2align 5 |
| ; P8BE-NEXT: .LBB0_1: # %bb6 |
| ; P8BE-NEXT: # |
| ; P8BE-NEXT: xxlxor f3, f3, f3 |
| ; P8BE-NEXT: xxlxor f4, f4, f4 |
| ; P8BE-NEXT: bl __gcc_qadd |
| ; P8BE-NEXT: nop |
| ; P8BE-NEXT: addi r30, r30, 4 |
| ; P8BE-NEXT: cmpldi r30, 0 |
| ; P8BE-NEXT: bne cr0, .LBB0_1 |
| ; P8BE-NEXT: # %bb.2: # %bb14 |
| ; P8BE-NEXT: ld r30, 112(r1) # 8-byte Folded Reload |
| ; P8BE-NEXT: addi r1, r1, 128 |
| ; P8BE-NEXT: ld r0, 16(r1) |
| ; P8BE-NEXT: mtlr r0 |
| ; P8BE-NEXT: blr |
| bb: |
| br label %bb6 |
| |
| bb6: ; preds = %bb6, %bb |
| %i = phi ppc_fp128 [ %i8, %bb6 ], [ 0xM00000000000000000000000000000000, %bb ] |
| %i7 = phi i64 [ %i9, %bb6 ], [ 0, %bb ] |
| %i8 = tail call ppc_fp128 @llvm.fmuladd.ppcf128(ppc_fp128 0xM00000000000000000000000000000000, ppc_fp128 0xM00000000000000000000000000000000, ppc_fp128 %i) #4 |
| %i9 = add i64 %i7, -4 |
| %i10 = icmp eq i64 %i9, 0 |
| br i1 %i10, label %bb14, label %bb6 |
| |
| bb14: ; preds = %bb6 |
| ret ppc_fp128 %i8 |
| } |