| ; Test strict multiplication of two f64s, producing an f128 result. |
| ; |
| ; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s |
| |
| declare fp128 @llvm.experimental.constrained.fmul.f128(fp128, fp128, metadata, metadata) |
| declare double @llvm.experimental.constrained.fadd.f64(double, double, metadata, metadata) |
| declare double @llvm.experimental.constrained.fptrunc.f64.f128(fp128, metadata, metadata) |
| declare fp128 @llvm.experimental.constrained.fpext.f128.f64(double, metadata) |
| |
| declare double @foo() |
| |
| ; Check register multiplication. "mxdbr %f0, %f2" is not valid from LLVM's |
| ; point of view, because %f2 is the low register of the FP128 %f0. Pass the |
| ; multiplier in %f4 instead. |
| define void @f1(double %f1, double %dummy, double %f2, ptr %dst) #0 { |
| ; CHECK-LABEL: f1: |
| ; CHECK: mxdbr %f0, %f4 |
| ; CHECK: std %f0, 0(%r2) |
| ; CHECK: std %f2, 8(%r2) |
| ; CHECK: br %r14 |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check the low end of the MXDB range. |
| define void @f2(double %f1, ptr %ptr, ptr %dst) #0 { |
| ; CHECK-LABEL: f2: |
| ; CHECK: mxdb %f0, 0(%r2) |
| ; CHECK: std %f0, 0(%r3) |
| ; CHECK: std %f2, 8(%r3) |
| ; CHECK: br %r14 |
| %f2 = load double, ptr %ptr |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check the high end of the aligned MXDB range. |
| define void @f3(double %f1, ptr %base, ptr %dst) #0 { |
| ; CHECK-LABEL: f3: |
| ; CHECK: mxdb %f0, 4088(%r2) |
| ; CHECK: std %f0, 0(%r3) |
| ; CHECK: std %f2, 8(%r3) |
| ; CHECK: br %r14 |
| %ptr = getelementptr double, ptr %base, i64 511 |
| %f2 = load double, ptr %ptr |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check the next doubleword up, which needs separate address logic. |
| ; Other sequences besides this one would be OK. |
| define void @f4(double %f1, ptr %base, ptr %dst) #0 { |
| ; CHECK-LABEL: f4: |
| ; CHECK: aghi %r2, 4096 |
| ; CHECK: mxdb %f0, 0(%r2) |
| ; CHECK: std %f0, 0(%r3) |
| ; CHECK: std %f2, 8(%r3) |
| ; CHECK: br %r14 |
| %ptr = getelementptr double, ptr %base, i64 512 |
| %f2 = load double, ptr %ptr |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check negative displacements, which also need separate address logic. |
| define void @f5(double %f1, ptr %base, ptr %dst) #0 { |
| ; CHECK-LABEL: f5: |
| ; CHECK: aghi %r2, -8 |
| ; CHECK: mxdb %f0, 0(%r2) |
| ; CHECK: std %f0, 0(%r3) |
| ; CHECK: std %f2, 8(%r3) |
| ; CHECK: br %r14 |
| %ptr = getelementptr double, ptr %base, i64 -1 |
| %f2 = load double, ptr %ptr |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check that MXDB allows indices. |
| define void @f6(double %f1, ptr %base, i64 %index, ptr %dst) #0 { |
| ; CHECK-LABEL: f6: |
| ; CHECK: sllg %r1, %r3, 3 |
| ; CHECK: mxdb %f0, 800(%r1,%r2) |
| ; CHECK: std %f0, 0(%r4) |
| ; CHECK: std %f2, 8(%r4) |
| ; CHECK: br %r14 |
| %ptr1 = getelementptr double, ptr %base, i64 %index |
| %ptr2 = getelementptr double, ptr %ptr1, i64 100 |
| %f2 = load double, ptr %ptr2 |
| %f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f1, |
| metadata !"fpexcept.strict") #0 |
| %f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %f2, |
| metadata !"fpexcept.strict") #0 |
| %res = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %f1x, fp128 %f2x, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| store fp128 %res, ptr %dst |
| ret void |
| } |
| |
| ; Check that multiplications of spilled values can use MXDB rather than MXDBR. |
| define double @f7(ptr %ptr0) #0 { |
| ; CHECK-LABEL: f7: |
| ; CHECK: brasl %r14, foo@PLT |
| ; CHECK: mxdb %f0, 160(%r15) |
| ; CHECK: br %r14 |
| %ptr1 = getelementptr double, ptr %ptr0, i64 2 |
| %ptr2 = getelementptr double, ptr %ptr0, i64 4 |
| %ptr3 = getelementptr double, ptr %ptr0, i64 6 |
| %ptr4 = getelementptr double, ptr %ptr0, i64 8 |
| %ptr5 = getelementptr double, ptr %ptr0, i64 10 |
| %ptr6 = getelementptr double, ptr %ptr0, i64 12 |
| %ptr7 = getelementptr double, ptr %ptr0, i64 14 |
| %ptr8 = getelementptr double, ptr %ptr0, i64 16 |
| %ptr9 = getelementptr double, ptr %ptr0, i64 18 |
| %ptr10 = getelementptr double, ptr %ptr0, i64 20 |
| |
| %val0 = load double, ptr %ptr0 |
| %val1 = load double, ptr %ptr1 |
| %val2 = load double, ptr %ptr2 |
| %val3 = load double, ptr %ptr3 |
| %val4 = load double, ptr %ptr4 |
| %val5 = load double, ptr %ptr5 |
| %val6 = load double, ptr %ptr6 |
| %val7 = load double, ptr %ptr7 |
| %val8 = load double, ptr %ptr8 |
| %val9 = load double, ptr %ptr9 |
| %val10 = load double, ptr %ptr10 |
| |
| %frob0 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val0, double %val0, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob1 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val1, double %val1, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob2 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val2, double %val2, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob3 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val3, double %val3, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob4 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val4, double %val4, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob5 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val5, double %val5, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob6 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val6, double %val6, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob7 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val7, double %val7, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob8 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val8, double %val8, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob9 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val9, double %val9, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %frob10 = call double @llvm.experimental.constrained.fadd.f64( |
| double %val10, double %val10, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| store double %frob0, ptr %ptr0 |
| store double %frob1, ptr %ptr1 |
| store double %frob2, ptr %ptr2 |
| store double %frob3, ptr %ptr3 |
| store double %frob4, ptr %ptr4 |
| store double %frob5, ptr %ptr5 |
| store double %frob6, ptr %ptr6 |
| store double %frob7, ptr %ptr7 |
| store double %frob8, ptr %ptr8 |
| store double %frob9, ptr %ptr9 |
| store double %frob10, ptr %ptr10 |
| |
| %ret = call double @foo() #0 |
| |
| %accext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %ret, |
| metadata !"fpexcept.strict") #0 |
| %ext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob0, |
| metadata !"fpexcept.strict") #0 |
| %mul0 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext0, fp128 %ext0, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra0 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul0, fp128 0xL00000000000000003fff000001000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc0 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra0, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc0, |
| metadata !"fpexcept.strict") #0 |
| %ext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob1, |
| metadata !"fpexcept.strict") #0 |
| %mul1 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext1, fp128 %ext1, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra1 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul1, fp128 0xL00000000000000003fff000002000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc1 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra1, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc1, |
| metadata !"fpexcept.strict") #0 |
| %ext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob2, |
| metadata !"fpexcept.strict") #0 |
| %mul2 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext2, fp128 %ext2, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra2 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul2, fp128 0xL00000000000000003fff000003000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc2 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra2, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc2, |
| metadata !"fpexcept.strict") #0 |
| %ext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob3, |
| metadata !"fpexcept.strict") #0 |
| %mul3 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext3, fp128 %ext3, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra3 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul3, fp128 0xL00000000000000003fff000004000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc3 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra3, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc3, |
| metadata !"fpexcept.strict") #0 |
| %ext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob4, |
| metadata !"fpexcept.strict") #0 |
| %mul4 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext4, fp128 %ext4, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra4 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul4, fp128 0xL00000000000000003fff000005000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc4 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra4, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc4, |
| metadata !"fpexcept.strict") #0 |
| %ext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob5, |
| metadata !"fpexcept.strict") #0 |
| %mul5 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext5, fp128 %ext5, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra5 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul5, fp128 0xL00000000000000003fff000006000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc5 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra5, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc5, |
| metadata !"fpexcept.strict") #0 |
| %ext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob6, |
| metadata !"fpexcept.strict") #0 |
| %mul6 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext6, fp128 %ext6, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra6 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul6, fp128 0xL00000000000000003fff000007000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc6 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra6, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc6, |
| metadata !"fpexcept.strict") #0 |
| %ext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob7, |
| metadata !"fpexcept.strict") #0 |
| %mul7 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext7, fp128 %ext7, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra7 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul7, fp128 0xL00000000000000003fff000008000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc7 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra7, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc7, |
| metadata !"fpexcept.strict") #0 |
| %ext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob8, |
| metadata !"fpexcept.strict") #0 |
| %mul8 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext8, fp128 %ext8, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra8 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul8, fp128 0xL00000000000000003fff000009000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc8 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra8, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| %accext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %trunc8, |
| metadata !"fpexcept.strict") #0 |
| %ext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64( |
| double %frob9, |
| metadata !"fpexcept.strict") #0 |
| %mul9 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %accext9, fp128 %ext9, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %extra9 = call fp128 @llvm.experimental.constrained.fmul.f128( |
| fp128 %mul9, fp128 0xL00000000000000003fff00000a000000, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| %trunc9 = call double @llvm.experimental.constrained.fptrunc.f64.f128( |
| fp128 %extra9, |
| metadata !"round.dynamic", |
| metadata !"fpexcept.strict") #0 |
| |
| ret double %trunc9 |
| } |
| |
| attributes #0 = { strictfp } |