| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 4 |
| ; Test vector add reduction intrinsic |
| ; |
| ; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z16 | FileCheck %s |
| |
| ; 1 vector length |
| declare i8 @llvm.vector.reduce.add.v16i8(<16 x i8> %a) |
| declare i16 @llvm.vector.reduce.add.v8i16(<8 x i16> %a) |
| declare i32 @llvm.vector.reduce.add.v4i32(<4 x i32> %a) |
| declare i64 @llvm.vector.reduce.add.v2i64(<2 x i64> %a) |
| declare i128 @llvm.vector.reduce.add.v1i128(<1 x i128> %a) |
| ; 2 vector lengths |
| declare i8 @llvm.vector.reduce.add.v32i8(<32 x i8> %a) |
| declare i16 @llvm.vector.reduce.add.v16i16(<16 x i16> %a) |
| declare i32 @llvm.vector.reduce.add.v8i32(<8 x i32> %a) |
| declare i64 @llvm.vector.reduce.add.v4i64(<4 x i64> %a) |
| declare i128 @llvm.vector.reduce.add.v2i128(<2 x i128> %a) |
| ; ; TODO |
| ; ; 4 vector lengths |
| declare i8 @llvm.vector.reduce.add.v64i8(<64 x i8> %a) |
| declare i16 @llvm.vector.reduce.add.v32i16(<32 x i16> %a) |
| declare i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %a) |
| declare i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %a) |
| declare i128 @llvm.vector.reduce.add.v4i128(<4 x i128> %a) |
| ; ; Subvector lengths |
| declare i8 @llvm.vector.reduce.add.v8i8(<8 x i8> %a) |
| declare i16 @llvm.vector.reduce.add.v4i16(<4 x i16> %a) |
| declare i32 @llvm.vector.reduce.add.v2i32(<2 x i32> %a) |
| declare i64 @llvm.vector.reduce.add.v1i64(<1 x i64> %a) |
| |
| define i8 @f1_1(<16 x i8> %a) { |
| ; CHECK-LABEL: f1_1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vgbm %v0, 0 |
| ; CHECK-NEXT: vsumb %v1, %v24, %v0 |
| ; CHECK-NEXT: vsumqf %v0, %v1, %v0 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i8 @llvm.vector.reduce.add.v16i8(<16 x i8> %a) |
| ret i8 %redadd |
| } |
| |
| define i16 @f1_2(<8 x i16> %a) { |
| ; CHECK-LABEL: f1_2: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vgbm %v0, 0 |
| ; CHECK-NEXT: vsumh %v1, %v24, %v0 |
| ; CHECK-NEXT: vsumqf %v0, %v1, %v0 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i16 @llvm.vector.reduce.add.v8i16(<8 x i16> %a) |
| ret i16 %redadd |
| } |
| |
| define i32 @f1_3(<4 x i32> %a) { |
| ; CHECK-LABEL: f1_3: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vgbm %v0, 0 |
| ; CHECK-NEXT: vsumqf %v0, %v24, %v0 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> %a) |
| ret i32 %redadd |
| } |
| |
| define i64 @f1_4(<2 x i64> %a) { |
| ; CHECK-LABEL: f1_4: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vrepg %v0, %v24, 1 |
| ; CHECK-NEXT: vag %v0, %v24, %v0 |
| ; CHECK-NEXT: vlgvg %r2, %v0, 0 |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i64 @llvm.vector.reduce.add.v2i64(<2 x i64> %a) |
| ret i64 %redadd |
| } |
| |
| define i128 @f1_5(<1 x i128> %a) { |
| ; CHECK-LABEL: f1_5: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vst %v24, 0(%r2), 3 |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i128 @llvm.vector.reduce.add.v1i128(<1 x i128> %a) |
| ret i128 %redadd |
| } |
| |
| define i8 @f2_1(<32 x i8> %a) { |
| ; CHECK-LABEL: f2_1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vab %v0, %v24, %v26 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumb %v0, %v0, %v1 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i8 @llvm.vector.reduce.add.v32i8(<32 x i8> %a) |
| ret i8 %redadd |
| } |
| |
| define i16 @f2_2(<16 x i16> %a) { |
| ; CHECK-LABEL: f2_2: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vah %v0, %v24, %v26 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumh %v0, %v0, %v1 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i16 @llvm.vector.reduce.add.v16i16(<16 x i16> %a) |
| ret i16 %redadd |
| } |
| |
| define i32 @f2_3(<8 x i32> %a) { |
| ; CHECK-LABEL: f2_3: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vaf %v0, %v24, %v26 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> %a) |
| ret i32 %redadd |
| } |
| |
| define i64 @f2_4(<4 x i64> %a) { |
| ; CHECK-LABEL: f2_4: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vag %v0, %v24, %v26 |
| ; CHECK-NEXT: vrepg %v1, %v0, 1 |
| ; CHECK-NEXT: vag %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvg %r2, %v0, 0 |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i64 @llvm.vector.reduce.add.v4i64(<4 x i64> %a) |
| ret i64 %redadd |
| } |
| |
| define i128 @f2_5(<2 x i128> %a) { |
| ; CHECK-LABEL: f2_5: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vl %v0, 16(%r3), 3 |
| ; CHECK-NEXT: vl %v1, 0(%r3), 3 |
| ; CHECK-NEXT: vaq %v0, %v1, %v0 |
| ; CHECK-NEXT: vst %v0, 0(%r2), 3 |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i128 @llvm.vector.reduce.add.v2i128(<2 x i128> %a) |
| ret i128 %redadd |
| } |
| |
| define i8 @f3_1(<64 x i8> %a) { |
| ; CHECK-LABEL: f3_1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vab %v0, %v26, %v30 |
| ; CHECK-NEXT: vab %v1, %v24, %v28 |
| ; CHECK-NEXT: vab %v0, %v1, %v0 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumb %v0, %v0, %v1 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i8 @llvm.vector.reduce.add.v64i8(<64 x i8> %a) |
| ret i8 %redadd |
| } |
| |
| define i16 @f3_2(<32 x i16> %a) { |
| ; CHECK-LABEL: f3_2: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vah %v0, %v26, %v30 |
| ; CHECK-NEXT: vah %v1, %v24, %v28 |
| ; CHECK-NEXT: vah %v0, %v1, %v0 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumh %v0, %v0, %v1 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i16 @llvm.vector.reduce.add.v32i16(<32 x i16> %a) |
| ret i16 %redadd |
| } |
| |
| define i32 @f3_3(<16 x i32> %a) { |
| ; CHECK-LABEL: f3_3: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vaf %v0, %v26, %v30 |
| ; CHECK-NEXT: vaf %v1, %v24, %v28 |
| ; CHECK-NEXT: vaf %v0, %v1, %v0 |
| ; CHECK-NEXT: vgbm %v1, 0 |
| ; CHECK-NEXT: vsumqf %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 3 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %a) |
| ret i32 %redadd |
| } |
| |
| define i64 @f3_4(<8 x i64> %a) { |
| ; CHECK-LABEL: f3_4: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vag %v0, %v26, %v30 |
| ; CHECK-NEXT: vag %v1, %v24, %v28 |
| ; CHECK-NEXT: vag %v0, %v1, %v0 |
| ; CHECK-NEXT: vrepg %v1, %v0, 1 |
| ; CHECK-NEXT: vag %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvg %r2, %v0, 0 |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %a) |
| ret i64 %redadd |
| } |
| |
| define i128 @f3_5(<4 x i128> %a) { |
| ; CHECK-LABEL: f3_5: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vl %v0, 32(%r3), 3 |
| ; CHECK-NEXT: vl %v1, 0(%r3), 3 |
| ; CHECK-NEXT: vl %v2, 48(%r3), 3 |
| ; CHECK-NEXT: vl %v3, 16(%r3), 3 |
| ; CHECK-NEXT: vaq %v2, %v3, %v2 |
| ; CHECK-NEXT: vaq %v0, %v1, %v0 |
| ; CHECK-NEXT: vaq %v0, %v0, %v2 |
| ; CHECK-NEXT: vst %v0, 0(%r2), 3 |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i128 @llvm.vector.reduce.add.v4i128(<4 x i128> %a) |
| ret i128 %redadd |
| } |
| |
| |
| define i8 @f4_1(<8 x i8> %a) { |
| ; CHECK-LABEL: f4_1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vpkg %v0, %v24, %v24 |
| ; CHECK-NEXT: vab %v0, %v24, %v0 |
| ; CHECK-NEXT: vpkf %v1, %v0, %v0 |
| ; CHECK-NEXT: vab %v0, %v0, %v1 |
| ; CHECK-NEXT: vrepb %v1, %v0, 1 |
| ; CHECK-NEXT: vab %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvb %r2, %v0, 0 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i8 @llvm.vector.reduce.add.v8i8(<8 x i8> %a) |
| ret i8 %redadd |
| } |
| |
| define i16 @f4_2(<4 x i16> %a) { |
| ; CHECK-LABEL: f4_2: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vpkg %v0, %v24, %v24 |
| ; CHECK-NEXT: vah %v0, %v24, %v0 |
| ; CHECK-NEXT: vreph %v1, %v0, 1 |
| ; CHECK-NEXT: vah %v0, %v0, %v1 |
| ; CHECK-NEXT: vlgvh %r2, %v0, 0 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| %redadd = call i16 @llvm.vector.reduce.add.v4i16(<4 x i16> %a) |
| ret i16 %redadd |
| } |
| |
| define i32 @f4_3(<2 x i32> %a) { |
| ; CHECK-LABEL: f4_3: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vrepf %v0, %v24, 1 |
| ; CHECK-NEXT: vaf %v0, %v24, %v0 |
| ; CHECK-NEXT: vlgvf %r2, %v0, 0 |
| ; CHECK-NEXT: # kill: def $r2l killed $r2l killed $r2d |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i32 @llvm.vector.reduce.add.v2i32(<2 x i32> %a) |
| ret i32 %redadd |
| } |
| |
| define i64 @f4_4(<1 x i64> %a) { |
| ; CHECK-LABEL: f4_4: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vlgvg %r2, %v24, 0 |
| ; CHECK-NEXT: br %r14 |
| |
| %redadd = call i64 @llvm.vector.reduce.add.v1i64(<1 x i64> %a) |
| ret i64 %redadd |
| } |