| # NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py |
| # RUN: llc -mtriple=aarch64-unknown-unknown -verify-machineinstrs -run-pass=legalizer %s -o - | FileCheck %s |
| # RUN: llc -mtriple=aarch64-unknown-unknown -verify-machineinstrs -run-pass=legalizer -mattr=+cssc %s -o - | FileCheck %s --check-prefix=CHECK-CSSC |
| ... |
| --- |
| name: v8s8_legal |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $d0 |
| ; CHECK-LABEL: name: v8s8_legal |
| ; CHECK: liveins: $d0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(<8 x s8>) = COPY $d0 |
| ; CHECK-NEXT: %ctpop:_(<8 x s8>) = G_CTPOP %copy(<8 x s8>) |
| ; CHECK-NEXT: $d0 = COPY %ctpop(<8 x s8>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $d0 |
| ; |
| ; CHECK-CSSC-LABEL: name: v8s8_legal |
| ; CHECK-CSSC: liveins: $d0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(<8 x s8>) = COPY $d0 |
| ; CHECK-CSSC-NEXT: %ctpop:_(<8 x s8>) = G_CTPOP %copy(<8 x s8>) |
| ; CHECK-CSSC-NEXT: $d0 = COPY %ctpop(<8 x s8>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $d0 |
| %copy:_(<8 x s8>) = COPY $d0 |
| %ctpop:_(<8 x s8>) = G_CTPOP %copy(<8 x s8>) |
| $d0 = COPY %ctpop(<8 x s8>) |
| RET_ReallyLR implicit $d0 |
| |
| ... |
| --- |
| name: v16s8_legal |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $q0 |
| ; CHECK-LABEL: name: v16s8_legal |
| ; CHECK: liveins: $q0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(<16 x s8>) = COPY $q0 |
| ; CHECK-NEXT: %ctpop:_(<16 x s8>) = G_CTPOP %copy(<16 x s8>) |
| ; CHECK-NEXT: $q0 = COPY %ctpop(<16 x s8>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $q0 |
| ; |
| ; CHECK-CSSC-LABEL: name: v16s8_legal |
| ; CHECK-CSSC: liveins: $q0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(<16 x s8>) = COPY $q0 |
| ; CHECK-CSSC-NEXT: %ctpop:_(<16 x s8>) = G_CTPOP %copy(<16 x s8>) |
| ; CHECK-CSSC-NEXT: $q0 = COPY %ctpop(<16 x s8>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $q0 |
| %copy:_(<16 x s8>) = COPY $q0 |
| %ctpop:_(<16 x s8>) = G_CTPOP %copy(<16 x s8>) |
| $q0 = COPY %ctpop(<16 x s8>) |
| RET_ReallyLR implicit $q0 |
| |
| ... |
| --- |
| name: s32_lower |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $w0 |
| ; CHECK-LABEL: name: s32_lower |
| ; CHECK: liveins: $w0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-NEXT: [[ZEXT:%[0-9]+]]:_(s64) = G_ZEXT %copy(s32) |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[ZEXT]](s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: %ctpop:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: $w0 = COPY %ctpop(s32) |
| ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| ; |
| ; CHECK-CSSC-LABEL: name: s32_lower |
| ; CHECK-CSSC: liveins: $w0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-CSSC-NEXT: %ctpop:_(s32) = G_CTPOP %copy(s32) |
| ; CHECK-CSSC-NEXT: $w0 = COPY %ctpop(s32) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $w0 |
| %copy:_(s32) = COPY $w0 |
| %ctpop:_(s32) = G_CTPOP %copy(s32) |
| $w0 = COPY %ctpop(s32) |
| RET_ReallyLR implicit $w0 |
| |
| ... |
| --- |
| name: s64_lower |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $x0 |
| ; CHECK-LABEL: name: s64_lower |
| ; CHECK: liveins: $x0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s64) = COPY $x0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST %copy(s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: %ctpop:_(s64) = G_ZEXT [[INT]](s32) |
| ; CHECK-NEXT: $x0 = COPY %ctpop(s64) |
| ; CHECK-NEXT: RET_ReallyLR implicit $x0 |
| ; |
| ; CHECK-CSSC-LABEL: name: s64_lower |
| ; CHECK-CSSC: liveins: $x0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s64) = COPY $x0 |
| ; CHECK-CSSC-NEXT: %ctpop:_(s64) = G_CTPOP %copy(s64) |
| ; CHECK-CSSC-NEXT: $x0 = COPY %ctpop(s64) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $x0 |
| %copy:_(s64) = COPY $x0 |
| %ctpop:_(s64) = G_CTPOP %copy(s64) |
| $x0 = COPY %ctpop(s64) |
| RET_ReallyLR implicit $x0 |
| |
| ... |
| --- |
| name: s128_lower |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $x0, $x1 |
| |
| ; CHECK-LABEL: name: s128_lower |
| ; CHECK: liveins: $x0, $x1 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0 |
| ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1 |
| ; CHECK-NEXT: [[MV:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[COPY]](s64), [[COPY1]](s64) |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[MV]](s128) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<16 x s8>) |
| ; CHECK-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 0 |
| ; CHECK-NEXT: [[MV1:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[INT]](s32), [[C]](s32) |
| ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| ; CHECK-NEXT: $x0 = COPY [[MV1]](s64) |
| ; CHECK-NEXT: $x1 = COPY [[C1]](s64) |
| ; CHECK-NEXT: RET_ReallyLR implicit $x0, implicit $x1 |
| ; |
| ; CHECK-CSSC-LABEL: name: s128_lower |
| ; CHECK-CSSC: liveins: $x0, $x1 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0 |
| ; CHECK-CSSC-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1 |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(s64) = G_CTPOP [[COPY]](s64) |
| ; CHECK-CSSC-NEXT: [[CTPOP1:%[0-9]+]]:_(s64) = G_CTPOP [[COPY1]](s64) |
| ; CHECK-CSSC-NEXT: [[ADD:%[0-9]+]]:_(s64) = G_ADD [[CTPOP]], [[CTPOP1]] |
| ; CHECK-CSSC-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| ; CHECK-CSSC-NEXT: $x0 = COPY [[ADD]](s64) |
| ; CHECK-CSSC-NEXT: $x1 = COPY [[C]](s64) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $x0, implicit $x1 |
| %1:_(s64) = COPY $x0 |
| %2:_(s64) = COPY $x1 |
| %0:_(s128) = G_MERGE_VALUES %1(s64), %2(s64) |
| %3:_(s128) = G_CTPOP %0(s128) |
| %4:_(s64), %5:_(s64) = G_UNMERGE_VALUES %3(s128) |
| $x0 = COPY %4(s64) |
| $x1 = COPY %5(s64) |
| RET_ReallyLR implicit $x0, implicit $x1 |
| |
| ... |
| --- |
| name: widen_s16 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $w0 |
| |
| ; CHECK-LABEL: name: widen_s16 |
| ; CHECK: liveins: $w0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-NEXT: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT %copy(s32) |
| ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 65535 |
| ; CHECK-NEXT: [[AND:%[0-9]+]]:_(s64) = G_AND [[ANYEXT]], [[C]] |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[AND]](s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[INT]](s32) |
| ; CHECK-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| ; |
| ; CHECK-CSSC-LABEL: name: widen_s16 |
| ; CHECK-CSSC: liveins: $w0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-CSSC-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 65535 |
| ; CHECK-CSSC-NEXT: [[AND:%[0-9]+]]:_(s32) = G_AND %copy, [[C]] |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(s32) = G_CTPOP [[AND]](s32) |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[CTPOP]](s32) |
| ; CHECK-CSSC-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $w0 |
| %copy:_(s32) = COPY $w0 |
| %trunc:_(s16) = G_TRUNC %copy(s32) |
| %ctpop:_(s16) = G_CTPOP %trunc(s16) |
| %ext:_(s32) = G_ANYEXT %ctpop(s16) |
| $w0 = COPY %ext(s32) |
| RET_ReallyLR implicit $w0 |
| |
| ... |
| --- |
| name: widen_s8 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $w0 |
| |
| ; CHECK-LABEL: name: widen_s8 |
| ; CHECK: liveins: $w0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-NEXT: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT %copy(s32) |
| ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 255 |
| ; CHECK-NEXT: [[AND:%[0-9]+]]:_(s64) = G_AND [[ANYEXT]], [[C]] |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[AND]](s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[INT]](s32) |
| ; CHECK-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| ; |
| ; CHECK-CSSC-LABEL: name: widen_s8 |
| ; CHECK-CSSC: liveins: $w0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-CSSC-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 255 |
| ; CHECK-CSSC-NEXT: [[AND:%[0-9]+]]:_(s32) = G_AND %copy, [[C]] |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(s32) = G_CTPOP [[AND]](s32) |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[CTPOP]](s32) |
| ; CHECK-CSSC-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $w0 |
| %copy:_(s32) = COPY $w0 |
| %trunc:_(s8) = G_TRUNC %copy(s32) |
| %ctpop:_(s8) = G_CTPOP %trunc(s8) |
| %ext:_(s32) = G_ANYEXT %ctpop(s8) |
| $w0 = COPY %ext(s32) |
| RET_ReallyLR implicit $w0 |
| |
| ... |
| --- |
| name: widen_s3 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $w0 |
| |
| ; CHECK-LABEL: name: widen_s3 |
| ; CHECK: liveins: $w0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-NEXT: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT %copy(s32) |
| ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 7 |
| ; CHECK-NEXT: [[AND:%[0-9]+]]:_(s64) = G_AND [[ANYEXT]], [[C]] |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[AND]](s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[INT]](s32) |
| ; CHECK-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| ; |
| ; CHECK-CSSC-LABEL: name: widen_s3 |
| ; CHECK-CSSC: liveins: $w0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-CSSC-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 7 |
| ; CHECK-CSSC-NEXT: [[AND:%[0-9]+]]:_(s32) = G_AND %copy, [[C]] |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(s32) = G_CTPOP [[AND]](s32) |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[CTPOP]](s32) |
| ; CHECK-CSSC-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $w0 |
| %copy:_(s32) = COPY $w0 |
| %trunc:_(s3) = G_TRUNC %copy(s32) |
| %ctpop:_(s3) = G_CTPOP %trunc(s3) |
| %ext:_(s32) = G_ANYEXT %ctpop(s3) |
| $w0 = COPY %ext(s32) |
| RET_ReallyLR implicit $w0 |
| |
| ... |
| --- |
| name: different_sizes |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $w0 |
| ; CHECK-LABEL: name: different_sizes |
| ; CHECK: liveins: $w0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-NEXT: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT %copy(s32) |
| ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 255 |
| ; CHECK-NEXT: [[AND:%[0-9]+]]:_(s64) = G_AND [[ANYEXT]], [[C]] |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[AND]](s64) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(s32) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlv), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[INT]](s32) |
| ; CHECK-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| ; |
| ; CHECK-CSSC-LABEL: name: different_sizes |
| ; CHECK-CSSC: liveins: $w0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: %copy:_(s32) = COPY $w0 |
| ; CHECK-CSSC-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 255 |
| ; CHECK-CSSC-NEXT: [[AND:%[0-9]+]]:_(s32) = G_AND %copy, [[C]] |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(s32) = G_CTPOP [[AND]](s32) |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(s32) = COPY [[CTPOP]](s32) |
| ; CHECK-CSSC-NEXT: $w0 = COPY [[COPY]](s32) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $w0 |
| %copy:_(s32) = COPY $w0 |
| %trunc:_(s8) = G_TRUNC %copy(s32) |
| %ctpop:_(s16) = G_CTPOP %trunc(s8) |
| %ext:_(s32) = G_ANYEXT %ctpop(s16) |
| $w0 = COPY %ext(s32) |
| RET_ReallyLR implicit $w0 |
| |
| ... |
| --- |
| name: custom_8x16 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $q0 |
| |
| ; CHECK-LABEL: name: custom_8x16 |
| ; CHECK: liveins: $q0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<8 x s16>) = COPY $q0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<8 x s16>) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-NEXT: $q0 = COPY [[INT]](<8 x s16>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $q0 |
| ; |
| ; CHECK-CSSC-LABEL: name: custom_8x16 |
| ; CHECK-CSSC: liveins: $q0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(<8 x s16>) = COPY $q0 |
| ; CHECK-CSSC-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<8 x s16>) |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: $q0 = COPY [[INT]](<8 x s16>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $q0 |
| %0:_(<8 x s16>) = COPY $q0 |
| %1:_(<8 x s16>) = G_CTPOP %0(<8 x s16>) |
| $q0 = COPY %1(<8 x s16>) |
| RET_ReallyLR implicit $q0 |
| |
| ... |
| --- |
| name: custom_4x32 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $q0 |
| |
| ; CHECK-LABEL: name: custom_4x32 |
| ; CHECK: liveins: $q0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<4 x s32>) = COPY $q0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<4 x s32>) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-NEXT: [[INT1:%[0-9]+]]:_(<4 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<8 x s16>) |
| ; CHECK-NEXT: $q0 = COPY [[INT1]](<4 x s32>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $q0 |
| ; |
| ; CHECK-CSSC-LABEL: name: custom_4x32 |
| ; CHECK-CSSC: liveins: $q0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(<4 x s32>) = COPY $q0 |
| ; CHECK-CSSC-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<4 x s32>) |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT1:%[0-9]+]]:_(<4 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<8 x s16>) |
| ; CHECK-CSSC-NEXT: $q0 = COPY [[INT1]](<4 x s32>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $q0 |
| %0:_(<4 x s32>) = COPY $q0 |
| %1:_(<4 x s32>) = G_CTPOP %0(<4 x s32>) |
| $q0 = COPY %1(<4 x s32>) |
| RET_ReallyLR implicit $q0 |
| |
| ... |
| --- |
| name: custom_2x64 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $q0 |
| |
| ; CHECK-LABEL: name: custom_2x64 |
| ; CHECK: liveins: $q0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<2 x s64>) = COPY $q0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<2 x s64>) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-NEXT: [[INT1:%[0-9]+]]:_(<4 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<8 x s16>) |
| ; CHECK-NEXT: [[INT2:%[0-9]+]]:_(<2 x s64>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT1]](<4 x s32>) |
| ; CHECK-NEXT: $q0 = COPY [[INT2]](<2 x s64>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $q0 |
| ; |
| ; CHECK-CSSC-LABEL: name: custom_2x64 |
| ; CHECK-CSSC: liveins: $q0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(<2 x s64>) = COPY $q0 |
| ; CHECK-CSSC-NEXT: [[BITCAST:%[0-9]+]]:_(<16 x s8>) = G_BITCAST [[COPY]](<2 x s64>) |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(<16 x s8>) = G_CTPOP [[BITCAST]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT:%[0-9]+]]:_(<8 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<16 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT1:%[0-9]+]]:_(<4 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<8 x s16>) |
| ; CHECK-CSSC-NEXT: [[INT2:%[0-9]+]]:_(<2 x s64>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT1]](<4 x s32>) |
| ; CHECK-CSSC-NEXT: $q0 = COPY [[INT2]](<2 x s64>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $q0 |
| %0:_(<2 x s64>) = COPY $q0 |
| %1:_(<2 x s64>) = G_CTPOP %0(<2 x s64>) |
| $q0 = COPY %1(<2 x s64>) |
| RET_ReallyLR implicit $q0 |
| |
| ... |
| --- |
| name: custom_4x16 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $d0 |
| |
| ; CHECK-LABEL: name: custom_4x16 |
| ; CHECK: liveins: $d0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<4 x s16>) = COPY $d0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[COPY]](<4 x s16>) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(<4 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: $d0 = COPY [[INT]](<4 x s16>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $d0 |
| ; |
| ; CHECK-CSSC-LABEL: name: custom_4x16 |
| ; CHECK-CSSC: liveins: $d0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(<4 x s16>) = COPY $d0 |
| ; CHECK-CSSC-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[COPY]](<4 x s16>) |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT:%[0-9]+]]:_(<4 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<8 x s8>) |
| ; CHECK-CSSC-NEXT: $d0 = COPY [[INT]](<4 x s16>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $d0 |
| %0:_(<4 x s16>) = COPY $d0 |
| %1:_(<4 x s16>) = G_CTPOP %0(<4 x s16>) |
| $d0 = COPY %1(<4 x s16>) |
| RET_ReallyLR implicit $d0 |
| |
| ... |
| --- |
| name: custom_2x32 |
| tracksRegLiveness: true |
| body: | |
| bb.0: |
| liveins: $d0 |
| |
| ; CHECK-LABEL: name: custom_2x32 |
| ; CHECK: liveins: $d0 |
| ; CHECK-NEXT: {{ $}} |
| ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<2 x s32>) = COPY $d0 |
| ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[COPY]](<2 x s32>) |
| ; CHECK-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-NEXT: [[INT:%[0-9]+]]:_(<4 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<8 x s8>) |
| ; CHECK-NEXT: [[INT1:%[0-9]+]]:_(<2 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<4 x s16>) |
| ; CHECK-NEXT: $d0 = COPY [[INT1]](<2 x s32>) |
| ; CHECK-NEXT: RET_ReallyLR implicit $d0 |
| ; |
| ; CHECK-CSSC-LABEL: name: custom_2x32 |
| ; CHECK-CSSC: liveins: $d0 |
| ; CHECK-CSSC-NEXT: {{ $}} |
| ; CHECK-CSSC-NEXT: [[COPY:%[0-9]+]]:_(<2 x s32>) = COPY $d0 |
| ; CHECK-CSSC-NEXT: [[BITCAST:%[0-9]+]]:_(<8 x s8>) = G_BITCAST [[COPY]](<2 x s32>) |
| ; CHECK-CSSC-NEXT: [[CTPOP:%[0-9]+]]:_(<8 x s8>) = G_CTPOP [[BITCAST]](<8 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT:%[0-9]+]]:_(<4 x s16>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[CTPOP]](<8 x s8>) |
| ; CHECK-CSSC-NEXT: [[INT1:%[0-9]+]]:_(<2 x s32>) = G_INTRINSIC intrinsic(@llvm.aarch64.neon.uaddlp), [[INT]](<4 x s16>) |
| ; CHECK-CSSC-NEXT: $d0 = COPY [[INT1]](<2 x s32>) |
| ; CHECK-CSSC-NEXT: RET_ReallyLR implicit $d0 |
| %0:_(<2 x s32>) = COPY $d0 |
| %1:_(<2 x s32>) = G_CTPOP %0(<2 x s32>) |
| $d0 = COPY %1(<2 x s32>) |
| RET_ReallyLR implicit $d0 |