blob: 1653d9157732f079c07b1c8bdabde8c5ef366be7 [file] [log] [blame]
# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py
# RUN: llc -mtriple aarch64-apple-darwin -debugify-and-strip-all-safe -run-pass=aarch64-prelegalizer-combiner -global-isel -verify-machineinstrs %s -o - | FileCheck %s
# Check that we propagate the G_SEXT to the sources of the phi operand.
---
name: sext_icst_through_phi
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: sext_icst_through_phi
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_4:_(s32) = G_CONSTANT i32 4
; CHECK: [[SEXT:%[0-9]+]]:_(s64) = G_SEXT %cst32_4(s32)
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: [[SEXT1:%[0-9]+]]:_(s64) = G_SEXT %cst32_10(s32)
; CHECK: bb.3:
; CHECK: %ext:_(s64) = G_PHI [[SEXT]](s64), %bb.1, [[SEXT1]](s64), %bb.2
; CHECK: $x0 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%cst32_4:_(s32) = G_CONSTANT i32 4
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %cst32_4(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_SEXT %phi
$x0 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...
# Check that we propagate the G_ZEXT to the sources of the phi operand.
---
name: zext_icst_through_phi
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: zext_icst_through_phi
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_4:_(s32) = G_CONSTANT i32 4
; CHECK: [[ZEXT:%[0-9]+]]:_(s64) = G_ZEXT %cst32_4(s32)
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: [[ZEXT1:%[0-9]+]]:_(s64) = G_ZEXT %cst32_10(s32)
; CHECK: bb.3:
; CHECK: %ext:_(s64) = G_PHI [[ZEXT]](s64), %bb.1, [[ZEXT1]](s64), %bb.2
; CHECK: $x0 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%cst32_4:_(s32) = G_CONSTANT i32 4
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %cst32_4(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_ZEXT %phi
$x0 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...
# Don't handle vectors because of potential cost issues.
---
name: sext_load_through_phi_vector
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: sext_load_through_phi_vector
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $x0, $q0, $q1
; CHECK: %ptr:_(p0) = COPY $x0
; CHECK: %cmp:_(s1) = G_IMPLICIT_DEF
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %ld1:_(<4 x s32>) = G_LOAD %ptr(p0) :: (load 16)
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %ld2:_(<4 x s32>) = G_LOAD %ptr(p0) :: (load 16)
; CHECK: bb.3:
; CHECK: %phi:_(<4 x s32>) = G_PHI %ld1(<4 x s32>), %bb.1, %ld2(<4 x s32>), %bb.2
; CHECK: %ext:_(<4 x s64>) = G_SEXT %phi(<4 x s32>)
; CHECK: G_STORE %ext(<4 x s64>), %ptr(p0) :: (store 16)
; CHECK: RET_ReallyLR
bb.1.entry:
liveins: $x0, $q0, $q1
%0:_(<4 x s32>) = COPY $q0
%1:_(<4 x s32>) = COPY $q1
%ptr:_(p0) = COPY $x0
%cmp:_(s1) = G_IMPLICIT_DEF
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%ld1:_(<4 x s32>) = G_LOAD %ptr(p0) :: (load 16)
G_BR %bb.4
bb.3:
%ld2:_(<4 x s32>) = G_LOAD %ptr(p0) :: (load 16)
bb.4:
%phi:_(<4 x s32>) = G_PHI %ld1(<4 x s32>), %bb.2, %ld2(<4 x s32>), %bb.3
%ext:_(<4 x s64>) = G_SEXT %phi
G_STORE %ext(<4 x s64>), %ptr(p0) :: (store 16)
RET_ReallyLR
...
# Check that we don't propagate if the extend is used by a G_PTR_ADD, which on
# AArch64 has a good chance of folding in the extend.
---
name: sext_icst_through_phi_used_by_ptradd
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: sext_icst_through_phi_used_by_ptradd
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1, $x2
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: %base:_(p0) = COPY $x2
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_4:_(s32) = G_CONSTANT i32 4
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: bb.3:
; CHECK: %phi:_(s32) = G_PHI %cst32_4(s32), %bb.1, %cst32_10(s32), %bb.2
; CHECK: %ext:_(s64) = G_SEXT %phi(s32)
; CHECK: %ptr:_(p0) = G_PTR_ADD %base, %ext(s64)
; CHECK: $x0 = COPY %ptr(p0)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1, $x2
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%base:_(p0) = COPY $x2
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%cst32_4:_(s32) = G_CONSTANT i32 4
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %cst32_4(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_SEXT %phi
%ptr:_(p0) = G_PTR_ADD %base, %ext
$x0 = COPY %ptr(p0)
RET_ReallyLR implicit $x0
...
# Same as above but we do it here because the extend has multiple users, so the
# it probably won't cost extra instructions if we remove it.
---
name: sext_icst_through_phi_used_by_ptradd_multiuse
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: sext_icst_through_phi_used_by_ptradd_multiuse
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1, $x2
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: %base:_(p0) = COPY $x2
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_4:_(s32) = G_CONSTANT i32 4
; CHECK: [[SEXT:%[0-9]+]]:_(s64) = G_SEXT %cst32_4(s32)
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: [[SEXT1:%[0-9]+]]:_(s64) = G_SEXT %cst32_10(s32)
; CHECK: bb.3:
; CHECK: %ext:_(s64) = G_PHI [[SEXT]](s64), %bb.1, [[SEXT1]](s64), %bb.2
; CHECK: %ptr:_(p0) = G_PTR_ADD %base, %ext(s64)
; CHECK: $x0 = COPY %ptr(p0)
; CHECK: $x1 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1, $x2
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%base:_(p0) = COPY $x2
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%cst32_4:_(s32) = G_CONSTANT i32 4
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %cst32_4(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_SEXT %phi
%ptr:_(p0) = G_PTR_ADD %base, %ext
$x0 = COPY %ptr(p0)
$x1 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...
# Check we don't propagate if there are more than 2 unique incoming values in the phi.
# Doing so might cause too much code bloat.
---
name: zext_icst_through_phi_too_many_incoming
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: zext_icst_through_phi_too_many_incoming
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x40000000), %bb.4(0x40000000)
; CHECK: %cst32_4:_(s32) = G_CONSTANT i32 4
; CHECK: %cond:_(s1) = G_IMPLICIT_DEF
; CHECK: G_BRCOND %cond(s1), %bb.3
; CHECK: G_BR %bb.4
; CHECK: bb.2:
; CHECK: successors: %bb.4(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: G_BR %bb.4
; CHECK: bb.3:
; CHECK: successors: %bb.4(0x80000000)
; CHECK: %cst32_42:_(s32) = G_CONSTANT i32 42
; CHECK: bb.4:
; CHECK: %phi:_(s32) = G_PHI %cst32_4(s32), %bb.1, %cst32_10(s32), %bb.2, %cst32_42(s32), %bb.3
; CHECK: %ext:_(s64) = G_ZEXT %phi(s32)
; CHECK: $x0 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%cst32_4:_(s32) = G_CONSTANT i32 4
%cond:_(s1) = G_IMPLICIT_DEF
G_BRCOND %cond, %bb.5
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
G_BR %bb.4
bb.5:
%cst32_42:_(s32) = G_CONSTANT i32 42
bb.4:
%phi:_(s32) = G_PHI %cst32_4(s32), %bb.2, %cst32_10(s32), %bb.3, %cst32_42(s32), %bb.5
%ext:_(s64) = G_ZEXT %phi
$x0 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...
# Check that we don't propagate if the extension would be of a non-allowed inst.
---
name: sext_add_through_phi
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: sext_add_through_phi
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: [[COPY1:%[0-9]+]]:_(s32) = COPY $w1
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %add:_(s32) = G_ADD [[COPY]], [[COPY1]]
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: bb.3:
; CHECK: %phi:_(s32) = G_PHI %add(s32), %bb.1, %cst32_10(s32), %bb.2
; CHECK: %ext:_(s64) = G_SEXT %phi(s32)
; CHECK: $x0 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%add:_(s32) = G_ADD %0, %1
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %add(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_SEXT %phi
$x0 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...
# Same as above but allowed with a G_ANYEXT.
---
name: anyext_add_through_phi
tracksRegLiveness: true
body: |
; CHECK-LABEL: name: anyext_add_through_phi
; CHECK: bb.0.entry:
; CHECK: successors: %bb.1(0x40000000), %bb.2(0x40000000)
; CHECK: liveins: $w0, $w1
; CHECK: [[COPY:%[0-9]+]]:_(s32) = COPY $w0
; CHECK: [[COPY1:%[0-9]+]]:_(s32) = COPY $w1
; CHECK: %one:_(s32) = G_CONSTANT i32 2
; CHECK: %cmp:_(s1) = G_ICMP intpred(sle), [[COPY]](s32), %one
; CHECK: G_BRCOND %cmp(s1), %bb.2
; CHECK: G_BR %bb.1
; CHECK: bb.1:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %add:_(s32) = G_ADD [[COPY]], [[COPY1]]
; CHECK: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT %add(s32)
; CHECK: G_BR %bb.3
; CHECK: bb.2:
; CHECK: successors: %bb.3(0x80000000)
; CHECK: %cst32_10:_(s32) = G_CONSTANT i32 10
; CHECK: [[ANYEXT1:%[0-9]+]]:_(s64) = G_ANYEXT %cst32_10(s32)
; CHECK: bb.3:
; CHECK: %ext:_(s64) = G_PHI [[ANYEXT]](s64), %bb.1, [[ANYEXT1]](s64), %bb.2
; CHECK: $x0 = COPY %ext(s64)
; CHECK: RET_ReallyLR implicit $x0
bb.1.entry:
liveins: $w0, $w1
%0:_(s32) = COPY $w0
%1:_(s32) = COPY $w1
%zero:_(s32) = G_CONSTANT i32 0
%one:_(s32) = G_CONSTANT i32 2
%cmp:_(s1) = G_ICMP intpred(sgt), %0(s32), %one
G_BRCOND %cmp(s1), %bb.2
G_BR %bb.3
bb.2:
%add:_(s32) = G_ADD %0, %1
G_BR %bb.4
bb.3:
%cst32_10:_(s32) = G_CONSTANT i32 10
bb.4:
%phi:_(s32) = G_PHI %add(s32), %bb.2, %cst32_10(s32), %bb.3
%ext:_(s64) = G_ANYEXT %phi
$x0 = COPY %ext(s64)
RET_ReallyLR implicit $x0
...