| ; RUN: opt < %s -passes='print<scalar-evolution>,loop(loop-rotate),invalidate<scalar-evolution>,print<scalar-evolution>' -disable-output 2>&1 | FileCheck -check-prefixes CHECK-SCEV %s |
| ; RUN: opt < %s -passes='print<scalar-evolution>,loop(loop-rotate),print<scalar-evolution>' -disable-output 2>&1 | FileCheck -check-prefixes CHECK-SCEV %s |
| ; RUN: opt < %s -passes='loop(canon-freeze),loop(loop-rotate),print<scalar-evolution>' -disable-output |
| |
| ; In the first two RUN lines print<scalar-evolution> is used to populate the |
| ; analysis cache before loop-rotate. That was enough to see the problem by |
| ; examining print<scalar-evolution> printouts after loop-rotate. However, the |
| ; crashes where only observed when using canon-freeze as a trigger to populate |
| ; the analysis cache, so that is why canon-freeze is used in the third RUN |
| ; line. |
| |
| ; Verify that we get the same SCEV expressions after loop-rotate, regardless |
| ; if we invalidate scalar-evolution before the final printing or not. |
| ; |
| ; This used to fail as described by PR51981 (some expressions still referred |
| ; to (trunc i32 %div210 to i16) but after the rotation it should be (trunc i32 |
| ; %div2102 to i16). |
| ; |
| ; CHECK-SCEV: Classifying expressions for: @test_function |
| ; CHECK-SCEV: %wide = load i32, ptr @offset, align 1 |
| ; CHECK-SCEV: --> %wide U: full-set S: full-set Exits: <<Unknown>> LoopDispositions: { %loop.outer.header: Variant, %loop.inner: Invariant } |
| ; CHECK-SCEV: %narrow = trunc i32 %wide to i16 |
| ; CHECK-SCEV: --> (trunc i32 %wide to i16) U: full-set S: full-set Exits: <<Unknown>> LoopDispositions: { %loop.outer.header: Variant, %loop.inner: Invariant } |
| ; CHECK-SCEV: %iv = phi i16 [ %narrow, %loop.inner.ph ], [ %iv.plus, %loop.inner ] |
| ; CHECK-SCEV: --> {(trunc i32 %wide to i16),+,1}<nw><%loop.inner> U: full-set S: full-set Exits: (-1 + (700 umax (1 + (trunc i32 %wide to i16)))) LoopDispositions: { %loop.inner: Computable, %loop.outer.header: Variant } |
| ; |
| ; CHECK-SCEV: Classifying expressions for: @test_function |
| ; CHECK-SCEV: %wide1 = load i32, ptr @offset, align 1 |
| ; CHECK-SCEV: --> %wide1 U: full-set S: full-set |
| ; CHECK-SCEV: %wide2 = phi i32 [ %wide1, %loop.inner.ph.lr.ph ], [ %wide, %loop.outer.latch ] |
| ; CHECK-SCEV: --> %wide2 U: full-set S: full-set Exits: <<Unknown>> LoopDispositions: { %loop.inner.ph: Variant, %loop.inner: Invariant } |
| ; CHECK-SCEV: %narrow = trunc i32 %wide2 to i16 |
| ; CHECK-SCEV: --> (trunc i32 %wide2 to i16) U: full-set S: full-set Exits: <<Unknown>> LoopDispositions: { %loop.inner.ph: Variant, %loop.inner: Invariant } |
| ; CHECK-SCEV: %iv = phi i16 [ %narrow, %loop.inner.ph ], [ %iv.plus, %loop.inner ] |
| ; CHECK-SCEV: --> {(trunc i32 %wide2 to i16),+,1}<nw><%loop.inner> U: full-set S: full-set Exits: (-1 + (700 umax (1 + (trunc i32 %wide2 to i16)))) LoopDispositions: { %loop.inner: Computable, %loop.inner.ph: Variant } |
| |
| |
| @offset = external dso_local global i32, align 1 |
| @array = internal global [11263 x i32] zeroinitializer, align 1 |
| |
| define void @test_function(i1 %cond) { |
| entry: |
| br label %loop.outer.header |
| |
| loop.outer.header: ; preds = %loop.outer.latch, %entry |
| %wide = load i32, ptr @offset, align 1 |
| br i1 %cond, label %exit, label %loop.inner.ph |
| |
| loop.inner.ph: ; preds = %loop.outer.header |
| %narrow = trunc i32 %wide to i16 |
| br label %loop.inner |
| |
| loop.inner: ; preds = %loop.inner, %loop.inner.ph |
| %iv = phi i16 [ %narrow, %loop.inner.ph ], [ %iv.plus, %loop.inner ] |
| %iv.promoted = zext i16 %iv to i32 |
| %gep = getelementptr inbounds [11263 x i32], ptr @array, i32 0, i32 %iv.promoted |
| store i32 7, ptr %gep, align 1 |
| %iv.plus = add i16 %iv, 1 |
| %cmp = icmp ult i16 %iv.plus, 700 |
| br i1 %cmp, label %loop.inner, label %loop.outer.latch |
| |
| loop.outer.latch: ; preds = %loop.inner |
| br label %loop.outer.header |
| |
| exit: ; preds = %loop.outer.header |
| ret void |
| } |
| |