blob: c343bb8dd6a32fcadbb64b6b63f69f17d14b2862 [file] [log] [blame]
; RUN: opt -basic-aa -scoped-noalias-aa -loop-vectorize -licm -force-vector-width=2 \
; RUN: -force-vector-interleave=1 -S < %s | FileCheck %s
target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128"
; In order to vectorize the inner loop, it needs to be versioned with
; memchecks between {A} x {B, C} first:
;
; for (i = 0; i < n; i++)
; for (j = 0; j < m; j++)
; A[j] += B[i] + C[j];
;
; Since in the versioned vector loop A and B can no longer alias, B[i] can be
; LICM'ed from the inner loop.
define void @f(i32* %a, i32* %b, i32* %c) {
entry:
br label %outer
outer:
%i.2 = phi i64 [ 0, %entry ], [ %i, %inner.end ]
%arrayidxB = getelementptr inbounds i32, i32* %b, i64 %i.2
br label %inner.ph
inner.ph:
; CHECK: vector.ph:
; CHECK: load i32, i32* %arrayidxB,
; CHECK: br label %vector.body
br label %inner
inner:
%j.2 = phi i64 [ 0, %inner.ph ], [ %j, %inner ]
%arrayidxA = getelementptr inbounds i32, i32* %a, i64 %j.2
%loadA = load i32, i32* %arrayidxA, align 4
%loadB = load i32, i32* %arrayidxB, align 4
%arrayidxC = getelementptr inbounds i32, i32* %c, i64 %j.2
%loadC = load i32, i32* %arrayidxC, align 4
%add = add nuw i32 %loadA, %loadB
%add2 = add nuw i32 %add, %loadC
store i32 %add2, i32* %arrayidxA, align 4
%j = add nuw nsw i64 %j.2, 1
%cond1 = icmp eq i64 %j, 20
br i1 %cond1, label %inner.end, label %inner
inner.end:
%i = add nuw nsw i64 %i.2, 1
%cond2 = icmp eq i64 %i, 30
br i1 %cond2, label %outer.end, label %outer
outer.end:
ret void
}