blob: 62e7f800cdd4ac22947c671da4559761f41fa2bb [file] [log] [blame]
; RUN: opt -loop-vectorize -hexagon-autohvx=1 -force-vector-width=64 -prefer-predicate-over-epilogue=predicate-dont-vectorize -S %s | FileCheck %s
target datalayout = "e-m:e-p:32:32:32-a:0-n16:32-i64:64:64-i32:32:32-i16:16:16-i1:8:8-f32:32:32-f64:64:64-v32:32:32-v64:64:64-v512:512:512-v1024:1024:1024-v2048:2048:2048"
target triple = "hexagon"
; Test for PR45572.
; Check that interleave groups and decisions based on them are correctly
; invalidated with tail-folding on platforms where masked interleaved accesses
; are disabled.
; Make sure a vector body has been created, 64 element vectors are used and a block predicate has been computed.
; Also make sure the loads are not widened.
; CHECK-LABEL: @test1
; CHECK: vector.body:
; CHECK: icmp ule <64 x i32> %vec.ind
; CHECK-NOT: load <{{.*}} x i32>
define void @test1(i32* %arg, i32 %N) #0 {
entry:
%tmp = alloca i8
br label %loop
loop: ; preds = %bb2, %bb
%iv = phi i32 [ %iv.next, %loop], [ 0, %entry ]
%idx.mul = mul nuw nsw i32 %iv, 7
%idx.start = add nuw nsw i32 %idx.mul, 1
%tmp6 = getelementptr inbounds i32, i32* %arg, i32 %idx.start
%tmp7 = load i32, i32* %tmp6, align 4
%tmp8 = add nuw nsw i32 %idx.start, 1
%tmp9 = getelementptr inbounds i32, i32* %arg, i32 %tmp8
%tmp10 = load i32, i32* %tmp9, align 4
%tmp11 = add nuw nsw i32 %idx.start, 2
%tmp12 = getelementptr inbounds i32, i32* %arg, i32 %tmp11
%tmp13 = load i32, i32* %tmp12, align 4
%tmp14 = add nuw nsw i32 %idx.start, 3
%tmp15 = getelementptr inbounds i32, i32* %arg, i32 %tmp14
%tmp16 = load i32, i32* %tmp15, align 4
%tmp18 = add nuw nsw i32 %idx.start, 4
%tmp19 = getelementptr inbounds i32, i32* %arg, i32 %tmp18
%tmp20 = load i32, i32* %tmp19, align 4
%tmp21 = add nuw nsw i32 %idx.start, 5
%tmp22 = getelementptr inbounds i32, i32* %arg, i32 %tmp21
%tmp23 = load i32, i32* %tmp22, align 4
%tmp25 = add nuw nsw i32 %idx.start, 6
%tmp26 = getelementptr inbounds i32, i32* %arg, i32 %tmp25
%tmp27 = load i32, i32* %tmp26, align 4
store i8 0, i8* %tmp, align 1
%iv.next= add nuw nsw i32 %iv, 1
%exit.cond = icmp eq i32 %iv.next, %N
br i1 %exit.cond, label %exit, label %loop
exit: ; preds = %loop
ret void
}
; The loop below only requires tail folding due to interleave groups with gaps.
; Make sure the loads are not widened.
; CHECK-LABEL: @test2
; CHECK: vector.body:
; CHECK-NOT: load <{{.*}} x i32>
define void @test2(i32* %arg) #1 {
entry:
%tmp = alloca i8
br label %loop
loop: ; preds = %bb2, %bb
%iv = phi i32 [ %iv.next, %loop], [ 0, %entry ]
%idx.start = mul nuw nsw i32 %iv, 5
%tmp6 = getelementptr inbounds i32, i32* %arg, i32 %idx.start
%tmp7 = load i32, i32* %tmp6, align 4
%tmp8 = add nuw nsw i32 %idx.start, 1
%tmp9 = getelementptr inbounds i32, i32* %arg, i32 %tmp8
%tmp10 = load i32, i32* %tmp9, align 4
%tmp11 = add nuw nsw i32 %idx.start, 2
%tmp12 = getelementptr inbounds i32, i32* %arg, i32 %tmp11
%tmp13 = load i32, i32* %tmp12, align 4
%tmp14 = add nuw nsw i32 %idx.start, 3
%tmp15 = getelementptr inbounds i32, i32* %arg, i32 %tmp14
%tmp16 = load i32, i32* %tmp15, align 4
store i8 0, i8* %tmp, align 1
%iv.next= add nuw nsw i32 %iv, 1
%exit.cond = icmp eq i32 %iv.next, 128
br i1 %exit.cond, label %exit, label %loop
exit: ; preds = %loop
ret void
}
attributes #0 = { "target-features"="+hvx,+hvx-length128b" }
attributes #1 = { optsize "target-features"="+hvx,+hvx-length128b" }