blob: 8e8fffedb271c5b61547c414b33dec6ace3525c1 [file] [log] [blame]
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt < %s -mtriple=x86_64-unknown -slp-vectorizer -S | FileCheck %s
; REQUIRES: asserts
; SLP crashed when tried to delete instruction with uses.
; It tried to match reduction subsequently on %i23, then %i22 etc
; When it reached %i18 it was still failing to match reduction but
; succeeded with its operands pair: %i17, %i11.
; Then it popped instruction %i17 from stack to make next attempt on
; matching reduction but the instruction was actually erased on prior
; iteration (it was matched and vectorized, which added a use of a deleted
; instruction)
define void @test() {
; CHECK-LABEL: @test(
; CHECK-NEXT: entry:
; CHECK-NEXT: br i1 undef, label [[IF_END:%.*]], label [[FOR_COND_PREHEADER:%.*]]
; CHECK: for.cond.preheader:
; CHECK-NEXT: [[I:%.*]] = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 2
; CHECK-NEXT: [[I1:%.*]] = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 3
; CHECK-NEXT: [[I2:%.*]] = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 4
; CHECK-NEXT: [[I3:%.*]] = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 5
; CHECK-NEXT: [[I4:%.*]] = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 6
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[I]] to <4 x i32>*
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, <4 x i32>* [[TMP0]], align 8
; CHECK-NEXT: [[I5:%.*]] = add i32 undef, undef
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> [[TMP1]])
; CHECK-NEXT: [[OP_EXTRA2:%.*]] = add i32 [[TMP2]], [[I5]]
; CHECK-NEXT: [[I10:%.*]] = add i32 [[OP_EXTRA2]], undef
; CHECK-NEXT: [[I11:%.*]] = add i32 [[OP_EXTRA2]], [[I10]]
; CHECK-NEXT: [[TMP3:%.*]] = bitcast i32* [[I1]] to <4 x i32>*
; CHECK-NEXT: [[TMP4:%.*]] = load <4 x i32>, <4 x i32>* [[TMP3]], align 4
; CHECK-NEXT: [[I12:%.*]] = add i32 undef, undef
; CHECK-NEXT: [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> [[TMP4]])
; CHECK-NEXT: [[OP_EXTRA:%.*]] = add i32 [[TMP5]], [[I12]]
; CHECK-NEXT: [[OP_EXTRA1:%.*]] = add i32 [[OP_EXTRA]], undef
; CHECK-NEXT: [[I18:%.*]] = add i32 [[OP_EXTRA1]], [[I11]]
; CHECK-NEXT: [[I19:%.*]] = add i32 [[OP_EXTRA1]], [[I18]]
; CHECK-NEXT: [[I20:%.*]] = add i32 undef, [[I19]]
; CHECK-NEXT: [[I21:%.*]] = add i32 undef, [[I20]]
; CHECK-NEXT: [[I22:%.*]] = add i32 undef, [[I21]]
; CHECK-NEXT: [[I23:%.*]] = add i32 undef, [[I22]]
; CHECK-NEXT: br label [[IF_END]]
; CHECK: if.end:
; CHECK-NEXT: [[R:%.*]] = phi i32 [ [[I23]], [[FOR_COND_PREHEADER]] ], [ undef, [[ENTRY:%.*]] ]
; CHECK-NEXT: ret void
;
entry:
br i1 undef, label %if.end, label %for.cond.preheader
for.cond.preheader: ; preds = %entry
%i = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 2
%i1 = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 3
%i2 = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 4
%i3 = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 5
%i4 = getelementptr inbounds [100 x i32], [100 x i32]* undef, i64 0, i64 6
%ld0 = load i32, i32* %i, align 8
%ld1 = load i32, i32* %i1, align 4
%ld2 = load i32, i32* %i2, align 16
%ld3 = load i32, i32* %i3, align 4
%i5 = add i32 undef, undef
%i6 = add i32 %i5, %ld3
%i7 = add i32 %i6, %ld2
%i8 = add i32 %i7, %ld1
%i9 = add i32 %i8, %ld0
%i10 = add i32 %i9, undef
%i11 = add i32 %i9, %i10
%ld4 = load i32, i32* %i1, align 4
%ld5 = load i32, i32* %i2, align 16
%ld6 = load i32, i32* %i3, align 4
%ld7 = load i32, i32* %i4, align 8
%i12 = add i32 undef, undef
%i13 = add i32 %i12, %ld7
%i14 = add i32 %i13, %ld6
%i15 = add i32 %i14, %ld5
%i16 = add i32 %i15, %ld4
%i17 = add i32 %i16, undef
%i18 = add i32 %i17, %i11
%i19 = add i32 %i17, %i18
%i20 = add i32 undef, %i19
%i21 = add i32 undef, %i20
%i22 = add i32 undef, %i21
%i23 = add i32 undef, %i22
br label %if.end
if.end: ; preds = %for.cond.preheader, %entry
%r = phi i32 [ %i23, %for.cond.preheader ], [ undef, %entry ]
ret void
}