| ; REQUIRES: asserts |
| ; RUN: opt %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s |
| ; RUN: opt %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -enable-interleaved-mem-accesses -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s --check-prefix=INTER |
| |
| target datalayout = "e-m:e-i64:64-i128:128-n32:64-S128" |
| |
| %pair = type { i32, i32 } |
| |
| ; CHECK-LABEL: consecutive_ptr_forward |
| ; |
| ; Check that a forward consecutive pointer is recognized as uniform and remains |
| ; uniform after vectorization. |
| ; |
| ; CHECK: LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK-NOT: getelementptr |
| ; CHECK: getelementptr inbounds i32, ptr %a, i64 %index |
| ; CHECK-NOT: getelementptr |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define i32 @consecutive_ptr_forward(ptr %a, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %tmp0 = phi i32 [ %tmp3, %for.body ], [ 0, %entry ] |
| %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i |
| %tmp2 = load i32, ptr %tmp1, align 8 |
| %tmp3 = add i32 %tmp0, %tmp2 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| %tmp4 = phi i32 [ %tmp3, %for.body ] |
| ret i32 %tmp4 |
| } |
| |
| ; CHECK-LABEL: consecutive_ptr_reverse |
| ; |
| ; Check that a reverse consecutive pointer is recognized as uniform and remains |
| ; uniform after vectorization. |
| ; |
| ; CHECK: LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: [[OFFSET_IDX:%.+]] = sub i64 %n, %index |
| ; CHECK-NOT: getelementptr |
| ; CHECK: %[[G0:.+]] = getelementptr inbounds i32, ptr %a, i64 [[OFFSET_IDX]] |
| ; CHECK: getelementptr inbounds i8, ptr %[[G0]], i64 -12 |
| ; CHECK-NOT: getelementptr |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define i32 @consecutive_ptr_reverse(ptr %a, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ %n, %entry ] |
| %tmp0 = phi i32 [ %tmp3, %for.body ], [ 0, %entry ] |
| %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i |
| %tmp2 = load i32, ptr %tmp1, align 8 |
| %tmp3 = add i32 %tmp0, %tmp2 |
| %i.next = add nsw i64 %i, -1 |
| %cond = icmp sgt i64 %i.next, 0 |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| %tmp4 = phi i32 [ %tmp3, %for.body ] |
| ret i32 %tmp4 |
| } |
| |
| ; CHECK-LABEL: interleaved_access_forward |
| ; INTER-LABEL: interleaved_access_forward |
| ; |
| ; Check that a consecutive-like pointer used by a forward interleaved group is |
| ; recognized as uniform and remains uniform after vectorization. When |
| ; interleaved memory accesses aren't enabled, the pointer should not be |
| ; recognized as uniform, and it should not be uniform after vectorization. |
| ; |
| ; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| ; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: %[[I1:.+]] = or disjoint i64 %index, 1 |
| ; CHECK: %[[I2:.+]] = or disjoint i64 %index, 2 |
| ; CHECK: %[[I3:.+]] = or disjoint i64 %index, 3 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %index, i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %index, i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 1 |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| ; INTER: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| ; INTER: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| ; INTER: vector.body |
| ; INTER: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; INTER-NOT: getelementptr |
| ; INTER: getelementptr inbounds %pair, ptr %p, i64 %index, i32 0 |
| ; INTER-NOT: getelementptr |
| ; INTER: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define i32 @interleaved_access_forward(ptr %p, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %tmp0 = phi i32 [ %tmp6, %for.body ], [ 0, %entry ] |
| %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| %tmp3 = load i32, ptr %tmp1, align 8 |
| %tmp4 = load i32, ptr %tmp2, align 8 |
| %tmp5 = add i32 %tmp3, %tmp4 |
| %tmp6 = add i32 %tmp0, %tmp5 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| %tmp14 = phi i32 [ %tmp6, %for.body ] |
| ret i32 %tmp14 |
| } |
| |
| ; CHECK-LABEL: interleaved_access_reverse |
| ; INTER-LABEL: interleaved_access_reverse |
| ; |
| ; Check that a consecutive-like pointer used by a reverse interleaved group is |
| ; recognized as uniform and remains uniform after vectorization. When |
| ; interleaved memory accesses aren't enabled, the pointer should not be |
| ; recognized as uniform, and it should not be uniform after vectorization. |
| ; |
| ; recognized as uniform, and it should not be uniform after vectorization. |
| ; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| ; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: [[OFFSET_IDX:%.+]] = sub i64 %n, %index |
| ; CHECK: %[[I1:.+]] = add i64 [[OFFSET_IDX]], -1 |
| ; CHECK: %[[I2:.+]] = add i64 [[OFFSET_IDX]], -2 |
| ; CHECK: %[[I3:.+]] = add i64 [[OFFSET_IDX]], -3 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 [[OFFSET_IDX]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 [[OFFSET_IDX]], i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 1 |
| ; CHECK: getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 1 |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| ; INTER: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| ; INTER: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| ; INTER: vector.body |
| ; INTER: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; INTER: [[OFFSET_IDX:%.+]] = sub i64 %n, %index |
| ; INTER-NOT: getelementptr |
| ; INTER: %[[G0:.+]] = getelementptr inbounds %pair, ptr %p, i64 [[OFFSET_IDX]], i32 0 |
| ; INTER: getelementptr inbounds i8, ptr %[[G0]], i64 -24 |
| ; INTER-NOT: getelementptr |
| ; INTER: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define i32 @interleaved_access_reverse(ptr %p, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ %n, %entry ] |
| %tmp0 = phi i32 [ %tmp6, %for.body ], [ 0, %entry ] |
| %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1 |
| %tmp3 = load i32, ptr %tmp1, align 8 |
| %tmp4 = load i32, ptr %tmp2, align 8 |
| %tmp5 = add i32 %tmp3, %tmp4 |
| %tmp6 = add i32 %tmp0, %tmp5 |
| %i.next = add nsw i64 %i, -1 |
| %cond = icmp sgt i64 %i.next, 0 |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| %tmp14 = phi i32 [ %tmp6, %for.body ] |
| ret i32 %tmp14 |
| } |
| |
| ; INTER-LABEL: predicated_store |
| ; |
| ; Check that a consecutive-like pointer used by a forward interleaved group and |
| ; scalarized store is not recognized as uniform and is not uniform after |
| ; vectorization. The store is scalarized because it's in a predicated block. |
| ; Even though the load in this example is vectorized and only uses the pointer |
| ; as if it were uniform, the store is scalarized, making the pointer |
| ; non-uniform. |
| ; |
| ; INTER-NOT: LV: Found uniform instruction: %tmp0 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| ; INTER: vector.body |
| ; INTER: %index = phi i64 [ 0, %vector.ph ], [ %index.next, {{.*}} ] |
| ; INTER: %[[G0:.+]] = getelementptr inbounds %pair, ptr %p, i64 %index, i32 0 |
| ; INTER: %wide.vec = load <8 x i32>, ptr %[[G0]], align 8 |
| ; INTER: %[[I1:.+]] = or disjoint i64 %index, 1 |
| ; INTER: getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0 |
| ; INTER: %[[I2:.+]] = or disjoint i64 %index, 2 |
| ; INTER: getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0 |
| ; INTER: %[[I3:.+]] = or disjoint i64 %index, 3 |
| ; INTER: getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0 |
| ; INTER: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @predicated_store(ptr %p, i32 %x, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %if.merge ], [ 0, %entry ] |
| %tmp0 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0 |
| %tmp1 = load i32, ptr %tmp0, align 8 |
| %tmp2 = icmp eq i32 %tmp1, %x |
| br i1 %tmp2, label %if.then, label %if.merge |
| |
| if.then: |
| store i32 %tmp1, ptr %tmp0, align 8 |
| br label %if.merge |
| |
| if.merge: |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: irregular_type |
| ; |
| ; Check that a consecutive pointer used by a scalarized store is not recognized |
| ; as uniform and is not uniform after vectorization. The store is scalarized |
| ; because the stored type may required padding. |
| ; |
| ; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds x86_fp80, ptr %a, i64 %i |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: %[[I1:.+]] = or disjoint i64 %index, 1 |
| ; CHECK: %[[I2:.+]] = or disjoint i64 %index, 2 |
| ; CHECK: %[[I3:.+]] = or disjoint i64 %index, 3 |
| ; CHECK: getelementptr inbounds x86_fp80, ptr %a, i64 %index |
| ; CHECK: getelementptr inbounds x86_fp80, ptr %a, i64 %[[I1]] |
| ; CHECK: getelementptr inbounds x86_fp80, ptr %a, i64 %[[I2]] |
| ; CHECK: getelementptr inbounds x86_fp80, ptr %a, i64 %[[I3]] |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @irregular_type(ptr %a, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] |
| %tmp0 = sitofp i32 1 to x86_fp80 |
| %tmp1 = getelementptr inbounds x86_fp80, ptr %a, i64 %i |
| store x86_fp80 %tmp0, ptr %tmp1, align 16 |
| %i.next = add i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: pointer_iv_uniform |
| ; |
| ; Check that a pointer induction variable is recognized as uniform and remains |
| ; uniform after vectorization. |
| ; |
| ; CHECK: LV: Found uniform instruction: %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ] |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK-NOT: getelementptr |
| ; CHECK: [[SHL:%.+]] = shl i64 %index, 2 |
| ; CHECK: %next.gep = getelementptr i8, ptr %a, i64 [[SHL]] |
| ; CHECK-NOT: getelementptr |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @pointer_iv_uniform(ptr %a, i32 %x, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ] |
| store i32 %x, ptr %p, align 8 |
| %tmp03 = getelementptr inbounds i32, ptr %p, i32 1 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; INTER-LABEL: pointer_iv_non_uniform_0 |
| ; |
| ; Check that a pointer induction variable with a non-uniform user is not |
| ; recognized as uniform and is not uniform after vectorization. The pointer |
| ; induction variable is used by getelementptr instructions that are non-uniform |
| ; due to scalarization of the stores. |
| ; |
| ; INTER-NOT: LV: Found uniform instruction: %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ] |
| ; INTER: vector.body |
| ; INTER: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; INTER: %[[I0:.+]] = shl i64 %index, 4 |
| ; INTER: %[[I1:.+]] = or disjoint i64 %[[I0]], 16 |
| ; INTER: %[[I2:.+]] = or disjoint i64 %[[I0]], 32 |
| ; INTER: %[[I3:.+]] = or disjoint i64 %[[I0]], 48 |
| ; INTER: %next.gep = getelementptr i8, ptr %a, i64 %[[I0]] |
| ; INTER-NEXT: = getelementptr i8, ptr %a, i64 %[[I1]] |
| ; INTER-NEXT: = getelementptr i8, ptr %a, i64 %[[I2]] |
| ; INTER-NEXT: = getelementptr i8, ptr %a, i64 %[[I3]] |
| ; INTER: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @pointer_iv_non_uniform_0(ptr %a, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ] |
| %tmp00 = load i32, ptr %p, align 8 |
| %tmp01 = getelementptr inbounds i32, ptr %p, i32 1 |
| %tmp02 = load i32, ptr %tmp01, align 8 |
| %tmp03 = getelementptr inbounds i32, ptr %p, i32 4 |
| %tmp04 = load i32, ptr %tmp03, align 8 |
| %tmp05 = getelementptr inbounds i32, ptr %p, i32 5 |
| %tmp06 = load i32, ptr %tmp05, align 8 |
| %tmp07 = sub i32 %tmp04, %tmp00 |
| %tmp08 = sub i32 %tmp02, %tmp02 |
| %tmp09 = getelementptr inbounds i32, ptr %p, i32 2 |
| store i32 %tmp07, ptr %tmp09, align 8 |
| %tmp10 = getelementptr inbounds i32, ptr %p, i32 3 |
| store i32 %tmp08, ptr %tmp10, align 8 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: pointer_iv_non_uniform_1 |
| ; |
| ; Check that a pointer induction variable with a non-uniform user is not |
| ; recognized as uniform and is not uniform after vectorization. The pointer |
| ; induction variable is used by a store that will be scalarized. |
| ; |
| ; CHECK-NOT: LV: Found uniform instruction: %p = phi ptr [%tmp1, %for.body], [%a, %entry] |
| ; CHECK: vector.body |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: [[SHL1:%.+]] = shl i64 %index, 4 |
| ; CHECK: %[[I1:.+]] = or disjoint i64 [[SHL1]], 16 |
| ; CHECK: %[[I2:.+]] = or disjoint i64 [[SHL1]], 32 |
| ; CHECK: %[[I3:.+]] = or disjoint i64 [[SHL1]], 48 |
| ; CHECK: %next.gep = getelementptr i8, ptr %a, i64 [[SHL1]] |
| ; CHECK: = getelementptr i8, ptr %a, i64 %[[I1]] |
| ; CHECK: = getelementptr i8, ptr %a, i64 %[[I2]] |
| ; CHECK: = getelementptr i8, ptr %a, i64 %[[I3]] |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @pointer_iv_non_uniform_1(ptr %a, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %p = phi ptr [%tmp1, %for.body], [%a, %entry] |
| %tmp0 = sitofp i32 1 to x86_fp80 |
| store x86_fp80 %tmp0, ptr %p, align 16 |
| %tmp1 = getelementptr inbounds x86_fp80, ptr %p, i32 1 |
| %i.next = add i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: pointer_iv_mixed |
| ; |
| ; Check multiple pointer induction variables where only one is recognized as |
| ; uniform and remains uniform after vectorization. The other pointer induction |
| ; variable is not recognized as uniform and is not uniform after vectorization |
| ; because it is stored to memory. |
| ; |
| ; CHECK-NOT: LV: Found uniform instruction: %p = phi ptr [ %tmp3, %for.body ], [ %a, %entry ] |
| ; CHECK: LV: Found uniform instruction: %q = phi ptr [ %tmp4, %for.body ], [ %b, %entry ] |
| ; CHECK: vector.body |
| ; CHECK: %pointer.phi = phi ptr [ %a, %vector.ph ], [ %ptr.ind, %vector.body ] |
| ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK: %[[PTRVEC:.+]] = getelementptr i8, ptr %pointer.phi, <4 x i64> <i64 0, i64 4, i64 8, i64 12> |
| ; CHECK: [[SHL:%.+]] = shl i64 %index, 3 |
| ; CHECK: %next.gep = getelementptr i8, ptr %b, i64 [[SHL]] |
| ; CHECK: store <4 x ptr> %[[PTRVEC]], ptr %next.gep, align 8 |
| ; CHECK: %ptr.ind = getelementptr i8, ptr %pointer.phi, i64 16 |
| ; CHECK: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define i32 @pointer_iv_mixed(ptr %a, ptr %b, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] |
| %p = phi ptr [ %tmp3, %for.body ], [ %a, %entry ] |
| %q = phi ptr [ %tmp4, %for.body ], [ %b, %entry ] |
| %tmp0 = phi i32 [ %tmp2, %for.body ], [ 0, %entry ] |
| %tmp1 = load i32, ptr %p, align 8 |
| %tmp2 = add i32 %tmp1, %tmp0 |
| store ptr %p, ptr %q, align 8 |
| %tmp3 = getelementptr inbounds i32, ptr %p, i32 1 |
| %tmp4 = getelementptr inbounds ptr, ptr %q, i32 1 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| %tmp5 = phi i32 [ %tmp2, %for.body ] |
| ret i32 %tmp5 |
| } |
| |
| ; INTER-LABEL: pointer_operand_geps_with_different_indexed_types |
| ; |
| ; Check that a pointer operand having a user other than a memory access is |
| ; recognized as uniform after vectorization. In this test case, %tmp0 is a |
| ; GEP that is used by a load and a getelementptr instruction (%tmp2). Once |
| ; %tmp2 is marked uniform, %tmp0 should be marked uniform as well. |
| ; |
| ; INTER: LV: Found uniform instruction: %cond = icmp slt i64 %i.next, %n |
| ; INTER-NEXT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds i8, ptr %tmp0, i64 3 |
| ; INTER-NEXT: LV: Found uniform instruction: %tmp6 = getelementptr inbounds i8, ptr %B, i64 %i |
| ; INTER-NEXT: LV: Found uniform instruction: %tmp0 = getelementptr inbounds i64, ptr %A, i64 %i |
| ; INTER-NEXT: LV: Found uniform instruction: %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] |
| ; INTER-NEXT: LV: Found uniform instruction: %i.next = add nuw nsw i64 %i, 1 |
| ; INTER: define void @pointer_operand_geps_with_different_indexed_types( |
| ; INTER: vector.body: |
| ; INTER-NEXT: [[INDEX:%.*]] = phi i64 [ 0, %vector.ph ], [ [[INDEX_NEXT:%.*]], %vector.body ] |
| ; INTER-NEXT: [[TMP4:%.*]] = getelementptr inbounds i64, ptr %A, i64 [[INDEX]] |
| ; INTER-NEXT: [[WIDE_VEC:%.*]] = load <32 x i8>, ptr [[TMP4]], align 1 |
| ; INTER-NEXT: [[STRIDED_VEC:%.*]] = shufflevector <32 x i8> [[WIDE_VEC]], <32 x i8> poison, <4 x i32> <i32 0, i32 8, i32 16, i32 24> |
| ; INTER-NEXT: [[STRIDED_VEC3:%.*]] = shufflevector <32 x i8> [[WIDE_VEC]], <32 x i8> poison, <4 x i32> <i32 3, i32 11, i32 19, i32 27> |
| ; INTER-NEXT: [[TMP5:%.*]] = xor <4 x i8> [[STRIDED_VEC3]], [[STRIDED_VEC]] |
| ; INTER-NEXT: [[TMP6:%.*]] = getelementptr inbounds i8, ptr %B, i64 [[INDEX]] |
| ; INTER-NEXT: store <4 x i8> [[TMP5]], ptr [[TMP6]], align 1 |
| ; INTER-NEXT: [[INDEX_NEXT]] = add nuw i64 [[INDEX]], 4 |
| ; INTER: br i1 {{.*}}, label %middle.block, label %vector.body |
| ; |
| define void @pointer_operand_geps_with_different_indexed_types(ptr %A, ptr %B, i64 %n) { |
| entry: |
| br label %for.body |
| |
| for.body: |
| %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ] |
| %tmp0 = getelementptr inbounds i64, ptr %A, i64 %i |
| %tmp2 = getelementptr inbounds i8, ptr %tmp0, i64 3 |
| %tmp3 = load i8, ptr %tmp2, align 1 |
| %tmp4 = load i8, ptr %tmp0, align 1 |
| %tmp5 = xor i8 %tmp3, %tmp4 |
| %tmp6 = getelementptr inbounds i8, ptr %B, i64 %i |
| store i8 %tmp5, ptr %tmp6 |
| %i.next = add nuw nsw i64 %i, 1 |
| %cond = icmp slt i64 %i.next, %n |
| br i1 %cond, label %for.body, label %for.end |
| |
| for.end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: pr61396_pointer_used_as_both_stored_value_and_pointer_operand_by_store |
| ; CHECK-NOT: LV: Found uniform instruction: %cur.ptr = getelementptr inbounds ptr, ptr %ary, i64 %iv |
| |
| ; CHECK: define void @pr61396_pointer_used_as_both_stored_value_and_pointer_operand_by_store( |
| ; CHECK: vector.body: |
| ; CHECK-NEXT: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ] |
| ; CHECK-NEXT: [[VEC_IND:%.+]] = phi <4 x i64> [ <i64 0, i64 1, i64 2, i64 3>, %vector.ph ], [ %vec.ind.next, %vector.body ] |
| ; CHECK-NEXT: [[GEP:%.+]] = getelementptr inbounds ptr, ptr %ary, <4 x i64> [[VEC_IND]] |
| ; CHECK-NEXT: [[EXT:%.+]] = extractelement <4 x ptr> [[GEP]], i64 0 |
| ; CHECK-NEXT: store <4 x ptr> [[GEP]], ptr [[EXT]], align 8 |
| ; |
| |
| define void @pr61396_pointer_used_as_both_stored_value_and_pointer_operand_by_store(ptr %ary) { |
| entry: |
| br label %loop |
| |
| loop: |
| %iv = phi i64 [ 0, %entry ], [ %iv.next, %loop ] |
| %cur.ptr = getelementptr inbounds ptr, ptr %ary, i64 %iv |
| store ptr %cur.ptr, ptr %cur.ptr, align 8 |
| %iv.next = add nuw nsw i64 %iv, 1 |
| %done = icmp eq i64 %iv, 10240 |
| br i1 %done, label %exit, label %loop |
| |
| exit: |
| ret void |
| } |