|  | ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py | 
|  | ; RUN: %if x86-registered-target %{ opt < %s -passes=slp-vectorizer -S -mtriple=x86_64-apple-macosx10.10.0 -mattr=+sse4.2 | FileCheck %s %} | 
|  | ; RUN: %if aarch64-registered-target %{ opt < %s -passes=slp-vectorizer -S -mtriple=aarch64-unknown-linux-gnu -mattr=+neon | FileCheck %s %} | 
|  |  | 
|  | ; PR28474 | 
|  |  | 
|  | ;void foo(); | 
|  | ; | 
|  | ;int test1(unsigned int *p) { | 
|  | ;  int sum = 0; | 
|  | ;  #pragma nounroll | 
|  | ;  for (int y = 0; y < 2; y++) { | 
|  | ;    // Inner loop gets unrolled | 
|  | ;    for (int x = 0; x < 8; x++) { | 
|  | ;      sum += pptr 42; | 
|  | ;    } | 
|  | ;    // Dummy call to keep outer loop alive | 
|  | ;    foo(); | 
|  | ;  } | 
|  | ;  return sum; | 
|  | ;} | 
|  |  | 
|  | define i32 @test(ptr nocapture readonly %p) { | 
|  | ; CHECK-LABEL: @test( | 
|  | ; CHECK-NEXT:  entry: | 
|  | ; CHECK-NEXT:    br label [[FOR_BODY:%.*]] | 
|  | ; CHECK:       for.body: | 
|  | ; CHECK-NEXT:    [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ] | 
|  | ; CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i32>, ptr [[P:%.*]], align 4 | 
|  | ; CHECK-NEXT:    [[TMP2:%.*]] = mul <8 x i32> [[TMP1]], splat (i32 42) | 
|  | ; CHECK-NEXT:    [[TMP3:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP2]]) | 
|  | ; CHECK-NEXT:    [[OP_RDX]] = add i32 [[TMP3]], [[SUM]] | 
|  | ; CHECK-NEXT:    br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]] | 
|  | ; CHECK:       for.end: | 
|  | ; CHECK-NEXT:    ret i32 [[OP_RDX]] | 
|  | ; | 
|  | entry: | 
|  | %arrayidx.1 = getelementptr inbounds i32, ptr %p, i64 1 | 
|  | %arrayidx.2 = getelementptr inbounds i32, ptr %p, i64 2 | 
|  | %arrayidx.3 = getelementptr inbounds i32, ptr %p, i64 3 | 
|  | %arrayidx.4 = getelementptr inbounds i32, ptr %p, i64 4 | 
|  | %arrayidx.5 = getelementptr inbounds i32, ptr %p, i64 5 | 
|  | %arrayidx.6 = getelementptr inbounds i32, ptr %p, i64 6 | 
|  | %arrayidx.7 = getelementptr inbounds i32, ptr %p, i64 7 | 
|  | br label %for.body | 
|  |  | 
|  | for.body: | 
|  | %sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ] | 
|  | %tmp = load i32, ptr %p, align 4 | 
|  | %mul = mul i32 %tmp, 42 | 
|  | %add = add i32 %mul, %sum | 
|  | %tmp5 = load i32, ptr %arrayidx.1, align 4 | 
|  | %mul.1 = mul i32 %tmp5, 42 | 
|  | %add.1 = add i32 %mul.1, %add | 
|  | %tmp6 = load i32, ptr %arrayidx.2, align 4 | 
|  | %mul.2 = mul i32 %tmp6, 42 | 
|  | %add.2 = add i32 %mul.2, %add.1 | 
|  | %tmp7 = load i32, ptr %arrayidx.3, align 4 | 
|  | %mul.3 = mul i32 %tmp7, 42 | 
|  | %add.3 = add i32 %mul.3, %add.2 | 
|  | %tmp8 = load i32, ptr %arrayidx.4, align 4 | 
|  | %mul.4 = mul i32 %tmp8, 42 | 
|  | %add.4 = add i32 %mul.4, %add.3 | 
|  | %tmp9 = load i32, ptr %arrayidx.5, align 4 | 
|  | %mul.5 = mul i32 %tmp9, 42 | 
|  | %add.5 = add i32 %mul.5, %add.4 | 
|  | %tmp10 = load i32, ptr %arrayidx.6, align 4 | 
|  | %mul.6 = mul i32 %tmp10, 42 | 
|  | %add.6 = add i32 %mul.6, %add.5 | 
|  | %tmp11 = load i32, ptr %arrayidx.7, align 4 | 
|  | %mul.7 = mul i32 %tmp11, 42 | 
|  | %add.7 = add i32 %mul.7, %add.6 | 
|  | br i1 true, label %for.end, label %for.body | 
|  |  | 
|  | for.end: | 
|  | ret i32 %add.7 | 
|  | } | 
|  |  | 
|  | ;void foo(); | 
|  | ; | 
|  | ;int test2(unsigned int *p, unsigned int *q) { | 
|  | ;  int sum = 0; | 
|  | ;  #pragma nounroll | 
|  | ;  for (int y = 0; y < 2; y++) { | 
|  | ;    // Inner loop gets unrolled | 
|  | ;    for (int x = 0; x < 8; x++) { | 
|  | ;      sum += pptr q[x]; | 
|  | ;    } | 
|  | ;    // Dummy call to keep outer loop alive | 
|  | ;    foo(); | 
|  | ;  } | 
|  | ;  return sum; | 
|  | ;} | 
|  |  | 
|  | define i32 @test2(ptr nocapture readonly %p, ptr nocapture readonly %q) { | 
|  | ; CHECK-LABEL: @test2( | 
|  | ; CHECK-NEXT:  entry: | 
|  | ; CHECK-NEXT:    br label [[FOR_BODY:%.*]] | 
|  | ; CHECK:       for.body: | 
|  | ; CHECK-NEXT:    [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ] | 
|  | ; CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i32>, ptr [[P:%.*]], align 4 | 
|  | ; CHECK-NEXT:    [[TMP3:%.*]] = load <8 x i32>, ptr [[Q:%.*]], align 4 | 
|  | ; CHECK-NEXT:    [[TMP4:%.*]] = mul <8 x i32> [[TMP1]], [[TMP3]] | 
|  | ; CHECK-NEXT:    [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP4]]) | 
|  | ; CHECK-NEXT:    [[OP_RDX]] = add i32 [[TMP5]], [[SUM]] | 
|  | ; CHECK-NEXT:    br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]] | 
|  | ; CHECK:       for.end: | 
|  | ; CHECK-NEXT:    ret i32 [[OP_RDX]] | 
|  | ; | 
|  | entry: | 
|  | %arrayidx.p.1 = getelementptr inbounds i32, ptr %p, i64 1 | 
|  | %arrayidx.p.2 = getelementptr inbounds i32, ptr %p, i64 2 | 
|  | %arrayidx.p.3 = getelementptr inbounds i32, ptr %p, i64 3 | 
|  | %arrayidx.p.4 = getelementptr inbounds i32, ptr %p, i64 4 | 
|  | %arrayidx.p.5 = getelementptr inbounds i32, ptr %p, i64 5 | 
|  | %arrayidx.p.6 = getelementptr inbounds i32, ptr %p, i64 6 | 
|  | %arrayidx.p.7 = getelementptr inbounds i32, ptr %p, i64 7 | 
|  |  | 
|  | %arrayidx.q.1 = getelementptr inbounds i32, ptr %q, i64 1 | 
|  | %arrayidx.q.2 = getelementptr inbounds i32, ptr %q, i64 2 | 
|  | %arrayidx.q.3 = getelementptr inbounds i32, ptr %q, i64 3 | 
|  | %arrayidx.q.4 = getelementptr inbounds i32, ptr %q, i64 4 | 
|  | %arrayidx.q.5 = getelementptr inbounds i32, ptr %q, i64 5 | 
|  | %arrayidx.q.6 = getelementptr inbounds i32, ptr %q, i64 6 | 
|  | %arrayidx.q.7 = getelementptr inbounds i32, ptr %q, i64 7 | 
|  | br label %for.body | 
|  |  | 
|  | for.body: | 
|  | %sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ] | 
|  | %tmpp = load i32, ptr %p, align 4 | 
|  | %tmpq = load i32, ptr %q, align 4 | 
|  | %mul = mul i32 %tmpp, %tmpq | 
|  | %add = add i32 %mul, %sum | 
|  | %tmp5p = load i32, ptr %arrayidx.p.1, align 4 | 
|  | %tmp5q = load i32, ptr %arrayidx.q.1, align 4 | 
|  | %mul.1 = mul i32 %tmp5p, %tmp5q | 
|  | %add.1 = add i32 %mul.1, %add | 
|  | %tmp6p = load i32, ptr %arrayidx.p.2, align 4 | 
|  | %tmp6q = load i32, ptr %arrayidx.q.2, align 4 | 
|  | %mul.2 = mul i32 %tmp6p, %tmp6q | 
|  | %add.2 = add i32 %mul.2, %add.1 | 
|  | %tmp7p = load i32, ptr %arrayidx.p.3, align 4 | 
|  | %tmp7q = load i32, ptr %arrayidx.q.3, align 4 | 
|  | %mul.3 = mul i32 %tmp7p, %tmp7q | 
|  | %add.3 = add i32 %mul.3, %add.2 | 
|  | %tmp8p = load i32, ptr %arrayidx.p.4, align 4 | 
|  | %tmp8q = load i32, ptr %arrayidx.q.4, align 4 | 
|  | %mul.4 = mul i32 %tmp8p, %tmp8q | 
|  | %add.4 = add i32 %mul.4, %add.3 | 
|  | %tmp9p = load i32, ptr %arrayidx.p.5, align 4 | 
|  | %tmp9q = load i32, ptr %arrayidx.q.5, align 4 | 
|  | %mul.5 = mul i32 %tmp9p, %tmp9q | 
|  | %add.5 = add i32 %mul.5, %add.4 | 
|  | %tmp10p = load i32, ptr %arrayidx.p.6, align 4 | 
|  | %tmp10q = load i32, ptr %arrayidx.q.6, align 4 | 
|  | %mul.6 = mul i32 %tmp10p, %tmp10q | 
|  | %add.6 = add i32 %mul.6, %add.5 | 
|  | %tmp11p = load i32, ptr %arrayidx.p.7, align 4 | 
|  | %tmp11q = load i32, ptr %arrayidx.q.7, align 4 | 
|  | %mul.7 = mul i32 %tmp11p, %tmp11q | 
|  | %add.7 = add i32 %mul.7, %add.6 | 
|  | br i1 true, label %for.end, label %for.body | 
|  |  | 
|  | for.end: | 
|  | ret i32 %add.7 | 
|  | } | 
|  |  | 
|  | ;void foo(); | 
|  | ; | 
|  | ;int test3(unsigned int *p, unsigned int *q) { | 
|  | ;  int sum = 0; | 
|  | ;  #pragma nounroll | 
|  | ;  for (int y = 0; y < 2; y++) { | 
|  | ;    // Inner loop gets unrolled | 
|  | ;    for (int x = 0; x < 8; x++) { | 
|  | ;      sum += pptr q[7-x]; | 
|  | ;    } | 
|  | ;    // Dummy call to keep outer loop alive | 
|  | ;    foo(); | 
|  | ;  } | 
|  | ;  return sum; | 
|  | ;} | 
|  |  | 
|  | define i32 @test3(ptr nocapture readonly %p, ptr nocapture readonly %q) { | 
|  | ; CHECK-LABEL: @test3( | 
|  | ; CHECK-NEXT:  entry: | 
|  | ; CHECK-NEXT:    br label [[FOR_BODY:%.*]] | 
|  | ; CHECK:       for.body: | 
|  | ; CHECK-NEXT:    [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ] | 
|  | ; CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i32>, ptr [[P:%.*]], align 4 | 
|  | ; CHECK-NEXT:    [[TMP3:%.*]] = load <8 x i32>, ptr [[Q:%.*]], align 4 | 
|  | ; CHECK-NEXT:    [[SHUFFLE:%.*]] = shufflevector <8 x i32> [[TMP3]], <8 x i32> poison, <8 x i32> <i32 7, i32 6, i32 5, i32 4, i32 3, i32 2, i32 1, i32 0> | 
|  | ; CHECK-NEXT:    [[TMP4:%.*]] = mul <8 x i32> [[TMP1]], [[SHUFFLE]] | 
|  | ; CHECK-NEXT:    [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP4]]) | 
|  | ; CHECK-NEXT:    [[OP_RDX]] = add i32 [[TMP5]], [[SUM]] | 
|  | ; CHECK-NEXT:    br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]] | 
|  | ; CHECK:       for.end: | 
|  | ; CHECK-NEXT:    ret i32 [[OP_RDX]] | 
|  | ; | 
|  | entry: | 
|  | %arrayidx.p.1 = getelementptr inbounds i32, ptr %p, i64 1 | 
|  | %arrayidx.p.2 = getelementptr inbounds i32, ptr %p, i64 2 | 
|  | %arrayidx.p.3 = getelementptr inbounds i32, ptr %p, i64 3 | 
|  | %arrayidx.p.4 = getelementptr inbounds i32, ptr %p, i64 4 | 
|  | %arrayidx.p.5 = getelementptr inbounds i32, ptr %p, i64 5 | 
|  | %arrayidx.p.6 = getelementptr inbounds i32, ptr %p, i64 6 | 
|  | %arrayidx.p.7 = getelementptr inbounds i32, ptr %p, i64 7 | 
|  |  | 
|  | %arrayidx.q.1 = getelementptr inbounds i32, ptr %q, i64 1 | 
|  | %arrayidx.q.2 = getelementptr inbounds i32, ptr %q, i64 2 | 
|  | %arrayidx.q.3 = getelementptr inbounds i32, ptr %q, i64 3 | 
|  | %arrayidx.q.4 = getelementptr inbounds i32, ptr %q, i64 4 | 
|  | %arrayidx.q.5 = getelementptr inbounds i32, ptr %q, i64 5 | 
|  | %arrayidx.q.6 = getelementptr inbounds i32, ptr %q, i64 6 | 
|  | %arrayidx.q.7 = getelementptr inbounds i32, ptr %q, i64 7 | 
|  | br label %for.body | 
|  |  | 
|  | for.body: | 
|  | %sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ] | 
|  | %tmpp = load i32, ptr %p, align 4 | 
|  | %tmpq = load i32, ptr %arrayidx.q.7, align 4 | 
|  | %mul = mul i32 %tmpp, %tmpq | 
|  | %add = add i32 %mul, %sum | 
|  | %tmp5p = load i32, ptr %arrayidx.p.1, align 4 | 
|  | %tmp5q = load i32, ptr %arrayidx.q.6, align 4 | 
|  | %mul.1 = mul i32 %tmp5p, %tmp5q | 
|  | %add.1 = add i32 %mul.1, %add | 
|  | %tmp6p = load i32, ptr %arrayidx.p.2, align 4 | 
|  | %tmp6q = load i32, ptr %arrayidx.q.5, align 4 | 
|  | %mul.2 = mul i32 %tmp6p, %tmp6q | 
|  | %add.2 = add i32 %mul.2, %add.1 | 
|  | %tmp7p = load i32, ptr %arrayidx.p.3, align 4 | 
|  | %tmp7q = load i32, ptr %arrayidx.q.4, align 4 | 
|  | %mul.3 = mul i32 %tmp7p, %tmp7q | 
|  | %add.3 = add i32 %mul.3, %add.2 | 
|  | %tmp8p = load i32, ptr %arrayidx.p.4, align 4 | 
|  | %tmp8q = load i32, ptr %arrayidx.q.3, align 4 | 
|  | %mul.4 = mul i32 %tmp8p, %tmp8q | 
|  | %add.4 = add i32 %mul.4, %add.3 | 
|  | %tmp9p = load i32, ptr %arrayidx.p.5, align 4 | 
|  | %tmp9q = load i32, ptr %arrayidx.q.2, align 4 | 
|  | %mul.5 = mul i32 %tmp9p, %tmp9q | 
|  | %add.5 = add i32 %mul.5, %add.4 | 
|  | %tmp10p = load i32, ptr %arrayidx.p.6, align 4 | 
|  | %tmp10q = load i32, ptr %arrayidx.q.1, align 4 | 
|  | %mul.6 = mul i32 %tmp10p, %tmp10q | 
|  | %add.6 = add i32 %mul.6, %add.5 | 
|  | %tmp11p = load i32, ptr %arrayidx.p.7, align 4 | 
|  | %tmp11q = load i32, ptr %q, align 4 | 
|  | %mul.7 = mul i32 %tmp11p, %tmp11q | 
|  | %add.7 = add i32 %mul.7, %add.6 | 
|  | br i1 true, label %for.end, label %for.body | 
|  |  | 
|  | for.end: | 
|  | ret i32 %add.7 | 
|  | } |