blob: 40ee849c6c2b35f27a09b9fcde32a4f31c542489 [file] [log] [blame]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -verify-machineinstrs < %s -mtriple=powerpc-unknown-linux-gnu \
; RUN: -mattr=+spe | FileCheck %s -check-prefixes=CHECK,SPE
target datalayout = "E-m:e-p:32:32-Fn32-i64:64-n32"
target triple = "ppc32"
%struct.cmplx = type { double, double }
; Function Attrs: noinline nounwind optnone uwtable
define dso_local i32 @main() #0 {
; CHECK-LABEL: main:
; CHECK: # %bb.0:
; CHECK-NEXT: mflr 0
; CHECK-NEXT: stwu 1, -48(1)
; CHECK-NEXT: stw 31, 44(1)
; CHECK-NEXT: stw 0, 52(1)
; CHECK-NEXT: .cfi_def_cfa_offset 48
; CHECK-NEXT: .cfi_offset r31, -4
; CHECK-NEXT: .cfi_offset lr, 4
; CHECK-NEXT: mr 31, 1
; CHECK-NEXT: .cfi_def_cfa_register r31
; CHECK-NEXT: li 3, 10
; CHECK-NEXT: stw 3, 40(31)
; CHECK-NEXT: li 3, 0
; CHECK-NEXT: stw 3, 28(31)
; CHECK-NEXT: lis 4, 16404
; CHECK-NEXT: stw 4, 24(31)
; CHECK-NEXT: stw 3, 36(31)
; CHECK-NEXT: lis 3, 16420
; CHECK-NEXT: stw 3, 32(31)
; CHECK-NEXT: lwz 3, 40(31)
; CHECK-NEXT: slwi 3, 3, 4
; CHECK-NEXT: bl malloc
; CHECK-NEXT: stw 3, 20(31)
; CHECK-NEXT: addi 7, 31, 24
; CHECK-NEXT: stw 7, 16(31)
; CHECK-NEXT: lwz 3, 20(31)
; CHECK-NEXT: stw 3, 12(31)
; CHECK-NEXT: lwz 5, 16(31)
; CHECK-NEXT: lwz 6, 12(31)
; CHECK-NEXT: li 3, 5
; CHECK-NEXT: li 4, 1
; CHECK-NEXT: li 8, 1
; CHECK-NEXT: bl pass11
; CHECK-NEXT: li 3, 0
; CHECK-NEXT: lwz 0, 52(1)
; CHECK-NEXT: lwz 31, 44(1)
; CHECK-NEXT: addi 1, 1, 48
; CHECK-NEXT: mtlr 0
; CHECK-NEXT: blr
%1 = alloca i32, align 4
%2 = alloca %struct.cmplx, align 8
%3 = alloca ptr, align 4
%4 = alloca ptr, align 4
%5 = alloca ptr, align 4
store i32 10, ptr %1, align 4
%6 = getelementptr inbounds %struct.cmplx, ptr %2, i32 0, i32 0
store double 5.000000e+00, ptr %6, align 8
%7 = getelementptr inbounds %struct.cmplx, ptr %2, i32 0, i32 1
store double 1.000000e+01, ptr %7, align 8
%8 = load i32, ptr %1, align 4
%9 = mul i32 %8, 16
%10 = call ptr @malloc(i32 noundef %9)
store ptr %10, ptr %3, align 4
store ptr %2, ptr %4, align 4
%11 = load ptr, ptr %3, align 4
store ptr %11, ptr %5, align 4
%12 = load ptr, ptr %4, align 4
%13 = load ptr, ptr %5, align 4
call void @pass11(i32 noundef 5, i32 noundef 1, ptr noundef %12, ptr noundef %13, ptr noundef %2, i32 noundef 1)
ret i32 0
}
declare dso_local ptr @malloc(i32 noundef) #1
; Function Attrs: noinline nounwind optnone uwtable
define internal void @pass11(i32 noundef %0, i32 noundef %1, ptr noalias noundef %2, ptr noalias noundef %3, ptr noalias noundef %4, i32 noundef %5) #0 {
; CHECK-LABEL: pass11:
; CHECK: # %bb.0:
; CHECK-NEXT: stwu 1, -1088(1)
; CHECK-NEXT: stw 31, 1084(1)
; CHECK-NEXT: .cfi_def_cfa_offset 1088
; CHECK-NEXT: .cfi_offset r31, -4
; CHECK-NEXT: mr 31, 1
; CHECK-NEXT: .cfi_def_cfa_register r31
; CHECK-NEXT: .cfi_offset r15, -68
; CHECK-NEXT: .cfi_offset r16, -208
; CHECK-NEXT: .cfi_offset r17, -200
; CHECK-NEXT: .cfi_offset r18, -192
; CHECK-NEXT: .cfi_offset r19, -184
; CHECK-NEXT: .cfi_offset r20, -176
; CHECK-NEXT: .cfi_offset r21, -168
; CHECK-NEXT: .cfi_offset r22, -160
; CHECK-NEXT: .cfi_offset r23, -152
; CHECK-NEXT: .cfi_offset r24, -144
; CHECK-NEXT: .cfi_offset r25, -136
; CHECK-NEXT: .cfi_offset r26, -128
; CHECK-NEXT: .cfi_offset r27, -120
; CHECK-NEXT: .cfi_offset r28, -112
; CHECK-NEXT: .cfi_offset r29, -104
; CHECK-NEXT: .cfi_offset r30, -8
; CHECK-NEXT: stw 15, 1020(31) # 4-byte Folded Spill
; CHECK-NEXT: li 9, 880
; CHECK-NEXT: evstddx 16, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 888
; CHECK-NEXT: evstddx 17, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 896
; CHECK-NEXT: evstddx 18, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 904
; CHECK-NEXT: evstddx 19, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 912
; CHECK-NEXT: evstddx 20, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 920
; CHECK-NEXT: evstddx 21, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 928
; CHECK-NEXT: evstddx 22, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 936
; CHECK-NEXT: evstddx 23, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 944
; CHECK-NEXT: evstddx 24, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 952
; CHECK-NEXT: evstddx 25, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 960
; CHECK-NEXT: evstddx 26, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 968
; CHECK-NEXT: evstddx 27, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 976
; CHECK-NEXT: evstddx 28, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: li 9, 984
; CHECK-NEXT: evstddx 29, 31, 9 # 8-byte Folded Spill
; CHECK-NEXT: stw 30, 1080(31) # 4-byte Folded Spill
; CHECK-NEXT: stw 3, 876(31)
; CHECK-NEXT: stw 4, 872(31)
; CHECK-NEXT: stw 5, 868(31)
; CHECK-NEXT: stw 6, 864(31)
; CHECK-NEXT: stw 7, 860(31)
; CHECK-NEXT: stw 8, 856(31)
; CHECK-NEXT: li 3, 11
; CHECK-NEXT: stw 3, 852(31)
; CHECK-NEXT: lis 3, -30876
; CHECK-NEXT: ori 3, 3, 61626
; CHECK-NEXT: stw 3, 844(31)
; CHECK-NEXT: lis 3, 16362
; CHECK-NEXT: ori 3, 3, 60300
; CHECK-NEXT: stw 3, 840(31)
; CHECK-NEXT: lwz 3, 856(31)
; CHECK-NEXT: efdcfsi 3, 3
; CHECK-NEXT: li 4, .LCPI1_0@l
; CHECK-NEXT: lis 5, .LCPI1_0@ha
; CHECK-NEXT: evlddx 4, 5, 4
; CHECK-NEXT: efdmul 3, 3, 4
; CHECK-NEXT: li 4, 832
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lis 3, -9785
; CHECK-NEXT: ori 3, 3, 4790
; CHECK-NEXT: stw 3, 828(31)
; CHECK-NEXT: lis 3, 16346
; CHECK-NEXT: ori 3, 3, 38440
; CHECK-NEXT: stw 3, 824(31)
; CHECK-NEXT: lwz 3, 856(31)
; CHECK-NEXT: efdcfsi 3, 3
; CHECK-NEXT: li 4, .LCPI1_1@l
; CHECK-NEXT: lis 5, .LCPI1_1@ha
; CHECK-NEXT: evlddx 4, 5, 4
; CHECK-NEXT: efdmul 3, 3, 4
; CHECK-NEXT: li 4, 816
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lis 3, 25615
; CHECK-NEXT: ori 3, 3, 17627
; CHECK-NEXT: stw 3, 812(31)
; CHECK-NEXT: lis 3, -16446
; CHECK-NEXT: ori 3, 3, 14175
; CHECK-NEXT: stw 3, 808(31)
; CHECK-NEXT: lwz 3, 856(31)
; CHECK-NEXT: efdcfsi 3, 3
; CHECK-NEXT: li 4, .LCPI1_2@l
; CHECK-NEXT: lis 5, .LCPI1_2@ha
; CHECK-NEXT: evlddx 4, 5, 4
; CHECK-NEXT: efdmul 3, 3, 4
; CHECK-NEXT: li 4, 800
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lis 3, 32631
; CHECK-NEXT: ori 3, 3, 22663
; CHECK-NEXT: stw 3, 796(31)
; CHECK-NEXT: lis 3, -16412
; CHECK-NEXT: ori 3, 3, 62622
; CHECK-NEXT: stw 3, 792(31)
; CHECK-NEXT: lwz 3, 856(31)
; CHECK-NEXT: efdcfsi 3, 3
; CHECK-NEXT: li 4, .LCPI1_3@l
; CHECK-NEXT: lis 5, .LCPI1_3@ha
; CHECK-NEXT: evlddx 4, 5, 4
; CHECK-NEXT: efdmul 3, 3, 4
; CHECK-NEXT: li 4, 784
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lis 3, -25651
; CHECK-NEXT: ori 3, 3, 20567
; CHECK-NEXT: stw 3, 780(31)
; CHECK-NEXT: lis 3, -16402
; CHECK-NEXT: ori 3, 3, 46122
; CHECK-NEXT: stw 3, 776(31)
; CHECK-NEXT: lwz 3, 856(31)
; CHECK-NEXT: efdcfsi 3, 3
; CHECK-NEXT: li 4, .LCPI1_4@l
; CHECK-NEXT: lis 5, .LCPI1_4@ha
; CHECK-NEXT: evlddx 4, 5, 4
; CHECK-NEXT: efdmul 3, 3, 4
; CHECK-NEXT: li 4, 768
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: li 3, 0
; CHECK-NEXT: stw 3, 764(31)
; CHECK-NEXT: b .LBB1_1
; CHECK-NEXT: .LBB1_1: # =>This Loop Header: Depth=1
; CHECK-NEXT: # Child Loop BB1_3 Depth 2
; CHECK-NEXT: lwz 3, 764(31)
; CHECK-NEXT: lwz 4, 872(31)
; CHECK-NEXT: cmplw 3, 4
; CHECK-NEXT: bge 0, .LBB1_8
; CHECK-NEXT: b .LBB1_2
; CHECK-NEXT: .LBB1_2:
; CHECK-NEXT: lwz 3, 868(31)
; CHECK-NEXT: lwz 4, 876(31)
; CHECK-NEXT: lwz 5, 764(31)
; CHECK-NEXT: mullw 4, 5, 4
; CHECK-NEXT: mulli 4, 4, 176
; CHECK-NEXT: lwzux 4, 3, 4
; CHECK-NEXT: stw 4, 744(31)
; CHECK-NEXT: lwz 4, 12(3)
; CHECK-NEXT: stw 4, 756(31)
; CHECK-NEXT: lwz 4, 8(3)
; CHECK-NEXT: stw 4, 752(31)
; CHECK-NEXT: lwz 3, 4(3)
; CHECK-NEXT: stw 3, 748(31)
; CHECK-NEXT: lwz 3, 868(31)
; CHECK-NEXT: lwz 4, 876(31)
; CHECK-NEXT: lwz 5, 764(31)
; CHECK-NEXT: mulli 5, 5, 11
; CHECK-NEXT: addi 6, 5, 1
; CHECK-NEXT: mullw 6, 4, 6
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: evlddx 6, 3, 6
; CHECK-NEXT: addi 5, 5, 10
; CHECK-NEXT: mullw 4, 4, 5
; CHECK-NEXT: slwi 4, 4, 4
; CHECK-NEXT: evlddx 3, 3, 4
; CHECK-NEXT: efdadd 3, 6, 3
; CHECK-NEXT: li 4, 728
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 5, 876(31)
; CHECK-NEXT: lwz 3, 764(31)
; CHECK-NEXT: mulli 6, 3, 11
; CHECK-NEXT: addi 3, 6, 1
; CHECK-NEXT: mullw 3, 5, 3
; CHECK-NEXT: slwi 3, 3, 4
; CHECK-NEXT: add 7, 4, 3
; CHECK-NEXT: li 3, 8
; CHECK-NEXT: evlddx 7, 7, 3
; CHECK-NEXT: addi 6, 6, 10
; CHECK-NEXT: mullw 5, 5, 6
; CHECK-NEXT: slwi 5, 5, 4
; CHECK-NEXT: add 4, 4, 5
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: efdadd 4, 7, 4
; CHECK-NEXT: addi 5, 31, 728
; CHECK-NEXT: evstddx 4, 5, 3
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 6, 876(31)
; CHECK-NEXT: lwz 7, 764(31)
; CHECK-NEXT: mulli 7, 7, 11
; CHECK-NEXT: addi 8, 7, 1
; CHECK-NEXT: mullw 8, 6, 8
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: evlddx 8, 4, 8
; CHECK-NEXT: addi 7, 7, 10
; CHECK-NEXT: mullw 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: evlddx 4, 4, 6
; CHECK-NEXT: efdsub 4, 8, 4
; CHECK-NEXT: li 6, 584
; CHECK-NEXT: evstddx 4, 31, 6
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 6, 876(31)
; CHECK-NEXT: lwz 7, 764(31)
; CHECK-NEXT: mulli 7, 7, 11
; CHECK-NEXT: addi 8, 7, 1
; CHECK-NEXT: mullw 8, 6, 8
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: add 8, 4, 8
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: addi 7, 7, 10
; CHECK-NEXT: mullw 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: add 4, 4, 6
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: efdsub 6, 8, 4
; CHECK-NEXT: addi 4, 31, 584
; CHECK-NEXT: evstddx 6, 4, 3
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: lwz 8, 764(31)
; CHECK-NEXT: mulli 8, 8, 11
; CHECK-NEXT: addi 9, 8, 2
; CHECK-NEXT: mullw 9, 7, 9
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: evlddx 9, 6, 9
; CHECK-NEXT: addi 8, 8, 9
; CHECK-NEXT: mullw 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: evlddx 6, 6, 7
; CHECK-NEXT: efdadd 6, 9, 6
; CHECK-NEXT: li 7, 712
; CHECK-NEXT: evstddx 6, 31, 7
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: lwz 8, 764(31)
; CHECK-NEXT: mulli 8, 8, 11
; CHECK-NEXT: addi 9, 8, 2
; CHECK-NEXT: mullw 9, 7, 9
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 9, 6, 9
; CHECK-NEXT: evlddx 9, 9, 3
; CHECK-NEXT: addi 8, 8, 9
; CHECK-NEXT: mullw 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdadd 6, 9, 6
; CHECK-NEXT: addi 7, 31, 712
; CHECK-NEXT: evstddx 6, 7, 3
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: mulli 9, 9, 11
; CHECK-NEXT: addi 10, 9, 2
; CHECK-NEXT: mullw 10, 8, 10
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: evlddx 10, 6, 10
; CHECK-NEXT: addi 9, 9, 9
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: evlddx 6, 6, 8
; CHECK-NEXT: efdsub 6, 10, 6
; CHECK-NEXT: li 8, 600
; CHECK-NEXT: evstddx 6, 31, 8
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: mulli 9, 9, 11
; CHECK-NEXT: addi 10, 9, 2
; CHECK-NEXT: mullw 10, 8, 10
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: add 10, 6, 10
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: addi 9, 9, 9
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: add 6, 6, 8
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdsub 8, 10, 6
; CHECK-NEXT: addi 6, 31, 600
; CHECK-NEXT: evstddx 8, 6, 3
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 9, 876(31)
; CHECK-NEXT: lwz 10, 764(31)
; CHECK-NEXT: mulli 10, 10, 11
; CHECK-NEXT: addi 11, 10, 3
; CHECK-NEXT: mullw 11, 9, 11
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: evlddx 11, 8, 11
; CHECK-NEXT: addi 10, 10, 8
; CHECK-NEXT: mullw 9, 9, 10
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: evlddx 8, 8, 9
; CHECK-NEXT: efdadd 8, 11, 8
; CHECK-NEXT: li 9, 696
; CHECK-NEXT: evstddx 8, 31, 9
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 9, 876(31)
; CHECK-NEXT: lwz 10, 764(31)
; CHECK-NEXT: mulli 10, 10, 11
; CHECK-NEXT: addi 11, 10, 3
; CHECK-NEXT: mullw 11, 9, 11
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: add 11, 8, 11
; CHECK-NEXT: evlddx 11, 11, 3
; CHECK-NEXT: addi 10, 10, 8
; CHECK-NEXT: mullw 9, 9, 10
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 8, 8, 9
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: efdadd 8, 11, 8
; CHECK-NEXT: addi 9, 31, 696
; CHECK-NEXT: evstddx 8, 9, 3
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 10, 876(31)
; CHECK-NEXT: lwz 11, 764(31)
; CHECK-NEXT: mulli 11, 11, 11
; CHECK-NEXT: addi 12, 11, 3
; CHECK-NEXT: mullw 12, 10, 12
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: evlddx 12, 8, 12
; CHECK-NEXT: addi 11, 11, 8
; CHECK-NEXT: mullw 10, 10, 11
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: evlddx 8, 8, 10
; CHECK-NEXT: efdsub 8, 12, 8
; CHECK-NEXT: li 10, 616
; CHECK-NEXT: evstddx 8, 31, 10
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 10, 876(31)
; CHECK-NEXT: lwz 11, 764(31)
; CHECK-NEXT: mulli 11, 11, 11
; CHECK-NEXT: addi 12, 11, 3
; CHECK-NEXT: mullw 12, 10, 12
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: add 12, 8, 12
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: addi 11, 11, 8
; CHECK-NEXT: mullw 10, 10, 11
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: add 8, 8, 10
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: efdsub 10, 12, 8
; CHECK-NEXT: addi 8, 31, 616
; CHECK-NEXT: evstddx 10, 8, 3
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 11, 876(31)
; CHECK-NEXT: lwz 12, 764(31)
; CHECK-NEXT: mulli 12, 12, 11
; CHECK-NEXT: addi 0, 12, 4
; CHECK-NEXT: mullw 0, 11, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 0, 10, 0
; CHECK-NEXT: addi 12, 12, 7
; CHECK-NEXT: mullw 11, 11, 12
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: evlddx 10, 10, 11
; CHECK-NEXT: efdadd 10, 0, 10
; CHECK-NEXT: li 11, 680
; CHECK-NEXT: evstddx 10, 31, 11
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 11, 876(31)
; CHECK-NEXT: lwz 12, 764(31)
; CHECK-NEXT: mulli 12, 12, 11
; CHECK-NEXT: addi 0, 12, 4
; CHECK-NEXT: mullw 0, 11, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 30, 10, 0
; CHECK-NEXT: evlddx 0, 30, 3
; CHECK-NEXT: addi 12, 12, 7
; CHECK-NEXT: mullw 11, 11, 12
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: add 10, 10, 11
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: efdadd 10, 0, 10
; CHECK-NEXT: addi 11, 31, 680
; CHECK-NEXT: evstddx 10, 11, 3
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 12, 876(31)
; CHECK-NEXT: lwz 0, 764(31)
; CHECK-NEXT: mulli 30, 0, 11
; CHECK-NEXT: addi 0, 30, 4
; CHECK-NEXT: mullw 0, 12, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 0, 10, 0
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 12, 12, 30
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: evlddx 10, 10, 12
; CHECK-NEXT: efdsub 10, 0, 10
; CHECK-NEXT: li 12, 632
; CHECK-NEXT: evstddx 10, 31, 12
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 12, 876(31)
; CHECK-NEXT: lwz 0, 764(31)
; CHECK-NEXT: mulli 30, 0, 11
; CHECK-NEXT: addi 0, 30, 4
; CHECK-NEXT: mullw 0, 12, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 29, 10, 0
; CHECK-NEXT: evlddx 0, 29, 3
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 12, 12, 30
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: add 10, 10, 12
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: efdsub 12, 0, 10
; CHECK-NEXT: addi 10, 31, 632
; CHECK-NEXT: evstddx 12, 10, 3
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 30, 764(31)
; CHECK-NEXT: mulli 30, 30, 11
; CHECK-NEXT: addi 29, 30, 5
; CHECK-NEXT: mullw 29, 0, 29
; CHECK-NEXT: slwi 29, 29, 4
; CHECK-NEXT: evlddx 29, 12, 29
; CHECK-NEXT: addi 30, 30, 6
; CHECK-NEXT: mullw 0, 0, 30
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 12, 12, 0
; CHECK-NEXT: efdadd 12, 29, 12
; CHECK-NEXT: li 30, 664
; CHECK-NEXT: evstddx 12, 31, 30
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 30, 764(31)
; CHECK-NEXT: mulli 30, 30, 11
; CHECK-NEXT: addi 29, 30, 5
; CHECK-NEXT: mullw 29, 0, 29
; CHECK-NEXT: slwi 29, 29, 4
; CHECK-NEXT: add 29, 12, 29
; CHECK-NEXT: evlddx 29, 29, 3
; CHECK-NEXT: addi 30, 30, 6
; CHECK-NEXT: mullw 0, 0, 30
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: efdadd 12, 29, 12
; CHECK-NEXT: addi 30, 31, 664
; CHECK-NEXT: evstddx 12, 30, 3
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 29, 764(31)
; CHECK-NEXT: mulli 29, 29, 11
; CHECK-NEXT: addi 28, 29, 5
; CHECK-NEXT: mullw 28, 0, 28
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: evlddx 28, 12, 28
; CHECK-NEXT: addi 29, 29, 6
; CHECK-NEXT: mullw 0, 0, 29
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 12, 12, 0
; CHECK-NEXT: efdsub 12, 28, 12
; CHECK-NEXT: li 29, 648
; CHECK-NEXT: evstddx 12, 31, 29
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 29, 764(31)
; CHECK-NEXT: mulli 29, 29, 11
; CHECK-NEXT: addi 28, 29, 5
; CHECK-NEXT: mullw 28, 0, 28
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: add 28, 12, 28
; CHECK-NEXT: evlddx 28, 28, 3
; CHECK-NEXT: addi 29, 29, 6
; CHECK-NEXT: mullw 0, 0, 29
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: efdsub 0, 28, 12
; CHECK-NEXT: addi 12, 31, 648
; CHECK-NEXT: evstddx 0, 12, 3
; CHECK-NEXT: li 29, 744
; CHECK-NEXT: evlddx 0, 31, 29
; CHECK-NEXT: li 29, 728
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 712
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 696
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 680
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 664
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: lwz 29, 864(31)
; CHECK-NEXT: lwz 28, 876(31)
; CHECK-NEXT: lwz 27, 764(31)
; CHECK-NEXT: mullw 28, 28, 27
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: evstddx 0, 29, 28
; CHECK-NEXT: addi 29, 31, 744
; CHECK-NEXT: evlddx 0, 29, 3
; CHECK-NEXT: evlddx 28, 5, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 7, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 9, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 11, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 30, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: lwz 28, 864(31)
; CHECK-NEXT: lwz 27, 876(31)
; CHECK-NEXT: lwz 26, 764(31)
; CHECK-NEXT: mullw 27, 27, 26
; CHECK-NEXT: slwi 27, 27, 4
; CHECK-NEXT: add 28, 28, 27
; CHECK-NEXT: evstddx 0, 28, 3
; CHECK-NEXT: li 28, 744
; CHECK-NEXT: evlddx 28, 31, 28
; CHECK-NEXT: li 27, 728
; CHECK-NEXT: evlddx 27, 31, 27
; CHECK-NEXT: li 0, .LCPI1_5@l
; CHECK-NEXT: lis 26, .LCPI1_5@ha
; CHECK-NEXT: evlddx 0, 26, 0
; CHECK-NEXT: efdmul 27, 27, 0
; CHECK-NEXT: efdadd 27, 27, 28
; CHECK-NEXT: li 28, 712
; CHECK-NEXT: evlddx 26, 31, 28
; CHECK-NEXT: li 28, .LCPI1_6@l
; CHECK-NEXT: lis 25, .LCPI1_6@ha
; CHECK-NEXT: evlddx 28, 25, 28
; CHECK-NEXT: efdmul 26, 26, 28
; CHECK-NEXT: efdadd 26, 26, 27
; CHECK-NEXT: li 27, 696
; CHECK-NEXT: evlddx 25, 31, 27
; CHECK-NEXT: li 27, .LCPI1_7@l
; CHECK-NEXT: lis 24, .LCPI1_7@ha
; CHECK-NEXT: evlddx 27, 24, 27
; CHECK-NEXT: efdmul 25, 25, 27
; CHECK-NEXT: efdadd 26, 25, 26
; CHECK-NEXT: li 25, 680
; CHECK-NEXT: evlddx 24, 31, 25
; CHECK-NEXT: li 25, .LCPI1_8@l
; CHECK-NEXT: lis 23, .LCPI1_8@ha
; CHECK-NEXT: evlddx 25, 23, 25
; CHECK-NEXT: efdmul 24, 24, 25
; CHECK-NEXT: efdadd 24, 24, 26
; CHECK-NEXT: li 26, 664
; CHECK-NEXT: evlddx 23, 31, 26
; CHECK-NEXT: li 26, .LCPI1_9@l
; CHECK-NEXT: lis 22, .LCPI1_9@ha
; CHECK-NEXT: evlddx 26, 22, 26
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 568
; CHECK-NEXT: evstddx 24, 31, 23
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 568
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 832
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 584
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 816
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 600
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: li 22, 800
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 616
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 784
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 632
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 768
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 648
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 22, 22, 23
; CHECK-NEXT: addi 23, 31, 552
; CHECK-NEXT: evstddx 22, 23, 3
; CHECK-NEXT: li 22, 832
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 816
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 22, 22, 20
; CHECK-NEXT: li 21, 800
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 22, 21, 22
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 22, 21, 22
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 22, 21, 22
; CHECK-NEXT: efdneg 22, 22
; CHECK-NEXT: li 21, 552
; CHECK-NEXT: evstddx 22, 31, 21
; CHECK-NEXT: li 22, 568
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 552
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 22, 24, 3
; CHECK-NEXT: evlddx 21, 23, 3
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: evstddx 22, 21, 3
; CHECK-NEXT: li 22, 568
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 552
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: mulli 18, 18, 10
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 23, 3
; CHECK-NEXT: efdsub 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 22, 876(31)
; CHECK-NEXT: lwz 21, 764(31)
; CHECK-NEXT: lwz 20, 872(31)
; CHECK-NEXT: mulli 20, 20, 10
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: mullw 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 744
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 728
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 712
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 696
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 680
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 664
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 536
; CHECK-NEXT: evstddx 24, 31, 23
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 536
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 816
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 584
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 600
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: li 22, 768
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 616
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 800
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 632
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 832
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 648
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 22, 23, 22
; CHECK-NEXT: addi 23, 31, 520
; CHECK-NEXT: evstddx 22, 23, 3
; CHECK-NEXT: li 22, 816
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 784
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 22, 22, 20
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: li 21, 800
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: li 21, 832
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: efdneg 22, 22
; CHECK-NEXT: li 21, 520
; CHECK-NEXT: evstddx 22, 31, 21
; CHECK-NEXT: li 22, 536
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 520
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: slwi 18, 18, 1
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 22, 24, 3
; CHECK-NEXT: evlddx 21, 23, 3
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: slwi 18, 18, 1
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: evstddx 22, 21, 3
; CHECK-NEXT: li 22, 536
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 520
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: mulli 18, 18, 9
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 23, 3
; CHECK-NEXT: efdsub 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 22, 876(31)
; CHECK-NEXT: lwz 21, 764(31)
; CHECK-NEXT: lwz 20, 872(31)
; CHECK-NEXT: mulli 20, 20, 9
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: mullw 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 744
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 728
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 712
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 696
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 680
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 664
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 504
; CHECK-NEXT: evstddx 24, 31, 23
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 504
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 800
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 584
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 600
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: li 22, 816
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 616
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 832
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 632
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 784
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 648
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 22, 22, 23
; CHECK-NEXT: addi 23, 31, 488
; CHECK-NEXT: evstddx 22, 23, 3
; CHECK-NEXT: li 22, 800
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 768
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 22, 22, 20
; CHECK-NEXT: li 21, 816
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: li 21, 832
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 22, 21, 22
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 22, 21, 22
; CHECK-NEXT: efdneg 22, 22
; CHECK-NEXT: li 21, 488
; CHECK-NEXT: evstddx 22, 31, 21
; CHECK-NEXT: li 22, 504
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 488
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: mulli 18, 18, 3
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 22, 24, 3
; CHECK-NEXT: evlddx 21, 23, 3
; CHECK-NEXT: efdadd 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: mulli 18, 18, 3
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: evstddx 22, 21, 3
; CHECK-NEXT: li 22, 504
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 488
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdsub 22, 22, 21
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: slwi 18, 18, 3
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: evstddx 22, 21, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 23, 3
; CHECK-NEXT: efdsub 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 22, 876(31)
; CHECK-NEXT: lwz 21, 764(31)
; CHECK-NEXT: lwz 20, 872(31)
; CHECK-NEXT: slwi 20, 20, 3
; CHECK-NEXT: add 21, 21, 20
; CHECK-NEXT: mullw 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 744
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 728
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 712
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 696
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 680
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 664
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 472
; CHECK-NEXT: evstddx 24, 31, 23
; CHECK-NEXT: evlddx 29, 29, 3
; CHECK-NEXT: evlddx 5, 5, 3
; CHECK-NEXT: efdmul 5, 5, 25
; CHECK-NEXT: efdadd 5, 5, 29
; CHECK-NEXT: evlddx 7, 7, 3
; CHECK-NEXT: efdmul 7, 7, 27
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 9, 3
; CHECK-NEXT: efdmul 7, 7, 0
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 11, 3
; CHECK-NEXT: efdmul 7, 7, 26
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 30, 3
; CHECK-NEXT: efdmul 7, 7, 28
; CHECK-NEXT: efdadd 7, 7, 5
; CHECK-NEXT: addi 5, 31, 472
; CHECK-NEXT: evstddx 7, 5, 3
; CHECK-NEXT: li 7, 784
; CHECK-NEXT: evlddx 7, 31, 7
; CHECK-NEXT: li 9, 584
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 800
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: li 30, 600
; CHECK-NEXT: evlddx 0, 31, 30
; CHECK-NEXT: efdmul 11, 11, 0
; CHECK-NEXT: efdmul 7, 7, 9
; CHECK-NEXT: efdsub 7, 7, 11
; CHECK-NEXT: li 9, 832
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 616
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdadd 7, 9, 7
; CHECK-NEXT: li 9, 768
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 632
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdadd 7, 9, 7
; CHECK-NEXT: li 9, 816
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 648
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdsub 9, 7, 9
; CHECK-NEXT: addi 7, 31, 456
; CHECK-NEXT: evstddx 9, 7, 3
; CHECK-NEXT: li 9, 784
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: li 11, 800
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdmul 6, 11, 6
; CHECK-NEXT: efdmul 4, 9, 4
; CHECK-NEXT: efdsub 4, 4, 6
; CHECK-NEXT: li 6, 832
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: efdmul 6, 6, 8
; CHECK-NEXT: efdadd 4, 6, 4
; CHECK-NEXT: li 6, 768
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 8, 10, 3
; CHECK-NEXT: efdmul 6, 6, 8
; CHECK-NEXT: efdadd 4, 6, 4
; CHECK-NEXT: li 6, 816
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 8, 12, 3
; CHECK-NEXT: efdmul 6, 6, 8
; CHECK-NEXT: efdsub 4, 4, 6
; CHECK-NEXT: efdneg 4, 4
; CHECK-NEXT: li 6, 456
; CHECK-NEXT: evstddx 4, 31, 6
; CHECK-NEXT: li 4, 472
; CHECK-NEXT: evlddx 4, 31, 4
; CHECK-NEXT: li 6, 456
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: efdadd 4, 4, 6
; CHECK-NEXT: lwz 6, 864(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: lwz 10, 872(31)
; CHECK-NEXT: slwi 10, 10, 2
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: evstddx 4, 6, 8
; CHECK-NEXT: evlddx 4, 5, 3
; CHECK-NEXT: evlddx 6, 7, 3
; CHECK-NEXT: efdadd 4, 4, 6
; CHECK-NEXT: lwz 6, 864(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: lwz 10, 872(31)
; CHECK-NEXT: slwi 10, 10, 2
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: add 6, 6, 8
; CHECK-NEXT: evstddx 4, 6, 3
; CHECK-NEXT: li 4, 472
; CHECK-NEXT: evlddx 4, 31, 4
; CHECK-NEXT: li 6, 456
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: efdsub 4, 4, 6
; CHECK-NEXT: lwz 6, 864(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: lwz 10, 872(31)
; CHECK-NEXT: mulli 10, 10, 7
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: evstddx 4, 6, 8
; CHECK-NEXT: evlddx 4, 5, 3
; CHECK-NEXT: evlddx 5, 7, 3
; CHECK-NEXT: efdsub 4, 4, 5
; CHECK-NEXT: lwz 5, 864(31)
; CHECK-NEXT: lwz 6, 876(31)
; CHECK-NEXT: lwz 7, 764(31)
; CHECK-NEXT: lwz 8, 872(31)
; CHECK-NEXT: mulli 8, 8, 7
; CHECK-NEXT: add 7, 7, 8
; CHECK-NEXT: mullw 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: add 5, 5, 6
; CHECK-NEXT: evstddx 4, 5, 3
; CHECK-NEXT: li 3, 1
; CHECK-NEXT: stw 3, 452(31)
; CHECK-NEXT: b .LBB1_3
; CHECK-NEXT: .LBB1_3: # Parent Loop BB1_1 Depth=1
; CHECK-NEXT: # => This Inner Loop Header: Depth=2
; CHECK-NEXT: lwz 3, 452(31)
; CHECK-NEXT: lwz 4, 876(31)
; CHECK-NEXT: cmplw 3, 4
; CHECK-NEXT: bge 0, .LBB1_6
; CHECK-NEXT: b .LBB1_4
; CHECK-NEXT: .LBB1_4:
; CHECK-NEXT: lwz 3, 868(31)
; CHECK-NEXT: lwz 4, 452(31)
; CHECK-NEXT: lwz 5, 876(31)
; CHECK-NEXT: lwz 6, 764(31)
; CHECK-NEXT: mullw 5, 6, 5
; CHECK-NEXT: mulli 5, 5, 11
; CHECK-NEXT: add 4, 4, 5
; CHECK-NEXT: slwi 4, 4, 4
; CHECK-NEXT: lwzux 4, 3, 4
; CHECK-NEXT: stw 4, 432(31)
; CHECK-NEXT: lwz 4, 12(3)
; CHECK-NEXT: stw 4, 444(31)
; CHECK-NEXT: lwz 4, 8(3)
; CHECK-NEXT: stw 4, 440(31)
; CHECK-NEXT: lwz 3, 4(3)
; CHECK-NEXT: stw 3, 436(31)
; CHECK-NEXT: lwz 3, 868(31)
; CHECK-NEXT: lwz 4, 452(31)
; CHECK-NEXT: lwz 5, 876(31)
; CHECK-NEXT: lwz 6, 764(31)
; CHECK-NEXT: mulli 6, 6, 11
; CHECK-NEXT: addi 7, 6, 1
; CHECK-NEXT: mullw 7, 5, 7
; CHECK-NEXT: add 7, 4, 7
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: evlddx 7, 3, 7
; CHECK-NEXT: addi 6, 6, 10
; CHECK-NEXT: mullw 5, 5, 6
; CHECK-NEXT: add 4, 4, 5
; CHECK-NEXT: slwi 4, 4, 4
; CHECK-NEXT: evlddx 3, 3, 4
; CHECK-NEXT: efdadd 3, 7, 3
; CHECK-NEXT: li 4, 416
; CHECK-NEXT: evstddx 3, 31, 4
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 5, 452(31)
; CHECK-NEXT: lwz 6, 876(31)
; CHECK-NEXT: lwz 3, 764(31)
; CHECK-NEXT: mulli 7, 3, 11
; CHECK-NEXT: addi 3, 7, 1
; CHECK-NEXT: mullw 3, 6, 3
; CHECK-NEXT: add 3, 5, 3
; CHECK-NEXT: slwi 3, 3, 4
; CHECK-NEXT: add 8, 4, 3
; CHECK-NEXT: li 3, 8
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: addi 7, 7, 10
; CHECK-NEXT: mullw 6, 6, 7
; CHECK-NEXT: add 5, 5, 6
; CHECK-NEXT: slwi 5, 5, 4
; CHECK-NEXT: add 4, 4, 5
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: efdadd 4, 8, 4
; CHECK-NEXT: addi 5, 31, 416
; CHECK-NEXT: evstddx 4, 5, 3
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 6, 452(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: lwz 8, 764(31)
; CHECK-NEXT: mulli 8, 8, 11
; CHECK-NEXT: addi 9, 8, 1
; CHECK-NEXT: mullw 9, 7, 9
; CHECK-NEXT: add 9, 6, 9
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: evlddx 9, 4, 9
; CHECK-NEXT: addi 8, 8, 10
; CHECK-NEXT: mullw 7, 7, 8
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: evlddx 4, 4, 6
; CHECK-NEXT: efdsub 4, 9, 4
; CHECK-NEXT: li 6, 272
; CHECK-NEXT: evstddx 4, 31, 6
; CHECK-NEXT: lwz 4, 868(31)
; CHECK-NEXT: lwz 6, 452(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: lwz 8, 764(31)
; CHECK-NEXT: mulli 8, 8, 11
; CHECK-NEXT: addi 9, 8, 1
; CHECK-NEXT: mullw 9, 7, 9
; CHECK-NEXT: add 9, 6, 9
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 9, 4, 9
; CHECK-NEXT: evlddx 9, 9, 3
; CHECK-NEXT: addi 8, 8, 10
; CHECK-NEXT: mullw 7, 7, 8
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: add 4, 4, 6
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: efdsub 6, 9, 4
; CHECK-NEXT: addi 4, 31, 272
; CHECK-NEXT: evstddx 6, 4, 3
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 7, 452(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: mulli 9, 9, 11
; CHECK-NEXT: addi 10, 9, 2
; CHECK-NEXT: mullw 10, 8, 10
; CHECK-NEXT: add 10, 7, 10
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: evlddx 10, 6, 10
; CHECK-NEXT: addi 9, 9, 9
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: add 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: evlddx 6, 6, 7
; CHECK-NEXT: efdadd 6, 10, 6
; CHECK-NEXT: li 7, 400
; CHECK-NEXT: evstddx 6, 31, 7
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 7, 452(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: mulli 9, 9, 11
; CHECK-NEXT: addi 10, 9, 2
; CHECK-NEXT: mullw 10, 8, 10
; CHECK-NEXT: add 10, 7, 10
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: add 10, 6, 10
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: addi 9, 9, 9
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: add 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdadd 6, 10, 6
; CHECK-NEXT: addi 7, 31, 400
; CHECK-NEXT: evstddx 6, 7, 3
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 8, 452(31)
; CHECK-NEXT: lwz 9, 876(31)
; CHECK-NEXT: lwz 10, 764(31)
; CHECK-NEXT: mulli 10, 10, 11
; CHECK-NEXT: addi 11, 10, 2
; CHECK-NEXT: mullw 11, 9, 11
; CHECK-NEXT: add 11, 8, 11
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: evlddx 11, 6, 11
; CHECK-NEXT: addi 10, 10, 9
; CHECK-NEXT: mullw 9, 9, 10
; CHECK-NEXT: add 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: evlddx 6, 6, 8
; CHECK-NEXT: efdsub 6, 11, 6
; CHECK-NEXT: li 8, 288
; CHECK-NEXT: evstddx 6, 31, 8
; CHECK-NEXT: lwz 6, 868(31)
; CHECK-NEXT: lwz 8, 452(31)
; CHECK-NEXT: lwz 9, 876(31)
; CHECK-NEXT: lwz 10, 764(31)
; CHECK-NEXT: mulli 10, 10, 11
; CHECK-NEXT: addi 11, 10, 2
; CHECK-NEXT: mullw 11, 9, 11
; CHECK-NEXT: add 11, 8, 11
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: add 11, 6, 11
; CHECK-NEXT: evlddx 11, 11, 3
; CHECK-NEXT: addi 10, 10, 9
; CHECK-NEXT: mullw 9, 9, 10
; CHECK-NEXT: add 8, 8, 9
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: add 6, 6, 8
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdsub 8, 11, 6
; CHECK-NEXT: addi 6, 31, 288
; CHECK-NEXT: evstddx 8, 6, 3
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 9, 452(31)
; CHECK-NEXT: lwz 10, 876(31)
; CHECK-NEXT: lwz 11, 764(31)
; CHECK-NEXT: mulli 11, 11, 11
; CHECK-NEXT: addi 12, 11, 3
; CHECK-NEXT: mullw 12, 10, 12
; CHECK-NEXT: add 12, 9, 12
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: evlddx 12, 8, 12
; CHECK-NEXT: addi 11, 11, 8
; CHECK-NEXT: mullw 10, 10, 11
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: evlddx 8, 8, 9
; CHECK-NEXT: efdadd 8, 12, 8
; CHECK-NEXT: li 9, 384
; CHECK-NEXT: evstddx 8, 31, 9
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 9, 452(31)
; CHECK-NEXT: lwz 10, 876(31)
; CHECK-NEXT: lwz 11, 764(31)
; CHECK-NEXT: mulli 11, 11, 11
; CHECK-NEXT: addi 12, 11, 3
; CHECK-NEXT: mullw 12, 10, 12
; CHECK-NEXT: add 12, 9, 12
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: add 12, 8, 12
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: addi 11, 11, 8
; CHECK-NEXT: mullw 10, 10, 11
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 8, 8, 9
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: efdadd 8, 12, 8
; CHECK-NEXT: addi 9, 31, 384
; CHECK-NEXT: evstddx 8, 9, 3
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 10, 452(31)
; CHECK-NEXT: lwz 11, 876(31)
; CHECK-NEXT: lwz 12, 764(31)
; CHECK-NEXT: mulli 12, 12, 11
; CHECK-NEXT: addi 0, 12, 3
; CHECK-NEXT: mullw 0, 11, 0
; CHECK-NEXT: add 0, 10, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 0, 8, 0
; CHECK-NEXT: addi 12, 12, 8
; CHECK-NEXT: mullw 11, 11, 12
; CHECK-NEXT: add 10, 10, 11
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: evlddx 8, 8, 10
; CHECK-NEXT: efdsub 8, 0, 8
; CHECK-NEXT: li 10, 304
; CHECK-NEXT: evstddx 8, 31, 10
; CHECK-NEXT: lwz 8, 868(31)
; CHECK-NEXT: lwz 10, 452(31)
; CHECK-NEXT: lwz 11, 876(31)
; CHECK-NEXT: lwz 12, 764(31)
; CHECK-NEXT: mulli 12, 12, 11
; CHECK-NEXT: addi 0, 12, 3
; CHECK-NEXT: mullw 0, 11, 0
; CHECK-NEXT: add 0, 10, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 30, 8, 0
; CHECK-NEXT: evlddx 0, 30, 3
; CHECK-NEXT: addi 12, 12, 8
; CHECK-NEXT: mullw 11, 11, 12
; CHECK-NEXT: add 10, 10, 11
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: add 8, 8, 10
; CHECK-NEXT: evlddx 8, 8, 3
; CHECK-NEXT: efdsub 10, 0, 8
; CHECK-NEXT: addi 8, 31, 304
; CHECK-NEXT: evstddx 10, 8, 3
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 11, 452(31)
; CHECK-NEXT: lwz 12, 876(31)
; CHECK-NEXT: lwz 0, 764(31)
; CHECK-NEXT: mulli 30, 0, 11
; CHECK-NEXT: addi 0, 30, 4
; CHECK-NEXT: mullw 0, 12, 0
; CHECK-NEXT: add 0, 11, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 0, 10, 0
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 12, 12, 30
; CHECK-NEXT: add 11, 11, 12
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: evlddx 10, 10, 11
; CHECK-NEXT: efdadd 10, 0, 10
; CHECK-NEXT: li 11, 368
; CHECK-NEXT: evstddx 10, 31, 11
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 11, 452(31)
; CHECK-NEXT: lwz 12, 876(31)
; CHECK-NEXT: lwz 0, 764(31)
; CHECK-NEXT: mulli 30, 0, 11
; CHECK-NEXT: addi 0, 30, 4
; CHECK-NEXT: mullw 0, 12, 0
; CHECK-NEXT: add 0, 11, 0
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 29, 10, 0
; CHECK-NEXT: evlddx 0, 29, 3
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 12, 12, 30
; CHECK-NEXT: add 11, 11, 12
; CHECK-NEXT: slwi 11, 11, 4
; CHECK-NEXT: add 10, 10, 11
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: efdadd 10, 0, 10
; CHECK-NEXT: addi 11, 31, 368
; CHECK-NEXT: evstddx 10, 11, 3
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 12, 452(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 30, 764(31)
; CHECK-NEXT: mulli 30, 30, 11
; CHECK-NEXT: addi 29, 30, 4
; CHECK-NEXT: mullw 29, 0, 29
; CHECK-NEXT: add 29, 12, 29
; CHECK-NEXT: slwi 29, 29, 4
; CHECK-NEXT: evlddx 29, 10, 29
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 0, 0, 30
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: evlddx 10, 10, 12
; CHECK-NEXT: efdsub 10, 29, 10
; CHECK-NEXT: li 12, 320
; CHECK-NEXT: evstddx 10, 31, 12
; CHECK-NEXT: lwz 10, 868(31)
; CHECK-NEXT: lwz 12, 452(31)
; CHECK-NEXT: lwz 0, 876(31)
; CHECK-NEXT: lwz 30, 764(31)
; CHECK-NEXT: mulli 30, 30, 11
; CHECK-NEXT: addi 29, 30, 4
; CHECK-NEXT: mullw 29, 0, 29
; CHECK-NEXT: add 29, 12, 29
; CHECK-NEXT: slwi 29, 29, 4
; CHECK-NEXT: add 29, 10, 29
; CHECK-NEXT: evlddx 29, 29, 3
; CHECK-NEXT: addi 30, 30, 7
; CHECK-NEXT: mullw 0, 0, 30
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: slwi 12, 12, 4
; CHECK-NEXT: add 10, 10, 12
; CHECK-NEXT: evlddx 10, 10, 3
; CHECK-NEXT: efdsub 12, 29, 10
; CHECK-NEXT: addi 10, 31, 320
; CHECK-NEXT: evstddx 12, 10, 3
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 452(31)
; CHECK-NEXT: lwz 30, 876(31)
; CHECK-NEXT: lwz 29, 764(31)
; CHECK-NEXT: mulli 29, 29, 11
; CHECK-NEXT: addi 28, 29, 5
; CHECK-NEXT: mullw 28, 30, 28
; CHECK-NEXT: add 28, 0, 28
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: evlddx 28, 12, 28
; CHECK-NEXT: addi 29, 29, 6
; CHECK-NEXT: mullw 30, 30, 29
; CHECK-NEXT: add 0, 0, 30
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 12, 12, 0
; CHECK-NEXT: efdadd 12, 28, 12
; CHECK-NEXT: li 30, 352
; CHECK-NEXT: evstddx 12, 31, 30
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 452(31)
; CHECK-NEXT: lwz 30, 876(31)
; CHECK-NEXT: lwz 29, 764(31)
; CHECK-NEXT: mulli 29, 29, 11
; CHECK-NEXT: addi 28, 29, 5
; CHECK-NEXT: mullw 28, 30, 28
; CHECK-NEXT: add 28, 0, 28
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: add 28, 12, 28
; CHECK-NEXT: evlddx 28, 28, 3
; CHECK-NEXT: addi 29, 29, 6
; CHECK-NEXT: mullw 30, 30, 29
; CHECK-NEXT: add 0, 0, 30
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: efdadd 12, 28, 12
; CHECK-NEXT: addi 30, 31, 352
; CHECK-NEXT: evstddx 12, 30, 3
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 452(31)
; CHECK-NEXT: lwz 29, 876(31)
; CHECK-NEXT: lwz 28, 764(31)
; CHECK-NEXT: mulli 28, 28, 11
; CHECK-NEXT: addi 27, 28, 5
; CHECK-NEXT: mullw 27, 29, 27
; CHECK-NEXT: add 27, 0, 27
; CHECK-NEXT: slwi 27, 27, 4
; CHECK-NEXT: evlddx 27, 12, 27
; CHECK-NEXT: addi 28, 28, 6
; CHECK-NEXT: mullw 29, 29, 28
; CHECK-NEXT: add 0, 0, 29
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: evlddx 12, 12, 0
; CHECK-NEXT: efdsub 12, 27, 12
; CHECK-NEXT: li 29, 336
; CHECK-NEXT: evstddx 12, 31, 29
; CHECK-NEXT: lwz 12, 868(31)
; CHECK-NEXT: lwz 0, 452(31)
; CHECK-NEXT: lwz 29, 876(31)
; CHECK-NEXT: lwz 28, 764(31)
; CHECK-NEXT: mulli 28, 28, 11
; CHECK-NEXT: addi 27, 28, 5
; CHECK-NEXT: mullw 27, 29, 27
; CHECK-NEXT: add 27, 0, 27
; CHECK-NEXT: slwi 27, 27, 4
; CHECK-NEXT: add 27, 12, 27
; CHECK-NEXT: evlddx 27, 27, 3
; CHECK-NEXT: addi 28, 28, 6
; CHECK-NEXT: mullw 29, 29, 28
; CHECK-NEXT: add 0, 0, 29
; CHECK-NEXT: slwi 0, 0, 4
; CHECK-NEXT: add 12, 12, 0
; CHECK-NEXT: evlddx 12, 12, 3
; CHECK-NEXT: efdsub 0, 27, 12
; CHECK-NEXT: addi 12, 31, 336
; CHECK-NEXT: evstddx 0, 12, 3
; CHECK-NEXT: li 29, 432
; CHECK-NEXT: evlddx 0, 31, 29
; CHECK-NEXT: li 29, 416
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 400
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 384
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 368
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: li 29, 352
; CHECK-NEXT: evlddx 29, 31, 29
; CHECK-NEXT: efdadd 0, 0, 29
; CHECK-NEXT: lwz 29, 864(31)
; CHECK-NEXT: lwz 28, 452(31)
; CHECK-NEXT: lwz 27, 876(31)
; CHECK-NEXT: lwz 26, 764(31)
; CHECK-NEXT: mullw 27, 27, 26
; CHECK-NEXT: add 28, 28, 27
; CHECK-NEXT: slwi 28, 28, 4
; CHECK-NEXT: evstddx 0, 29, 28
; CHECK-NEXT: addi 29, 31, 432
; CHECK-NEXT: evlddx 0, 29, 3
; CHECK-NEXT: evlddx 28, 5, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 7, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 9, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 11, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: evlddx 28, 30, 3
; CHECK-NEXT: efdadd 0, 0, 28
; CHECK-NEXT: lwz 28, 864(31)
; CHECK-NEXT: lwz 27, 452(31)
; CHECK-NEXT: lwz 26, 876(31)
; CHECK-NEXT: lwz 25, 764(31)
; CHECK-NEXT: mullw 26, 26, 25
; CHECK-NEXT: add 27, 27, 26
; CHECK-NEXT: slwi 27, 27, 4
; CHECK-NEXT: add 28, 28, 27
; CHECK-NEXT: evstddx 0, 28, 3
; CHECK-NEXT: li 28, 432
; CHECK-NEXT: evlddx 28, 31, 28
; CHECK-NEXT: li 27, 416
; CHECK-NEXT: evlddx 27, 31, 27
; CHECK-NEXT: li 0, .LCPI1_5@l
; CHECK-NEXT: lis 26, .LCPI1_5@ha
; CHECK-NEXT: evlddx 0, 26, 0
; CHECK-NEXT: efdmul 27, 27, 0
; CHECK-NEXT: efdadd 27, 27, 28
; CHECK-NEXT: li 28, 400
; CHECK-NEXT: evlddx 26, 31, 28
; CHECK-NEXT: li 28, .LCPI1_6@l
; CHECK-NEXT: lis 25, .LCPI1_6@ha
; CHECK-NEXT: evlddx 28, 25, 28
; CHECK-NEXT: efdmul 26, 26, 28
; CHECK-NEXT: efdadd 26, 26, 27
; CHECK-NEXT: li 27, 384
; CHECK-NEXT: evlddx 25, 31, 27
; CHECK-NEXT: li 27, .LCPI1_7@l
; CHECK-NEXT: lis 24, .LCPI1_7@ha
; CHECK-NEXT: evlddx 27, 24, 27
; CHECK-NEXT: efdmul 25, 25, 27
; CHECK-NEXT: efdadd 26, 25, 26
; CHECK-NEXT: li 25, 368
; CHECK-NEXT: evlddx 24, 31, 25
; CHECK-NEXT: li 25, .LCPI1_8@l
; CHECK-NEXT: lis 23, .LCPI1_8@ha
; CHECK-NEXT: evlddx 25, 23, 25
; CHECK-NEXT: efdmul 24, 24, 25
; CHECK-NEXT: efdadd 24, 24, 26
; CHECK-NEXT: li 26, 352
; CHECK-NEXT: evlddx 23, 31, 26
; CHECK-NEXT: li 26, .LCPI1_9@l
; CHECK-NEXT: lis 22, .LCPI1_9@ha
; CHECK-NEXT: evlddx 26, 22, 26
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evstdd 24, 224(31)
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 224
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 832
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 272
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 816
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 288
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: li 22, 800
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 304
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 784
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 320
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 768
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 336
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: addi 22, 31, 208
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: li 23, 832
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 816
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 23, 23, 21
; CHECK-NEXT: efdadd 23, 23, 20
; CHECK-NEXT: li 21, 800
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: efdneg 23, 23
; CHECK-NEXT: evstdd 23, 208(31)
; CHECK-NEXT: evldd 23, 224(31)
; CHECK-NEXT: evldd 21, 208(31)
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: li 21, 256
; CHECK-NEXT: evstddx 23, 31, 21
; CHECK-NEXT: evlddx 23, 24, 3
; CHECK-NEXT: evlddx 21, 22, 3
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: addi 21, 31, 256
; CHECK-NEXT: evstddx 23, 21, 3
; CHECK-NEXT: evldd 23, 224(31)
; CHECK-NEXT: evldd 20, 208(31)
; CHECK-NEXT: efdsub 23, 23, 20
; CHECK-NEXT: evstdd 23, 240(31)
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 22, 3
; CHECK-NEXT: efdsub 23, 24, 23
; CHECK-NEXT: addi 24, 31, 240
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: slwi 20, 22, 4
; CHECK-NEXT: lwz 19, 876(31)
; CHECK-NEXT: add 23, 20, 23
; CHECK-NEXT: li 20, -16
; CHECK-NEXT: evlddx 18, 23, 20
; CHECK-NEXT: li 17, 256
; CHECK-NEXT: evlddx 17, 31, 17
; CHECK-NEXT: lwz 16, 856(31)
; CHECK-NEXT: efdcfsi 16, 16
; CHECK-NEXT: li 15, -8
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 16, 23
; CHECK-NEXT: evlddx 16, 21, 3
; CHECK-NEXT: efdmul 23, 23, 16
; CHECK-NEXT: efdmul 18, 18, 17
; CHECK-NEXT: efdsub 23, 18, 23
; CHECK-NEXT: lwz 18, 864(31)
; CHECK-NEXT: lwz 17, 764(31)
; CHECK-NEXT: lwz 16, 872(31)
; CHECK-NEXT: add 17, 17, 16
; CHECK-NEXT: mullw 19, 19, 17
; CHECK-NEXT: add 22, 22, 19
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 18, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: slwi 19, 22, 4
; CHECK-NEXT: lwz 18, 876(31)
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: evlddx 20, 23, 20
; CHECK-NEXT: evlddx 21, 21, 3
; CHECK-NEXT: lwz 19, 856(31)
; CHECK-NEXT: efdcfsi 19, 19
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 19, 23
; CHECK-NEXT: li 19, 256
; CHECK-NEXT: evlddx 19, 31, 19
; CHECK-NEXT: efdmul 23, 23, 19
; CHECK-NEXT: efdmul 21, 20, 21
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 20, 764(31)
; CHECK-NEXT: lwz 19, 872(31)
; CHECK-NEXT: add 20, 20, 19
; CHECK-NEXT: mullw 20, 18, 20
; CHECK-NEXT: add 22, 22, 20
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 22, 21, 22
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: mulli 20, 21, 9
; CHECK-NEXT: add 20, 20, 22
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 23, 20, 23
; CHECK-NEXT: li 20, -160
; CHECK-NEXT: evlddx 19, 23, 20
; CHECK-NEXT: evldd 18, 240(31)
; CHECK-NEXT: lwz 17, 856(31)
; CHECK-NEXT: efdcfsi 17, 17
; CHECK-NEXT: li 16, -152
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 17, 23
; CHECK-NEXT: evlddx 17, 24, 3
; CHECK-NEXT: efdmul 23, 23, 17
; CHECK-NEXT: efdmul 19, 19, 18
; CHECK-NEXT: efdsub 23, 19, 23
; CHECK-NEXT: lwz 19, 864(31)
; CHECK-NEXT: lwz 18, 764(31)
; CHECK-NEXT: lwz 17, 872(31)
; CHECK-NEXT: mulli 17, 17, 10
; CHECK-NEXT: add 18, 18, 17
; CHECK-NEXT: mullw 21, 21, 18
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 19, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: mulli 19, 21, 9
; CHECK-NEXT: add 19, 19, 22
; CHECK-NEXT: slwi 19, 19, 4
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: evlddx 20, 23, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: lwz 19, 856(31)
; CHECK-NEXT: efdcfsi 19, 19
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 19, 23
; CHECK-NEXT: evldd 19, 240(31)
; CHECK-NEXT: efdmul 23, 23, 19
; CHECK-NEXT: efdmul 24, 20, 24
; CHECK-NEXT: efdadd 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 20, 764(31)
; CHECK-NEXT: lwz 19, 872(31)
; CHECK-NEXT: mulli 19, 19, 10
; CHECK-NEXT: add 20, 20, 19
; CHECK-NEXT: mullw 21, 21, 20
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 432
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 416
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 400
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 384
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 368
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 352
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evstdd 24, 160(31)
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 160
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 816
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 272
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 288
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: li 22, 768
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 304
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 800
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 320
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 832
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 336
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: addi 22, 31, 144
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: li 23, 816
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 784
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 23, 23, 21
; CHECK-NEXT: efdadd 23, 23, 20
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: li 21, 800
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: li 21, 832
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: efdneg 23, 23
; CHECK-NEXT: evstdd 23, 144(31)
; CHECK-NEXT: evldd 23, 160(31)
; CHECK-NEXT: evldd 21, 144(31)
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: evstdd 23, 192(31)
; CHECK-NEXT: evlddx 23, 24, 3
; CHECK-NEXT: evlddx 21, 22, 3
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: addi 21, 31, 192
; CHECK-NEXT: evstddx 23, 21, 3
; CHECK-NEXT: evldd 23, 160(31)
; CHECK-NEXT: evldd 20, 144(31)
; CHECK-NEXT: efdsub 23, 23, 20
; CHECK-NEXT: evstdd 23, 176(31)
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 22, 3
; CHECK-NEXT: efdsub 23, 24, 23
; CHECK-NEXT: addi 24, 31, 176
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: add 19, 20, 22
; CHECK-NEXT: slwi 19, 19, 4
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: li 19, -32
; CHECK-NEXT: evlddx 18, 23, 19
; CHECK-NEXT: evldd 17, 192(31)
; CHECK-NEXT: lwz 16, 856(31)
; CHECK-NEXT: efdcfsi 16, 16
; CHECK-NEXT: li 15, -24
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 16, 23
; CHECK-NEXT: evlddx 16, 21, 3
; CHECK-NEXT: efdmul 23, 23, 16
; CHECK-NEXT: efdmul 18, 18, 17
; CHECK-NEXT: efdsub 23, 18, 23
; CHECK-NEXT: lwz 18, 864(31)
; CHECK-NEXT: lwz 17, 764(31)
; CHECK-NEXT: lwz 16, 872(31)
; CHECK-NEXT: slwi 16, 16, 1
; CHECK-NEXT: add 17, 17, 16
; CHECK-NEXT: mullw 20, 20, 17
; CHECK-NEXT: add 22, 22, 20
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 18, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: add 18, 20, 22
; CHECK-NEXT: slwi 18, 18, 4
; CHECK-NEXT: add 23, 18, 23
; CHECK-NEXT: evlddx 19, 23, 19
; CHECK-NEXT: evlddx 21, 21, 3
; CHECK-NEXT: lwz 18, 856(31)
; CHECK-NEXT: efdcfsi 18, 18
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 18, 23
; CHECK-NEXT: evldd 18, 192(31)
; CHECK-NEXT: efdmul 23, 23, 18
; CHECK-NEXT: efdmul 21, 19, 21
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: slwi 18, 18, 1
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: add 22, 22, 20
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 22, 21, 22
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: slwi 20, 21, 3
; CHECK-NEXT: add 20, 20, 22
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 23, 20, 23
; CHECK-NEXT: li 20, -144
; CHECK-NEXT: evlddx 19, 23, 20
; CHECK-NEXT: evldd 18, 176(31)
; CHECK-NEXT: lwz 17, 856(31)
; CHECK-NEXT: efdcfsi 17, 17
; CHECK-NEXT: li 16, -136
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 17, 23
; CHECK-NEXT: evlddx 17, 24, 3
; CHECK-NEXT: efdmul 23, 23, 17
; CHECK-NEXT: efdmul 19, 19, 18
; CHECK-NEXT: efdsub 23, 19, 23
; CHECK-NEXT: lwz 19, 864(31)
; CHECK-NEXT: lwz 18, 764(31)
; CHECK-NEXT: lwz 17, 872(31)
; CHECK-NEXT: mulli 17, 17, 9
; CHECK-NEXT: add 18, 18, 17
; CHECK-NEXT: mullw 21, 21, 18
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 19, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: slwi 19, 21, 3
; CHECK-NEXT: add 19, 19, 22
; CHECK-NEXT: slwi 19, 19, 4
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: evlddx 20, 23, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: lwz 19, 856(31)
; CHECK-NEXT: efdcfsi 19, 19
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 19, 23
; CHECK-NEXT: evldd 19, 176(31)
; CHECK-NEXT: efdmul 23, 23, 19
; CHECK-NEXT: efdmul 24, 20, 24
; CHECK-NEXT: efdadd 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 20, 764(31)
; CHECK-NEXT: lwz 19, 872(31)
; CHECK-NEXT: mulli 19, 19, 9
; CHECK-NEXT: add 20, 20, 19
; CHECK-NEXT: mullw 21, 21, 20
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 432
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 416
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 400
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 384
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 368
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 352
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evstdd 24, 96(31)
; CHECK-NEXT: evlddx 24, 29, 3
; CHECK-NEXT: evlddx 23, 5, 3
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 7, 3
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 9, 3
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 11, 3
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evlddx 23, 30, 3
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 23, 23, 24
; CHECK-NEXT: addi 24, 31, 96
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: li 23, 800
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: li 22, 272
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 768
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: li 20, 288
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdmul 23, 23, 22
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: li 22, 816
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 304
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdsub 23, 23, 22
; CHECK-NEXT: li 22, 832
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 320
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: li 22, 784
; CHECK-NEXT: evlddx 22, 31, 22
; CHECK-NEXT: li 21, 336
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: efdmul 22, 22, 21
; CHECK-NEXT: efdadd 23, 22, 23
; CHECK-NEXT: addi 22, 31, 80
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: li 23, 800
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: evlddx 21, 4, 3
; CHECK-NEXT: li 20, 768
; CHECK-NEXT: evlddx 20, 31, 20
; CHECK-NEXT: evlddx 19, 6, 3
; CHECK-NEXT: efdmul 20, 20, 19
; CHECK-NEXT: efdmul 23, 23, 21
; CHECK-NEXT: efdsub 23, 23, 20
; CHECK-NEXT: li 21, 816
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 8, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdsub 23, 23, 21
; CHECK-NEXT: li 21, 832
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 10, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: li 21, 784
; CHECK-NEXT: evlddx 21, 31, 21
; CHECK-NEXT: evlddx 20, 12, 3
; CHECK-NEXT: efdmul 21, 21, 20
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: efdneg 23, 23
; CHECK-NEXT: evstdd 23, 80(31)
; CHECK-NEXT: evldd 23, 96(31)
; CHECK-NEXT: evldd 21, 80(31)
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: evstdd 23, 128(31)
; CHECK-NEXT: evlddx 23, 24, 3
; CHECK-NEXT: evlddx 21, 22, 3
; CHECK-NEXT: efdadd 23, 23, 21
; CHECK-NEXT: addi 21, 31, 128
; CHECK-NEXT: evstddx 23, 21, 3
; CHECK-NEXT: evldd 23, 96(31)
; CHECK-NEXT: evldd 20, 80(31)
; CHECK-NEXT: efdsub 23, 23, 20
; CHECK-NEXT: evstdd 23, 112(31)
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: evlddx 23, 22, 3
; CHECK-NEXT: efdsub 23, 24, 23
; CHECK-NEXT: addi 24, 31, 112
; CHECK-NEXT: evstddx 23, 24, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: slwi 19, 20, 1
; CHECK-NEXT: add 19, 19, 22
; CHECK-NEXT: slwi 19, 19, 4
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: li 19, -48
; CHECK-NEXT: evlddx 18, 23, 19
; CHECK-NEXT: evldd 17, 128(31)
; CHECK-NEXT: lwz 16, 856(31)
; CHECK-NEXT: efdcfsi 16, 16
; CHECK-NEXT: li 15, -40
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 16, 23
; CHECK-NEXT: evlddx 16, 21, 3
; CHECK-NEXT: efdmul 23, 23, 16
; CHECK-NEXT: efdmul 18, 18, 17
; CHECK-NEXT: efdsub 23, 18, 23
; CHECK-NEXT: lwz 18, 864(31)
; CHECK-NEXT: lwz 17, 764(31)
; CHECK-NEXT: lwz 16, 872(31)
; CHECK-NEXT: mulli 16, 16, 3
; CHECK-NEXT: add 17, 17, 16
; CHECK-NEXT: mullw 20, 20, 17
; CHECK-NEXT: add 22, 22, 20
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 18, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 20, 876(31)
; CHECK-NEXT: slwi 18, 20, 1
; CHECK-NEXT: add 18, 18, 22
; CHECK-NEXT: slwi 18, 18, 4
; CHECK-NEXT: add 23, 18, 23
; CHECK-NEXT: evlddx 19, 23, 19
; CHECK-NEXT: evlddx 21, 21, 3
; CHECK-NEXT: lwz 18, 856(31)
; CHECK-NEXT: efdcfsi 18, 18
; CHECK-NEXT: evlddx 23, 23, 15
; CHECK-NEXT: efdmul 23, 18, 23
; CHECK-NEXT: evldd 18, 128(31)
; CHECK-NEXT: efdmul 23, 23, 18
; CHECK-NEXT: efdmul 21, 19, 21
; CHECK-NEXT: efdadd 23, 21, 23
; CHECK-NEXT: lwz 21, 864(31)
; CHECK-NEXT: lwz 19, 764(31)
; CHECK-NEXT: lwz 18, 872(31)
; CHECK-NEXT: mulli 18, 18, 3
; CHECK-NEXT: add 19, 19, 18
; CHECK-NEXT: mullw 20, 20, 19
; CHECK-NEXT: add 22, 22, 20
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 22, 21, 22
; CHECK-NEXT: evstddx 23, 22, 3
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: mulli 20, 21, 7
; CHECK-NEXT: add 20, 20, 22
; CHECK-NEXT: slwi 20, 20, 4
; CHECK-NEXT: add 23, 20, 23
; CHECK-NEXT: li 20, -128
; CHECK-NEXT: evlddx 19, 23, 20
; CHECK-NEXT: evldd 18, 112(31)
; CHECK-NEXT: lwz 17, 856(31)
; CHECK-NEXT: efdcfsi 17, 17
; CHECK-NEXT: li 16, -120
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 17, 23
; CHECK-NEXT: evlddx 17, 24, 3
; CHECK-NEXT: efdmul 23, 23, 17
; CHECK-NEXT: efdmul 19, 19, 18
; CHECK-NEXT: efdsub 23, 19, 23
; CHECK-NEXT: lwz 19, 864(31)
; CHECK-NEXT: lwz 18, 764(31)
; CHECK-NEXT: lwz 17, 872(31)
; CHECK-NEXT: slwi 17, 17, 3
; CHECK-NEXT: add 18, 18, 17
; CHECK-NEXT: mullw 21, 21, 18
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: evstddx 23, 19, 22
; CHECK-NEXT: lwz 23, 860(31)
; CHECK-NEXT: lwz 22, 452(31)
; CHECK-NEXT: lwz 21, 876(31)
; CHECK-NEXT: mulli 19, 21, 7
; CHECK-NEXT: add 19, 19, 22
; CHECK-NEXT: slwi 19, 19, 4
; CHECK-NEXT: add 23, 19, 23
; CHECK-NEXT: evlddx 20, 23, 20
; CHECK-NEXT: evlddx 24, 24, 3
; CHECK-NEXT: lwz 19, 856(31)
; CHECK-NEXT: efdcfsi 19, 19
; CHECK-NEXT: evlddx 23, 23, 16
; CHECK-NEXT: efdmul 23, 19, 23
; CHECK-NEXT: evldd 19, 112(31)
; CHECK-NEXT: efdmul 23, 23, 19
; CHECK-NEXT: efdmul 24, 20, 24
; CHECK-NEXT: efdadd 24, 24, 23
; CHECK-NEXT: lwz 23, 864(31)
; CHECK-NEXT: lwz 20, 764(31)
; CHECK-NEXT: lwz 19, 872(31)
; CHECK-NEXT: slwi 19, 19, 3
; CHECK-NEXT: add 20, 20, 19
; CHECK-NEXT: mullw 21, 21, 20
; CHECK-NEXT: add 22, 22, 21
; CHECK-NEXT: slwi 22, 22, 4
; CHECK-NEXT: add 23, 23, 22
; CHECK-NEXT: evstddx 24, 23, 3
; CHECK-NEXT: li 24, 432
; CHECK-NEXT: evlddx 24, 31, 24
; CHECK-NEXT: li 23, 416
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 25
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 400
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 27
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 384
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 0
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 368
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 26
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: li 23, 352
; CHECK-NEXT: evlddx 23, 31, 23
; CHECK-NEXT: efdmul 23, 23, 28
; CHECK-NEXT: efdadd 24, 23, 24
; CHECK-NEXT: evstdd 24, 32(31)
; CHECK-NEXT: evlddx 29, 29, 3
; CHECK-NEXT: evlddx 5, 5, 3
; CHECK-NEXT: efdmul 5, 5, 25
; CHECK-NEXT: efdadd 5, 5, 29
; CHECK-NEXT: evlddx 7, 7, 3
; CHECK-NEXT: efdmul 7, 7, 27
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 9, 3
; CHECK-NEXT: efdmul 7, 7, 0
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 11, 3
; CHECK-NEXT: efdmul 7, 7, 26
; CHECK-NEXT: efdadd 5, 7, 5
; CHECK-NEXT: evlddx 7, 30, 3
; CHECK-NEXT: efdmul 7, 7, 28
; CHECK-NEXT: efdadd 7, 7, 5
; CHECK-NEXT: addi 5, 31, 32
; CHECK-NEXT: evstddx 7, 5, 3
; CHECK-NEXT: li 7, 784
; CHECK-NEXT: evlddx 7, 31, 7
; CHECK-NEXT: li 9, 272
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 800
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: li 30, 288
; CHECK-NEXT: evlddx 0, 31, 30
; CHECK-NEXT: efdmul 11, 11, 0
; CHECK-NEXT: efdmul 7, 7, 9
; CHECK-NEXT: efdsub 7, 7, 11
; CHECK-NEXT: li 9, 832
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 304
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdadd 7, 9, 7
; CHECK-NEXT: li 9, 768
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 320
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdadd 7, 9, 7
; CHECK-NEXT: li 9, 816
; CHECK-NEXT: evlddx 9, 31, 9
; CHECK-NEXT: li 11, 336
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: efdmul 9, 9, 11
; CHECK-NEXT: efdsub 7, 7, 9
; CHECK-NEXT: addi 9, 31, 16
; CHECK-NEXT: evstddx 7, 9, 3
; CHECK-NEXT: li 7, 784
; CHECK-NEXT: evlddx 7, 31, 7
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: li 11, 800
; CHECK-NEXT: evlddx 11, 31, 11
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: efdmul 6, 11, 6
; CHECK-NEXT: efdmul 4, 7, 4
; CHECK-NEXT: efdsub 4, 4, 6
; CHECK-NEXT: li 6, 832
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 7, 8, 3
; CHECK-NEXT: efdmul 6, 6, 7
; CHECK-NEXT: efdadd 4, 6, 4
; CHECK-NEXT: li 6, 768
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 7, 10, 3
; CHECK-NEXT: efdmul 6, 6, 7
; CHECK-NEXT: efdadd 4, 6, 4
; CHECK-NEXT: li 6, 816
; CHECK-NEXT: evlddx 6, 31, 6
; CHECK-NEXT: evlddx 7, 12, 3
; CHECK-NEXT: efdmul 6, 6, 7
; CHECK-NEXT: efdsub 4, 4, 6
; CHECK-NEXT: efdneg 4, 4
; CHECK-NEXT: evstdd 4, 16(31)
; CHECK-NEXT: evldd 4, 32(31)
; CHECK-NEXT: evldd 6, 16(31)
; CHECK-NEXT: efdadd 4, 4, 6
; CHECK-NEXT: evstdd 4, 64(31)
; CHECK-NEXT: evlddx 4, 5, 3
; CHECK-NEXT: evlddx 6, 9, 3
; CHECK-NEXT: efdadd 4, 4, 6
; CHECK-NEXT: addi 6, 31, 64
; CHECK-NEXT: evstddx 4, 6, 3
; CHECK-NEXT: evldd 4, 32(31)
; CHECK-NEXT: evldd 7, 16(31)
; CHECK-NEXT: efdsub 4, 4, 7
; CHECK-NEXT: evstdd 4, 48(31)
; CHECK-NEXT: evlddx 4, 5, 3
; CHECK-NEXT: evlddx 5, 9, 3
; CHECK-NEXT: efdsub 5, 4, 5
; CHECK-NEXT: addi 4, 31, 48
; CHECK-NEXT: evstddx 5, 4, 3
; CHECK-NEXT: lwz 5, 860(31)
; CHECK-NEXT: lwz 7, 452(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: mulli 9, 8, 3
; CHECK-NEXT: add 9, 9, 7
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 5, 9, 5
; CHECK-NEXT: li 9, -64
; CHECK-NEXT: evlddx 10, 5, 9
; CHECK-NEXT: evldd 11, 64(31)
; CHECK-NEXT: lwz 12, 856(31)
; CHECK-NEXT: efdcfsi 12, 12
; CHECK-NEXT: li 0, -56
; CHECK-NEXT: evlddx 5, 5, 0
; CHECK-NEXT: efdmul 5, 12, 5
; CHECK-NEXT: evlddx 12, 6, 3
; CHECK-NEXT: efdmul 5, 5, 12
; CHECK-NEXT: efdmul 10, 10, 11
; CHECK-NEXT: efdsub 5, 10, 5
; CHECK-NEXT: lwz 10, 864(31)
; CHECK-NEXT: lwz 11, 764(31)
; CHECK-NEXT: lwz 12, 872(31)
; CHECK-NEXT: slwi 12, 12, 2
; CHECK-NEXT: add 11, 11, 12
; CHECK-NEXT: mullw 8, 8, 11
; CHECK-NEXT: add 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: evstddx 5, 10, 7
; CHECK-NEXT: lwz 5, 860(31)
; CHECK-NEXT: lwz 7, 452(31)
; CHECK-NEXT: lwz 8, 876(31)
; CHECK-NEXT: mulli 10, 8, 3
; CHECK-NEXT: add 10, 10, 7
; CHECK-NEXT: slwi 10, 10, 4
; CHECK-NEXT: add 5, 10, 5
; CHECK-NEXT: evlddx 9, 5, 9
; CHECK-NEXT: evlddx 6, 6, 3
; CHECK-NEXT: lwz 10, 856(31)
; CHECK-NEXT: efdcfsi 10, 10
; CHECK-NEXT: evlddx 5, 5, 0
; CHECK-NEXT: efdmul 5, 10, 5
; CHECK-NEXT: evldd 10, 64(31)
; CHECK-NEXT: efdmul 5, 5, 10
; CHECK-NEXT: efdmul 6, 9, 6
; CHECK-NEXT: efdadd 5, 6, 5
; CHECK-NEXT: lwz 6, 864(31)
; CHECK-NEXT: lwz 9, 764(31)
; CHECK-NEXT: lwz 10, 872(31)
; CHECK-NEXT: slwi 10, 10, 2
; CHECK-NEXT: add 9, 9, 10
; CHECK-NEXT: mullw 8, 8, 9
; CHECK-NEXT: add 7, 7, 8
; CHECK-NEXT: slwi 7, 7, 4
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: evstddx 5, 6, 3
; CHECK-NEXT: lwz 5, 860(31)
; CHECK-NEXT: lwz 6, 452(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: mulli 8, 7, 6
; CHECK-NEXT: add 8, 8, 6
; CHECK-NEXT: slwi 8, 8, 4
; CHECK-NEXT: add 5, 8, 5
; CHECK-NEXT: li 8, -112
; CHECK-NEXT: evlddx 9, 5, 8
; CHECK-NEXT: evldd 10, 48(31)
; CHECK-NEXT: lwz 11, 856(31)
; CHECK-NEXT: efdcfsi 11, 11
; CHECK-NEXT: li 12, -104
; CHECK-NEXT: evlddx 5, 5, 12
; CHECK-NEXT: efdmul 5, 11, 5
; CHECK-NEXT: evlddx 11, 4, 3
; CHECK-NEXT: efdmul 5, 5, 11
; CHECK-NEXT: efdmul 9, 9, 10
; CHECK-NEXT: efdsub 5, 9, 5
; CHECK-NEXT: lwz 9, 864(31)
; CHECK-NEXT: lwz 10, 764(31)
; CHECK-NEXT: lwz 11, 872(31)
; CHECK-NEXT: mulli 11, 11, 7
; CHECK-NEXT: add 10, 10, 11
; CHECK-NEXT: mullw 7, 7, 10
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: evstddx 5, 9, 6
; CHECK-NEXT: lwz 5, 860(31)
; CHECK-NEXT: lwz 6, 452(31)
; CHECK-NEXT: lwz 7, 876(31)
; CHECK-NEXT: mulli 9, 7, 6
; CHECK-NEXT: add 9, 9, 6
; CHECK-NEXT: slwi 9, 9, 4
; CHECK-NEXT: add 5, 9, 5
; CHECK-NEXT: evlddx 8, 5, 8
; CHECK-NEXT: evlddx 4, 4, 3
; CHECK-NEXT: lwz 9, 856(31)
; CHECK-NEXT: efdcfsi 9, 9
; CHECK-NEXT: evlddx 5, 5, 12
; CHECK-NEXT: efdmul 5, 9, 5
; CHECK-NEXT: evldd 9, 48(31)
; CHECK-NEXT: efdmul 5, 5, 9
; CHECK-NEXT: efdmul 4, 8, 4
; CHECK-NEXT: efdadd 4, 4, 5
; CHECK-NEXT: lwz 5, 864(31)
; CHECK-NEXT: lwz 8, 764(31)
; CHECK-NEXT: lwz 9, 872(31)
; CHECK-NEXT: mulli 9, 9, 7
; CHECK-NEXT: add 8, 8, 9
; CHECK-NEXT: mullw 7, 7, 8
; CHECK-NEXT: add 6, 6, 7
; CHECK-NEXT: slwi 6, 6, 4
; CHECK-NEXT: add 5, 5, 6
; CHECK-NEXT: evstddx 4, 5, 3
; CHECK-NEXT: b .LBB1_5
; CHECK-NEXT: .LBB1_5:
; CHECK-NEXT: lwz 3, 452(31)
; CHECK-NEXT: addi 3, 3, 1
; CHECK-NEXT: stw 3, 452(31)
; CHECK-NEXT: b .LBB1_3
; CHECK-NEXT: .LBB1_6:
; CHECK-NEXT: b .LBB1_7
; CHECK-NEXT: .LBB1_7:
; CHECK-NEXT: lwz 3, 764(31)
; CHECK-NEXT: addi 3, 3, 1
; CHECK-NEXT: stw 3, 764(31)
; CHECK-NEXT: b .LBB1_1
; CHECK-NEXT: .LBB1_8:
; CHECK-NEXT: lwz 30, 1080(31) # 4-byte Folded Reload
; CHECK-NEXT: li 3, 984
; CHECK-NEXT: evlddx 29, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 976
; CHECK-NEXT: evlddx 28, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 968
; CHECK-NEXT: evlddx 27, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 960
; CHECK-NEXT: evlddx 26, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 952
; CHECK-NEXT: evlddx 25, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 944
; CHECK-NEXT: evlddx 24, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 936
; CHECK-NEXT: evlddx 23, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 928
; CHECK-NEXT: evlddx 22, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 920
; CHECK-NEXT: evlddx 21, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 912
; CHECK-NEXT: evlddx 20, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 904
; CHECK-NEXT: evlddx 19, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 896
; CHECK-NEXT: evlddx 18, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 888
; CHECK-NEXT: evlddx 17, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: li 3, 880
; CHECK-NEXT: evlddx 16, 31, 3 # 8-byte Folded Reload
; CHECK-NEXT: lwz 15, 1020(31) # 4-byte Folded Reload
; CHECK-NEXT: lwz 31, 1084(1)
; CHECK-NEXT: addi 1, 1, 1088
; CHECK-NEXT: blr
%7 = alloca i32, align 4
%8 = alloca i32, align 4
%9 = alloca ptr, align 4
%10 = alloca ptr, align 4
%11 = alloca ptr, align 4
%12 = alloca i32, align 4
%13 = alloca i32, align 4
%14 = alloca double, align 8
%15 = alloca double, align 8
%16 = alloca double, align 8
%17 = alloca double, align 8
%18 = alloca double, align 8
%19 = alloca double, align 8
%20 = alloca double, align 8
%21 = alloca double, align 8
%22 = alloca double, align 8
%23 = alloca double, align 8
%24 = alloca i32, align 4
%25 = alloca %struct.cmplx, align 8
%26 = alloca %struct.cmplx, align 8
%27 = alloca %struct.cmplx, align 8
%28 = alloca %struct.cmplx, align 8
%29 = alloca %struct.cmplx, align 8
%30 = alloca %struct.cmplx, align 8
%31 = alloca %struct.cmplx, align 8
%32 = alloca %struct.cmplx, align 8
%33 = alloca %struct.cmplx, align 8
%34 = alloca %struct.cmplx, align 8
%35 = alloca %struct.cmplx, align 8
%36 = alloca %struct.cmplx, align 8
%37 = alloca %struct.cmplx, align 8
%38 = alloca %struct.cmplx, align 8
%39 = alloca %struct.cmplx, align 8
%40 = alloca %struct.cmplx, align 8
%41 = alloca %struct.cmplx, align 8
%42 = alloca %struct.cmplx, align 8
%43 = alloca %struct.cmplx, align 8
%44 = alloca i32, align 4
%45 = alloca %struct.cmplx, align 8
%46 = alloca %struct.cmplx, align 8
%47 = alloca %struct.cmplx, align 8
%48 = alloca %struct.cmplx, align 8
%49 = alloca %struct.cmplx, align 8
%50 = alloca %struct.cmplx, align 8
%51 = alloca %struct.cmplx, align 8
%52 = alloca %struct.cmplx, align 8
%53 = alloca %struct.cmplx, align 8
%54 = alloca %struct.cmplx, align 8
%55 = alloca %struct.cmplx, align 8
%56 = alloca %struct.cmplx, align 8
%57 = alloca %struct.cmplx, align 8
%58 = alloca %struct.cmplx, align 8
%59 = alloca %struct.cmplx, align 8
%60 = alloca %struct.cmplx, align 8
%61 = alloca %struct.cmplx, align 8
%62 = alloca %struct.cmplx, align 8
%63 = alloca %struct.cmplx, align 8
%64 = alloca %struct.cmplx, align 8
%65 = alloca %struct.cmplx, align 8
%66 = alloca %struct.cmplx, align 8
%67 = alloca %struct.cmplx, align 8
%68 = alloca %struct.cmplx, align 8
%69 = alloca %struct.cmplx, align 8
%70 = alloca %struct.cmplx, align 8
%71 = alloca %struct.cmplx, align 8
store i32 %0, ptr %7, align 4
store i32 %1, ptr %8, align 4
store ptr %2, ptr %9, align 4
store ptr %3, ptr %10, align 4
store ptr %4, ptr %11, align 4
store i32 %5, ptr %12, align 4
store i32 11, ptr %13, align 4
store double 0x3FEAEB8C8764F0BA, ptr %14, align 8
%72 = load i32, ptr %12, align 4
%73 = sitofp i32 %72 to double
%74 = fmul double %73, 0x3FE14CEDF8BB580B
store double %74, ptr %15, align 8
store double 0x3FDA9628D9C712B6, ptr %16, align 8
%75 = load i32, ptr %12, align 4
%76 = sitofp i32 %75 to double
%77 = fmul double %76, 0x3FED1BB48EEE2C13
store double %77, ptr %17, align 8
store double 0xBFC2375F640F44DB, ptr %18, align 8
%78 = load i32, ptr %12, align 4
%79 = sitofp i32 %78 to double
%80 = fmul double %79, 0x3FEFAC9E043842EF
store double %80, ptr %19, align 8
store double 0xBFE4F49E7F775887, ptr %20, align 8
%81 = load i32, ptr %12, align 4
%82 = sitofp i32 %81 to double
%83 = fmul double %82, 0x3FE82F19BB3A28A1
store double %83, ptr %21, align 8
store double 0xBFEEB42A9BCD5057, ptr %22, align 8
%84 = load i32, ptr %12, align 4
%85 = sitofp i32 %84 to double
%86 = fmul double %85, 0x3FD207E7FD768DBF
store double %86, ptr %23, align 8
store i32 0, ptr %24, align 4
br label %87
87: ; preds = %2792, %6
%88 = load i32, ptr %24, align 4
%89 = load i32, ptr %8, align 4
%90 = icmp ult i32 %88, %89
br i1 %90, label %91, label %2795
91: ; preds = %87
%92 = load ptr, ptr %9, align 4
%93 = load i32, ptr %7, align 4
%94 = load i32, ptr %24, align 4
%95 = mul i32 11, %94
%96 = add i32 0, %95
%97 = mul i32 %93, %96
%98 = add i32 0, %97
%99 = getelementptr inbounds %struct.cmplx, ptr %92, i32 %98
call void @llvm.memcpy.p0.p0.i32(ptr align 8 %25, ptr align 8 %99, i32 16, i1 false)
%100 = load ptr, ptr %9, align 4
%101 = load i32, ptr %7, align 4
%102 = load i32, ptr %24, align 4
%103 = mul i32 11, %102
%104 = add i32 1, %103
%105 = mul i32 %101, %104
%106 = add i32 0, %105
%107 = getelementptr inbounds %struct.cmplx, ptr %100, i32 %106
%108 = getelementptr inbounds %struct.cmplx, ptr %107, i32 0, i32 0
%109 = load double, ptr %108, align 8
%110 = load ptr, ptr %9, align 4
%111 = load i32, ptr %7, align 4
%112 = load i32, ptr %24, align 4
%113 = mul i32 11, %112
%114 = add i32 10, %113
%115 = mul i32 %111, %114
%116 = add i32 0, %115
%117 = getelementptr inbounds %struct.cmplx, ptr %110, i32 %116
%118 = getelementptr inbounds %struct.cmplx, ptr %117, i32 0, i32 0
%119 = load double, ptr %118, align 8
%120 = fadd double %109, %119
%121 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
store double %120, ptr %121, align 8
%122 = load ptr, ptr %9, align 4
%123 = load i32, ptr %7, align 4
%124 = load i32, ptr %24, align 4
%125 = mul i32 11, %124
%126 = add i32 1, %125
%127 = mul i32 %123, %126
%128 = add i32 0, %127
%129 = getelementptr inbounds %struct.cmplx, ptr %122, i32 %128
%130 = getelementptr inbounds %struct.cmplx, ptr %129, i32 0, i32 1
%131 = load double, ptr %130, align 8
%132 = load ptr, ptr %9, align 4
%133 = load i32, ptr %7, align 4
%134 = load i32, ptr %24, align 4
%135 = mul i32 11, %134
%136 = add i32 10, %135
%137 = mul i32 %133, %136
%138 = add i32 0, %137
%139 = getelementptr inbounds %struct.cmplx, ptr %132, i32 %138
%140 = getelementptr inbounds %struct.cmplx, ptr %139, i32 0, i32 1
%141 = load double, ptr %140, align 8
%142 = fadd double %131, %141
%143 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
store double %142, ptr %143, align 8
%144 = load ptr, ptr %9, align 4
%145 = load i32, ptr %7, align 4
%146 = load i32, ptr %24, align 4
%147 = mul i32 11, %146
%148 = add i32 1, %147
%149 = mul i32 %145, %148
%150 = add i32 0, %149
%151 = getelementptr inbounds %struct.cmplx, ptr %144, i32 %150
%152 = getelementptr inbounds %struct.cmplx, ptr %151, i32 0, i32 0
%153 = load double, ptr %152, align 8
%154 = load ptr, ptr %9, align 4
%155 = load i32, ptr %7, align 4
%156 = load i32, ptr %24, align 4
%157 = mul i32 11, %156
%158 = add i32 10, %157
%159 = mul i32 %155, %158
%160 = add i32 0, %159
%161 = getelementptr inbounds %struct.cmplx, ptr %154, i32 %160
%162 = getelementptr inbounds %struct.cmplx, ptr %161, i32 0, i32 0
%163 = load double, ptr %162, align 8
%164 = fsub double %153, %163
%165 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 0
store double %164, ptr %165, align 8
%166 = load ptr, ptr %9, align 4
%167 = load i32, ptr %7, align 4
%168 = load i32, ptr %24, align 4
%169 = mul i32 11, %168
%170 = add i32 1, %169
%171 = mul i32 %167, %170
%172 = add i32 0, %171
%173 = getelementptr inbounds %struct.cmplx, ptr %166, i32 %172
%174 = getelementptr inbounds %struct.cmplx, ptr %173, i32 0, i32 1
%175 = load double, ptr %174, align 8
%176 = load ptr, ptr %9, align 4
%177 = load i32, ptr %7, align 4
%178 = load i32, ptr %24, align 4
%179 = mul i32 11, %178
%180 = add i32 10, %179
%181 = mul i32 %177, %180
%182 = add i32 0, %181
%183 = getelementptr inbounds %struct.cmplx, ptr %176, i32 %182
%184 = getelementptr inbounds %struct.cmplx, ptr %183, i32 0, i32 1
%185 = load double, ptr %184, align 8
%186 = fsub double %175, %185
%187 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 1
store double %186, ptr %187, align 8
%188 = load ptr, ptr %9, align 4
%189 = load i32, ptr %7, align 4
%190 = load i32, ptr %24, align 4
%191 = mul i32 11, %190
%192 = add i32 2, %191
%193 = mul i32 %189, %192
%194 = add i32 0, %193
%195 = getelementptr inbounds %struct.cmplx, ptr %188, i32 %194
%196 = getelementptr inbounds %struct.cmplx, ptr %195, i32 0, i32 0
%197 = load double, ptr %196, align 8
%198 = load ptr, ptr %9, align 4
%199 = load i32, ptr %7, align 4
%200 = load i32, ptr %24, align 4
%201 = mul i32 11, %200
%202 = add i32 9, %201
%203 = mul i32 %199, %202
%204 = add i32 0, %203
%205 = getelementptr inbounds %struct.cmplx, ptr %198, i32 %204
%206 = getelementptr inbounds %struct.cmplx, ptr %205, i32 0, i32 0
%207 = load double, ptr %206, align 8
%208 = fadd double %197, %207
%209 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
store double %208, ptr %209, align 8
%210 = load ptr, ptr %9, align 4
%211 = load i32, ptr %7, align 4
%212 = load i32, ptr %24, align 4
%213 = mul i32 11, %212
%214 = add i32 2, %213
%215 = mul i32 %211, %214
%216 = add i32 0, %215
%217 = getelementptr inbounds %struct.cmplx, ptr %210, i32 %216
%218 = getelementptr inbounds %struct.cmplx, ptr %217, i32 0, i32 1
%219 = load double, ptr %218, align 8
%220 = load ptr, ptr %9, align 4
%221 = load i32, ptr %7, align 4
%222 = load i32, ptr %24, align 4
%223 = mul i32 11, %222
%224 = add i32 9, %223
%225 = mul i32 %221, %224
%226 = add i32 0, %225
%227 = getelementptr inbounds %struct.cmplx, ptr %220, i32 %226
%228 = getelementptr inbounds %struct.cmplx, ptr %227, i32 0, i32 1
%229 = load double, ptr %228, align 8
%230 = fadd double %219, %229
%231 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
store double %230, ptr %231, align 8
%232 = load ptr, ptr %9, align 4
%233 = load i32, ptr %7, align 4
%234 = load i32, ptr %24, align 4
%235 = mul i32 11, %234
%236 = add i32 2, %235
%237 = mul i32 %233, %236
%238 = add i32 0, %237
%239 = getelementptr inbounds %struct.cmplx, ptr %232, i32 %238
%240 = getelementptr inbounds %struct.cmplx, ptr %239, i32 0, i32 0
%241 = load double, ptr %240, align 8
%242 = load ptr, ptr %9, align 4
%243 = load i32, ptr %7, align 4
%244 = load i32, ptr %24, align 4
%245 = mul i32 11, %244
%246 = add i32 9, %245
%247 = mul i32 %243, %246
%248 = add i32 0, %247
%249 = getelementptr inbounds %struct.cmplx, ptr %242, i32 %248
%250 = getelementptr inbounds %struct.cmplx, ptr %249, i32 0, i32 0
%251 = load double, ptr %250, align 8
%252 = fsub double %241, %251
%253 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 0
store double %252, ptr %253, align 8
%254 = load ptr, ptr %9, align 4
%255 = load i32, ptr %7, align 4
%256 = load i32, ptr %24, align 4
%257 = mul i32 11, %256
%258 = add i32 2, %257
%259 = mul i32 %255, %258
%260 = add i32 0, %259
%261 = getelementptr inbounds %struct.cmplx, ptr %254, i32 %260
%262 = getelementptr inbounds %struct.cmplx, ptr %261, i32 0, i32 1
%263 = load double, ptr %262, align 8
%264 = load ptr, ptr %9, align 4
%265 = load i32, ptr %7, align 4
%266 = load i32, ptr %24, align 4
%267 = mul i32 11, %266
%268 = add i32 9, %267
%269 = mul i32 %265, %268
%270 = add i32 0, %269
%271 = getelementptr inbounds %struct.cmplx, ptr %264, i32 %270
%272 = getelementptr inbounds %struct.cmplx, ptr %271, i32 0, i32 1
%273 = load double, ptr %272, align 8
%274 = fsub double %263, %273
%275 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 1
store double %274, ptr %275, align 8
%276 = load ptr, ptr %9, align 4
%277 = load i32, ptr %7, align 4
%278 = load i32, ptr %24, align 4
%279 = mul i32 11, %278
%280 = add i32 3, %279
%281 = mul i32 %277, %280
%282 = add i32 0, %281
%283 = getelementptr inbounds %struct.cmplx, ptr %276, i32 %282
%284 = getelementptr inbounds %struct.cmplx, ptr %283, i32 0, i32 0
%285 = load double, ptr %284, align 8
%286 = load ptr, ptr %9, align 4
%287 = load i32, ptr %7, align 4
%288 = load i32, ptr %24, align 4
%289 = mul i32 11, %288
%290 = add i32 8, %289
%291 = mul i32 %287, %290
%292 = add i32 0, %291
%293 = getelementptr inbounds %struct.cmplx, ptr %286, i32 %292
%294 = getelementptr inbounds %struct.cmplx, ptr %293, i32 0, i32 0
%295 = load double, ptr %294, align 8
%296 = fadd double %285, %295
%297 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
store double %296, ptr %297, align 8
%298 = load ptr, ptr %9, align 4
%299 = load i32, ptr %7, align 4
%300 = load i32, ptr %24, align 4
%301 = mul i32 11, %300
%302 = add i32 3, %301
%303 = mul i32 %299, %302
%304 = add i32 0, %303
%305 = getelementptr inbounds %struct.cmplx, ptr %298, i32 %304
%306 = getelementptr inbounds %struct.cmplx, ptr %305, i32 0, i32 1
%307 = load double, ptr %306, align 8
%308 = load ptr, ptr %9, align 4
%309 = load i32, ptr %7, align 4
%310 = load i32, ptr %24, align 4
%311 = mul i32 11, %310
%312 = add i32 8, %311
%313 = mul i32 %309, %312
%314 = add i32 0, %313
%315 = getelementptr inbounds %struct.cmplx, ptr %308, i32 %314
%316 = getelementptr inbounds %struct.cmplx, ptr %315, i32 0, i32 1
%317 = load double, ptr %316, align 8
%318 = fadd double %307, %317
%319 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
store double %318, ptr %319, align 8
%320 = load ptr, ptr %9, align 4
%321 = load i32, ptr %7, align 4
%322 = load i32, ptr %24, align 4
%323 = mul i32 11, %322
%324 = add i32 3, %323
%325 = mul i32 %321, %324
%326 = add i32 0, %325
%327 = getelementptr inbounds %struct.cmplx, ptr %320, i32 %326
%328 = getelementptr inbounds %struct.cmplx, ptr %327, i32 0, i32 0
%329 = load double, ptr %328, align 8
%330 = load ptr, ptr %9, align 4
%331 = load i32, ptr %7, align 4
%332 = load i32, ptr %24, align 4
%333 = mul i32 11, %332
%334 = add i32 8, %333
%335 = mul i32 %331, %334
%336 = add i32 0, %335
%337 = getelementptr inbounds %struct.cmplx, ptr %330, i32 %336
%338 = getelementptr inbounds %struct.cmplx, ptr %337, i32 0, i32 0
%339 = load double, ptr %338, align 8
%340 = fsub double %329, %339
%341 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 0
store double %340, ptr %341, align 8
%342 = load ptr, ptr %9, align 4
%343 = load i32, ptr %7, align 4
%344 = load i32, ptr %24, align 4
%345 = mul i32 11, %344
%346 = add i32 3, %345
%347 = mul i32 %343, %346
%348 = add i32 0, %347
%349 = getelementptr inbounds %struct.cmplx, ptr %342, i32 %348
%350 = getelementptr inbounds %struct.cmplx, ptr %349, i32 0, i32 1
%351 = load double, ptr %350, align 8
%352 = load ptr, ptr %9, align 4
%353 = load i32, ptr %7, align 4
%354 = load i32, ptr %24, align 4
%355 = mul i32 11, %354
%356 = add i32 8, %355
%357 = mul i32 %353, %356
%358 = add i32 0, %357
%359 = getelementptr inbounds %struct.cmplx, ptr %352, i32 %358
%360 = getelementptr inbounds %struct.cmplx, ptr %359, i32 0, i32 1
%361 = load double, ptr %360, align 8
%362 = fsub double %351, %361
%363 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 1
store double %362, ptr %363, align 8
%364 = load ptr, ptr %9, align 4
%365 = load i32, ptr %7, align 4
%366 = load i32, ptr %24, align 4
%367 = mul i32 11, %366
%368 = add i32 4, %367
%369 = mul i32 %365, %368
%370 = add i32 0, %369
%371 = getelementptr inbounds %struct.cmplx, ptr %364, i32 %370
%372 = getelementptr inbounds %struct.cmplx, ptr %371, i32 0, i32 0
%373 = load double, ptr %372, align 8
%374 = load ptr, ptr %9, align 4
%375 = load i32, ptr %7, align 4
%376 = load i32, ptr %24, align 4
%377 = mul i32 11, %376
%378 = add i32 7, %377
%379 = mul i32 %375, %378
%380 = add i32 0, %379
%381 = getelementptr inbounds %struct.cmplx, ptr %374, i32 %380
%382 = getelementptr inbounds %struct.cmplx, ptr %381, i32 0, i32 0
%383 = load double, ptr %382, align 8
%384 = fadd double %373, %383
%385 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
store double %384, ptr %385, align 8
%386 = load ptr, ptr %9, align 4
%387 = load i32, ptr %7, align 4
%388 = load i32, ptr %24, align 4
%389 = mul i32 11, %388
%390 = add i32 4, %389
%391 = mul i32 %387, %390
%392 = add i32 0, %391
%393 = getelementptr inbounds %struct.cmplx, ptr %386, i32 %392
%394 = getelementptr inbounds %struct.cmplx, ptr %393, i32 0, i32 1
%395 = load double, ptr %394, align 8
%396 = load ptr, ptr %9, align 4
%397 = load i32, ptr %7, align 4
%398 = load i32, ptr %24, align 4
%399 = mul i32 11, %398
%400 = add i32 7, %399
%401 = mul i32 %397, %400
%402 = add i32 0, %401
%403 = getelementptr inbounds %struct.cmplx, ptr %396, i32 %402
%404 = getelementptr inbounds %struct.cmplx, ptr %403, i32 0, i32 1
%405 = load double, ptr %404, align 8
%406 = fadd double %395, %405
%407 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
store double %406, ptr %407, align 8
%408 = load ptr, ptr %9, align 4
%409 = load i32, ptr %7, align 4
%410 = load i32, ptr %24, align 4
%411 = mul i32 11, %410
%412 = add i32 4, %411
%413 = mul i32 %409, %412
%414 = add i32 0, %413
%415 = getelementptr inbounds %struct.cmplx, ptr %408, i32 %414
%416 = getelementptr inbounds %struct.cmplx, ptr %415, i32 0, i32 0
%417 = load double, ptr %416, align 8
%418 = load ptr, ptr %9, align 4
%419 = load i32, ptr %7, align 4
%420 = load i32, ptr %24, align 4
%421 = mul i32 11, %420
%422 = add i32 7, %421
%423 = mul i32 %419, %422
%424 = add i32 0, %423
%425 = getelementptr inbounds %struct.cmplx, ptr %418, i32 %424
%426 = getelementptr inbounds %struct.cmplx, ptr %425, i32 0, i32 0
%427 = load double, ptr %426, align 8
%428 = fsub double %417, %427
%429 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 0
store double %428, ptr %429, align 8
%430 = load ptr, ptr %9, align 4
%431 = load i32, ptr %7, align 4
%432 = load i32, ptr %24, align 4
%433 = mul i32 11, %432
%434 = add i32 4, %433
%435 = mul i32 %431, %434
%436 = add i32 0, %435
%437 = getelementptr inbounds %struct.cmplx, ptr %430, i32 %436
%438 = getelementptr inbounds %struct.cmplx, ptr %437, i32 0, i32 1
%439 = load double, ptr %438, align 8
%440 = load ptr, ptr %9, align 4
%441 = load i32, ptr %7, align 4
%442 = load i32, ptr %24, align 4
%443 = mul i32 11, %442
%444 = add i32 7, %443
%445 = mul i32 %441, %444
%446 = add i32 0, %445
%447 = getelementptr inbounds %struct.cmplx, ptr %440, i32 %446
%448 = getelementptr inbounds %struct.cmplx, ptr %447, i32 0, i32 1
%449 = load double, ptr %448, align 8
%450 = fsub double %439, %449
%451 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 1
store double %450, ptr %451, align 8
%452 = load ptr, ptr %9, align 4
%453 = load i32, ptr %7, align 4
%454 = load i32, ptr %24, align 4
%455 = mul i32 11, %454
%456 = add i32 5, %455
%457 = mul i32 %453, %456
%458 = add i32 0, %457
%459 = getelementptr inbounds %struct.cmplx, ptr %452, i32 %458
%460 = getelementptr inbounds %struct.cmplx, ptr %459, i32 0, i32 0
%461 = load double, ptr %460, align 8
%462 = load ptr, ptr %9, align 4
%463 = load i32, ptr %7, align 4
%464 = load i32, ptr %24, align 4
%465 = mul i32 11, %464
%466 = add i32 6, %465
%467 = mul i32 %463, %466
%468 = add i32 0, %467
%469 = getelementptr inbounds %struct.cmplx, ptr %462, i32 %468
%470 = getelementptr inbounds %struct.cmplx, ptr %469, i32 0, i32 0
%471 = load double, ptr %470, align 8
%472 = fadd double %461, %471
%473 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
store double %472, ptr %473, align 8
%474 = load ptr, ptr %9, align 4
%475 = load i32, ptr %7, align 4
%476 = load i32, ptr %24, align 4
%477 = mul i32 11, %476
%478 = add i32 5, %477
%479 = mul i32 %475, %478
%480 = add i32 0, %479
%481 = getelementptr inbounds %struct.cmplx, ptr %474, i32 %480
%482 = getelementptr inbounds %struct.cmplx, ptr %481, i32 0, i32 1
%483 = load double, ptr %482, align 8
%484 = load ptr, ptr %9, align 4
%485 = load i32, ptr %7, align 4
%486 = load i32, ptr %24, align 4
%487 = mul i32 11, %486
%488 = add i32 6, %487
%489 = mul i32 %485, %488
%490 = add i32 0, %489
%491 = getelementptr inbounds %struct.cmplx, ptr %484, i32 %490
%492 = getelementptr inbounds %struct.cmplx, ptr %491, i32 0, i32 1
%493 = load double, ptr %492, align 8
%494 = fadd double %483, %493
%495 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
store double %494, ptr %495, align 8
%496 = load ptr, ptr %9, align 4
%497 = load i32, ptr %7, align 4
%498 = load i32, ptr %24, align 4
%499 = mul i32 11, %498
%500 = add i32 5, %499
%501 = mul i32 %497, %500
%502 = add i32 0, %501
%503 = getelementptr inbounds %struct.cmplx, ptr %496, i32 %502
%504 = getelementptr inbounds %struct.cmplx, ptr %503, i32 0, i32 0
%505 = load double, ptr %504, align 8
%506 = load ptr, ptr %9, align 4
%507 = load i32, ptr %7, align 4
%508 = load i32, ptr %24, align 4
%509 = mul i32 11, %508
%510 = add i32 6, %509
%511 = mul i32 %507, %510
%512 = add i32 0, %511
%513 = getelementptr inbounds %struct.cmplx, ptr %506, i32 %512
%514 = getelementptr inbounds %struct.cmplx, ptr %513, i32 0, i32 0
%515 = load double, ptr %514, align 8
%516 = fsub double %505, %515
%517 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 0
store double %516, ptr %517, align 8
%518 = load ptr, ptr %9, align 4
%519 = load i32, ptr %7, align 4
%520 = load i32, ptr %24, align 4
%521 = mul i32 11, %520
%522 = add i32 5, %521
%523 = mul i32 %519, %522
%524 = add i32 0, %523
%525 = getelementptr inbounds %struct.cmplx, ptr %518, i32 %524
%526 = getelementptr inbounds %struct.cmplx, ptr %525, i32 0, i32 1
%527 = load double, ptr %526, align 8
%528 = load ptr, ptr %9, align 4
%529 = load i32, ptr %7, align 4
%530 = load i32, ptr %24, align 4
%531 = mul i32 11, %530
%532 = add i32 6, %531
%533 = mul i32 %529, %532
%534 = add i32 0, %533
%535 = getelementptr inbounds %struct.cmplx, ptr %528, i32 %534
%536 = getelementptr inbounds %struct.cmplx, ptr %535, i32 0, i32 1
%537 = load double, ptr %536, align 8
%538 = fsub double %527, %537
%539 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 1
store double %538, ptr %539, align 8
%540 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 0
%541 = load double, ptr %540, align 8
%542 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
%543 = load double, ptr %542, align 8
%544 = fadd double %541, %543
%545 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
%546 = load double, ptr %545, align 8
%547 = fadd double %544, %546
%548 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
%549 = load double, ptr %548, align 8
%550 = fadd double %547, %549
%551 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
%552 = load double, ptr %551, align 8
%553 = fadd double %550, %552
%554 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
%555 = load double, ptr %554, align 8
%556 = fadd double %553, %555
%557 = load ptr, ptr %10, align 4
%558 = load i32, ptr %7, align 4
%559 = load i32, ptr %24, align 4
%560 = load i32, ptr %8, align 4
%561 = mul i32 %560, 0
%562 = add i32 %559, %561
%563 = mul i32 %558, %562
%564 = add i32 0, %563
%565 = getelementptr inbounds %struct.cmplx, ptr %557, i32 %564
%566 = getelementptr inbounds %struct.cmplx, ptr %565, i32 0, i32 0
store double %556, ptr %566, align 8
%567 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 1
%568 = load double, ptr %567, align 8
%569 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
%570 = load double, ptr %569, align 8
%571 = fadd double %568, %570
%572 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
%573 = load double, ptr %572, align 8
%574 = fadd double %571, %573
%575 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
%576 = load double, ptr %575, align 8
%577 = fadd double %574, %576
%578 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
%579 = load double, ptr %578, align 8
%580 = fadd double %577, %579
%581 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
%582 = load double, ptr %581, align 8
%583 = fadd double %580, %582
%584 = load ptr, ptr %10, align 4
%585 = load i32, ptr %7, align 4
%586 = load i32, ptr %24, align 4
%587 = load i32, ptr %8, align 4
%588 = mul i32 %587, 0
%589 = add i32 %586, %588
%590 = mul i32 %585, %589
%591 = add i32 0, %590
%592 = getelementptr inbounds %struct.cmplx, ptr %584, i32 %591
%593 = getelementptr inbounds %struct.cmplx, ptr %592, i32 0, i32 1
store double %583, ptr %593, align 8
%594 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 0
%595 = load double, ptr %594, align 8
%596 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
%597 = load double, ptr %596, align 8
%598 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %597, double %595)
%599 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
%600 = load double, ptr %599, align 8
%601 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %600, double %598)
%602 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
%603 = load double, ptr %602, align 8
%604 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %603, double %601)
%605 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
%606 = load double, ptr %605, align 8
%607 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %606, double %604)
%608 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
%609 = load double, ptr %608, align 8
%610 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %609, double %607)
%611 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 0
store double %610, ptr %611, align 8
%612 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 1
%613 = load double, ptr %612, align 8
%614 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
%615 = load double, ptr %614, align 8
%616 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %615, double %613)
%617 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
%618 = load double, ptr %617, align 8
%619 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %618, double %616)
%620 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
%621 = load double, ptr %620, align 8
%622 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %621, double %619)
%623 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
%624 = load double, ptr %623, align 8
%625 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %624, double %622)
%626 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
%627 = load double, ptr %626, align 8
%628 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %627, double %625)
%629 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 1
store double %628, ptr %629, align 8
%630 = load double, ptr %15, align 8
%631 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 0
%632 = load double, ptr %631, align 8
%633 = load double, ptr %17, align 8
%634 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 0
%635 = load double, ptr %634, align 8
%636 = fmul double %633, %635
%637 = call double @llvm.fmuladd.f64(double %630, double %632, double %636)
%638 = load double, ptr %19, align 8
%639 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 0
%640 = load double, ptr %639, align 8
%641 = call double @llvm.fmuladd.f64(double %638, double %640, double %637)
%642 = load double, ptr %21, align 8
%643 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 0
%644 = load double, ptr %643, align 8
%645 = call double @llvm.fmuladd.f64(double %642, double %644, double %641)
%646 = load double, ptr %23, align 8
%647 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 0
%648 = load double, ptr %647, align 8
%649 = call double @llvm.fmuladd.f64(double %646, double %648, double %645)
%650 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 1
store double %649, ptr %650, align 8
%651 = load double, ptr %15, align 8
%652 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 1
%653 = load double, ptr %652, align 8
%654 = load double, ptr %17, align 8
%655 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 1
%656 = load double, ptr %655, align 8
%657 = fmul double %654, %656
%658 = call double @llvm.fmuladd.f64(double %651, double %653, double %657)
%659 = load double, ptr %19, align 8
%660 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 1
%661 = load double, ptr %660, align 8
%662 = call double @llvm.fmuladd.f64(double %659, double %661, double %658)
%663 = load double, ptr %21, align 8
%664 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 1
%665 = load double, ptr %664, align 8
%666 = call double @llvm.fmuladd.f64(double %663, double %665, double %662)
%667 = load double, ptr %23, align 8
%668 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 1
%669 = load double, ptr %668, align 8
%670 = call double @llvm.fmuladd.f64(double %667, double %669, double %666)
%671 = fneg double %670
%672 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 0
store double %671, ptr %672, align 8
%673 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 0
%674 = load double, ptr %673, align 8
%675 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 0
%676 = load double, ptr %675, align 8
%677 = fadd double %674, %676
%678 = load ptr, ptr %10, align 4
%679 = load i32, ptr %7, align 4
%680 = load i32, ptr %24, align 4
%681 = load i32, ptr %8, align 4
%682 = mul i32 %681, 1
%683 = add i32 %680, %682
%684 = mul i32 %679, %683
%685 = add i32 0, %684
%686 = getelementptr inbounds %struct.cmplx, ptr %678, i32 %685
%687 = getelementptr inbounds %struct.cmplx, ptr %686, i32 0, i32 0
store double %677, ptr %687, align 8
%688 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 1
%689 = load double, ptr %688, align 8
%690 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 1
%691 = load double, ptr %690, align 8
%692 = fadd double %689, %691
%693 = load ptr, ptr %10, align 4
%694 = load i32, ptr %7, align 4
%695 = load i32, ptr %24, align 4
%696 = load i32, ptr %8, align 4
%697 = mul i32 %696, 1
%698 = add i32 %695, %697
%699 = mul i32 %694, %698
%700 = add i32 0, %699
%701 = getelementptr inbounds %struct.cmplx, ptr %693, i32 %700
%702 = getelementptr inbounds %struct.cmplx, ptr %701, i32 0, i32 1
store double %692, ptr %702, align 8
%703 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 0
%704 = load double, ptr %703, align 8
%705 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 0
%706 = load double, ptr %705, align 8
%707 = fsub double %704, %706
%708 = load ptr, ptr %10, align 4
%709 = load i32, ptr %7, align 4
%710 = load i32, ptr %24, align 4
%711 = load i32, ptr %8, align 4
%712 = mul i32 %711, 10
%713 = add i32 %710, %712
%714 = mul i32 %709, %713
%715 = add i32 0, %714
%716 = getelementptr inbounds %struct.cmplx, ptr %708, i32 %715
%717 = getelementptr inbounds %struct.cmplx, ptr %716, i32 0, i32 0
store double %707, ptr %717, align 8
%718 = getelementptr inbounds %struct.cmplx, ptr %36, i32 0, i32 1
%719 = load double, ptr %718, align 8
%720 = getelementptr inbounds %struct.cmplx, ptr %37, i32 0, i32 1
%721 = load double, ptr %720, align 8
%722 = fsub double %719, %721
%723 = load ptr, ptr %10, align 4
%724 = load i32, ptr %7, align 4
%725 = load i32, ptr %24, align 4
%726 = load i32, ptr %8, align 4
%727 = mul i32 %726, 10
%728 = add i32 %725, %727
%729 = mul i32 %724, %728
%730 = add i32 0, %729
%731 = getelementptr inbounds %struct.cmplx, ptr %723, i32 %730
%732 = getelementptr inbounds %struct.cmplx, ptr %731, i32 0, i32 1
store double %722, ptr %732, align 8
%733 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 0
%734 = load double, ptr %733, align 8
%735 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
%736 = load double, ptr %735, align 8
%737 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %736, double %734)
%738 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
%739 = load double, ptr %738, align 8
%740 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %739, double %737)
%741 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
%742 = load double, ptr %741, align 8
%743 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %742, double %740)
%744 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
%745 = load double, ptr %744, align 8
%746 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %745, double %743)
%747 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
%748 = load double, ptr %747, align 8
%749 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %748, double %746)
%750 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 0
store double %749, ptr %750, align 8
%751 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 1
%752 = load double, ptr %751, align 8
%753 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
%754 = load double, ptr %753, align 8
%755 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %754, double %752)
%756 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
%757 = load double, ptr %756, align 8
%758 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %757, double %755)
%759 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
%760 = load double, ptr %759, align 8
%761 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %760, double %758)
%762 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
%763 = load double, ptr %762, align 8
%764 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %763, double %761)
%765 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
%766 = load double, ptr %765, align 8
%767 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %766, double %764)
%768 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 1
store double %767, ptr %768, align 8
%769 = load double, ptr %17, align 8
%770 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 0
%771 = load double, ptr %770, align 8
%772 = load double, ptr %21, align 8
%773 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 0
%774 = load double, ptr %773, align 8
%775 = fmul double %772, %774
%776 = call double @llvm.fmuladd.f64(double %769, double %771, double %775)
%777 = load double, ptr %23, align 8
%778 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 0
%779 = load double, ptr %778, align 8
%780 = fneg double %777
%781 = call double @llvm.fmuladd.f64(double %780, double %779, double %776)
%782 = load double, ptr %19, align 8
%783 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 0
%784 = load double, ptr %783, align 8
%785 = fneg double %782
%786 = call double @llvm.fmuladd.f64(double %785, double %784, double %781)
%787 = load double, ptr %15, align 8
%788 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 0
%789 = load double, ptr %788, align 8
%790 = fneg double %787
%791 = call double @llvm.fmuladd.f64(double %790, double %789, double %786)
%792 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 1
store double %791, ptr %792, align 8
%793 = load double, ptr %17, align 8
%794 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 1
%795 = load double, ptr %794, align 8
%796 = load double, ptr %21, align 8
%797 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 1
%798 = load double, ptr %797, align 8
%799 = fmul double %796, %798
%800 = call double @llvm.fmuladd.f64(double %793, double %795, double %799)
%801 = load double, ptr %23, align 8
%802 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 1
%803 = load double, ptr %802, align 8
%804 = fneg double %801
%805 = call double @llvm.fmuladd.f64(double %804, double %803, double %800)
%806 = load double, ptr %19, align 8
%807 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 1
%808 = load double, ptr %807, align 8
%809 = fneg double %806
%810 = call double @llvm.fmuladd.f64(double %809, double %808, double %805)
%811 = load double, ptr %15, align 8
%812 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 1
%813 = load double, ptr %812, align 8
%814 = fneg double %811
%815 = call double @llvm.fmuladd.f64(double %814, double %813, double %810)
%816 = fneg double %815
%817 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 0
store double %816, ptr %817, align 8
%818 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 0
%819 = load double, ptr %818, align 8
%820 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 0
%821 = load double, ptr %820, align 8
%822 = fadd double %819, %821
%823 = load ptr, ptr %10, align 4
%824 = load i32, ptr %7, align 4
%825 = load i32, ptr %24, align 4
%826 = load i32, ptr %8, align 4
%827 = mul i32 %826, 2
%828 = add i32 %825, %827
%829 = mul i32 %824, %828
%830 = add i32 0, %829
%831 = getelementptr inbounds %struct.cmplx, ptr %823, i32 %830
%832 = getelementptr inbounds %struct.cmplx, ptr %831, i32 0, i32 0
store double %822, ptr %832, align 8
%833 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 1
%834 = load double, ptr %833, align 8
%835 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 1
%836 = load double, ptr %835, align 8
%837 = fadd double %834, %836
%838 = load ptr, ptr %10, align 4
%839 = load i32, ptr %7, align 4
%840 = load i32, ptr %24, align 4
%841 = load i32, ptr %8, align 4
%842 = mul i32 %841, 2
%843 = add i32 %840, %842
%844 = mul i32 %839, %843
%845 = add i32 0, %844
%846 = getelementptr inbounds %struct.cmplx, ptr %838, i32 %845
%847 = getelementptr inbounds %struct.cmplx, ptr %846, i32 0, i32 1
store double %837, ptr %847, align 8
%848 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 0
%849 = load double, ptr %848, align 8
%850 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 0
%851 = load double, ptr %850, align 8
%852 = fsub double %849, %851
%853 = load ptr, ptr %10, align 4
%854 = load i32, ptr %7, align 4
%855 = load i32, ptr %24, align 4
%856 = load i32, ptr %8, align 4
%857 = mul i32 %856, 9
%858 = add i32 %855, %857
%859 = mul i32 %854, %858
%860 = add i32 0, %859
%861 = getelementptr inbounds %struct.cmplx, ptr %853, i32 %860
%862 = getelementptr inbounds %struct.cmplx, ptr %861, i32 0, i32 0
store double %852, ptr %862, align 8
%863 = getelementptr inbounds %struct.cmplx, ptr %38, i32 0, i32 1
%864 = load double, ptr %863, align 8
%865 = getelementptr inbounds %struct.cmplx, ptr %39, i32 0, i32 1
%866 = load double, ptr %865, align 8
%867 = fsub double %864, %866
%868 = load ptr, ptr %10, align 4
%869 = load i32, ptr %7, align 4
%870 = load i32, ptr %24, align 4
%871 = load i32, ptr %8, align 4
%872 = mul i32 %871, 9
%873 = add i32 %870, %872
%874 = mul i32 %869, %873
%875 = add i32 0, %874
%876 = getelementptr inbounds %struct.cmplx, ptr %868, i32 %875
%877 = getelementptr inbounds %struct.cmplx, ptr %876, i32 0, i32 1
store double %867, ptr %877, align 8
%878 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 0
%879 = load double, ptr %878, align 8
%880 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
%881 = load double, ptr %880, align 8
%882 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %881, double %879)
%883 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
%884 = load double, ptr %883, align 8
%885 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %884, double %882)
%886 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
%887 = load double, ptr %886, align 8
%888 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %887, double %885)
%889 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
%890 = load double, ptr %889, align 8
%891 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %890, double %888)
%892 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
%893 = load double, ptr %892, align 8
%894 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %893, double %891)
%895 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 0
store double %894, ptr %895, align 8
%896 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 1
%897 = load double, ptr %896, align 8
%898 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
%899 = load double, ptr %898, align 8
%900 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %899, double %897)
%901 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
%902 = load double, ptr %901, align 8
%903 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %902, double %900)
%904 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
%905 = load double, ptr %904, align 8
%906 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %905, double %903)
%907 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
%908 = load double, ptr %907, align 8
%909 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %908, double %906)
%910 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
%911 = load double, ptr %910, align 8
%912 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %911, double %909)
%913 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 1
store double %912, ptr %913, align 8
%914 = load double, ptr %19, align 8
%915 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 0
%916 = load double, ptr %915, align 8
%917 = load double, ptr %23, align 8
%918 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 0
%919 = load double, ptr %918, align 8
%920 = fmul double %917, %919
%921 = fneg double %920
%922 = call double @llvm.fmuladd.f64(double %914, double %916, double %921)
%923 = load double, ptr %17, align 8
%924 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 0
%925 = load double, ptr %924, align 8
%926 = fneg double %923
%927 = call double @llvm.fmuladd.f64(double %926, double %925, double %922)
%928 = load double, ptr %15, align 8
%929 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 0
%930 = load double, ptr %929, align 8
%931 = call double @llvm.fmuladd.f64(double %928, double %930, double %927)
%932 = load double, ptr %21, align 8
%933 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 0
%934 = load double, ptr %933, align 8
%935 = call double @llvm.fmuladd.f64(double %932, double %934, double %931)
%936 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 1
store double %935, ptr %936, align 8
%937 = load double, ptr %19, align 8
%938 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 1
%939 = load double, ptr %938, align 8
%940 = load double, ptr %23, align 8
%941 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 1
%942 = load double, ptr %941, align 8
%943 = fmul double %940, %942
%944 = fneg double %943
%945 = call double @llvm.fmuladd.f64(double %937, double %939, double %944)
%946 = load double, ptr %17, align 8
%947 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 1
%948 = load double, ptr %947, align 8
%949 = fneg double %946
%950 = call double @llvm.fmuladd.f64(double %949, double %948, double %945)
%951 = load double, ptr %15, align 8
%952 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 1
%953 = load double, ptr %952, align 8
%954 = call double @llvm.fmuladd.f64(double %951, double %953, double %950)
%955 = load double, ptr %21, align 8
%956 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 1
%957 = load double, ptr %956, align 8
%958 = call double @llvm.fmuladd.f64(double %955, double %957, double %954)
%959 = fneg double %958
%960 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 0
store double %959, ptr %960, align 8
%961 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 0
%962 = load double, ptr %961, align 8
%963 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 0
%964 = load double, ptr %963, align 8
%965 = fadd double %962, %964
%966 = load ptr, ptr %10, align 4
%967 = load i32, ptr %7, align 4
%968 = load i32, ptr %24, align 4
%969 = load i32, ptr %8, align 4
%970 = mul i32 %969, 3
%971 = add i32 %968, %970
%972 = mul i32 %967, %971
%973 = add i32 0, %972
%974 = getelementptr inbounds %struct.cmplx, ptr %966, i32 %973
%975 = getelementptr inbounds %struct.cmplx, ptr %974, i32 0, i32 0
store double %965, ptr %975, align 8
%976 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 1
%977 = load double, ptr %976, align 8
%978 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 1
%979 = load double, ptr %978, align 8
%980 = fadd double %977, %979
%981 = load ptr, ptr %10, align 4
%982 = load i32, ptr %7, align 4
%983 = load i32, ptr %24, align 4
%984 = load i32, ptr %8, align 4
%985 = mul i32 %984, 3
%986 = add i32 %983, %985
%987 = mul i32 %982, %986
%988 = add i32 0, %987
%989 = getelementptr inbounds %struct.cmplx, ptr %981, i32 %988
%990 = getelementptr inbounds %struct.cmplx, ptr %989, i32 0, i32 1
store double %980, ptr %990, align 8
%991 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 0
%992 = load double, ptr %991, align 8
%993 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 0
%994 = load double, ptr %993, align 8
%995 = fsub double %992, %994
%996 = load ptr, ptr %10, align 4
%997 = load i32, ptr %7, align 4
%998 = load i32, ptr %24, align 4
%999 = load i32, ptr %8, align 4
%1000 = mul i32 %999, 8
%1001 = add i32 %998, %1000
%1002 = mul i32 %997, %1001
%1003 = add i32 0, %1002
%1004 = getelementptr inbounds %struct.cmplx, ptr %996, i32 %1003
%1005 = getelementptr inbounds %struct.cmplx, ptr %1004, i32 0, i32 0
store double %995, ptr %1005, align 8
%1006 = getelementptr inbounds %struct.cmplx, ptr %40, i32 0, i32 1
%1007 = load double, ptr %1006, align 8
%1008 = getelementptr inbounds %struct.cmplx, ptr %41, i32 0, i32 1
%1009 = load double, ptr %1008, align 8
%1010 = fsub double %1007, %1009
%1011 = load ptr, ptr %10, align 4
%1012 = load i32, ptr %7, align 4
%1013 = load i32, ptr %24, align 4
%1014 = load i32, ptr %8, align 4
%1015 = mul i32 %1014, 8
%1016 = add i32 %1013, %1015
%1017 = mul i32 %1012, %1016
%1018 = add i32 0, %1017
%1019 = getelementptr inbounds %struct.cmplx, ptr %1011, i32 %1018
%1020 = getelementptr inbounds %struct.cmplx, ptr %1019, i32 0, i32 1
store double %1010, ptr %1020, align 8
%1021 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 0
%1022 = load double, ptr %1021, align 8
%1023 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 0
%1024 = load double, ptr %1023, align 8
%1025 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %1024, double %1022)
%1026 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 0
%1027 = load double, ptr %1026, align 8
%1028 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %1027, double %1025)
%1029 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 0
%1030 = load double, ptr %1029, align 8
%1031 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %1030, double %1028)
%1032 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 0
%1033 = load double, ptr %1032, align 8
%1034 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %1033, double %1031)
%1035 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 0
%1036 = load double, ptr %1035, align 8
%1037 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %1036, double %1034)
%1038 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 0
store double %1037, ptr %1038, align 8
%1039 = getelementptr inbounds %struct.cmplx, ptr %25, i32 0, i32 1
%1040 = load double, ptr %1039, align 8
%1041 = getelementptr inbounds %struct.cmplx, ptr %26, i32 0, i32 1
%1042 = load double, ptr %1041, align 8
%1043 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %1042, double %1040)
%1044 = getelementptr inbounds %struct.cmplx, ptr %27, i32 0, i32 1
%1045 = load double, ptr %1044, align 8
%1046 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %1045, double %1043)
%1047 = getelementptr inbounds %struct.cmplx, ptr %28, i32 0, i32 1
%1048 = load double, ptr %1047, align 8
%1049 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %1048, double %1046)
%1050 = getelementptr inbounds %struct.cmplx, ptr %29, i32 0, i32 1
%1051 = load double, ptr %1050, align 8
%1052 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %1051, double %1049)
%1053 = getelementptr inbounds %struct.cmplx, ptr %30, i32 0, i32 1
%1054 = load double, ptr %1053, align 8
%1055 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %1054, double %1052)
%1056 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 1
store double %1055, ptr %1056, align 8
%1057 = load double, ptr %21, align 8
%1058 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 0
%1059 = load double, ptr %1058, align 8
%1060 = load double, ptr %19, align 8
%1061 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 0
%1062 = load double, ptr %1061, align 8
%1063 = fmul double %1060, %1062
%1064 = fneg double %1063
%1065 = call double @llvm.fmuladd.f64(double %1057, double %1059, double %1064)
%1066 = load double, ptr %15, align 8
%1067 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 0
%1068 = load double, ptr %1067, align 8
%1069 = call double @llvm.fmuladd.f64(double %1066, double %1068, double %1065)
%1070 = load double, ptr %23, align 8
%1071 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 0
%1072 = load double, ptr %1071, align 8
%1073 = call double @llvm.fmuladd.f64(double %1070, double %1072, double %1069)
%1074 = load double, ptr %17, align 8
%1075 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 0
%1076 = load double, ptr %1075, align 8
%1077 = fneg double %1074
%1078 = call double @llvm.fmuladd.f64(double %1077, double %1076, double %1073)
%1079 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 1
store double %1078, ptr %1079, align 8
%1080 = load double, ptr %21, align 8
%1081 = getelementptr inbounds %struct.cmplx, ptr %35, i32 0, i32 1
%1082 = load double, ptr %1081, align 8
%1083 = load double, ptr %19, align 8
%1084 = getelementptr inbounds %struct.cmplx, ptr %34, i32 0, i32 1
%1085 = load double, ptr %1084, align 8
%1086 = fmul double %1083, %1085
%1087 = fneg double %1086
%1088 = call double @llvm.fmuladd.f64(double %1080, double %1082, double %1087)
%1089 = load double, ptr %15, align 8
%1090 = getelementptr inbounds %struct.cmplx, ptr %33, i32 0, i32 1
%1091 = load double, ptr %1090, align 8
%1092 = call double @llvm.fmuladd.f64(double %1089, double %1091, double %1088)
%1093 = load double, ptr %23, align 8
%1094 = getelementptr inbounds %struct.cmplx, ptr %32, i32 0, i32 1
%1095 = load double, ptr %1094, align 8
%1096 = call double @llvm.fmuladd.f64(double %1093, double %1095, double %1092)
%1097 = load double, ptr %17, align 8
%1098 = getelementptr inbounds %struct.cmplx, ptr %31, i32 0, i32 1
%1099 = load double, ptr %1098, align 8
%1100 = fneg double %1097
%1101 = call double @llvm.fmuladd.f64(double %1100, double %1099, double %1096)
%1102 = fneg double %1101
%1103 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 0
store double %1102, ptr %1103, align 8
%1104 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 0
%1105 = load double, ptr %1104, align 8
%1106 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 0
%1107 = load double, ptr %1106, align 8
%1108 = fadd double %1105, %1107
%1109 = load ptr, ptr %10, align 4
%1110 = load i32, ptr %7, align 4
%1111 = load i32, ptr %24, align 4
%1112 = load i32, ptr %8, align 4
%1113 = mul i32 %1112, 4
%1114 = add i32 %1111, %1113
%1115 = mul i32 %1110, %1114
%1116 = add i32 0, %1115
%1117 = getelementptr inbounds %struct.cmplx, ptr %1109, i32 %1116
%1118 = getelementptr inbounds %struct.cmplx, ptr %1117, i32 0, i32 0
store double %1108, ptr %1118, align 8
%1119 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 1
%1120 = load double, ptr %1119, align 8
%1121 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 1
%1122 = load double, ptr %1121, align 8
%1123 = fadd double %1120, %1122
%1124 = load ptr, ptr %10, align 4
%1125 = load i32, ptr %7, align 4
%1126 = load i32, ptr %24, align 4
%1127 = load i32, ptr %8, align 4
%1128 = mul i32 %1127, 4
%1129 = add i32 %1126, %1128
%1130 = mul i32 %1125, %1129
%1131 = add i32 0, %1130
%1132 = getelementptr inbounds %struct.cmplx, ptr %1124, i32 %1131
%1133 = getelementptr inbounds %struct.cmplx, ptr %1132, i32 0, i32 1
store double %1123, ptr %1133, align 8
%1134 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 0
%1135 = load double, ptr %1134, align 8
%1136 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 0
%1137 = load double, ptr %1136, align 8
%1138 = fsub double %1135, %1137
%1139 = load ptr, ptr %10, align 4
%1140 = load i32, ptr %7, align 4
%1141 = load i32, ptr %24, align 4
%1142 = load i32, ptr %8, align 4
%1143 = mul i32 %1142, 7
%1144 = add i32 %1141, %1143
%1145 = mul i32 %1140, %1144
%1146 = add i32 0, %1145
%1147 = getelementptr inbounds %struct.cmplx, ptr %1139, i32 %1146
%1148 = getelementptr inbounds %struct.cmplx, ptr %1147, i32 0, i32 0
store double %1138, ptr %1148, align 8
%1149 = getelementptr inbounds %struct.cmplx, ptr %42, i32 0, i32 1
%1150 = load double, ptr %1149, align 8
%1151 = getelementptr inbounds %struct.cmplx, ptr %43, i32 0, i32 1
%1152 = load double, ptr %1151, align 8
%1153 = fsub double %1150, %1152
%1154 = load ptr, ptr %10, align 4
%1155 = load i32, ptr %7, align 4
%1156 = load i32, ptr %24, align 4
%1157 = load i32, ptr %8, align 4
%1158 = mul i32 %1157, 7
%1159 = add i32 %1156, %1158
%1160 = mul i32 %1155, %1159
%1161 = add i32 0, %1160
%1162 = getelementptr inbounds %struct.cmplx, ptr %1154, i32 %1161
%1163 = getelementptr inbounds %struct.cmplx, ptr %1162, i32 0, i32 1
store double %1153, ptr %1163, align 8
store i32 1, ptr %44, align 4
br label %1164
1164: ; preds = %2788, %91
%1165 = load i32, ptr %44, align 4
%1166 = load i32, ptr %7, align 4
%1167 = icmp ult i32 %1165, %1166
br i1 %1167, label %1168, label %2791
1168: ; preds = %1164
%1169 = load ptr, ptr %9, align 4
%1170 = load i32, ptr %44, align 4
%1171 = load i32, ptr %7, align 4
%1172 = load i32, ptr %24, align 4
%1173 = mul i32 11, %1172
%1174 = add i32 0, %1173
%1175 = mul i32 %1171, %1174
%1176 = add i32 %1170, %1175
%1177 = getelementptr inbounds %struct.cmplx, ptr %1169, i32 %1176
call void @llvm.memcpy.p0.p0.i32(ptr align 8 %45, ptr align 8 %1177, i32 16, i1 false)
%1178 = load ptr, ptr %9, align 4
%1179 = load i32, ptr %44, align 4
%1180 = load i32, ptr %7, align 4
%1181 = load i32, ptr %24, align 4
%1182 = mul i32 11, %1181
%1183 = add i32 1, %1182
%1184 = mul i32 %1180, %1183
%1185 = add i32 %1179, %1184
%1186 = getelementptr inbounds %struct.cmplx, ptr %1178, i32 %1185
%1187 = getelementptr inbounds %struct.cmplx, ptr %1186, i32 0, i32 0
%1188 = load double, ptr %1187, align 8
%1189 = load ptr, ptr %9, align 4
%1190 = load i32, ptr %44, align 4
%1191 = load i32, ptr %7, align 4
%1192 = load i32, ptr %24, align 4
%1193 = mul i32 11, %1192
%1194 = add i32 10, %1193
%1195 = mul i32 %1191, %1194
%1196 = add i32 %1190, %1195
%1197 = getelementptr inbounds %struct.cmplx, ptr %1189, i32 %1196
%1198 = getelementptr inbounds %struct.cmplx, ptr %1197, i32 0, i32 0
%1199 = load double, ptr %1198, align 8
%1200 = fadd double %1188, %1199
%1201 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
store double %1200, ptr %1201, align 8
%1202 = load ptr, ptr %9, align 4
%1203 = load i32, ptr %44, align 4
%1204 = load i32, ptr %7, align 4
%1205 = load i32, ptr %24, align 4
%1206 = mul i32 11, %1205
%1207 = add i32 1, %1206
%1208 = mul i32 %1204, %1207
%1209 = add i32 %1203, %1208
%1210 = getelementptr inbounds %struct.cmplx, ptr %1202, i32 %1209
%1211 = getelementptr inbounds %struct.cmplx, ptr %1210, i32 0, i32 1
%1212 = load double, ptr %1211, align 8
%1213 = load ptr, ptr %9, align 4
%1214 = load i32, ptr %44, align 4
%1215 = load i32, ptr %7, align 4
%1216 = load i32, ptr %24, align 4
%1217 = mul i32 11, %1216
%1218 = add i32 10, %1217
%1219 = mul i32 %1215, %1218
%1220 = add i32 %1214, %1219
%1221 = getelementptr inbounds %struct.cmplx, ptr %1213, i32 %1220
%1222 = getelementptr inbounds %struct.cmplx, ptr %1221, i32 0, i32 1
%1223 = load double, ptr %1222, align 8
%1224 = fadd double %1212, %1223
%1225 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
store double %1224, ptr %1225, align 8
%1226 = load ptr, ptr %9, align 4
%1227 = load i32, ptr %44, align 4
%1228 = load i32, ptr %7, align 4
%1229 = load i32, ptr %24, align 4
%1230 = mul i32 11, %1229
%1231 = add i32 1, %1230
%1232 = mul i32 %1228, %1231
%1233 = add i32 %1227, %1232
%1234 = getelementptr inbounds %struct.cmplx, ptr %1226, i32 %1233
%1235 = getelementptr inbounds %struct.cmplx, ptr %1234, i32 0, i32 0
%1236 = load double, ptr %1235, align 8
%1237 = load ptr, ptr %9, align 4
%1238 = load i32, ptr %44, align 4
%1239 = load i32, ptr %7, align 4
%1240 = load i32, ptr %24, align 4
%1241 = mul i32 11, %1240
%1242 = add i32 10, %1241
%1243 = mul i32 %1239, %1242
%1244 = add i32 %1238, %1243
%1245 = getelementptr inbounds %struct.cmplx, ptr %1237, i32 %1244
%1246 = getelementptr inbounds %struct.cmplx, ptr %1245, i32 0, i32 0
%1247 = load double, ptr %1246, align 8
%1248 = fsub double %1236, %1247
%1249 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 0
store double %1248, ptr %1249, align 8
%1250 = load ptr, ptr %9, align 4
%1251 = load i32, ptr %44, align 4
%1252 = load i32, ptr %7, align 4
%1253 = load i32, ptr %24, align 4
%1254 = mul i32 11, %1253
%1255 = add i32 1, %1254
%1256 = mul i32 %1252, %1255
%1257 = add i32 %1251, %1256
%1258 = getelementptr inbounds %struct.cmplx, ptr %1250, i32 %1257
%1259 = getelementptr inbounds %struct.cmplx, ptr %1258, i32 0, i32 1
%1260 = load double, ptr %1259, align 8
%1261 = load ptr, ptr %9, align 4
%1262 = load i32, ptr %44, align 4
%1263 = load i32, ptr %7, align 4
%1264 = load i32, ptr %24, align 4
%1265 = mul i32 11, %1264
%1266 = add i32 10, %1265
%1267 = mul i32 %1263, %1266
%1268 = add i32 %1262, %1267
%1269 = getelementptr inbounds %struct.cmplx, ptr %1261, i32 %1268
%1270 = getelementptr inbounds %struct.cmplx, ptr %1269, i32 0, i32 1
%1271 = load double, ptr %1270, align 8
%1272 = fsub double %1260, %1271
%1273 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 1
store double %1272, ptr %1273, align 8
%1274 = load ptr, ptr %9, align 4
%1275 = load i32, ptr %44, align 4
%1276 = load i32, ptr %7, align 4
%1277 = load i32, ptr %24, align 4
%1278 = mul i32 11, %1277
%1279 = add i32 2, %1278
%1280 = mul i32 %1276, %1279
%1281 = add i32 %1275, %1280
%1282 = getelementptr inbounds %struct.cmplx, ptr %1274, i32 %1281
%1283 = getelementptr inbounds %struct.cmplx, ptr %1282, i32 0, i32 0
%1284 = load double, ptr %1283, align 8
%1285 = load ptr, ptr %9, align 4
%1286 = load i32, ptr %44, align 4
%1287 = load i32, ptr %7, align 4
%1288 = load i32, ptr %24, align 4
%1289 = mul i32 11, %1288
%1290 = add i32 9, %1289
%1291 = mul i32 %1287, %1290
%1292 = add i32 %1286, %1291
%1293 = getelementptr inbounds %struct.cmplx, ptr %1285, i32 %1292
%1294 = getelementptr inbounds %struct.cmplx, ptr %1293, i32 0, i32 0
%1295 = load double, ptr %1294, align 8
%1296 = fadd double %1284, %1295
%1297 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
store double %1296, ptr %1297, align 8
%1298 = load ptr, ptr %9, align 4
%1299 = load i32, ptr %44, align 4
%1300 = load i32, ptr %7, align 4
%1301 = load i32, ptr %24, align 4
%1302 = mul i32 11, %1301
%1303 = add i32 2, %1302
%1304 = mul i32 %1300, %1303
%1305 = add i32 %1299, %1304
%1306 = getelementptr inbounds %struct.cmplx, ptr %1298, i32 %1305
%1307 = getelementptr inbounds %struct.cmplx, ptr %1306, i32 0, i32 1
%1308 = load double, ptr %1307, align 8
%1309 = load ptr, ptr %9, align 4
%1310 = load i32, ptr %44, align 4
%1311 = load i32, ptr %7, align 4
%1312 = load i32, ptr %24, align 4
%1313 = mul i32 11, %1312
%1314 = add i32 9, %1313
%1315 = mul i32 %1311, %1314
%1316 = add i32 %1310, %1315
%1317 = getelementptr inbounds %struct.cmplx, ptr %1309, i32 %1316
%1318 = getelementptr inbounds %struct.cmplx, ptr %1317, i32 0, i32 1
%1319 = load double, ptr %1318, align 8
%1320 = fadd double %1308, %1319
%1321 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
store double %1320, ptr %1321, align 8
%1322 = load ptr, ptr %9, align 4
%1323 = load i32, ptr %44, align 4
%1324 = load i32, ptr %7, align 4
%1325 = load i32, ptr %24, align 4
%1326 = mul i32 11, %1325
%1327 = add i32 2, %1326
%1328 = mul i32 %1324, %1327
%1329 = add i32 %1323, %1328
%1330 = getelementptr inbounds %struct.cmplx, ptr %1322, i32 %1329
%1331 = getelementptr inbounds %struct.cmplx, ptr %1330, i32 0, i32 0
%1332 = load double, ptr %1331, align 8
%1333 = load ptr, ptr %9, align 4
%1334 = load i32, ptr %44, align 4
%1335 = load i32, ptr %7, align 4
%1336 = load i32, ptr %24, align 4
%1337 = mul i32 11, %1336
%1338 = add i32 9, %1337
%1339 = mul i32 %1335, %1338
%1340 = add i32 %1334, %1339
%1341 = getelementptr inbounds %struct.cmplx, ptr %1333, i32 %1340
%1342 = getelementptr inbounds %struct.cmplx, ptr %1341, i32 0, i32 0
%1343 = load double, ptr %1342, align 8
%1344 = fsub double %1332, %1343
%1345 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 0
store double %1344, ptr %1345, align 8
%1346 = load ptr, ptr %9, align 4
%1347 = load i32, ptr %44, align 4
%1348 = load i32, ptr %7, align 4
%1349 = load i32, ptr %24, align 4
%1350 = mul i32 11, %1349
%1351 = add i32 2, %1350
%1352 = mul i32 %1348, %1351
%1353 = add i32 %1347, %1352
%1354 = getelementptr inbounds %struct.cmplx, ptr %1346, i32 %1353
%1355 = getelementptr inbounds %struct.cmplx, ptr %1354, i32 0, i32 1
%1356 = load double, ptr %1355, align 8
%1357 = load ptr, ptr %9, align 4
%1358 = load i32, ptr %44, align 4
%1359 = load i32, ptr %7, align 4
%1360 = load i32, ptr %24, align 4
%1361 = mul i32 11, %1360
%1362 = add i32 9, %1361
%1363 = mul i32 %1359, %1362
%1364 = add i32 %1358, %1363
%1365 = getelementptr inbounds %struct.cmplx, ptr %1357, i32 %1364
%1366 = getelementptr inbounds %struct.cmplx, ptr %1365, i32 0, i32 1
%1367 = load double, ptr %1366, align 8
%1368 = fsub double %1356, %1367
%1369 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 1
store double %1368, ptr %1369, align 8
%1370 = load ptr, ptr %9, align 4
%1371 = load i32, ptr %44, align 4
%1372 = load i32, ptr %7, align 4
%1373 = load i32, ptr %24, align 4
%1374 = mul i32 11, %1373
%1375 = add i32 3, %1374
%1376 = mul i32 %1372, %1375
%1377 = add i32 %1371, %1376
%1378 = getelementptr inbounds %struct.cmplx, ptr %1370, i32 %1377
%1379 = getelementptr inbounds %struct.cmplx, ptr %1378, i32 0, i32 0
%1380 = load double, ptr %1379, align 8
%1381 = load ptr, ptr %9, align 4
%1382 = load i32, ptr %44, align 4
%1383 = load i32, ptr %7, align 4
%1384 = load i32, ptr %24, align 4
%1385 = mul i32 11, %1384
%1386 = add i32 8, %1385
%1387 = mul i32 %1383, %1386
%1388 = add i32 %1382, %1387
%1389 = getelementptr inbounds %struct.cmplx, ptr %1381, i32 %1388
%1390 = getelementptr inbounds %struct.cmplx, ptr %1389, i32 0, i32 0
%1391 = load double, ptr %1390, align 8
%1392 = fadd double %1380, %1391
%1393 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
store double %1392, ptr %1393, align 8
%1394 = load ptr, ptr %9, align 4
%1395 = load i32, ptr %44, align 4
%1396 = load i32, ptr %7, align 4
%1397 = load i32, ptr %24, align 4
%1398 = mul i32 11, %1397
%1399 = add i32 3, %1398
%1400 = mul i32 %1396, %1399
%1401 = add i32 %1395, %1400
%1402 = getelementptr inbounds %struct.cmplx, ptr %1394, i32 %1401
%1403 = getelementptr inbounds %struct.cmplx, ptr %1402, i32 0, i32 1
%1404 = load double, ptr %1403, align 8
%1405 = load ptr, ptr %9, align 4
%1406 = load i32, ptr %44, align 4
%1407 = load i32, ptr %7, align 4
%1408 = load i32, ptr %24, align 4
%1409 = mul i32 11, %1408
%1410 = add i32 8, %1409
%1411 = mul i32 %1407, %1410
%1412 = add i32 %1406, %1411
%1413 = getelementptr inbounds %struct.cmplx, ptr %1405, i32 %1412
%1414 = getelementptr inbounds %struct.cmplx, ptr %1413, i32 0, i32 1
%1415 = load double, ptr %1414, align 8
%1416 = fadd double %1404, %1415
%1417 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
store double %1416, ptr %1417, align 8
%1418 = load ptr, ptr %9, align 4
%1419 = load i32, ptr %44, align 4
%1420 = load i32, ptr %7, align 4
%1421 = load i32, ptr %24, align 4
%1422 = mul i32 11, %1421
%1423 = add i32 3, %1422
%1424 = mul i32 %1420, %1423
%1425 = add i32 %1419, %1424
%1426 = getelementptr inbounds %struct.cmplx, ptr %1418, i32 %1425
%1427 = getelementptr inbounds %struct.cmplx, ptr %1426, i32 0, i32 0
%1428 = load double, ptr %1427, align 8
%1429 = load ptr, ptr %9, align 4
%1430 = load i32, ptr %44, align 4
%1431 = load i32, ptr %7, align 4
%1432 = load i32, ptr %24, align 4
%1433 = mul i32 11, %1432
%1434 = add i32 8, %1433
%1435 = mul i32 %1431, %1434
%1436 = add i32 %1430, %1435
%1437 = getelementptr inbounds %struct.cmplx, ptr %1429, i32 %1436
%1438 = getelementptr inbounds %struct.cmplx, ptr %1437, i32 0, i32 0
%1439 = load double, ptr %1438, align 8
%1440 = fsub double %1428, %1439
%1441 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 0
store double %1440, ptr %1441, align 8
%1442 = load ptr, ptr %9, align 4
%1443 = load i32, ptr %44, align 4
%1444 = load i32, ptr %7, align 4
%1445 = load i32, ptr %24, align 4
%1446 = mul i32 11, %1445
%1447 = add i32 3, %1446
%1448 = mul i32 %1444, %1447
%1449 = add i32 %1443, %1448
%1450 = getelementptr inbounds %struct.cmplx, ptr %1442, i32 %1449
%1451 = getelementptr inbounds %struct.cmplx, ptr %1450, i32 0, i32 1
%1452 = load double, ptr %1451, align 8
%1453 = load ptr, ptr %9, align 4
%1454 = load i32, ptr %44, align 4
%1455 = load i32, ptr %7, align 4
%1456 = load i32, ptr %24, align 4
%1457 = mul i32 11, %1456
%1458 = add i32 8, %1457
%1459 = mul i32 %1455, %1458
%1460 = add i32 %1454, %1459
%1461 = getelementptr inbounds %struct.cmplx, ptr %1453, i32 %1460
%1462 = getelementptr inbounds %struct.cmplx, ptr %1461, i32 0, i32 1
%1463 = load double, ptr %1462, align 8
%1464 = fsub double %1452, %1463
%1465 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 1
store double %1464, ptr %1465, align 8
%1466 = load ptr, ptr %9, align 4
%1467 = load i32, ptr %44, align 4
%1468 = load i32, ptr %7, align 4
%1469 = load i32, ptr %24, align 4
%1470 = mul i32 11, %1469
%1471 = add i32 4, %1470
%1472 = mul i32 %1468, %1471
%1473 = add i32 %1467, %1472
%1474 = getelementptr inbounds %struct.cmplx, ptr %1466, i32 %1473
%1475 = getelementptr inbounds %struct.cmplx, ptr %1474, i32 0, i32 0
%1476 = load double, ptr %1475, align 8
%1477 = load ptr, ptr %9, align 4
%1478 = load i32, ptr %44, align 4
%1479 = load i32, ptr %7, align 4
%1480 = load i32, ptr %24, align 4
%1481 = mul i32 11, %1480
%1482 = add i32 7, %1481
%1483 = mul i32 %1479, %1482
%1484 = add i32 %1478, %1483
%1485 = getelementptr inbounds %struct.cmplx, ptr %1477, i32 %1484
%1486 = getelementptr inbounds %struct.cmplx, ptr %1485, i32 0, i32 0
%1487 = load double, ptr %1486, align 8
%1488 = fadd double %1476, %1487
%1489 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
store double %1488, ptr %1489, align 8
%1490 = load ptr, ptr %9, align 4
%1491 = load i32, ptr %44, align 4
%1492 = load i32, ptr %7, align 4
%1493 = load i32, ptr %24, align 4
%1494 = mul i32 11, %1493
%1495 = add i32 4, %1494
%1496 = mul i32 %1492, %1495
%1497 = add i32 %1491, %1496
%1498 = getelementptr inbounds %struct.cmplx, ptr %1490, i32 %1497
%1499 = getelementptr inbounds %struct.cmplx, ptr %1498, i32 0, i32 1
%1500 = load double, ptr %1499, align 8
%1501 = load ptr, ptr %9, align 4
%1502 = load i32, ptr %44, align 4
%1503 = load i32, ptr %7, align 4
%1504 = load i32, ptr %24, align 4
%1505 = mul i32 11, %1504
%1506 = add i32 7, %1505
%1507 = mul i32 %1503, %1506
%1508 = add i32 %1502, %1507
%1509 = getelementptr inbounds %struct.cmplx, ptr %1501, i32 %1508
%1510 = getelementptr inbounds %struct.cmplx, ptr %1509, i32 0, i32 1
%1511 = load double, ptr %1510, align 8
%1512 = fadd double %1500, %1511
%1513 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
store double %1512, ptr %1513, align 8
%1514 = load ptr, ptr %9, align 4
%1515 = load i32, ptr %44, align 4
%1516 = load i32, ptr %7, align 4
%1517 = load i32, ptr %24, align 4
%1518 = mul i32 11, %1517
%1519 = add i32 4, %1518
%1520 = mul i32 %1516, %1519
%1521 = add i32 %1515, %1520
%1522 = getelementptr inbounds %struct.cmplx, ptr %1514, i32 %1521
%1523 = getelementptr inbounds %struct.cmplx, ptr %1522, i32 0, i32 0
%1524 = load double, ptr %1523, align 8
%1525 = load ptr, ptr %9, align 4
%1526 = load i32, ptr %44, align 4
%1527 = load i32, ptr %7, align 4
%1528 = load i32, ptr %24, align 4
%1529 = mul i32 11, %1528
%1530 = add i32 7, %1529
%1531 = mul i32 %1527, %1530
%1532 = add i32 %1526, %1531
%1533 = getelementptr inbounds %struct.cmplx, ptr %1525, i32 %1532
%1534 = getelementptr inbounds %struct.cmplx, ptr %1533, i32 0, i32 0
%1535 = load double, ptr %1534, align 8
%1536 = fsub double %1524, %1535
%1537 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 0
store double %1536, ptr %1537, align 8
%1538 = load ptr, ptr %9, align 4
%1539 = load i32, ptr %44, align 4
%1540 = load i32, ptr %7, align 4
%1541 = load i32, ptr %24, align 4
%1542 = mul i32 11, %1541
%1543 = add i32 4, %1542
%1544 = mul i32 %1540, %1543
%1545 = add i32 %1539, %1544
%1546 = getelementptr inbounds %struct.cmplx, ptr %1538, i32 %1545
%1547 = getelementptr inbounds %struct.cmplx, ptr %1546, i32 0, i32 1
%1548 = load double, ptr %1547, align 8
%1549 = load ptr, ptr %9, align 4
%1550 = load i32, ptr %44, align 4
%1551 = load i32, ptr %7, align 4
%1552 = load i32, ptr %24, align 4
%1553 = mul i32 11, %1552
%1554 = add i32 7, %1553
%1555 = mul i32 %1551, %1554
%1556 = add i32 %1550, %1555
%1557 = getelementptr inbounds %struct.cmplx, ptr %1549, i32 %1556
%1558 = getelementptr inbounds %struct.cmplx, ptr %1557, i32 0, i32 1
%1559 = load double, ptr %1558, align 8
%1560 = fsub double %1548, %1559
%1561 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 1
store double %1560, ptr %1561, align 8
%1562 = load ptr, ptr %9, align 4
%1563 = load i32, ptr %44, align 4
%1564 = load i32, ptr %7, align 4
%1565 = load i32, ptr %24, align 4
%1566 = mul i32 11, %1565
%1567 = add i32 5, %1566
%1568 = mul i32 %1564, %1567
%1569 = add i32 %1563, %1568
%1570 = getelementptr inbounds %struct.cmplx, ptr %1562, i32 %1569
%1571 = getelementptr inbounds %struct.cmplx, ptr %1570, i32 0, i32 0
%1572 = load double, ptr %1571, align 8
%1573 = load ptr, ptr %9, align 4
%1574 = load i32, ptr %44, align 4
%1575 = load i32, ptr %7, align 4
%1576 = load i32, ptr %24, align 4
%1577 = mul i32 11, %1576
%1578 = add i32 6, %1577
%1579 = mul i32 %1575, %1578
%1580 = add i32 %1574, %1579
%1581 = getelementptr inbounds %struct.cmplx, ptr %1573, i32 %1580
%1582 = getelementptr inbounds %struct.cmplx, ptr %1581, i32 0, i32 0
%1583 = load double, ptr %1582, align 8
%1584 = fadd double %1572, %1583
%1585 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
store double %1584, ptr %1585, align 8
%1586 = load ptr, ptr %9, align 4
%1587 = load i32, ptr %44, align 4
%1588 = load i32, ptr %7, align 4
%1589 = load i32, ptr %24, align 4
%1590 = mul i32 11, %1589
%1591 = add i32 5, %1590
%1592 = mul i32 %1588, %1591
%1593 = add i32 %1587, %1592
%1594 = getelementptr inbounds %struct.cmplx, ptr %1586, i32 %1593
%1595 = getelementptr inbounds %struct.cmplx, ptr %1594, i32 0, i32 1
%1596 = load double, ptr %1595, align 8
%1597 = load ptr, ptr %9, align 4
%1598 = load i32, ptr %44, align 4
%1599 = load i32, ptr %7, align 4
%1600 = load i32, ptr %24, align 4
%1601 = mul i32 11, %1600
%1602 = add i32 6, %1601
%1603 = mul i32 %1599, %1602
%1604 = add i32 %1598, %1603
%1605 = getelementptr inbounds %struct.cmplx, ptr %1597, i32 %1604
%1606 = getelementptr inbounds %struct.cmplx, ptr %1605, i32 0, i32 1
%1607 = load double, ptr %1606, align 8
%1608 = fadd double %1596, %1607
%1609 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
store double %1608, ptr %1609, align 8
%1610 = load ptr, ptr %9, align 4
%1611 = load i32, ptr %44, align 4
%1612 = load i32, ptr %7, align 4
%1613 = load i32, ptr %24, align 4
%1614 = mul i32 11, %1613
%1615 = add i32 5, %1614
%1616 = mul i32 %1612, %1615
%1617 = add i32 %1611, %1616
%1618 = getelementptr inbounds %struct.cmplx, ptr %1610, i32 %1617
%1619 = getelementptr inbounds %struct.cmplx, ptr %1618, i32 0, i32 0
%1620 = load double, ptr %1619, align 8
%1621 = load ptr, ptr %9, align 4
%1622 = load i32, ptr %44, align 4
%1623 = load i32, ptr %7, align 4
%1624 = load i32, ptr %24, align 4
%1625 = mul i32 11, %1624
%1626 = add i32 6, %1625
%1627 = mul i32 %1623, %1626
%1628 = add i32 %1622, %1627
%1629 = getelementptr inbounds %struct.cmplx, ptr %1621, i32 %1628
%1630 = getelementptr inbounds %struct.cmplx, ptr %1629, i32 0, i32 0
%1631 = load double, ptr %1630, align 8
%1632 = fsub double %1620, %1631
%1633 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 0
store double %1632, ptr %1633, align 8
%1634 = load ptr, ptr %9, align 4
%1635 = load i32, ptr %44, align 4
%1636 = load i32, ptr %7, align 4
%1637 = load i32, ptr %24, align 4
%1638 = mul i32 11, %1637
%1639 = add i32 5, %1638
%1640 = mul i32 %1636, %1639
%1641 = add i32 %1635, %1640
%1642 = getelementptr inbounds %struct.cmplx, ptr %1634, i32 %1641
%1643 = getelementptr inbounds %struct.cmplx, ptr %1642, i32 0, i32 1
%1644 = load double, ptr %1643, align 8
%1645 = load ptr, ptr %9, align 4
%1646 = load i32, ptr %44, align 4
%1647 = load i32, ptr %7, align 4
%1648 = load i32, ptr %24, align 4
%1649 = mul i32 11, %1648
%1650 = add i32 6, %1649
%1651 = mul i32 %1647, %1650
%1652 = add i32 %1646, %1651
%1653 = getelementptr inbounds %struct.cmplx, ptr %1645, i32 %1652
%1654 = getelementptr inbounds %struct.cmplx, ptr %1653, i32 0, i32 1
%1655 = load double, ptr %1654, align 8
%1656 = fsub double %1644, %1655
%1657 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 1
store double %1656, ptr %1657, align 8
%1658 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 0
%1659 = load double, ptr %1658, align 8
%1660 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
%1661 = load double, ptr %1660, align 8
%1662 = fadd double %1659, %1661
%1663 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
%1664 = load double, ptr %1663, align 8
%1665 = fadd double %1662, %1664
%1666 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
%1667 = load double, ptr %1666, align 8
%1668 = fadd double %1665, %1667
%1669 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
%1670 = load double, ptr %1669, align 8
%1671 = fadd double %1668, %1670
%1672 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
%1673 = load double, ptr %1672, align 8
%1674 = fadd double %1671, %1673
%1675 = load ptr, ptr %10, align 4
%1676 = load i32, ptr %44, align 4
%1677 = load i32, ptr %7, align 4
%1678 = load i32, ptr %24, align 4
%1679 = load i32, ptr %8, align 4
%1680 = mul i32 %1679, 0
%1681 = add i32 %1678, %1680
%1682 = mul i32 %1677, %1681
%1683 = add i32 %1676, %1682
%1684 = getelementptr inbounds %struct.cmplx, ptr %1675, i32 %1683
%1685 = getelementptr inbounds %struct.cmplx, ptr %1684, i32 0, i32 0
store double %1674, ptr %1685, align 8
%1686 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 1
%1687 = load double, ptr %1686, align 8
%1688 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
%1689 = load double, ptr %1688, align 8
%1690 = fadd double %1687, %1689
%1691 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
%1692 = load double, ptr %1691, align 8
%1693 = fadd double %1690, %1692
%1694 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
%1695 = load double, ptr %1694, align 8
%1696 = fadd double %1693, %1695
%1697 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
%1698 = load double, ptr %1697, align 8
%1699 = fadd double %1696, %1698
%1700 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
%1701 = load double, ptr %1700, align 8
%1702 = fadd double %1699, %1701
%1703 = load ptr, ptr %10, align 4
%1704 = load i32, ptr %44, align 4
%1705 = load i32, ptr %7, align 4
%1706 = load i32, ptr %24, align 4
%1707 = load i32, ptr %8, align 4
%1708 = mul i32 %1707, 0
%1709 = add i32 %1706, %1708
%1710 = mul i32 %1705, %1709
%1711 = add i32 %1704, %1710
%1712 = getelementptr inbounds %struct.cmplx, ptr %1703, i32 %1711
%1713 = getelementptr inbounds %struct.cmplx, ptr %1712, i32 0, i32 1
store double %1702, ptr %1713, align 8
%1714 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 0
%1715 = load double, ptr %1714, align 8
%1716 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
%1717 = load double, ptr %1716, align 8
%1718 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %1717, double %1715)
%1719 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
%1720 = load double, ptr %1719, align 8
%1721 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %1720, double %1718)
%1722 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
%1723 = load double, ptr %1722, align 8
%1724 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %1723, double %1721)
%1725 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
%1726 = load double, ptr %1725, align 8
%1727 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %1726, double %1724)
%1728 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
%1729 = load double, ptr %1728, align 8
%1730 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %1729, double %1727)
%1731 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 0
store double %1730, ptr %1731, align 8
%1732 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 1
%1733 = load double, ptr %1732, align 8
%1734 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
%1735 = load double, ptr %1734, align 8
%1736 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %1735, double %1733)
%1737 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
%1738 = load double, ptr %1737, align 8
%1739 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %1738, double %1736)
%1740 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
%1741 = load double, ptr %1740, align 8
%1742 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %1741, double %1739)
%1743 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
%1744 = load double, ptr %1743, align 8
%1745 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %1744, double %1742)
%1746 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
%1747 = load double, ptr %1746, align 8
%1748 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %1747, double %1745)
%1749 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 1
store double %1748, ptr %1749, align 8
%1750 = load double, ptr %15, align 8
%1751 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 0
%1752 = load double, ptr %1751, align 8
%1753 = load double, ptr %17, align 8
%1754 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 0
%1755 = load double, ptr %1754, align 8
%1756 = fmul double %1753, %1755
%1757 = call double @llvm.fmuladd.f64(double %1750, double %1752, double %1756)
%1758 = load double, ptr %19, align 8
%1759 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 0
%1760 = load double, ptr %1759, align 8
%1761 = call double @llvm.fmuladd.f64(double %1758, double %1760, double %1757)
%1762 = load double, ptr %21, align 8
%1763 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 0
%1764 = load double, ptr %1763, align 8
%1765 = call double @llvm.fmuladd.f64(double %1762, double %1764, double %1761)
%1766 = load double, ptr %23, align 8
%1767 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 0
%1768 = load double, ptr %1767, align 8
%1769 = call double @llvm.fmuladd.f64(double %1766, double %1768, double %1765)
%1770 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 1
store double %1769, ptr %1770, align 8
%1771 = load double, ptr %15, align 8
%1772 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 1
%1773 = load double, ptr %1772, align 8
%1774 = load double, ptr %17, align 8
%1775 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 1
%1776 = load double, ptr %1775, align 8
%1777 = fmul double %1774, %1776
%1778 = call double @llvm.fmuladd.f64(double %1771, double %1773, double %1777)
%1779 = load double, ptr %19, align 8
%1780 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 1
%1781 = load double, ptr %1780, align 8
%1782 = call double @llvm.fmuladd.f64(double %1779, double %1781, double %1778)
%1783 = load double, ptr %21, align 8
%1784 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 1
%1785 = load double, ptr %1784, align 8
%1786 = call double @llvm.fmuladd.f64(double %1783, double %1785, double %1782)
%1787 = load double, ptr %23, align 8
%1788 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 1
%1789 = load double, ptr %1788, align 8
%1790 = call double @llvm.fmuladd.f64(double %1787, double %1789, double %1786)
%1791 = fneg double %1790
%1792 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 0
store double %1791, ptr %1792, align 8
%1793 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 0
%1794 = load double, ptr %1793, align 8
%1795 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 0
%1796 = load double, ptr %1795, align 8
%1797 = fadd double %1794, %1796
%1798 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 0
store double %1797, ptr %1798, align 8
%1799 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 1
%1800 = load double, ptr %1799, align 8
%1801 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 1
%1802 = load double, ptr %1801, align 8
%1803 = fadd double %1800, %1802
%1804 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 1
store double %1803, ptr %1804, align 8
%1805 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 0
%1806 = load double, ptr %1805, align 8
%1807 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 0
%1808 = load double, ptr %1807, align 8
%1809 = fsub double %1806, %1808
%1810 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 0
store double %1809, ptr %1810, align 8
%1811 = getelementptr inbounds %struct.cmplx, ptr %58, i32 0, i32 1
%1812 = load double, ptr %1811, align 8
%1813 = getelementptr inbounds %struct.cmplx, ptr %59, i32 0, i32 1
%1814 = load double, ptr %1813, align 8
%1815 = fsub double %1812, %1814
%1816 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 1
store double %1815, ptr %1816, align 8
%1817 = load ptr, ptr %11, align 4
%1818 = load i32, ptr %44, align 4
%1819 = sub i32 %1818, 1
%1820 = load i32, ptr %7, align 4
%1821 = sub i32 %1820, 1
%1822 = mul i32 0, %1821
%1823 = add i32 %1819, %1822
%1824 = getelementptr inbounds %struct.cmplx, ptr %1817, i32 %1823
%1825 = getelementptr inbounds %struct.cmplx, ptr %1824, i32 0, i32 0
%1826 = load double, ptr %1825, align 8
%1827 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 0
%1828 = load double, ptr %1827, align 8
%1829 = load i32, ptr %12, align 4
%1830 = sitofp i32 %1829 to double
%1831 = load ptr, ptr %11, align 4
%1832 = load i32, ptr %44, align 4
%1833 = sub i32 %1832, 1
%1834 = load i32, ptr %7, align 4
%1835 = sub i32 %1834, 1
%1836 = mul i32 0, %1835
%1837 = add i32 %1833, %1836
%1838 = getelementptr inbounds %struct.cmplx, ptr %1831, i32 %1837
%1839 = getelementptr inbounds %struct.cmplx, ptr %1838, i32 0, i32 1
%1840 = load double, ptr %1839, align 8
%1841 = fmul double %1830, %1840
%1842 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 1
%1843 = load double, ptr %1842, align 8
%1844 = fmul double %1841, %1843
%1845 = fneg double %1844
%1846 = call double @llvm.fmuladd.f64(double %1826, double %1828, double %1845)
%1847 = load ptr, ptr %10, align 4
%1848 = load i32, ptr %44, align 4
%1849 = load i32, ptr %7, align 4
%1850 = load i32, ptr %24, align 4
%1851 = load i32, ptr %8, align 4
%1852 = mul i32 %1851, 1
%1853 = add i32 %1850, %1852
%1854 = mul i32 %1849, %1853
%1855 = add i32 %1848, %1854
%1856 = getelementptr inbounds %struct.cmplx, ptr %1847, i32 %1855
%1857 = getelementptr inbounds %struct.cmplx, ptr %1856, i32 0, i32 0
store double %1846, ptr %1857, align 8
%1858 = load ptr, ptr %11, align 4
%1859 = load i32, ptr %44, align 4
%1860 = sub i32 %1859, 1
%1861 = load i32, ptr %7, align 4
%1862 = sub i32 %1861, 1
%1863 = mul i32 0, %1862
%1864 = add i32 %1860, %1863
%1865 = getelementptr inbounds %struct.cmplx, ptr %1858, i32 %1864
%1866 = getelementptr inbounds %struct.cmplx, ptr %1865, i32 0, i32 0
%1867 = load double, ptr %1866, align 8
%1868 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 1
%1869 = load double, ptr %1868, align 8
%1870 = load i32, ptr %12, align 4
%1871 = sitofp i32 %1870 to double
%1872 = load ptr, ptr %11, align 4
%1873 = load i32, ptr %44, align 4
%1874 = sub i32 %1873, 1
%1875 = load i32, ptr %7, align 4
%1876 = sub i32 %1875, 1
%1877 = mul i32 0, %1876
%1878 = add i32 %1874, %1877
%1879 = getelementptr inbounds %struct.cmplx, ptr %1872, i32 %1878
%1880 = getelementptr inbounds %struct.cmplx, ptr %1879, i32 0, i32 1
%1881 = load double, ptr %1880, align 8
%1882 = fmul double %1871, %1881
%1883 = getelementptr inbounds %struct.cmplx, ptr %56, i32 0, i32 0
%1884 = load double, ptr %1883, align 8
%1885 = fmul double %1882, %1884
%1886 = call double @llvm.fmuladd.f64(double %1867, double %1869, double %1885)
%1887 = load ptr, ptr %10, align 4
%1888 = load i32, ptr %44, align 4
%1889 = load i32, ptr %7, align 4
%1890 = load i32, ptr %24, align 4
%1891 = load i32, ptr %8, align 4
%1892 = mul i32 %1891, 1
%1893 = add i32 %1890, %1892
%1894 = mul i32 %1889, %1893
%1895 = add i32 %1888, %1894
%1896 = getelementptr inbounds %struct.cmplx, ptr %1887, i32 %1895
%1897 = getelementptr inbounds %struct.cmplx, ptr %1896, i32 0, i32 1
store double %1886, ptr %1897, align 8
%1898 = load ptr, ptr %11, align 4
%1899 = load i32, ptr %44, align 4
%1900 = sub i32 %1899, 1
%1901 = load i32, ptr %7, align 4
%1902 = sub i32 %1901, 1
%1903 = mul i32 9, %1902
%1904 = add i32 %1900, %1903
%1905 = getelementptr inbounds %struct.cmplx, ptr %1898, i32 %1904
%1906 = getelementptr inbounds %struct.cmplx, ptr %1905, i32 0, i32 0
%1907 = load double, ptr %1906, align 8
%1908 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 0
%1909 = load double, ptr %1908, align 8
%1910 = load i32, ptr %12, align 4
%1911 = sitofp i32 %1910 to double
%1912 = load ptr, ptr %11, align 4
%1913 = load i32, ptr %44, align 4
%1914 = sub i32 %1913, 1
%1915 = load i32, ptr %7, align 4
%1916 = sub i32 %1915, 1
%1917 = mul i32 9, %1916
%1918 = add i32 %1914, %1917
%1919 = getelementptr inbounds %struct.cmplx, ptr %1912, i32 %1918
%1920 = getelementptr inbounds %struct.cmplx, ptr %1919, i32 0, i32 1
%1921 = load double, ptr %1920, align 8
%1922 = fmul double %1911, %1921
%1923 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 1
%1924 = load double, ptr %1923, align 8
%1925 = fmul double %1922, %1924
%1926 = fneg double %1925
%1927 = call double @llvm.fmuladd.f64(double %1907, double %1909, double %1926)
%1928 = load ptr, ptr %10, align 4
%1929 = load i32, ptr %44, align 4
%1930 = load i32, ptr %7, align 4
%1931 = load i32, ptr %24, align 4
%1932 = load i32, ptr %8, align 4
%1933 = mul i32 %1932, 10
%1934 = add i32 %1931, %1933
%1935 = mul i32 %1930, %1934
%1936 = add i32 %1929, %1935
%1937 = getelementptr inbounds %struct.cmplx, ptr %1928, i32 %1936
%1938 = getelementptr inbounds %struct.cmplx, ptr %1937, i32 0, i32 0
store double %1927, ptr %1938, align 8
%1939 = load ptr, ptr %11, align 4
%1940 = load i32, ptr %44, align 4
%1941 = sub i32 %1940, 1
%1942 = load i32, ptr %7, align 4
%1943 = sub i32 %1942, 1
%1944 = mul i32 9, %1943
%1945 = add i32 %1941, %1944
%1946 = getelementptr inbounds %struct.cmplx, ptr %1939, i32 %1945
%1947 = getelementptr inbounds %struct.cmplx, ptr %1946, i32 0, i32 0
%1948 = load double, ptr %1947, align 8
%1949 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 1
%1950 = load double, ptr %1949, align 8
%1951 = load i32, ptr %12, align 4
%1952 = sitofp i32 %1951 to double
%1953 = load ptr, ptr %11, align 4
%1954 = load i32, ptr %44, align 4
%1955 = sub i32 %1954, 1
%1956 = load i32, ptr %7, align 4
%1957 = sub i32 %1956, 1
%1958 = mul i32 9, %1957
%1959 = add i32 %1955, %1958
%1960 = getelementptr inbounds %struct.cmplx, ptr %1953, i32 %1959
%1961 = getelementptr inbounds %struct.cmplx, ptr %1960, i32 0, i32 1
%1962 = load double, ptr %1961, align 8
%1963 = fmul double %1952, %1962
%1964 = getelementptr inbounds %struct.cmplx, ptr %57, i32 0, i32 0
%1965 = load double, ptr %1964, align 8
%1966 = fmul double %1963, %1965
%1967 = call double @llvm.fmuladd.f64(double %1948, double %1950, double %1966)
%1968 = load ptr, ptr %10, align 4
%1969 = load i32, ptr %44, align 4
%1970 = load i32, ptr %7, align 4
%1971 = load i32, ptr %24, align 4
%1972 = load i32, ptr %8, align 4
%1973 = mul i32 %1972, 10
%1974 = add i32 %1971, %1973
%1975 = mul i32 %1970, %1974
%1976 = add i32 %1969, %1975
%1977 = getelementptr inbounds %struct.cmplx, ptr %1968, i32 %1976
%1978 = getelementptr inbounds %struct.cmplx, ptr %1977, i32 0, i32 1
store double %1967, ptr %1978, align 8
%1979 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 0
%1980 = load double, ptr %1979, align 8
%1981 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
%1982 = load double, ptr %1981, align 8
%1983 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %1982, double %1980)
%1984 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
%1985 = load double, ptr %1984, align 8
%1986 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %1985, double %1983)
%1987 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
%1988 = load double, ptr %1987, align 8
%1989 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %1988, double %1986)
%1990 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
%1991 = load double, ptr %1990, align 8
%1992 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %1991, double %1989)
%1993 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
%1994 = load double, ptr %1993, align 8
%1995 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %1994, double %1992)
%1996 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 0
store double %1995, ptr %1996, align 8
%1997 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 1
%1998 = load double, ptr %1997, align 8
%1999 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
%2000 = load double, ptr %1999, align 8
%2001 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %2000, double %1998)
%2002 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
%2003 = load double, ptr %2002, align 8
%2004 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %2003, double %2001)
%2005 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
%2006 = load double, ptr %2005, align 8
%2007 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %2006, double %2004)
%2008 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
%2009 = load double, ptr %2008, align 8
%2010 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %2009, double %2007)
%2011 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
%2012 = load double, ptr %2011, align 8
%2013 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %2012, double %2010)
%2014 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 1
store double %2013, ptr %2014, align 8
%2015 = load double, ptr %17, align 8
%2016 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 0
%2017 = load double, ptr %2016, align 8
%2018 = load double, ptr %21, align 8
%2019 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 0
%2020 = load double, ptr %2019, align 8
%2021 = fmul double %2018, %2020
%2022 = call double @llvm.fmuladd.f64(double %2015, double %2017, double %2021)
%2023 = load double, ptr %23, align 8
%2024 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 0
%2025 = load double, ptr %2024, align 8
%2026 = fneg double %2023
%2027 = call double @llvm.fmuladd.f64(double %2026, double %2025, double %2022)
%2028 = load double, ptr %19, align 8
%2029 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 0
%2030 = load double, ptr %2029, align 8
%2031 = fneg double %2028
%2032 = call double @llvm.fmuladd.f64(double %2031, double %2030, double %2027)
%2033 = load double, ptr %15, align 8
%2034 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 0
%2035 = load double, ptr %2034, align 8
%2036 = fneg double %2033
%2037 = call double @llvm.fmuladd.f64(double %2036, double %2035, double %2032)
%2038 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 1
store double %2037, ptr %2038, align 8
%2039 = load double, ptr %17, align 8
%2040 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 1
%2041 = load double, ptr %2040, align 8
%2042 = load double, ptr %21, align 8
%2043 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 1
%2044 = load double, ptr %2043, align 8
%2045 = fmul double %2042, %2044
%2046 = call double @llvm.fmuladd.f64(double %2039, double %2041, double %2045)
%2047 = load double, ptr %23, align 8
%2048 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 1
%2049 = load double, ptr %2048, align 8
%2050 = fneg double %2047
%2051 = call double @llvm.fmuladd.f64(double %2050, double %2049, double %2046)
%2052 = load double, ptr %19, align 8
%2053 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 1
%2054 = load double, ptr %2053, align 8
%2055 = fneg double %2052
%2056 = call double @llvm.fmuladd.f64(double %2055, double %2054, double %2051)
%2057 = load double, ptr %15, align 8
%2058 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 1
%2059 = load double, ptr %2058, align 8
%2060 = fneg double %2057
%2061 = call double @llvm.fmuladd.f64(double %2060, double %2059, double %2056)
%2062 = fneg double %2061
%2063 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 0
store double %2062, ptr %2063, align 8
%2064 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 0
%2065 = load double, ptr %2064, align 8
%2066 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 0
%2067 = load double, ptr %2066, align 8
%2068 = fadd double %2065, %2067
%2069 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 0
store double %2068, ptr %2069, align 8
%2070 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 1
%2071 = load double, ptr %2070, align 8
%2072 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 1
%2073 = load double, ptr %2072, align 8
%2074 = fadd double %2071, %2073
%2075 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 1
store double %2074, ptr %2075, align 8
%2076 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 0
%2077 = load double, ptr %2076, align 8
%2078 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 0
%2079 = load double, ptr %2078, align 8
%2080 = fsub double %2077, %2079
%2081 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 0
store double %2080, ptr %2081, align 8
%2082 = getelementptr inbounds %struct.cmplx, ptr %62, i32 0, i32 1
%2083 = load double, ptr %2082, align 8
%2084 = getelementptr inbounds %struct.cmplx, ptr %63, i32 0, i32 1
%2085 = load double, ptr %2084, align 8
%2086 = fsub double %2083, %2085
%2087 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 1
store double %2086, ptr %2087, align 8
%2088 = load ptr, ptr %11, align 4
%2089 = load i32, ptr %44, align 4
%2090 = sub i32 %2089, 1
%2091 = load i32, ptr %7, align 4
%2092 = sub i32 %2091, 1
%2093 = mul i32 1, %2092
%2094 = add i32 %2090, %2093
%2095 = getelementptr inbounds %struct.cmplx, ptr %2088, i32 %2094
%2096 = getelementptr inbounds %struct.cmplx, ptr %2095, i32 0, i32 0
%2097 = load double, ptr %2096, align 8
%2098 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 0
%2099 = load double, ptr %2098, align 8
%2100 = load i32, ptr %12, align 4
%2101 = sitofp i32 %2100 to double
%2102 = load ptr, ptr %11, align 4
%2103 = load i32, ptr %44, align 4
%2104 = sub i32 %2103, 1
%2105 = load i32, ptr %7, align 4
%2106 = sub i32 %2105, 1
%2107 = mul i32 1, %2106
%2108 = add i32 %2104, %2107
%2109 = getelementptr inbounds %struct.cmplx, ptr %2102, i32 %2108
%2110 = getelementptr inbounds %struct.cmplx, ptr %2109, i32 0, i32 1
%2111 = load double, ptr %2110, align 8
%2112 = fmul double %2101, %2111
%2113 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 1
%2114 = load double, ptr %2113, align 8
%2115 = fmul double %2112, %2114
%2116 = fneg double %2115
%2117 = call double @llvm.fmuladd.f64(double %2097, double %2099, double %2116)
%2118 = load ptr, ptr %10, align 4
%2119 = load i32, ptr %44, align 4
%2120 = load i32, ptr %7, align 4
%2121 = load i32, ptr %24, align 4
%2122 = load i32, ptr %8, align 4
%2123 = mul i32 %2122, 2
%2124 = add i32 %2121, %2123
%2125 = mul i32 %2120, %2124
%2126 = add i32 %2119, %2125
%2127 = getelementptr inbounds %struct.cmplx, ptr %2118, i32 %2126
%2128 = getelementptr inbounds %struct.cmplx, ptr %2127, i32 0, i32 0
store double %2117, ptr %2128, align 8
%2129 = load ptr, ptr %11, align 4
%2130 = load i32, ptr %44, align 4
%2131 = sub i32 %2130, 1
%2132 = load i32, ptr %7, align 4
%2133 = sub i32 %2132, 1
%2134 = mul i32 1, %2133
%2135 = add i32 %2131, %2134
%2136 = getelementptr inbounds %struct.cmplx, ptr %2129, i32 %2135
%2137 = getelementptr inbounds %struct.cmplx, ptr %2136, i32 0, i32 0
%2138 = load double, ptr %2137, align 8
%2139 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 1
%2140 = load double, ptr %2139, align 8
%2141 = load i32, ptr %12, align 4
%2142 = sitofp i32 %2141 to double
%2143 = load ptr, ptr %11, align 4
%2144 = load i32, ptr %44, align 4
%2145 = sub i32 %2144, 1
%2146 = load i32, ptr %7, align 4
%2147 = sub i32 %2146, 1
%2148 = mul i32 1, %2147
%2149 = add i32 %2145, %2148
%2150 = getelementptr inbounds %struct.cmplx, ptr %2143, i32 %2149
%2151 = getelementptr inbounds %struct.cmplx, ptr %2150, i32 0, i32 1
%2152 = load double, ptr %2151, align 8
%2153 = fmul double %2142, %2152
%2154 = getelementptr inbounds %struct.cmplx, ptr %60, i32 0, i32 0
%2155 = load double, ptr %2154, align 8
%2156 = fmul double %2153, %2155
%2157 = call double @llvm.fmuladd.f64(double %2138, double %2140, double %2156)
%2158 = load ptr, ptr %10, align 4
%2159 = load i32, ptr %44, align 4
%2160 = load i32, ptr %7, align 4
%2161 = load i32, ptr %24, align 4
%2162 = load i32, ptr %8, align 4
%2163 = mul i32 %2162, 2
%2164 = add i32 %2161, %2163
%2165 = mul i32 %2160, %2164
%2166 = add i32 %2159, %2165
%2167 = getelementptr inbounds %struct.cmplx, ptr %2158, i32 %2166
%2168 = getelementptr inbounds %struct.cmplx, ptr %2167, i32 0, i32 1
store double %2157, ptr %2168, align 8
%2169 = load ptr, ptr %11, align 4
%2170 = load i32, ptr %44, align 4
%2171 = sub i32 %2170, 1
%2172 = load i32, ptr %7, align 4
%2173 = sub i32 %2172, 1
%2174 = mul i32 8, %2173
%2175 = add i32 %2171, %2174
%2176 = getelementptr inbounds %struct.cmplx, ptr %2169, i32 %2175
%2177 = getelementptr inbounds %struct.cmplx, ptr %2176, i32 0, i32 0
%2178 = load double, ptr %2177, align 8
%2179 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 0
%2180 = load double, ptr %2179, align 8
%2181 = load i32, ptr %12, align 4
%2182 = sitofp i32 %2181 to double
%2183 = load ptr, ptr %11, align 4
%2184 = load i32, ptr %44, align 4
%2185 = sub i32 %2184, 1
%2186 = load i32, ptr %7, align 4
%2187 = sub i32 %2186, 1
%2188 = mul i32 8, %2187
%2189 = add i32 %2185, %2188
%2190 = getelementptr inbounds %struct.cmplx, ptr %2183, i32 %2189
%2191 = getelementptr inbounds %struct.cmplx, ptr %2190, i32 0, i32 1
%2192 = load double, ptr %2191, align 8
%2193 = fmul double %2182, %2192
%2194 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 1
%2195 = load double, ptr %2194, align 8
%2196 = fmul double %2193, %2195
%2197 = fneg double %2196
%2198 = call double @llvm.fmuladd.f64(double %2178, double %2180, double %2197)
%2199 = load ptr, ptr %10, align 4
%2200 = load i32, ptr %44, align 4
%2201 = load i32, ptr %7, align 4
%2202 = load i32, ptr %24, align 4
%2203 = load i32, ptr %8, align 4
%2204 = mul i32 %2203, 9
%2205 = add i32 %2202, %2204
%2206 = mul i32 %2201, %2205
%2207 = add i32 %2200, %2206
%2208 = getelementptr inbounds %struct.cmplx, ptr %2199, i32 %2207
%2209 = getelementptr inbounds %struct.cmplx, ptr %2208, i32 0, i32 0
store double %2198, ptr %2209, align 8
%2210 = load ptr, ptr %11, align 4
%2211 = load i32, ptr %44, align 4
%2212 = sub i32 %2211, 1
%2213 = load i32, ptr %7, align 4
%2214 = sub i32 %2213, 1
%2215 = mul i32 8, %2214
%2216 = add i32 %2212, %2215
%2217 = getelementptr inbounds %struct.cmplx, ptr %2210, i32 %2216
%2218 = getelementptr inbounds %struct.cmplx, ptr %2217, i32 0, i32 0
%2219 = load double, ptr %2218, align 8
%2220 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 1
%2221 = load double, ptr %2220, align 8
%2222 = load i32, ptr %12, align 4
%2223 = sitofp i32 %2222 to double
%2224 = load ptr, ptr %11, align 4
%2225 = load i32, ptr %44, align 4
%2226 = sub i32 %2225, 1
%2227 = load i32, ptr %7, align 4
%2228 = sub i32 %2227, 1
%2229 = mul i32 8, %2228
%2230 = add i32 %2226, %2229
%2231 = getelementptr inbounds %struct.cmplx, ptr %2224, i32 %2230
%2232 = getelementptr inbounds %struct.cmplx, ptr %2231, i32 0, i32 1
%2233 = load double, ptr %2232, align 8
%2234 = fmul double %2223, %2233
%2235 = getelementptr inbounds %struct.cmplx, ptr %61, i32 0, i32 0
%2236 = load double, ptr %2235, align 8
%2237 = fmul double %2234, %2236
%2238 = call double @llvm.fmuladd.f64(double %2219, double %2221, double %2237)
%2239 = load ptr, ptr %10, align 4
%2240 = load i32, ptr %44, align 4
%2241 = load i32, ptr %7, align 4
%2242 = load i32, ptr %24, align 4
%2243 = load i32, ptr %8, align 4
%2244 = mul i32 %2243, 9
%2245 = add i32 %2242, %2244
%2246 = mul i32 %2241, %2245
%2247 = add i32 %2240, %2246
%2248 = getelementptr inbounds %struct.cmplx, ptr %2239, i32 %2247
%2249 = getelementptr inbounds %struct.cmplx, ptr %2248, i32 0, i32 1
store double %2238, ptr %2249, align 8
%2250 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 0
%2251 = load double, ptr %2250, align 8
%2252 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
%2253 = load double, ptr %2252, align 8
%2254 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %2253, double %2251)
%2255 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
%2256 = load double, ptr %2255, align 8
%2257 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %2256, double %2254)
%2258 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
%2259 = load double, ptr %2258, align 8
%2260 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %2259, double %2257)
%2261 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
%2262 = load double, ptr %2261, align 8
%2263 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %2262, double %2260)
%2264 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
%2265 = load double, ptr %2264, align 8
%2266 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %2265, double %2263)
%2267 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 0
store double %2266, ptr %2267, align 8
%2268 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 1
%2269 = load double, ptr %2268, align 8
%2270 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
%2271 = load double, ptr %2270, align 8
%2272 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %2271, double %2269)
%2273 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
%2274 = load double, ptr %2273, align 8
%2275 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %2274, double %2272)
%2276 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
%2277 = load double, ptr %2276, align 8
%2278 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %2277, double %2275)
%2279 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
%2280 = load double, ptr %2279, align 8
%2281 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %2280, double %2278)
%2282 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
%2283 = load double, ptr %2282, align 8
%2284 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %2283, double %2281)
%2285 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 1
store double %2284, ptr %2285, align 8
%2286 = load double, ptr %19, align 8
%2287 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 0
%2288 = load double, ptr %2287, align 8
%2289 = load double, ptr %23, align 8
%2290 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 0
%2291 = load double, ptr %2290, align 8
%2292 = fmul double %2289, %2291
%2293 = fneg double %2292
%2294 = call double @llvm.fmuladd.f64(double %2286, double %2288, double %2293)
%2295 = load double, ptr %17, align 8
%2296 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 0
%2297 = load double, ptr %2296, align 8
%2298 = fneg double %2295
%2299 = call double @llvm.fmuladd.f64(double %2298, double %2297, double %2294)
%2300 = load double, ptr %15, align 8
%2301 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 0
%2302 = load double, ptr %2301, align 8
%2303 = call double @llvm.fmuladd.f64(double %2300, double %2302, double %2299)
%2304 = load double, ptr %21, align 8
%2305 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 0
%2306 = load double, ptr %2305, align 8
%2307 = call double @llvm.fmuladd.f64(double %2304, double %2306, double %2303)
%2308 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 1
store double %2307, ptr %2308, align 8
%2309 = load double, ptr %19, align 8
%2310 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 1
%2311 = load double, ptr %2310, align 8
%2312 = load double, ptr %23, align 8
%2313 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 1
%2314 = load double, ptr %2313, align 8
%2315 = fmul double %2312, %2314
%2316 = fneg double %2315
%2317 = call double @llvm.fmuladd.f64(double %2309, double %2311, double %2316)
%2318 = load double, ptr %17, align 8
%2319 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 1
%2320 = load double, ptr %2319, align 8
%2321 = fneg double %2318
%2322 = call double @llvm.fmuladd.f64(double %2321, double %2320, double %2317)
%2323 = load double, ptr %15, align 8
%2324 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 1
%2325 = load double, ptr %2324, align 8
%2326 = call double @llvm.fmuladd.f64(double %2323, double %2325, double %2322)
%2327 = load double, ptr %21, align 8
%2328 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 1
%2329 = load double, ptr %2328, align 8
%2330 = call double @llvm.fmuladd.f64(double %2327, double %2329, double %2326)
%2331 = fneg double %2330
%2332 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 0
store double %2331, ptr %2332, align 8
%2333 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 0
%2334 = load double, ptr %2333, align 8
%2335 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 0
%2336 = load double, ptr %2335, align 8
%2337 = fadd double %2334, %2336
%2338 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 0
store double %2337, ptr %2338, align 8
%2339 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 1
%2340 = load double, ptr %2339, align 8
%2341 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 1
%2342 = load double, ptr %2341, align 8
%2343 = fadd double %2340, %2342
%2344 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 1
store double %2343, ptr %2344, align 8
%2345 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 0
%2346 = load double, ptr %2345, align 8
%2347 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 0
%2348 = load double, ptr %2347, align 8
%2349 = fsub double %2346, %2348
%2350 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 0
store double %2349, ptr %2350, align 8
%2351 = getelementptr inbounds %struct.cmplx, ptr %66, i32 0, i32 1
%2352 = load double, ptr %2351, align 8
%2353 = getelementptr inbounds %struct.cmplx, ptr %67, i32 0, i32 1
%2354 = load double, ptr %2353, align 8
%2355 = fsub double %2352, %2354
%2356 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 1
store double %2355, ptr %2356, align 8
%2357 = load ptr, ptr %11, align 4
%2358 = load i32, ptr %44, align 4
%2359 = sub i32 %2358, 1
%2360 = load i32, ptr %7, align 4
%2361 = sub i32 %2360, 1
%2362 = mul i32 2, %2361
%2363 = add i32 %2359, %2362
%2364 = getelementptr inbounds %struct.cmplx, ptr %2357, i32 %2363
%2365 = getelementptr inbounds %struct.cmplx, ptr %2364, i32 0, i32 0
%2366 = load double, ptr %2365, align 8
%2367 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 0
%2368 = load double, ptr %2367, align 8
%2369 = load i32, ptr %12, align 4
%2370 = sitofp i32 %2369 to double
%2371 = load ptr, ptr %11, align 4
%2372 = load i32, ptr %44, align 4
%2373 = sub i32 %2372, 1
%2374 = load i32, ptr %7, align 4
%2375 = sub i32 %2374, 1
%2376 = mul i32 2, %2375
%2377 = add i32 %2373, %2376
%2378 = getelementptr inbounds %struct.cmplx, ptr %2371, i32 %2377
%2379 = getelementptr inbounds %struct.cmplx, ptr %2378, i32 0, i32 1
%2380 = load double, ptr %2379, align 8
%2381 = fmul double %2370, %2380
%2382 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 1
%2383 = load double, ptr %2382, align 8
%2384 = fmul double %2381, %2383
%2385 = fneg double %2384
%2386 = call double @llvm.fmuladd.f64(double %2366, double %2368, double %2385)
%2387 = load ptr, ptr %10, align 4
%2388 = load i32, ptr %44, align 4
%2389 = load i32, ptr %7, align 4
%2390 = load i32, ptr %24, align 4
%2391 = load i32, ptr %8, align 4
%2392 = mul i32 %2391, 3
%2393 = add i32 %2390, %2392
%2394 = mul i32 %2389, %2393
%2395 = add i32 %2388, %2394
%2396 = getelementptr inbounds %struct.cmplx, ptr %2387, i32 %2395
%2397 = getelementptr inbounds %struct.cmplx, ptr %2396, i32 0, i32 0
store double %2386, ptr %2397, align 8
%2398 = load ptr, ptr %11, align 4
%2399 = load i32, ptr %44, align 4
%2400 = sub i32 %2399, 1
%2401 = load i32, ptr %7, align 4
%2402 = sub i32 %2401, 1
%2403 = mul i32 2, %2402
%2404 = add i32 %2400, %2403
%2405 = getelementptr inbounds %struct.cmplx, ptr %2398, i32 %2404
%2406 = getelementptr inbounds %struct.cmplx, ptr %2405, i32 0, i32 0
%2407 = load double, ptr %2406, align 8
%2408 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 1
%2409 = load double, ptr %2408, align 8
%2410 = load i32, ptr %12, align 4
%2411 = sitofp i32 %2410 to double
%2412 = load ptr, ptr %11, align 4
%2413 = load i32, ptr %44, align 4
%2414 = sub i32 %2413, 1
%2415 = load i32, ptr %7, align 4
%2416 = sub i32 %2415, 1
%2417 = mul i32 2, %2416
%2418 = add i32 %2414, %2417
%2419 = getelementptr inbounds %struct.cmplx, ptr %2412, i32 %2418
%2420 = getelementptr inbounds %struct.cmplx, ptr %2419, i32 0, i32 1
%2421 = load double, ptr %2420, align 8
%2422 = fmul double %2411, %2421
%2423 = getelementptr inbounds %struct.cmplx, ptr %64, i32 0, i32 0
%2424 = load double, ptr %2423, align 8
%2425 = fmul double %2422, %2424
%2426 = call double @llvm.fmuladd.f64(double %2407, double %2409, double %2425)
%2427 = load ptr, ptr %10, align 4
%2428 = load i32, ptr %44, align 4
%2429 = load i32, ptr %7, align 4
%2430 = load i32, ptr %24, align 4
%2431 = load i32, ptr %8, align 4
%2432 = mul i32 %2431, 3
%2433 = add i32 %2430, %2432
%2434 = mul i32 %2429, %2433
%2435 = add i32 %2428, %2434
%2436 = getelementptr inbounds %struct.cmplx, ptr %2427, i32 %2435
%2437 = getelementptr inbounds %struct.cmplx, ptr %2436, i32 0, i32 1
store double %2426, ptr %2437, align 8
%2438 = load ptr, ptr %11, align 4
%2439 = load i32, ptr %44, align 4
%2440 = sub i32 %2439, 1
%2441 = load i32, ptr %7, align 4
%2442 = sub i32 %2441, 1
%2443 = mul i32 7, %2442
%2444 = add i32 %2440, %2443
%2445 = getelementptr inbounds %struct.cmplx, ptr %2438, i32 %2444
%2446 = getelementptr inbounds %struct.cmplx, ptr %2445, i32 0, i32 0
%2447 = load double, ptr %2446, align 8
%2448 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 0
%2449 = load double, ptr %2448, align 8
%2450 = load i32, ptr %12, align 4
%2451 = sitofp i32 %2450 to double
%2452 = load ptr, ptr %11, align 4
%2453 = load i32, ptr %44, align 4
%2454 = sub i32 %2453, 1
%2455 = load i32, ptr %7, align 4
%2456 = sub i32 %2455, 1
%2457 = mul i32 7, %2456
%2458 = add i32 %2454, %2457
%2459 = getelementptr inbounds %struct.cmplx, ptr %2452, i32 %2458
%2460 = getelementptr inbounds %struct.cmplx, ptr %2459, i32 0, i32 1
%2461 = load double, ptr %2460, align 8
%2462 = fmul double %2451, %2461
%2463 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 1
%2464 = load double, ptr %2463, align 8
%2465 = fmul double %2462, %2464
%2466 = fneg double %2465
%2467 = call double @llvm.fmuladd.f64(double %2447, double %2449, double %2466)
%2468 = load ptr, ptr %10, align 4
%2469 = load i32, ptr %44, align 4
%2470 = load i32, ptr %7, align 4
%2471 = load i32, ptr %24, align 4
%2472 = load i32, ptr %8, align 4
%2473 = mul i32 %2472, 8
%2474 = add i32 %2471, %2473
%2475 = mul i32 %2470, %2474
%2476 = add i32 %2469, %2475
%2477 = getelementptr inbounds %struct.cmplx, ptr %2468, i32 %2476
%2478 = getelementptr inbounds %struct.cmplx, ptr %2477, i32 0, i32 0
store double %2467, ptr %2478, align 8
%2479 = load ptr, ptr %11, align 4
%2480 = load i32, ptr %44, align 4
%2481 = sub i32 %2480, 1
%2482 = load i32, ptr %7, align 4
%2483 = sub i32 %2482, 1
%2484 = mul i32 7, %2483
%2485 = add i32 %2481, %2484
%2486 = getelementptr inbounds %struct.cmplx, ptr %2479, i32 %2485
%2487 = getelementptr inbounds %struct.cmplx, ptr %2486, i32 0, i32 0
%2488 = load double, ptr %2487, align 8
%2489 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 1
%2490 = load double, ptr %2489, align 8
%2491 = load i32, ptr %12, align 4
%2492 = sitofp i32 %2491 to double
%2493 = load ptr, ptr %11, align 4
%2494 = load i32, ptr %44, align 4
%2495 = sub i32 %2494, 1
%2496 = load i32, ptr %7, align 4
%2497 = sub i32 %2496, 1
%2498 = mul i32 7, %2497
%2499 = add i32 %2495, %2498
%2500 = getelementptr inbounds %struct.cmplx, ptr %2493, i32 %2499
%2501 = getelementptr inbounds %struct.cmplx, ptr %2500, i32 0, i32 1
%2502 = load double, ptr %2501, align 8
%2503 = fmul double %2492, %2502
%2504 = getelementptr inbounds %struct.cmplx, ptr %65, i32 0, i32 0
%2505 = load double, ptr %2504, align 8
%2506 = fmul double %2503, %2505
%2507 = call double @llvm.fmuladd.f64(double %2488, double %2490, double %2506)
%2508 = load ptr, ptr %10, align 4
%2509 = load i32, ptr %44, align 4
%2510 = load i32, ptr %7, align 4
%2511 = load i32, ptr %24, align 4
%2512 = load i32, ptr %8, align 4
%2513 = mul i32 %2512, 8
%2514 = add i32 %2511, %2513
%2515 = mul i32 %2510, %2514
%2516 = add i32 %2509, %2515
%2517 = getelementptr inbounds %struct.cmplx, ptr %2508, i32 %2516
%2518 = getelementptr inbounds %struct.cmplx, ptr %2517, i32 0, i32 1
store double %2507, ptr %2518, align 8
%2519 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 0
%2520 = load double, ptr %2519, align 8
%2521 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 0
%2522 = load double, ptr %2521, align 8
%2523 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %2522, double %2520)
%2524 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 0
%2525 = load double, ptr %2524, align 8
%2526 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %2525, double %2523)
%2527 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 0
%2528 = load double, ptr %2527, align 8
%2529 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %2528, double %2526)
%2530 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 0
%2531 = load double, ptr %2530, align 8
%2532 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %2531, double %2529)
%2533 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 0
%2534 = load double, ptr %2533, align 8
%2535 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %2534, double %2532)
%2536 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 0
store double %2535, ptr %2536, align 8
%2537 = getelementptr inbounds %struct.cmplx, ptr %45, i32 0, i32 1
%2538 = load double, ptr %2537, align 8
%2539 = getelementptr inbounds %struct.cmplx, ptr %46, i32 0, i32 1
%2540 = load double, ptr %2539, align 8
%2541 = call double @llvm.fmuladd.f64(double 0xBFE4F49E7F775887, double %2540, double %2538)
%2542 = getelementptr inbounds %struct.cmplx, ptr %47, i32 0, i32 1
%2543 = load double, ptr %2542, align 8
%2544 = call double @llvm.fmuladd.f64(double 0xBFC2375F640F44DB, double %2543, double %2541)
%2545 = getelementptr inbounds %struct.cmplx, ptr %48, i32 0, i32 1
%2546 = load double, ptr %2545, align 8
%2547 = call double @llvm.fmuladd.f64(double 0x3FEAEB8C8764F0BA, double %2546, double %2544)
%2548 = getelementptr inbounds %struct.cmplx, ptr %49, i32 0, i32 1
%2549 = load double, ptr %2548, align 8
%2550 = call double @llvm.fmuladd.f64(double 0xBFEEB42A9BCD5057, double %2549, double %2547)
%2551 = getelementptr inbounds %struct.cmplx, ptr %50, i32 0, i32 1
%2552 = load double, ptr %2551, align 8
%2553 = call double @llvm.fmuladd.f64(double 0x3FDA9628D9C712B6, double %2552, double %2550)
%2554 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 1
store double %2553, ptr %2554, align 8
%2555 = load double, ptr %21, align 8
%2556 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 0
%2557 = load double, ptr %2556, align 8
%2558 = load double, ptr %19, align 8
%2559 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 0
%2560 = load double, ptr %2559, align 8
%2561 = fmul double %2558, %2560
%2562 = fneg double %2561
%2563 = call double @llvm.fmuladd.f64(double %2555, double %2557, double %2562)
%2564 = load double, ptr %15, align 8
%2565 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 0
%2566 = load double, ptr %2565, align 8
%2567 = call double @llvm.fmuladd.f64(double %2564, double %2566, double %2563)
%2568 = load double, ptr %23, align 8
%2569 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 0
%2570 = load double, ptr %2569, align 8
%2571 = call double @llvm.fmuladd.f64(double %2568, double %2570, double %2567)
%2572 = load double, ptr %17, align 8
%2573 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 0
%2574 = load double, ptr %2573, align 8
%2575 = fneg double %2572
%2576 = call double @llvm.fmuladd.f64(double %2575, double %2574, double %2571)
%2577 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 1
store double %2576, ptr %2577, align 8
%2578 = load double, ptr %21, align 8
%2579 = getelementptr inbounds %struct.cmplx, ptr %55, i32 0, i32 1
%2580 = load double, ptr %2579, align 8
%2581 = load double, ptr %19, align 8
%2582 = getelementptr inbounds %struct.cmplx, ptr %54, i32 0, i32 1
%2583 = load double, ptr %2582, align 8
%2584 = fmul double %2581, %2583
%2585 = fneg double %2584
%2586 = call double @llvm.fmuladd.f64(double %2578, double %2580, double %2585)
%2587 = load double, ptr %15, align 8
%2588 = getelementptr inbounds %struct.cmplx, ptr %53, i32 0, i32 1
%2589 = load double, ptr %2588, align 8
%2590 = call double @llvm.fmuladd.f64(double %2587, double %2589, double %2586)
%2591 = load double, ptr %23, align 8
%2592 = getelementptr inbounds %struct.cmplx, ptr %52, i32 0, i32 1
%2593 = load double, ptr %2592, align 8
%2594 = call double @llvm.fmuladd.f64(double %2591, double %2593, double %2590)
%2595 = load double, ptr %17, align 8
%2596 = getelementptr inbounds %struct.cmplx, ptr %51, i32 0, i32 1
%2597 = load double, ptr %2596, align 8
%2598 = fneg double %2595
%2599 = call double @llvm.fmuladd.f64(double %2598, double %2597, double %2594)
%2600 = fneg double %2599
%2601 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 0
store double %2600, ptr %2601, align 8
%2602 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 0
%2603 = load double, ptr %2602, align 8
%2604 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 0
%2605 = load double, ptr %2604, align 8
%2606 = fadd double %2603, %2605
%2607 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 0
store double %2606, ptr %2607, align 8
%2608 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 1
%2609 = load double, ptr %2608, align 8
%2610 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 1
%2611 = load double, ptr %2610, align 8
%2612 = fadd double %2609, %2611
%2613 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 1
store double %2612, ptr %2613, align 8
%2614 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 0
%2615 = load double, ptr %2614, align 8
%2616 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 0
%2617 = load double, ptr %2616, align 8
%2618 = fsub double %2615, %2617
%2619 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 0
store double %2618, ptr %2619, align 8
%2620 = getelementptr inbounds %struct.cmplx, ptr %70, i32 0, i32 1
%2621 = load double, ptr %2620, align 8
%2622 = getelementptr inbounds %struct.cmplx, ptr %71, i32 0, i32 1
%2623 = load double, ptr %2622, align 8
%2624 = fsub double %2621, %2623
%2625 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 1
store double %2624, ptr %2625, align 8
%2626 = load ptr, ptr %11, align 4
%2627 = load i32, ptr %44, align 4
%2628 = sub i32 %2627, 1
%2629 = load i32, ptr %7, align 4
%2630 = sub i32 %2629, 1
%2631 = mul i32 3, %2630
%2632 = add i32 %2628, %2631
%2633 = getelementptr inbounds %struct.cmplx, ptr %2626, i32 %2632
%2634 = getelementptr inbounds %struct.cmplx, ptr %2633, i32 0, i32 0
%2635 = load double, ptr %2634, align 8
%2636 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 0
%2637 = load double, ptr %2636, align 8
%2638 = load i32, ptr %12, align 4
%2639 = sitofp i32 %2638 to double
%2640 = load ptr, ptr %11, align 4
%2641 = load i32, ptr %44, align 4
%2642 = sub i32 %2641, 1
%2643 = load i32, ptr %7, align 4
%2644 = sub i32 %2643, 1
%2645 = mul i32 3, %2644
%2646 = add i32 %2642, %2645
%2647 = getelementptr inbounds %struct.cmplx, ptr %2640, i32 %2646
%2648 = getelementptr inbounds %struct.cmplx, ptr %2647, i32 0, i32 1
%2649 = load double, ptr %2648, align 8
%2650 = fmul double %2639, %2649
%2651 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 1
%2652 = load double, ptr %2651, align 8
%2653 = fmul double %2650, %2652
%2654 = fneg double %2653
%2655 = call double @llvm.fmuladd.f64(double %2635, double %2637, double %2654)
%2656 = load ptr, ptr %10, align 4
%2657 = load i32, ptr %44, align 4
%2658 = load i32, ptr %7, align 4
%2659 = load i32, ptr %24, align 4
%2660 = load i32, ptr %8, align 4
%2661 = mul i32 %2660, 4
%2662 = add i32 %2659, %2661
%2663 = mul i32 %2658, %2662
%2664 = add i32 %2657, %2663
%2665 = getelementptr inbounds %struct.cmplx, ptr %2656, i32 %2664
%2666 = getelementptr inbounds %struct.cmplx, ptr %2665, i32 0, i32 0
store double %2655, ptr %2666, align 8
%2667 = load ptr, ptr %11, align 4
%2668 = load i32, ptr %44, align 4
%2669 = sub i32 %2668, 1
%2670 = load i32, ptr %7, align 4
%2671 = sub i32 %2670, 1
%2672 = mul i32 3, %2671
%2673 = add i32 %2669, %2672
%2674 = getelementptr inbounds %struct.cmplx, ptr %2667, i32 %2673
%2675 = getelementptr inbounds %struct.cmplx, ptr %2674, i32 0, i32 0
%2676 = load double, ptr %2675, align 8
%2677 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 1
%2678 = load double, ptr %2677, align 8
%2679 = load i32, ptr %12, align 4
%2680 = sitofp i32 %2679 to double
%2681 = load ptr, ptr %11, align 4
%2682 = load i32, ptr %44, align 4
%2683 = sub i32 %2682, 1
%2684 = load i32, ptr %7, align 4
%2685 = sub i32 %2684, 1
%2686 = mul i32 3, %2685
%2687 = add i32 %2683, %2686
%2688 = getelementptr inbounds %struct.cmplx, ptr %2681, i32 %2687
%2689 = getelementptr inbounds %struct.cmplx, ptr %2688, i32 0, i32 1
%2690 = load double, ptr %2689, align 8
%2691 = fmul double %2680, %2690
%2692 = getelementptr inbounds %struct.cmplx, ptr %68, i32 0, i32 0
%2693 = load double, ptr %2692, align 8
%2694 = fmul double %2691, %2693
%2695 = call double @llvm.fmuladd.f64(double %2676, double %2678, double %2694)
%2696 = load ptr, ptr %10, align 4
%2697 = load i32, ptr %44, align 4
%2698 = load i32, ptr %7, align 4
%2699 = load i32, ptr %24, align 4
%2700 = load i32, ptr %8, align 4
%2701 = mul i32 %2700, 4
%2702 = add i32 %2699, %2701
%2703 = mul i32 %2698, %2702
%2704 = add i32 %2697, %2703
%2705 = getelementptr inbounds %struct.cmplx, ptr %2696, i32 %2704
%2706 = getelementptr inbounds %struct.cmplx, ptr %2705, i32 0, i32 1
store double %2695, ptr %2706, align 8
%2707 = load ptr, ptr %11, align 4
%2708 = load i32, ptr %44, align 4
%2709 = sub i32 %2708, 1
%2710 = load i32, ptr %7, align 4
%2711 = sub i32 %2710, 1
%2712 = mul i32 6, %2711
%2713 = add i32 %2709, %2712
%2714 = getelementptr inbounds %struct.cmplx, ptr %2707, i32 %2713
%2715 = getelementptr inbounds %struct.cmplx, ptr %2714, i32 0, i32 0
%2716 = load double, ptr %2715, align 8
%2717 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 0
%2718 = load double, ptr %2717, align 8
%2719 = load i32, ptr %12, align 4
%2720 = sitofp i32 %2719 to double
%2721 = load ptr, ptr %11, align 4
%2722 = load i32, ptr %44, align 4
%2723 = sub i32 %2722, 1
%2724 = load i32, ptr %7, align 4
%2725 = sub i32 %2724, 1
%2726 = mul i32 6, %2725
%2727 = add i32 %2723, %2726
%2728 = getelementptr inbounds %struct.cmplx, ptr %2721, i32 %2727
%2729 = getelementptr inbounds %struct.cmplx, ptr %2728, i32 0, i32 1
%2730 = load double, ptr %2729, align 8
%2731 = fmul double %2720, %2730
%2732 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 1
%2733 = load double, ptr %2732, align 8
%2734 = fmul double %2731, %2733
%2735 = fneg double %2734
%2736 = call double @llvm.fmuladd.f64(double %2716, double %2718, double %2735)
%2737 = load ptr, ptr %10, align 4
%2738 = load i32, ptr %44, align 4
%2739 = load i32, ptr %7, align 4
%2740 = load i32, ptr %24, align 4
%2741 = load i32, ptr %8, align 4
%2742 = mul i32 %2741, 7
%2743 = add i32 %2740, %2742
%2744 = mul i32 %2739, %2743
%2745 = add i32 %2738, %2744
%2746 = getelementptr inbounds %struct.cmplx, ptr %2737, i32 %2745
%2747 = getelementptr inbounds %struct.cmplx, ptr %2746, i32 0, i32 0
store double %2736, ptr %2747, align 8
%2748 = load ptr, ptr %11, align 4
%2749 = load i32, ptr %44, align 4
%2750 = sub i32 %2749, 1
%2751 = load i32, ptr %7, align 4
%2752 = sub i32 %2751, 1
%2753 = mul i32 6, %2752
%2754 = add i32 %2750, %2753
%2755 = getelementptr inbounds %struct.cmplx, ptr %2748, i32 %2754
%2756 = getelementptr inbounds %struct.cmplx, ptr %2755, i32 0, i32 0
%2757 = load double, ptr %2756, align 8
%2758 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 1
%2759 = load double, ptr %2758, align 8
%2760 = load i32, ptr %12, align 4
%2761 = sitofp i32 %2760 to double
%2762 = load ptr, ptr %11, align 4
%2763 = load i32, ptr %44, align 4
%2764 = sub i32 %2763, 1
%2765 = load i32, ptr %7, align 4
%2766 = sub i32 %2765, 1
%2767 = mul i32 6, %2766
%2768 = add i32 %2764, %2767
%2769 = getelementptr inbounds %struct.cmplx, ptr %2762, i32 %2768
%2770 = getelementptr inbounds %struct.cmplx, ptr %2769, i32 0, i32 1
%2771 = load double, ptr %2770, align 8
%2772 = fmul double %2761, %2771
%2773 = getelementptr inbounds %struct.cmplx, ptr %69, i32 0, i32 0
%2774 = load double, ptr %2773, align 8
%2775 = fmul double %2772, %2774
%2776 = call double @llvm.fmuladd.f64(double %2757, double %2759, double %2775)
%2777 = load ptr, ptr %10, align 4
%2778 = load i32, ptr %44, align 4
%2779 = load i32, ptr %7, align 4
%2780 = load i32, ptr %24, align 4
%2781 = load i32, ptr %8, align 4
%2782 = mul i32 %2781, 7
%2783 = add i32 %2780, %2782
%2784 = mul i32 %2779, %2783
%2785 = add i32 %2778, %2784
%2786 = getelementptr inbounds %struct.cmplx, ptr %2777, i32 %2785
%2787 = getelementptr inbounds %struct.cmplx, ptr %2786, i32 0, i32 1
store double %2776, ptr %2787, align 8
br label %2788
2788: ; preds = %1168
%2789 = load i32, ptr %44, align 4
%2790 = add i32 %2789, 1
store i32 %2790, ptr %44, align 4
br label %1164, !llvm.loop !4
2791: ; preds = %1164
br label %2792
2792: ; preds = %2791
%2793 = load i32, ptr %24, align 4
%2794 = add i32 %2793, 1
store i32 %2794, ptr %24, align 4
br label %87, !llvm.loop !6
2795: ; preds = %87
ret void
}
; Function Attrs: nocallback nofree nounwind willreturn memory(argmem: readwrite)
declare void @llvm.memcpy.p0.p0.i32(ptr noalias nocapture writeonly, ptr noalias nocapture readonly, i32, i1 immarg) #2
; Function Attrs: nocallback nofree nosync nounwind speculatable willreturn memory(none)
declare double @llvm.fmuladd.f64(double, double, double) #3
attributes #0 = { noinline nounwind optnone uwtable "frame-pointer"="all" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="e500" "target-features"="+spe,-altivec,-bpermd,-crbits,-crypto,-direct-move,-extdiv,-htm,-isa-v206-instructions,-isa-v207-instructions,-isa-v30-instructions,-power8-vector,-power9-vector,-privileged,-quadword-atomics,-rop-protect,-vsx" }
attributes #1 = { "frame-pointer"="all" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="e500" "target-features"="+spe,-altivec,-bpermd,-crbits,-crypto,-direct-move,-extdiv,-htm,-isa-v206-instructions,-isa-v207-instructions,-isa-v30-instructions,-power8-vector,-power9-vector,-privileged,-quadword-atomics,-rop-protect,-vsx" }
attributes #2 = { nocallback nofree nounwind willreturn memory(argmem: readwrite) }
attributes #3 = { nocallback nofree nosync nounwind speculatable willreturn memory(none) }
!llvm.module.flags = !{!0, !1, !2}
!llvm.ident = !{!3}
!0 = !{i32 1, !"wchar_size", i32 4}
!1 = !{i32 7, !"uwtable", i32 2}
!2 = !{i32 7, !"frame-pointer", i32 2}
!3 = !{!"clang version 17.0.0 (https://github.com/llvm/llvm-project.git 69db592f762ade86508826a7b3c9d5434c4837e2)"}
!4 = distinct !{!4, !5}
!5 = !{!"llvm.loop.mustprogress"}
!6 = distinct !{!6, !5}
;; NOTE: These prefixes are unused and the list is autogenerated. Do not add tests below this line:
; SPE: {{.*}}