blob: 61af9a4baf46d22394fa4c6b1f77f291ba6cb35e [file] [log] [blame]
// RUN: mlir-opt -pass-pipeline='builtin.module(func.func(affine-loop-fusion{mode=producer fusion-maximal}))' %s | FileCheck %s
// Test fusion of affine nests inside other region-holding ops (scf.for in the
// test case below).
// CHECK-LABEL: func @fusion_inner_simple
func.func @fusion_inner_simple(%A : memref<10xf32>) {
%cst = arith.constant 0.0 : f32
affine.for %i = 0 to 100 {
%B = memref.alloc() : memref<10xf32>
%C = memref.alloc() : memref<10xf32>
affine.for %j = 0 to 10 {
%v = affine.load %A[%j] : memref<10xf32>
affine.store %v, %B[%j] : memref<10xf32>
}
affine.for %j = 0 to 10 {
%v = affine.load %B[%j] : memref<10xf32>
affine.store %v, %C[%j] : memref<10xf32>
}
}
// CHECK: affine.for %{{.*}} = 0 to 100
// CHECK-NEXT: memref.alloc
// CHECK-NEXT: memref.alloc
// CHECK-NEXT: affine.for %{{.*}} = 0 to 10
// CHECK-NOT: affine.for
return
}
// CHECK-LABEL: func @fusion_inner_simple_scf
func.func @fusion_inner_simple_scf(%A : memref<10xf32>) {
%c0 = arith.constant 0 : index
%c1 = arith.constant 1 : index
%c100 = arith.constant 100 : index
%cst = arith.constant 0.0 : f32
scf.for %i = %c0 to %c100 step %c1 {
%B = memref.alloc() : memref<10xf32>
%C = memref.alloc() : memref<10xf32>
affine.for %j = 0 to 10 {
%v = affine.load %A[%j] : memref<10xf32>
affine.store %v, %B[%j] : memref<10xf32>
}
affine.for %j = 0 to 10 {
%v = affine.load %B[%j] : memref<10xf32>
affine.store %v, %C[%j] : memref<10xf32>
}
}
// CHECK: scf.for
// CHECK-NEXT: memref.alloc
// CHECK-NEXT: memref.alloc
// CHECK-NEXT: affine.for %{{.*}} = 0 to 10
// CHECK-NOT: affine.for
return
}
// CHECK-LABEL: func @fusion_inner_multiple_nests
func.func @fusion_inner_multiple_nests() {
%alloc_5 = memref.alloc() {alignment = 64 : i64} : memref<4x4xi8>
%alloc_10 = memref.alloc() : memref<8x4xi32>
affine.for %arg8 = 0 to 4 {
%alloc_14 = memref.alloc() : memref<4xi8>
%alloc_15 = memref.alloc() : memref<8x4xi8>
affine.for %arg9 = 0 to 4 {
%0 = affine.load %alloc_5[%arg9, %arg8] : memref<4x4xi8>
affine.store %0, %alloc_14[%arg9] : memref<4xi8>
}
%alloc_16 = memref.alloc() : memref<4xi8>
affine.for %arg9 = 0 to 4 {
%0 = affine.load %alloc_14[%arg9] : memref<4xi8>
affine.store %0, %alloc_16[%arg9] : memref<4xi8>
}
affine.for %arg9 = 0 to 2 {
%0 = affine.load %alloc_15[%arg9 * 4, 0] : memref<8x4xi8>
%1 = affine.load %alloc_16[0] : memref<4xi8>
%2 = affine.load %alloc_10[%arg9 * 4, %arg8] : memref<8x4xi32>
%3 = arith.muli %0, %1 : i8
%4 = arith.extsi %3 : i8 to i32
%5 = arith.addi %4, %2 : i32
affine.store %5, %alloc_10[%arg9 * 4 + 3, %arg8] : memref<8x4xi32>
}
memref.dealloc %alloc_16 : memref<4xi8>
}
// CHECK: affine.for %{{.*}} = 0 to 4 {
// Everything inside fused into two nests (the second will be DCE'd).
// CHECK-NEXT: memref.alloc() : memref<4xi8>
// CHECK-NEXT: memref.alloc() : memref<1xi8>
// CHECK-NEXT: memref.alloc() : memref<1xi8>
// CHECK-NEXT: memref.alloc() : memref<8x4xi8>
// CHECK-NEXT: memref.alloc() : memref<4xi8>
// CHECK-NEXT: affine.for %{{.*}} = 0 to 2 {
// CHECK: }
// CHECK: affine.for %{{.*}} = 0 to 4 {
// CHECK: }
// CHECK-NEXT: memref.dealloc
// CHECK-NEXT: }
// CHECK-NEXT: return
return
}
// CHECK-LABEL: func @fusion_inside_scf_while
func.func @fusion_inside_scf_while(%A : memref<10xf32>) {
%c0 = arith.constant 0 : index
%c1 = arith.constant 1 : index
%c100 = arith.constant 100 : index
%cst = arith.constant 0.0 : f32
%0 = scf.while (%arg3 = %cst) : (f32) -> (f32) {
%1 = arith.cmpf ult, %arg3, %cst : f32
scf.condition(%1) %arg3 : f32
} do {
^bb0(%arg5: f32):
%B = memref.alloc() : memref<10xf32>
%C = memref.alloc() : memref<10xf32>
affine.for %j = 0 to 10 {
%v = affine.load %A[%j] : memref<10xf32>
affine.store %v, %B[%j] : memref<10xf32>
}
affine.for %j = 0 to 10 {
%v = affine.load %B[%j] : memref<10xf32>
affine.store %v, %C[%j] : memref<10xf32>
}
%1 = arith.mulf %arg5, %cst : f32
scf.yield %1 : f32
}
// CHECK: scf.while
// CHECK: affine.for %{{.*}} = 0 to 10
// CHECK-NOT: affine.for
// CHECK: scf.yield
return
}
memref.global "private" constant @__constant_10x2xf32 : memref<10x2xf32> = dense<0.000000e+00>
// CHECK-LABEL: func @fusion_inner_long
func.func @fusion_inner_long(%arg0: memref<10x2xf32>, %arg1: memref<10x10xf32>, %arg2: memref<10x2xf32>, %s: index) {
%c0 = arith.constant 0 : index
%cst_0 = arith.constant 1.000000e-03 : f32
%c9 = arith.constant 9 : index
%c10_i32 = arith.constant 10 : i32
%c1_i32 = arith.constant 1 : i32
%c100_i32 = arith.constant 100 : i32
%c0_i32 = arith.constant 0 : i32
%0 = memref.get_global @__constant_10x2xf32 : memref<10x2xf32>
%1 = scf.for %arg3 = %c0_i32 to %c100_i32 step %c1_i32 iter_args(%arg4 = %arg0) -> (memref<10x2xf32>) : i32 {
%alloc = memref.alloc() {alignment = 64 : i64} : memref<10xi32>
affine.for %arg5 = 0 to 10 {
%3 = arith.index_cast %arg5 : index to i32
affine.store %3, %alloc[%arg5] : memref<10xi32>
}
%2 = scf.for %arg5 = %c0_i32 to %c10_i32 step %c1_i32 iter_args(%arg6 = %0) -> (memref<10x2xf32>) : i32 {
%alloc_5 = memref.alloc() : memref<2xf32>
affine.for %arg7 = 0 to 2 {
%16 = affine.load %arg4[%s, %arg7] : memref<10x2xf32>
affine.store %16, %alloc_5[%arg7] : memref<2xf32>
}
%alloc_6 = memref.alloc() {alignment = 64 : i64} : memref<1x2xf32>
affine.for %arg7 = 0 to 2 {
%16 = affine.load %alloc_5[%arg7] : memref<2xf32>
affine.store %16, %alloc_6[0, %arg7] : memref<1x2xf32>
}
%alloc_7 = memref.alloc() {alignment = 64 : i64} : memref<10x2xf32>
affine.for %arg7 = 0 to 10 {
affine.for %arg8 = 0 to 2 {
%16 = affine.load %alloc_6[0, %arg8] : memref<1x2xf32>
affine.store %16, %alloc_7[%arg7, %arg8] : memref<10x2xf32>
}
}
%alloc_8 = memref.alloc() {alignment = 64 : i64} : memref<10x2xf32>
affine.for %arg7 = 0 to 10 {
affine.for %arg8 = 0 to 2 {
%16 = affine.load %alloc_7[%arg7, %arg8] : memref<10x2xf32>
%17 = affine.load %arg4[%arg7, %arg8] : memref<10x2xf32>
%18 = arith.subf %16, %17 : f32
affine.store %18, %alloc_8[%arg7, %arg8] : memref<10x2xf32>
}
}
scf.yield %alloc_8 : memref<10x2xf32>
// CHECK: scf.for
// CHECK: scf.for
// CHECK: affine.for %{{.*}} = 0 to 10
// CHECK-NEXT: affine.for %{{.*}} = 0 to 2
// CHECK-NOT: affine.for
// CHECK: scf.yield
}
%alloc_2 = memref.alloc() {alignment = 64 : i64} : memref<10x2xf32>
affine.for %arg5 = 0 to 10 {
affine.for %arg6 = 0 to 2 {
affine.store %cst_0, %alloc_2[%arg5, %arg6] : memref<10x2xf32>
}
}
%alloc_3 = memref.alloc() {alignment = 64 : i64} : memref<10x2xf32>
affine.for %arg5 = 0 to 10 {
affine.for %arg6 = 0 to 2 {
%3 = affine.load %alloc_2[%arg5, %arg6] : memref<10x2xf32>
%4 = affine.load %2[%arg5, %arg6] : memref<10x2xf32>
%5 = arith.mulf %3, %4 : f32
affine.store %5, %alloc_3[%arg5, %arg6] : memref<10x2xf32>
}
}
scf.yield %alloc_3 : memref<10x2xf32>
// The nests above will be fused as well.
// CHECK: affine.for %{{.*}} = 0 to 10
// CHECK-NEXT: affine.for %{{.*}} = 0 to 2
// CHECK-NOT: affine.for
// CHECK: scf.yield
}
affine.for %arg3 = 0 to 10 {
affine.for %arg4 = 0 to 2 {
%2 = affine.load %1[%arg3, %arg4] : memref<10x2xf32>
affine.store %2, %arg2[%arg3, %arg4] : memref<10x2xf32>
}
}
return
}