// RUN: mlir-opt %s -affine-loop-invariant-code-motion -split-input-file | FileCheck %s func.func @nested_loops_both_having_invariant_code() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 affine.for %arg1 = 0 to 10 { affine.store %v0, %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst_0]] : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.store return } // ----- // The store-load forwarding can see through affine apply's since it relies on // dependence information. // CHECK-LABEL: func @store_affine_apply func.func @store_affine_apply() -> memref<10xf32> { %cf7 = arith.constant 7.0 : f32 %m = memref.alloc() : memref<10xf32> affine.for %arg0 = 0 to 10 { %t0 = affine.apply affine_map<(d1) -> (d1 + 1)>(%arg0) affine.store %cf7, %m[%t0] : memref<10xf32> } return %m : memref<10xf32> // CHECK: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[VAR_0:.*]] = memref.alloc() : memref<10xf32> // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.apply // CHECK-NEXT: affine.store %[[cst]] // CHECK-NEXT: } // CHECK-NEXT: return %[[VAR_0]] : memref<10xf32> } // ----- func.func @nested_loops_code_invariant_to_both() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst_0]] : f32 return } // ----- // CHECK-LABEL: func @nested_loops_inner_loops_invariant_to_outermost_loop func.func @nested_loops_inner_loops_invariant_to_outermost_loop(%m : memref<10xindex>) { affine.for %arg0 = 0 to 20 { affine.for %arg1 = 0 to 30 { %v0 = affine.for %arg2 = 0 to 10 iter_args (%prevAccum = %arg1) -> index { %v1 = affine.load %m[%arg2] : memref<10xindex> %newAccum = arith.addi %prevAccum, %v1 : index affine.yield %newAccum : index } } } // CHECK: affine.for %{{.*}} = 0 to 30 { // CHECK-NEXT: %{{.*}} = affine.for %{{.*}} = 0 to 10 iter_args(%{{.*}} = %{{.*}}) -> (index) { // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}} : memref<10xindex> // CHECK-NEXT: %{{.*}} = arith.addi %{{.*}}, %{{.*}} : index // CHECK-NEXT: affine.yield %{{.*}} : index // CHECK-NEXT: } // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 20 { // CHECK-NEXT: } return } // ----- func.func @single_loop_nothing_invariant() { %m1 = memref.alloc() : memref<10xf32> %m2 = memref.alloc() : memref<11xf32> affine.for %arg0 = 0 to 10 { %v0 = affine.load %m1[%arg0] : memref<10xf32> %v1 = affine.load %m2[%arg0] : memref<11xf32> %v2 = arith.addf %v0, %v1 : f32 affine.store %v2, %m1[%arg0] : memref<10xf32> } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: memref.alloc() : memref<11xf32> // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.load %{{.*}} : memref<10xf32> // CHECK-NEXT: affine.load %{{.*}} : memref<11xf32> // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.store %{{.*}} : memref<10xf32> return } // ----- func.func @invariant_code_inside_affine_if() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %t0 = affine.apply affine_map<(d1) -> (d1 + 1)>(%arg0) affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %t0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.apply #map{{[0-9]*}}(%arg0) // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store // CHECK-NEXT: } return } // ----- func.func @dependent_stores() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 affine.for %arg1 = 0 to 10 { %v1 = arith.mulf %cf7, %cf7 : f32 affine.store %v1, %m[%arg1] : memref<10xf32> affine.store %v0, %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst_0]] : f32 // CHECK-NEXT: %[[mul:.*]] = arith.mulf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.store %[[mul]] // CHECK-NEXT: affine.store return } // ----- func.func @independent_stores() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 affine.for %arg1 = 0 to 10 { %v1 = arith.mulf %cf7, %cf7 : f32 affine.store %v0, %m[%arg0] : memref<10xf32> affine.store %v1, %m[%arg1] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: %[[add:.*]] = arith.addf %[[cst]], %[[cst_0]] : f32 // CHECK-NEXT: %[[mul:.*]] = arith.mulf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.store %[[add]] // CHECK-NEXT: affine.store %[[mul]] // CHECK-NEXT: } return } // ----- func.func @load_dependent_store() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 affine.for %arg1 = 0 to 10 { %v1 = arith.addf %cf7, %cf7 : f32 affine.store %v0, %m[%arg1] : memref<10xf32> %v2 = affine.load %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst_0]] : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.for // CHECK-NEXT: affine.store // CHECK-NEXT: affine.load return } // ----- func.func @load_after_load() { %m = memref.alloc() : memref<10xf32> %cf7 = arith.constant 7.0 : f32 %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { %v0 = arith.addf %cf7, %cf8 : f32 affine.for %arg1 = 0 to 10 { %v1 = arith.addf %cf7, %cf7 : f32 %v3 = affine.load %m[%arg1] : memref<10xf32> %v2 = affine.load %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 7.000000e+00 : f32 // CHECK-NEXT: %[[cst_0:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst_0]] : f32 // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.load // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}}] : memref<10xf32> return } // ----- func.func @invariant_affine_if() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg0] : memref<10xf32> } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store // CHECK-NEXT: } return } // ----- func.func @invariant_affine_if2() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg1] : memref<10xf32> } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_affine_nested_if() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg0] : memref<10xf32> affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { affine.store %cf9, %m[%arg1] : memref<10xf32> } } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %[[arg0:.*]] = 0 to 10 { // CHECK-NEXT: affine.for %[[arg1:.*]] = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: affine.if // CHECK-NEXT: affine.store {{.*}}[%[[arg1]]] : memref<10xf32> // CHECK-NEXT: } // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_affine_nested_if_else() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg0] : memref<10xf32> affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { affine.store %cf9, %m[%arg0] : memref<10xf32> } else { affine.store %cf9, %m[%arg1] : memref<10xf32> } } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %[[arg0:.*]] = 0 to 10 { // CHECK-NEXT: affine.for %[[arg1:.*]] = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: affine.if // CHECK-NEXT: affine.store {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: } else { // CHECK-NEXT: affine.store {{.*}}[%[[arg1]]] : memref<10xf32> // CHECK-NEXT: } // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_affine_nested_if_else2() { %m = memref.alloc() : memref<10xf32> %m2 = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 %tload1 = affine.load %m[%arg0] : memref<10xf32> affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { affine.store %cf9, %m2[%arg0] : memref<10xf32> } else { %tload2 = affine.load %m[%arg0] : memref<10xf32> } } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %[[arg0:.*]] = 0 to 10 { // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.load {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: affine.if // CHECK-NEXT: affine.store {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: } else { // CHECK-NEXT: affine.load {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_affine_nested_if2() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 %v1 = affine.load %m[%arg0] : memref<10xf32> affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %v2 = affine.load %m[%arg0] : memref<10xf32> } } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: } // CHECK-NEXT: affine.for %[[arg0:.*]] = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.load {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: affine.if // CHECK-NEXT: affine.load {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_affine_for_inside_affine_if() { %m = memref.alloc() : memref<10xf32> %cf8 = arith.constant 8.0 : f32 affine.for %arg0 = 0 to 10 { affine.for %arg1 = 0 to 10 { affine.if affine_set<(d0, d1) : (d1 - d0 >= 0)> (%arg0, %arg0) { %cf9 = arith.addf %cf8, %cf8 : f32 affine.store %cf9, %m[%arg0] : memref<10xf32> affine.for %arg2 = 0 to 10 { affine.store %cf9, %m[%arg2] : memref<10xf32> } } } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: affine.for %[[arg0:.*]] = 0 to 10 { // CHECK-NEXT: affine.for %[[arg1:.*]] = 0 to 10 { // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf %[[cst]], %[[cst]] : f32 // CHECK-NEXT: affine.store {{.*}}[%[[arg0]]] : memref<10xf32> // CHECK-NEXT: affine.for %[[arg2:.*]] = 0 to 10 { // CHECK-NEXT: affine.store {{.*}}[%[[arg2]]] : memref<10xf32> // CHECK-NEXT: } // CHECK-NEXT: } // CHECK-NEXT: } return } // ----- func.func @invariant_constant_and_load() { %m = memref.alloc() : memref<100xf32> %m2 = memref.alloc() : memref<100xf32> affine.for %arg0 = 0 to 5 { %c0 = arith.constant 0 : index %v = affine.load %m2[%c0] : memref<100xf32> affine.store %v, %m[%arg0] : memref<100xf32> } // CHECK: memref.alloc() : memref<100xf32> // CHECK-NEXT: memref.alloc() : memref<100xf32> // CHECK-NEXT: arith.constant 0 : index // CHECK-NEXT: affine.load // CHECK-NEXT: affine.for %{{.*}} = 0 to 5 { // CHECK-NEXT: affine.store return } // ----- func.func @nested_load_store_same_memref() { %m = memref.alloc() : memref<10xf32> %cst = arith.constant 8.0 : f32 %c0 = arith.constant 0 : index affine.for %arg0 = 0 to 10 { %v0 = affine.load %m[%c0] : memref<10xf32> affine.for %arg1 = 0 to 10 { affine.store %cst, %m[%arg1] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.constant 0 : index // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.load // CHECK-NEXT: affine.for // CHECK-NEXT: affine.store %[[cst]] return } // ----- func.func @nested_load_store_same_memref2() { %m = memref.alloc() : memref<10xf32> %cst = arith.constant 8.0 : f32 %c0 = arith.constant 0 : index affine.for %arg0 = 0 to 10 { affine.store %cst, %m[%c0] : memref<10xf32> affine.for %arg1 = 0 to 10 { %v0 = affine.load %m[%arg0] : memref<10xf32> } } // CHECK: memref.alloc() : memref<10xf32> // CHECK-NEXT: %[[cst:.*]] = arith.constant 8.000000e+00 : f32 // CHECK-NEXT: arith.constant 0 : index // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: } // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.store %[[cst]] // CHECK-NEXT: affine.load return } // ----- // CHECK-LABEL: func @do_not_hoist_dependent_side_effect_free_op func.func @do_not_hoist_dependent_side_effect_free_op(%arg0: memref<10x512xf32>) { %0 = memref.alloca() : memref<1xf32> %cst = arith.constant 8.0 : f32 affine.for %i = 0 to 512 { affine.for %j = 0 to 10 { %5 = affine.load %arg0[%i, %j] : memref<10x512xf32> %6 = affine.load %0[0] : memref<1xf32> %add = arith.addf %5, %6 : f32 affine.store %add, %0[0] : memref<1xf32> } %3 = affine.load %0[0] : memref<1xf32> %4 = arith.mulf %3, %cst : f32 // It shouldn't be hoisted. } return } // CHECK: affine.for // CHECK-NEXT: affine.for // CHECK-NEXT: affine.load // CHECK-NEXT: affine.load // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.store // CHECK-NEXT: } // CHECK-NEXT: affine.load // CHECK-NEXT: arith.mulf // CHECK-NEXT: } // ----- // CHECK-LABEL: func @vector_loop_nothing_invariant func.func @vector_loop_nothing_invariant() { %m1 = memref.alloc() : memref<40xf32> %m2 = memref.alloc() : memref<40xf32> affine.for %arg0 = 0 to 10 { %v0 = affine.vector_load %m1[%arg0*4] : memref<40xf32>, vector<4xf32> %v1 = affine.vector_load %m2[%arg0*4] : memref<40xf32>, vector<4xf32> %v2 = arith.addf %v0, %v1 : vector<4xf32> affine.vector_store %v2, %m1[%arg0*4] : memref<40xf32>, vector<4xf32> } return } // CHECK: affine.for // CHECK-NEXT: affine.vector_load // CHECK-NEXT: affine.vector_load // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.vector_store // CHECK-NEXT: } // ----- // CHECK-LABEL: func @vector_loop_all_invariant func.func @vector_loop_all_invariant() { %m1 = memref.alloc() : memref<4xf32> %m2 = memref.alloc() : memref<4xf32> %m3 = memref.alloc() : memref<4xf32> affine.for %arg0 = 0 to 10 { %v0 = affine.vector_load %m1[0] : memref<4xf32>, vector<4xf32> %v1 = affine.vector_load %m2[0] : memref<4xf32>, vector<4xf32> %v2 = arith.addf %v0, %v1 : vector<4xf32> affine.vector_store %v2, %m3[0] : memref<4xf32>, vector<4xf32> } return } // CHECK: memref.alloc() // CHECK-NEXT: memref.alloc() // CHECK-NEXT: memref.alloc() // CHECK-NEXT: affine.vector_load // CHECK-NEXT: affine.vector_load // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.vector_store // CHECK-NEXT: affine.for // ----- #set = affine_set<(d0): (d0 - 10 >= 0)> // CHECK-LABEL: func @affine_if_not_invariant( func.func @affine_if_not_invariant(%buffer: memref<1024xf32>) -> f32 { %sum_init_0 = arith.constant 0.0 : f32 %sum_init_1 = arith.constant 1.0 : f32 %res = affine.for %i = 0 to 10 step 2 iter_args(%sum_iter = %sum_init_0) -> f32 { %t = affine.load %buffer[%i] : memref<1024xf32> %sum_next = affine.if #set(%i) -> (f32) { %new_sum = arith.addf %sum_iter, %t : f32 affine.yield %new_sum : f32 } else { affine.yield %sum_iter : f32 } %modified_sum = arith.addf %sum_next, %sum_init_1 : f32 affine.yield %modified_sum : f32 } return %res : f32 } // CHECK: arith.constant 0.000000e+00 : f32 // CHECK-NEXT: arith.constant 1.000000e+00 : f32 // CHECK-NEXT: affine.for // CHECK-NEXT: affine.load // CHECK-NEXT: affine.if // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.yield // CHECK-NEXT: } else { // CHECK-NEXT: affine.yield // CHECK-NEXT: } // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.yield // CHECK-NEXT: } // ----- // CHECK-LABEL: func @affine_for_not_invariant( func.func @affine_for_not_invariant(%in : memref<30x512xf32, 1>, %out : memref<30x1xf32, 1>) { %sum_0 = arith.constant 0.0 : f32 %cst_0 = arith.constant 1.1 : f32 affine.for %j = 0 to 30 { %sum = affine.for %i = 0 to 512 iter_args(%sum_iter = %sum_0) -> (f32) { %t = affine.load %in[%j,%i] : memref<30x512xf32,1> %sum_next = arith.addf %sum_iter, %t : f32 affine.yield %sum_next : f32 } %mod_sum = arith.mulf %sum, %cst_0 : f32 affine.store %mod_sum, %out[%j, 0] : memref<30x1xf32, 1> } return } // CHECK: arith.constant 0.000000e+00 : f32 // CHECK-NEXT: arith.constant 1.100000e+00 : f32 // CHECK-NEXT: affine.for // CHECK-NEXT: affine.for // CHECK-NEXT: affine.load // CHECK-NEXT: arith.addf // CHECK-NEXT: affine.yield // CHECK-NEXT: } // CHECK-NEXT: arith.mulf // CHECK-NEXT: affine.store // ----- // CHECK-LABEL: func @use_of_iter_operands_invariant func.func @use_of_iter_operands_invariant(%m : memref<10xindex>) { %sum_1 = arith.constant 0 : index %v0 = affine.for %arg1 = 0 to 11 iter_args (%prevAccum = %sum_1) -> index { %prod = arith.muli %sum_1, %sum_1 : index %newAccum = arith.addi %prevAccum, %prod : index affine.yield %newAccum : index } return } // CHECK: constant // CHECK-NEXT: muli // CHECK-NEXT: affine.for // CHECK-NEXT: addi // CHECK-NEXT: affine.yield // ----- // CHECK-LABEL: func @use_of_iter_args_not_invariant func.func @use_of_iter_args_not_invariant(%m : memref<10xindex>) { %sum_1 = arith.constant 0 : index %v0 = affine.for %arg1 = 0 to 11 iter_args (%prevAccum = %sum_1) -> index { %newAccum = arith.addi %prevAccum, %sum_1 : index affine.yield %newAccum : index } return } // CHECK: arith.constant // CHECK-NEXT: affine.for // CHECK-NEXT: arith.addi // CHECK-NEXT: affine.yield #map = affine_map<(d0) -> (64, d0 * -64 + 1020)> // CHECK-LABEL: func.func @affine_parallel func.func @affine_parallel(%memref_8: memref<4090x2040xf32>, %x: index) { %cst = arith.constant 0.000000e+00 : f32 affine.parallel (%arg3) = (0) to (32) { affine.for %arg4 = 0 to 16 { affine.parallel (%arg5, %arg6) = (0, 0) to (min(128, 122), min(64, %arg3 * -64 + 2040)) { affine.for %arg7 = 0 to min #map(%arg4) { affine.store %cst, %memref_8[%arg5 + 3968, %arg6 + %arg3 * 64] : memref<4090x2040xf32> } } } } // CHECK: affine.parallel // CHECK-NEXT: affine.for // CHECK-NEXT: affine.parallel // CHECK-NEXT: affine.store // CHECK-NEXT: affine.for %c0 = arith.constant 0 : index %c1 = arith.constant 1 : index %c32 = arith.constant 32 : index scf.parallel (%arg3) = (%c0) to (%c32) step (%c1) { affine.for %arg4 = 0 to 16 { affine.parallel (%arg5, %arg6) = (0, 0) to (min(128, 122), min(64, %x * -64 + 2040)) { affine.for %arg7 = 0 to min #map(%arg4) { affine.store %cst, %memref_8[%arg5 + 3968, %arg6] : memref<4090x2040xf32> } } } } // CHECK: scf.parallel // CHECK-NEXT: affine.for // CHECK-NEXT: affine.parallel // CHECK-NEXT: affine.store // CHECK-NEXT: affine.for affine.for %arg3 = 0 to 32 { affine.for %arg4 = 0 to 16 { affine.parallel (%arg5, %arg6) = (0, 0) to (min(128, 122), min(64, %arg3 * -64 + 2040)) { // Unknown region-holding op for this pass. scf.for %arg7 = %c0 to %x step %c1 { affine.store %cst, %memref_8[%arg5 + 3968, %arg6 + %arg3 * 64] : memref<4090x2040xf32> } } } } // CHECK: affine.for // CHECK-NEXT: affine.for // CHECK-NEXT: affine.parallel // CHECK-NEXT: scf.for // CHECK-NEXT: affine.store return } // ----- // CHECK-LABEL: func.func @affine_invariant_use_after_dma #map = affine_map<(d0) -> (d0 * 163840)> func.func @affine_invariant_use_after_dma(%arg0: memref<10485760xi32>, %arg1: memref<1xi32>, %arg2: memref<10485760xi32>) { %c320 = arith.constant 320 : index %c0 = arith.constant 0 : index %c1 = arith.constant 1 : index %alloc = memref.alloc() {alignment = 16 : i64} : memref<0xi32, 2> %alloc_0 = memref.alloc() : memref<1xi32, 2> affine.for %arg3 = 0 to 64 { %0 = affine.apply #map(%arg3) %alloc_1 = memref.alloc() {alignment = 16 : i64} : memref<0xi32, 2> %alloc_2 = memref.alloc() : memref<320xi32, 2> affine.dma_start %arg0[%0], %alloc_2[%c0], %alloc_1[%c0], %c320 : memref<10485760xi32>, memref<320xi32, 2>, memref<0xi32, 2> affine.dma_start %arg1[%c0], %alloc_0[%c0], %alloc[%c0], %c1 : memref<1xi32>, memref<1xi32, 2>, memref<0xi32, 2> affine.dma_wait %alloc_1[%c0], %c320 : memref<0xi32, 2> affine.dma_wait %alloc[%c0], %c1 : memref<0xi32, 2> %1 = affine.apply #map(%arg3) %alloc_3 = memref.alloc() {alignment = 16 : i64} : memref<0xi32, 2> %alloc_4 = memref.alloc() : memref<320xi32, 2> affine.for %arg4 = 0 to 320 { %2 = affine.load %alloc_2[%arg4] : memref<320xi32, 2> %3 = affine.load %alloc_0[0] : memref<1xi32, 2> %4 = arith.addi %2, %3 : i32 %5 = arith.addi %4, %2 : i32 affine.store %5, %alloc_4[%arg4] : memref<320xi32, 2> } affine.dma_start %alloc_4[%c0], %arg2[%1], %alloc_3[%c0], %c320 : memref<320xi32, 2>, memref<10485760xi32>, memref<0xi32, 2> affine.dma_wait %alloc_3[%c0], %c320 : memref<0xi32, 2> } return } // CHECK: %[[zero:.*]] = arith.constant 0 : index // CHECK: %[[scalar_mem:.*]] = memref.alloc() : memref<1xi32, 2> // CHECK: affine.dma_start %arg1[%[[zero]]], %alloc_0[%[[zero]]], %alloc[%[[zero]]], %c1 // CHECK: affine.load %[[scalar_mem]][0] // ----- // CHECK-LABEL: func @affine_prefetch_invariant func.func @affine_prefetch_invariant() { %0 = memref.alloc() : memref<10x10xf32> affine.for %i0 = 0 to 10 { affine.for %i1 = 0 to 10 { %1 = affine.load %0[%i0, %i1] : memref<10x10xf32> affine.prefetch %0[%i0, %i0], write, locality<0>, data : memref<10x10xf32> } } // CHECK: memref.alloc() : memref<10x10xf32> // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: affine.prefetch // CHECK-NEXT: affine.for %{{.*}} = 0 to 10 { // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}} : memref<10x10xf32> // CHECK-NEXT: } // CHECK-NEXT: } return } // Side-effecting ops shouldn't be hoisted. // CHECK-LABEL: func @side_effecting_ops func.func @side_effecting_ops() { %cst = arith.constant 0.0 : f32 %m0 = memref.alloc(): memref<1x512x16x16xf32> %0 = gpu.wait async affine.for %arg783 = 0 to 14 { affine.for %arg784 = 0 to 14 { affine.parallel (%arg785) = (0) to (512) { affine.for %arg786 = 0 to 1 { affine.for %arg787 = 0 to 1 { affine.for %arg788 = 0 to 1 { %m1 = memref.alloc() : memref<1xf32, 3> %m2 = memref.alloc() : memref<1xf32, 3> affine.store %cst, %m1[0] : memref<1xf32, 3> affine.store %cst, %m2[0] : memref<1xf32, 3> %memref_2897, %asyncToken_2898 = gpu.alloc async [%0] () : memref<1x512x16x16xf32> %2432 = gpu.memcpy async [%0] %memref_2897, %m0 : memref<1x512x16x16xf32>, memref<1x512x16x16xf32> affine.for %arg789 = 0 to 16 { affine.for %arg790 = 0 to 16 { affine.store %cst, %memref_2897[0, %arg785 + %arg788, %arg789, %arg790] : memref<1x512x16x16xf32> } } memref.dealloc %m2 : memref<1xf32, 3> memref.dealloc %m1 : memref<1xf32, 3> %2433 = gpu.memcpy async [%0] %m0, %memref_2897 : memref<1x512x16x16xf32>, memref<1x512x16x16xf32> %2434 = gpu.dealloc async [%asyncToken_2898] %memref_2897 : memref<1x512x16x16xf32> } } } } } } // CHECK: affine.for %{{.*}} = 0 to 1 // CHECK-NEXT: affine.for %{{.*}} = 0 to 1 // CHECK: memref.alloc // CHECK: memref.alloc // CHECK: gpu.memcpy // CHECK: affine.for %{{.*}} = 0 to 16 // CHECK: affine.for %{{.*}} = 0 to 16 // CHECK: memref.dealloc return }