// RUN: mlir-opt %s -split-input-file -async-parallel-for=async-dispatch=true \ // RUN: | FileCheck %s --dump-input=always // CHECK-LABEL: @loop_1d( // CHECK-SAME: %[[LB:.*]]: index, %[[UB:.*]]: index, %[[STEP:.*]]: index func.func @loop_1d(%arg0: index, %arg1: index, %arg2: index, %arg3: memref) { // CHECK: %[[C0:.*]] = arith.constant 0 : index // CHECK: %[[RANGE:.*]] = arith.subi %[[UB]], %[[LB]] // CHECK: %[[TRIP_CNT:.*]] = arith.ceildivsi %[[RANGE]], %[[STEP]] // CHECK: %[[IS_NOOP:.*]] = arith.cmpi eq, %[[TRIP_CNT]], %[[C0]] : index // CHECK: scf.if %[[IS_NOOP]] { // CHECK-NEXT: } else { // CHECK: scf.if {{.*}} { // CHECK: call @parallel_compute_fn(%[[C0]] // CHECK: } else { // CHECK: %[[GROUP:.*]] = async.create_group // CHECK: call @async_dispatch_fn // CHECK: async.await_all %[[GROUP]] // CHECK: } // CHECK: } scf.parallel (%i) = (%arg0) to (%arg1) step (%arg2) { %one = arith.constant 1.0 : f32 memref.store %one, %arg3[%i] : memref } return } // CHECK-LABEL: func private @parallel_compute_fn // CHECK: scf.for // CHECK: memref.store // CHECK-LABEL: func private @async_dispatch_fn // CHECK-SAME: ( // CHECK-SAME: %[[GROUP:arg0]]: !async.group, // CHECK-SAME: %[[BLOCK_START:arg1]]: index // CHECK-SAME: %[[BLOCK_END:arg2]]: index // CHECK-SAME: ) // CHECK: %[[C1:.*]] = arith.constant 1 : index // CHECK: %[[C2:.*]] = arith.constant 2 : index // CHECK: scf.while (%[[S0:.*]] = %[[BLOCK_START]], // CHECK-SAME: %[[E0:.*]] = %[[BLOCK_END]]) // While loop `before` block decides if we need to dispatch more tasks. // CHECK: { // CHECK: %[[DIFF0:.*]] = arith.subi %[[E0]], %[[S0]] // CHECK: %[[COND:.*]] = arith.cmpi sgt, %[[DIFF0]], %[[C1]] // CHECK: scf.condition(%[[COND]]) // While loop `after` block splits the range in half and submits async task // to process the second half using the call to the same dispatch function. // CHECK: } do { // CHECK: ^bb0(%[[S1:.*]]: index, %[[E1:.*]]: index): // CHECK: %[[DIFF1:.*]] = arith.subi %[[E1]], %[[S1]] // CHECK: %[[HALF:.*]] = arith.divsi %[[DIFF1]], %[[C2]] // CHECK: %[[MID:.*]] = arith.addi %[[S1]], %[[HALF]] // CHECK: %[[TOKEN:.*]] = async.execute // CHECK: call @async_dispatch_fn // CHECK: async.add_to_group // CHECK: scf.yield %[[S1]], %[[MID]] // CHECK: } // After async dispatch the first block processed in the caller thread. // CHECK: call @parallel_compute_fn(%[[BLOCK_START]] // ----- // CHECK-LABEL: @loop_2d func.func @loop_2d(%arg0: index, %arg1: index, %arg2: index, // lb, ub, step %arg3: index, %arg4: index, %arg5: index, // lb, ub, step %arg6: memref) { // CHECK: %[[GROUP:.*]] = async.create_group // CHECK: call @async_dispatch_fn // CHECK: async.await_all %[[GROUP]] scf.parallel (%i0, %i1) = (%arg0, %arg3) to (%arg1, %arg4) step (%arg2, %arg5) { %one = arith.constant 1.0 : f32 memref.store %one, %arg6[%i0, %i1] : memref } return } // CHECK-LABEL: func private @parallel_compute_fn // CHECK: scf.for // CHECK: scf.for // CHECK: memref.store // CHECK-LABEL: func private @async_dispatch_fn