| // RUN: mlir-opt %s -split-input-file -async-parallel-for=num-workers=-1 \ |
| // RUN: | FileCheck %s --dump-input=always |
| |
| // CHECK-LABEL: @num_worker_threads( |
| // CHECK: %[[MEMREF:.*]]: memref<?xf32> |
| func.func @num_worker_threads(%arg0: memref<?xf32>) { |
| |
| // CHECK-DAG: %[[scalingCstInit:.*]] = arith.constant 8.000000e+00 : f32 |
| // CHECK-DAG: %[[bracketLowerBound4:.*]] = arith.constant 4 : index |
| // CHECK-DAG: %[[scalingCst4:.*]] = arith.constant 4.000000e+00 : f32 |
| // CHECK-DAG: %[[bracketLowerBound8:.*]] = arith.constant 8 : index |
| // CHECK-DAG: %[[scalingCst8:.*]] = arith.constant 2.000000e+00 : f32 |
| // CHECK-DAG: %[[bracketLowerBound16:.*]] = arith.constant 16 : index |
| // CHECK-DAG: %[[scalingCst16:.*]] = arith.constant 1.000000e+00 : f32 |
| // CHECK-DAG: %[[bracketLowerBound32:.*]] = arith.constant 32 : index |
| // CHECK-DAG: %[[scalingCst32:.*]] = arith.constant 8.000000e-01 : f32 |
| // CHECK-DAG: %[[bracketLowerBound64:.*]] = arith.constant 64 : index |
| // CHECK-DAG: %[[scalingCst64:.*]] = arith.constant 6.000000e-01 : f32 |
| // CHECK: %[[workersIndex:.*]] = async.runtime.num_worker_threads : index |
| // CHECK: %[[inBracket4:.*]] = arith.cmpi sgt, %[[workersIndex]], %[[bracketLowerBound4]] : index |
| // CHECK: %[[scalingFactor4:.*]] = arith.select %[[inBracket4]], %[[scalingCst4]], %[[scalingCstInit]] : f32 |
| // CHECK: %[[inBracket8:.*]] = arith.cmpi sgt, %[[workersIndex]], %[[bracketLowerBound8]] : index |
| // CHECK: %[[scalingFactor8:.*]] = arith.select %[[inBracket8]], %[[scalingCst8]], %[[scalingFactor4]] : f32 |
| // CHECK: %[[inBracket16:.*]] = arith.cmpi sgt, %[[workersIndex]], %[[bracketLowerBound16]] : index |
| // CHECK: %[[scalingFactor16:.*]] = arith.select %[[inBracket16]], %[[scalingCst16]], %[[scalingFactor8]] : f32 |
| // CHECK: %[[inBracket32:.*]] = arith.cmpi sgt, %[[workersIndex]], %[[bracketLowerBound32]] : index |
| // CHECK: %[[scalingFactor32:.*]] = arith.select %[[inBracket32]], %[[scalingCst32]], %[[scalingFactor16]] : f32 |
| // CHECK: %[[inBracket64:.*]] = arith.cmpi sgt, %[[workersIndex]], %[[bracketLowerBound64]] : index |
| // CHECK: %[[scalingFactor64:.*]] = arith.select %[[inBracket64]], %[[scalingCst64]], %[[scalingFactor32]] : f32 |
| // CHECK: %[[workersInt:.*]] = arith.index_cast %[[workersIndex]] : index to i32 |
| // CHECK: %[[workersFloat:.*]] = arith.sitofp %[[workersInt]] : i32 to f32 |
| // CHECK: %[[scaledFloat:.*]] = arith.mulf %[[scalingFactor64]], %[[workersFloat]] : f32 |
| // CHECK: %[[scaledInt:.*]] = arith.fptosi %[[scaledFloat]] : f32 to i32 |
| // CHECK: %[[scaledIndex:.*]] = arith.index_cast %[[scaledInt]] : i32 to index |
| |
| %lb = arith.constant 0 : index |
| %ub = arith.constant 100 : index |
| %st = arith.constant 1 : index |
| scf.parallel (%i) = (%lb) to (%ub) step (%st) { |
| %one = arith.constant 1.0 : f32 |
| memref.store %one, %arg0[%i] : memref<?xf32> |
| } |
| |
| return |
| } |