// RUN: mlir-opt -convert-gpu-to-spirv %s -o - | FileCheck %s module attributes { gpu.container_module, spv.target_env = #spv.target_env< #spv.vce, {}> } { func @loop(%arg0 : memref<10xf32>, %arg1 : memref<10xf32>) { %c0 = constant 1 : index gpu.launch_func @kernels::@loop_kernel blocks in (%c0, %c0, %c0) threads in (%c0, %c0, %c0) args(%arg0 : memref<10xf32>, %arg1 : memref<10xf32>) return } gpu.module @kernels { gpu.func @loop_kernel(%arg2 : memref<10xf32>, %arg3 : memref<10xf32>) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { // CHECK: %[[LB:.*]] = spv.constant 4 : i32 %lb = constant 4 : index // CHECK: %[[UB:.*]] = spv.constant 42 : i32 %ub = constant 42 : index // CHECK: %[[STEP:.*]] = spv.constant 2 : i32 %step = constant 2 : index // CHECK: spv.loop { // CHECK-NEXT: spv.Branch ^[[HEADER:.*]](%[[LB]] : i32) // CHECK: ^[[HEADER]](%[[INDVAR:.*]]: i32): // CHECK: %[[CMP:.*]] = spv.SLessThan %[[INDVAR]], %[[UB]] : i32 // CHECK: spv.BranchConditional %[[CMP]], ^[[BODY:.*]], ^[[MERGE:.*]] // CHECK: ^[[BODY]]: // CHECK: %[[ZERO1:.*]] = spv.constant 0 : i32 // CHECK: %[[OFFSET1:.*]] = spv.constant 0 : i32 // CHECK: %[[STRIDE1:.*]] = spv.constant 1 : i32 // CHECK: %[[UPDATE1:.*]] = spv.IMul %[[STRIDE1]], %[[INDVAR]] : i32 // CHECK: %[[INDEX1:.*]] = spv.IAdd %[[OFFSET1]], %[[UPDATE1]] : i32 // CHECK: spv.AccessChain {{%.*}}{{\[}}%[[ZERO1]], %[[INDEX1]]{{\]}} // CHECK: %[[ZERO2:.*]] = spv.constant 0 : i32 // CHECK: %[[OFFSET2:.*]] = spv.constant 0 : i32 // CHECK: %[[STRIDE2:.*]] = spv.constant 1 : i32 // CHECK: %[[UPDATE2:.*]] = spv.IMul %[[STRIDE2]], %[[INDVAR]] : i32 // CHECK: %[[INDEX2:.*]] = spv.IAdd %[[OFFSET2]], %[[UPDATE2]] : i32 // CHECK: spv.AccessChain {{%.*}}[%[[ZERO2]], %[[INDEX2]]] // CHECK: %[[INCREMENT:.*]] = spv.IAdd %[[INDVAR]], %[[STEP]] : i32 // CHECK: spv.Branch ^[[HEADER]](%[[INCREMENT]] : i32) // CHECK: ^[[MERGE]] // CHECK: spv.mlir.merge // CHECK: } scf.for %arg4 = %lb to %ub step %step { %1 = load %arg2[%arg4] : memref<10xf32> store %1, %arg3[%arg4] : memref<10xf32> } gpu.return } // CHECK-LABEL: @loop_yield gpu.func @loop_yield(%arg2 : memref<10xf32>, %arg3 : memref<10xf32>) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { // CHECK: %[[LB:.*]] = spv.constant 4 : i32 %lb = constant 4 : index // CHECK: %[[UB:.*]] = spv.constant 42 : i32 %ub = constant 42 : index // CHECK: %[[STEP:.*]] = spv.constant 2 : i32 %step = constant 2 : index // CHECK: %[[INITVAR1:.*]] = spv.constant 0.000000e+00 : f32 %s0 = constant 0.0 : f32 // CHECK: %[[INITVAR2:.*]] = spv.constant 1.000000e+00 : f32 %s1 = constant 1.0 : f32 // CHECK: %[[VAR1:.*]] = spv.Variable : !spv.ptr // CHECK: %[[VAR2:.*]] = spv.Variable : !spv.ptr // CHECK: spv.loop { // CHECK: spv.Branch ^[[HEADER:.*]](%[[LB]], %[[INITVAR1]], %[[INITVAR2]] : i32, f32, f32) // CHECK: ^[[HEADER]](%[[INDVAR:.*]]: i32, %[[CARRIED1:.*]]: f32, %[[CARRIED2:.*]]: f32): // CHECK: %[[CMP:.*]] = spv.SLessThan %[[INDVAR]], %[[UB]] : i32 // CHECK: spv.BranchConditional %[[CMP]], ^[[BODY:.*]], ^[[MERGE:.*]] // CHECK: ^[[BODY]]: // CHECK: %[[UPDATED:.*]] = spv.FAdd %[[CARRIED1]], %[[CARRIED1]] : f32 // CHECK-DAG: %[[INCREMENT:.*]] = spv.IAdd %[[INDVAR]], %[[STEP]] : i32 // CHECK-DAG: spv.Store "Function" %[[VAR1]], %[[UPDATED]] : f32 // CHECK-DAG: spv.Store "Function" %[[VAR2]], %[[UPDATED]] : f32 // CHECK: spv.Branch ^[[HEADER]](%[[INCREMENT]], %[[UPDATED]], %[[UPDATED]] : i32, f32, f32) // CHECK: ^[[MERGE]]: // CHECK: spv.mlir.merge // CHECK: } %result:2 = scf.for %i0 = %lb to %ub step %step iter_args(%si = %s0, %sj = %s1) -> (f32, f32) { %sn = addf %si, %si : f32 scf.yield %sn, %sn : f32, f32 } // CHECK-DAG: %[[OUT1:.*]] = spv.Load "Function" %[[VAR1]] : f32 // CHECK-DAG: %[[OUT2:.*]] = spv.Load "Function" %[[VAR2]] : f32 // CHECK: spv.Store "StorageBuffer" {{%.*}}, %[[OUT1]] : f32 // CHECK: spv.Store "StorageBuffer" {{%.*}}, %[[OUT2]] : f32 store %result#0, %arg3[%lb] : memref<10xf32> store %result#1, %arg3[%ub] : memref<10xf32> gpu.return } } }