// RUN: mlir-opt -convert-gpu-to-spirv %s -o - | FileCheck %s module attributes { gpu.container_module, spv.target_env = #spv.target_env< #spv.vce, {}> } { func @main(%arg0 : memref<10xf32>, %arg1 : i1) { %c0 = constant 1 : index gpu.launch_func @kernels::@kernel_simple_selection blocks in (%c0, %c0, %c0) threads in (%c0, %c0, %c0) args(%arg0 : memref<10xf32>, %arg1 : i1) return } gpu.module @kernels { // CHECK-LABEL: @kernel_simple_selection gpu.func @kernel_simple_selection(%arg2 : memref<10xf32>, %arg3 : i1) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { %value = constant 0.0 : f32 %i = constant 0 : index // CHECK: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE:\^.*]], [[MERGE:\^.*]] // CHECK-NEXT: [[TRUE]]: // CHECK: spv.Branch [[MERGE]] // CHECK-NEXT: [[MERGE]]: // CHECK-NEXT: spv.mlir.merge // CHECK-NEXT: } // CHECK-NEXT: spv.Return scf.if %arg3 { store %value, %arg2[%i] : memref<10xf32> } gpu.return } // CHECK-LABEL: @kernel_nested_selection gpu.func @kernel_nested_selection(%arg3 : memref<10xf32>, %arg4 : memref<10xf32>, %arg5 : i1, %arg6 : i1) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { %i = constant 0 : index %j = constant 9 : index // CHECK: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE_TOP:\^.*]], [[FALSE_TOP:\^.*]] // CHECK-NEXT: [[TRUE_TOP]]: // CHECK-NEXT: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE_NESTED_TRUE_PATH:\^.*]], [[FALSE_NESTED_TRUE_PATH:\^.*]] // CHECK-NEXT: [[TRUE_NESTED_TRUE_PATH]]: // CHECK: spv.Branch [[MERGE_NESTED_TRUE_PATH:\^.*]] // CHECK-NEXT: [[FALSE_NESTED_TRUE_PATH]]: // CHECK: spv.Branch [[MERGE_NESTED_TRUE_PATH]] // CHECK-NEXT: [[MERGE_NESTED_TRUE_PATH]]: // CHECK-NEXT: spv.mlir.merge // CHECK-NEXT: } // CHECK-NEXT: spv.Branch [[MERGE_TOP:\^.*]] // CHECK-NEXT: [[FALSE_TOP]]: // CHECK-NEXT: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE_NESTED_FALSE_PATH:\^.*]], [[FALSE_NESTED_FALSE_PATH:\^.*]] // CHECK-NEXT: [[TRUE_NESTED_FALSE_PATH]]: // CHECK: spv.Branch [[MERGE_NESTED_FALSE_PATH:\^.*]] // CHECK-NEXT: [[FALSE_NESTED_FALSE_PATH]]: // CHECK: spv.Branch [[MERGE_NESTED_FALSE_PATH]] // CHECK: [[MERGE_NESTED_FALSE_PATH]]: // CHECK-NEXT: spv.mlir.merge // CHECK-NEXT: } // CHECK-NEXT: spv.Branch [[MERGE_TOP]] // CHECK-NEXT: [[MERGE_TOP]]: // CHECK-NEXT: spv.mlir.merge // CHECK-NEXT: } // CHECK-NEXT: spv.Return scf.if %arg5 { scf.if %arg6 { %value = load %arg3[%i] : memref<10xf32> store %value, %arg4[%i] : memref<10xf32> } else { %value = load %arg4[%i] : memref<10xf32> store %value, %arg3[%i] : memref<10xf32> } } else { scf.if %arg6 { %value = load %arg3[%j] : memref<10xf32> store %value, %arg4[%j] : memref<10xf32> } else { %value = load %arg4[%j] : memref<10xf32> store %value, %arg3[%j] : memref<10xf32> } } gpu.return } // CHECK-LABEL: @simple_if_yield gpu.func @simple_if_yield(%arg2 : memref<10xf32>, %arg3 : i1) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { // CHECK: %[[VAR1:.*]] = spv.Variable : !spv.ptr // CHECK: %[[VAR2:.*]] = spv.Variable : !spv.ptr // CHECK: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE:\^.*]], [[FALSE:\^.*]] // CHECK-NEXT: [[TRUE]]: // CHECK: %[[RET1TRUE:.*]] = spv.constant 0.000000e+00 : f32 // CHECK: %[[RET2TRUE:.*]] = spv.constant 1.000000e+00 : f32 // CHECK-DAG: spv.Store "Function" %[[VAR1]], %[[RET1TRUE]] : f32 // CHECK-DAG: spv.Store "Function" %[[VAR2]], %[[RET2TRUE]] : f32 // CHECK: spv.Branch ^[[MERGE:.*]] // CHECK-NEXT: [[FALSE]]: // CHECK: %[[RET2FALSE:.*]] = spv.constant 2.000000e+00 : f32 // CHECK: %[[RET1FALSE:.*]] = spv.constant 3.000000e+00 : f32 // CHECK-DAG: spv.Store "Function" %[[VAR1]], %[[RET1FALSE]] : f32 // CHECK-DAG: spv.Store "Function" %[[VAR2]], %[[RET2FALSE]] : f32 // CHECK: spv.Branch ^[[MERGE]] // CHECK-NEXT: ^[[MERGE]]: // CHECK: spv.mlir.merge // CHECK-NEXT: } // CHECK-DAG: %[[OUT1:.*]] = spv.Load "Function" %[[VAR1]] : f32 // CHECK-DAG: %[[OUT2:.*]] = spv.Load "Function" %[[VAR2]] : f32 // CHECK: spv.Store "StorageBuffer" {{%.*}}, %[[OUT1]] : f32 // CHECK: spv.Store "StorageBuffer" {{%.*}}, %[[OUT2]] : f32 // CHECK: spv.Return %0:2 = scf.if %arg3 -> (f32, f32) { %c0 = constant 0.0 : f32 %c1 = constant 1.0 : f32 scf.yield %c0, %c1 : f32, f32 } else { %c0 = constant 2.0 : f32 %c1 = constant 3.0 : f32 scf.yield %c1, %c0 : f32, f32 } %i = constant 0 : index %j = constant 1 : index store %0#0, %arg2[%i] : memref<10xf32> store %0#1, %arg2[%j] : memref<10xf32> gpu.return } // TODO: The transformation should only be legal if // VariablePointer capability is supported. This test is still useful to // make sure we can handle scf op result with type change. // CHECK-LABEL: @simple_if_yield_type_change // CHECK: %[[VAR:.*]] = spv.Variable : !spv.ptr [0])>, StorageBuffer>, Function> // CHECK: spv.selection { // CHECK-NEXT: spv.BranchConditional {{%.*}}, [[TRUE:\^.*]], [[FALSE:\^.*]] // CHECK-NEXT: [[TRUE]]: // CHECK: spv.Store "Function" %[[VAR]], {{%.*}} : !spv.ptr [0])>, StorageBuffer> // CHECK: spv.Branch ^[[MERGE:.*]] // CHECK-NEXT: [[FALSE]]: // CHECK: spv.Store "Function" %[[VAR]], {{%.*}} : !spv.ptr [0])>, StorageBuffer> // CHECK: spv.Branch ^[[MERGE]] // CHECK-NEXT: ^[[MERGE]]: // CHECK: spv.mlir.merge // CHECK-NEXT: } // CHECK: %[[OUT:.*]] = spv.Load "Function" %[[VAR]] : !spv.ptr [0])>, StorageBuffer> // CHECK: %[[ADD:.*]] = spv.AccessChain %[[OUT]][{{%.*}}, {{%.*}}] : !spv.ptr [0])>, StorageBuffer> // CHECK: spv.Store "StorageBuffer" %[[ADD]], {{%.*}} : f32 // CHECK: spv.Return gpu.func @simple_if_yield_type_change(%arg2 : memref<10xf32>, %arg3 : memref<10xf32>, %arg4 : i1) kernel attributes {spv.entry_point_abi = {local_size = dense<[16, 1, 1]>: vector<3xi32>}} { %i = constant 0 : index %value = constant 0.0 : f32 %0 = scf.if %arg4 -> (memref<10xf32>) { scf.yield %arg2 : memref<10xf32> } else { scf.yield %arg3 : memref<10xf32> } store %value, %0[%i] : memref<10xf32> gpu.return } } }