// RUN: mlir-opt -one-shot-bufferize="bufferize-function-boundaries" -split-input-file %s | FileCheck %s
// RUN: mlir-opt -one-shot-bufferize -split-input-file %s | FileCheck %s --check-prefix=CHECK-NO-FUNC
// CHECK-NO-FUNC-LABEL: func @br(
// CHECK-NO-FUNC-SAME: %[[t:.*]]: tensor<5xf32>)
// CHECK-NO-FUNC: %[[m:.*]] = bufferization.to_memref %[[t]] : memref<5xf32, strided<[?], offset: ?>>
// CHECK-NO-FUNC: %[[r:.*]] = scf.execute_region -> memref<5xf32, strided<[?], offset: ?>> {
// CHECK-NO-FUNC: cf.br ^[[block:.*]](%[[m]]
// CHECK-NO-FUNC: ^[[block]](%[[arg1:.*]]: memref<5xf32, strided<[?], offset: ?>>):
// CHECK-NO-FUNC: scf.yield %[[arg1]]
// CHECK-NO-FUNC: }
// CHECK-NO-FUNC: return
func.func @br(%t: tensor<5xf32>) {
%0 = scf.execute_region -> tensor<5xf32> {
cf.br ^bb1(%t : tensor<5xf32>)
^bb1(%arg1 : tensor<5xf32>):
scf.yield %arg1 : tensor<5xf32>
}
return
}
// -----
// CHECK-NO-FUNC-LABEL: func @cond_br(
// CHECK-NO-FUNC-SAME: %[[t1:.*]]: tensor<5xf32>,
// CHECK-NO-FUNC: %[[m1:.*]] = bufferization.to_memref %[[t1]] : memref<5xf32, strided<[?], offset: ?>>
// CHECK-NO-FUNC: %[[alloc:.*]] = memref.alloc() {{.*}} : memref<5xf32>
// CHECK-NO-FUNC: %[[r:.*]] = scf.execute_region -> memref<5xf32, strided<[?], offset: ?>> {
// CHECK-NO-FUNC: cf.cond_br %{{.*}}, ^[[block1:.*]](%[[m1]] : {{.*}}), ^[[block2:.*]](%[[alloc]] : {{.*}})
// CHECK-NO-FUNC: ^[[block1]](%[[arg1:.*]]: memref<5xf32, strided<[?], offset: ?>>):
// CHECK-NO-FUNC: scf.yield %[[arg1]]
// CHECK-NO-FUNC: ^[[block2]](%[[arg2:.*]]: memref<5xf32>):
// CHECK-NO-FUNC: %[[cast:.*]] = memref.cast %[[arg2]] : memref<5xf32> to memref<5xf32, strided<[?], offset: ?>
// CHECK-NO-FUNC: cf.br ^[[block1]](%[[cast]] : {{.*}})
// CHECK-NO-FUNC: }
// CHECK-NO-FUNC: return
func.func @cond_br(%t1: tensor<5xf32>, %c: i1) {
// Use an alloc for the second block instead of a function block argument.
// A cast must be inserted because the two will have different layout maps.
%t0 = bufferization.alloc_tensor() : tensor<5xf32>
%0 = scf.execute_region -> tensor<5xf32> {
cf.cond_br %c, ^bb1(%t1 : tensor<5xf32>), ^bb2(%t0 : tensor<5xf32>)
^bb1(%arg1 : tensor<5xf32>):
scf.yield %arg1 : tensor<5xf32>
^bb2(%arg2 : tensor<5xf32>):
cf.br ^bb1(%arg2 : tensor<5xf32>)
}
return
}
// -----
// CHECK-LABEL: func @looping_branches(
func.func @looping_branches() -> tensor<5xf32> {
// CHECK: %[[alloc:.*]] = memref.alloc
%0 = bufferization.alloc_tensor() : tensor<5xf32>
// CHECK: cf.br {{.*}}(%[[alloc]]
cf.br ^bb1(%0: tensor<5xf32>)
// CHECK: ^{{.*}}(%[[arg1:.*]]: memref<5xf32>):
^bb1(%arg1: tensor<5xf32>):
%pos = "test.foo"() : () -> (index)
%val = "test.bar"() : () -> (f32)
// CHECK: memref.store %{{.*}}, %[[arg1]]
%inserted = tensor.insert %val into %arg1[%pos] : tensor<5xf32>
%cond = "test.qux"() : () -> (i1)
// CHECK: cf.cond_br {{.*}}(%[[arg1]] {{.*}}(%[[arg1]]
cf.cond_br %cond, ^bb1(%inserted: tensor<5xf32>), ^bb2(%inserted: tensor<5xf32>)
// CHECK: ^{{.*}}(%[[arg2:.*]]: memref<5xf32>):
^bb2(%arg2: tensor<5xf32>):
// CHECK: return %[[arg2]]
func.return %arg2 : tensor<5xf32>
}