//===-- Passes.td - Sparse tensor pass definition file -----*- tablegen -*-===//
//
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
// See https://llvm.org/LICENSE.txt for license information.
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
//
//===----------------------------------------------------------------------===//
#ifndef MLIR_DIALECT_SPARSETENSOR_TRANSFORMS_PASSES
#define MLIR_DIALECT_SPARSETENSOR_TRANSFORMS_PASSES
include "mlir/Pass/PassBase.td"
def SparseAssembler : Pass<"sparse-assembler", "ModuleOp"> {
let summary = "Add [dis]assemble operations on external sparse tensors";
let description = [{
Unlike dense tensors, MLIR does **not** provide a direct `_mlir_ciface_`
ABI for passing sparse tensors as arguments from and to external methods
(within MLIR-generated methods, sparse tensors can be freely passed
around, but this eventually uses a bespoke parameter passing format
that is subject to change; like opaque pointers when the sparse runtime
support library is used or the constituent arrays and structs for
direct IR codegen). The sparse assembler pass, however, can be used
to obtain a stable `_mlir_ciface_` API for passing sparse tensors
from and to an external environment, such as Python, PyTorch, or JAX.
The pass converts public entry methods that use sparse tensors as
input parameters and/or output return values into wrapper methods
that [dis]assemble the individual tensors that constitute the actual
storage used externally into MLIR sparse tensors. This pass can be used
to prepare the public entry methods of a program that is compiled by the
MLIR sparsifier to interface with an external runtime, e.g., when passing
sparse tensors as numpy arrays from and to Python. Note that eventual
bufferization decisions (e.g. who [de]allocates the underlying memory)
should be resolved in agreement with the external runtime.
By default, the pass uses the [dis]assemble operations to input and output
sparse tensors. When the direct-out option is set, however, the output
directly returns the MLIR allocated buffers to the external runtime.
The pass should always run before the actual sparsification passes.
}];
let constructor = "mlir::createSparseAssembler()";
let dependentDialects = [
"bufferization::BufferizationDialect",
"sparse_tensor::SparseTensorDialect",
"tensor::TensorDialect",
];
let options = [
Option<"directOut", "direct-out", "bool",
"false", "Directly returns buffers externally">,
];
}
def SparseReinterpretMap : Pass<"sparse-reinterpret-map", "ModuleOp"> {
let summary = "Reinterprets sparse tensor type mappings";
let description = [{
A pass that reinterprets the mappings in all sparse tensor types in a
way that enables subsequent sparsification. This involves expressing all
`linalg.generic` operations in terms of level coordinates (rather than
the dimension coordinates of the input tensors) to align the iteration
space with the potentially remapped level space as well as resolving cycles
in the resulting iteration graphs with explicit sparse tensor conversions
where needed.
}];
let constructor = "mlir::createSparseReinterpretMapPass()";
let dependentDialects = [
"affine::AffineDialect",
"linalg::LinalgDialect",
"sparse_tensor::SparseTensorDialect",
];
let options = [
Option<"scope", "scope", "mlir::ReinterpretMapScope",
"mlir::ReinterpretMapScope::kAll",
"Set the reiterpretation scope", [{llvm::cl::values(
clEnumValN(mlir::ReinterpretMapScope::kAll, "all",
"Run on every applicable operations."),
clEnumValN(mlir::ReinterpretMapScope::kGenericOnly,
"only-generic",
"Run only on linalg.generic operations."),
clEnumValN(mlir::ReinterpretMapScope::kExceptGeneric,
"except-generic",
"Run on operations expect linalg.generic (e.g., foreach)"))}]>,
];
}
def PreSparsificationRewrite : Pass<"pre-sparsification-rewrite", "ModuleOp"> {
let summary = "Applies sparse tensor rewriting rules prior to sparsification";
let description = [{
A pass that applies rewriting rules to sparse tensor operations prior
to running the actual sparsification pass.
}];
let constructor = "mlir::createPreSparsificationRewritePass()";
let dependentDialects = [
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
}
def SparsificationPass : Pass<"sparsification", "ModuleOp"> {
let summary = "Automatically generate sparse tensor code from sparse tensor types";
let description = [{
A pass that implements the core functionality of a **sparsifier**.
Each Linalg operation (MLIR's tensor index notation) that operates on
sparse tensor types is converted into code in which the sparsity is
explicit both in terms of co-iterating looping logic as well as
selected sparse storage schemes.
See the `SparseTensor` dialect documentation for more background.
Example input:
```mlir
#matvec = {
indexing_maps = [
affine_map<(i,j) -> (i,j)>, // A
affine_map<(i,j) -> (j)>, // b
affine_map<(i,j) -> (i)> // x (out)
],
iterator_types = ["parallel", "reduction"],
doc = "X(i) += A(i,j) * B(j)"
}
// Multiply a sparse matrix A with a dense vector b into a dense vector x.
func.func @kernel_matvec(%arga: tensor<?x?xf64, #SparseMatrix>,
%argb: tensor<?xf64>,
%argx: tensor<?xf64>) -> tensor<?xf64> {
%0 = linalg.generic #matvec
ins(%arga, %argb: tensor<?x?xf64, #SparseMatrix>, tensor<?xf64>)
outs(%argx: tensor<?xf64>) {
^bb(%a: f64, %b: f64, %x: f64):
%0 = arith.mulf %a, %b : f64
%1 = arith.addf %x, %0 : f64
linalg.yield %1 : f64
} -> tensor<?xf64>
return %0 : tensor<?xf64>
}
```
}];
let constructor = "mlir::createSparsificationPass()";
let dependentDialects = [
"affine::AffineDialect",
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"LLVM::LLVMDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
// TODO(57514): These enum options are duplicated in Passes.h.
let options = [
Option<"parallelization", "parallelization-strategy", "mlir::SparseParallelizationStrategy",
"mlir::SparseParallelizationStrategy::kNone",
"Set the parallelization strategy", [{llvm::cl::values(
clEnumValN(mlir::SparseParallelizationStrategy::kNone, "none",
"Turn off sparse parallelization."),
clEnumValN(mlir::SparseParallelizationStrategy::kDenseOuterLoop,
"dense-outer-loop",
"Enable dense outer loop sparse parallelization."),
clEnumValN(mlir::SparseParallelizationStrategy::kAnyStorageOuterLoop,
"any-storage-outer-loop",
"Enable sparse parallelization regardless of storage for the outer loop."),
clEnumValN(mlir::SparseParallelizationStrategy::kDenseAnyLoop,
"dense-any-loop",
"Enable dense parallelization for any loop."),
clEnumValN(mlir::SparseParallelizationStrategy::kAnyStorageAnyLoop,
"any-storage-any-loop",
"Enable sparse parallelization for any storage and loop."))}]>,
Option<"sparseEmitStrategy", "sparse-emit-strategy", "mlir::SparseEmitStrategy",
"mlir::SparseEmitStrategy::kFunctional",
"Emit functional code or interfaces (to debug) for sparse loops", [{llvm::cl::values(
clEnumValN(mlir::SparseEmitStrategy::kFunctional, "functional",
"Emit functional code (with scf.for/while)."),
clEnumValN(mlir::SparseEmitStrategy::kSparseIterator, "sparse-iterator",
"Emit (experimental) loops (with sparse.iterate)."),
clEnumValN(mlir::SparseEmitStrategy::kDebugInterface, "debug-interface",
"Emit non-functional but easy-to-read interfaces to debug."))}]>,
Option<"enableRuntimeLibrary", "enable-runtime-library", "bool",
"true", "Enable runtime library for manipulating sparse tensors">,
];
}
def StageSparseOperations : Pass<"stage-sparse-ops", "func::FuncOp"> {
let summary = "Decompose a complex sparse operation into multiple stages";
let description = [{
A pass that decomposes a complex sparse operation into multiple stages.
E.g., CSR -> CSC is staged into CSR -> COO (unordered) -> sort -> CSC.
}];
let constructor = "mlir::createStageSparseOperationsPass()";
let dependentDialects = [
"sparse_tensor::SparseTensorDialect",
];
}
def LowerSparseOpsToForeach : Pass<"lower-sparse-ops-to-foreach", "ModuleOp"> {
let summary = "Applies sparse tensor rewriting rules after sparsification";
let description = [{
A pass that lowers high-level sparse operations to sparse_tensor.foreach.
}];
let constructor = "mlir::createLowerSparseOpsToForeachPass()";
let dependentDialects = [
"affine::AffineDialect",
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
let options = [
Option<"enableRuntimeLibrary", "enable-runtime-library", "bool",
"true", "Enable runtime library for manipulating sparse tensors">,
Option<"enableConvert", "enable-convert", "bool",
"true", "Enable rewriting rules for the convert operator">,
];
}
def LowerForeachToSCF : Pass<"lower-sparse-foreach-to-scf", "func::FuncOp"> {
let summary = "Decompose a complex sparse operation into multiple stages";
let description = [{
A pass that lowers sparse_tensor.foreach operation to scf dialect.
}];
let constructor = "mlir::createLowerForeachToSCFPass()";
let dependentDialects = [
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
}
def SparseTensorConversionPass : Pass<"sparse-tensor-conversion", "ModuleOp"> {
let summary = "Convert sparse tensors and primitives to library calls";
let description = [{
A pass that converts sparse tensor primitives into calls into a runtime
support library. Sparse tensor types are converted into opaque pointers
to the underlying sparse storage schemes.
The use of opaque pointers together with runtime support library keeps
the conversion relatively simple, but at the expense of IR opacity,
which obscures opportunities for subsequent optimization of the IR.
An alternative is provided by the SparseTensorCodegen pass.
Example of the conversion:
```mlir
Before:
func.func @foo(%arg0: tensor<8x8xf32, #CSR>) -> memref<?xindex> {
%0 = sparse_tensor.pointers %arg0 {dimension = 1 : index}
: tensor<8x8xf32, #CSR> to memref<?xindex>
return %0 : memref<?xindex>
}
After:
func.func @foo(%arg0: !llvm.ptr) -> memref<?xindex> {
%c1 = arith.constant 1 : index
%0 = call @sparsePointers0(%arg0, %c1)
: (!llvm.ptr, index) -> memref<?xindex>
return %0 : memref<?xindex>
}
```
}];
let constructor = "mlir::createSparseTensorConversionPass()";
let dependentDialects = [
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"LLVM::LLVMDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
}
def SparseTensorCodegen : Pass<"sparse-tensor-codegen", "ModuleOp"> {
let summary = "Convert sparse tensors and primitives to actual code";
let description = [{
A pass that converts sparse tensor types and primitives to actual
compiler visible buffers and compiler IR that implements these
primitives on the selected sparse tensor storage schemes.
This pass provides an alternative to the SparseTensorConversion pass,
eliminating the dependence on a runtime support library, and providing
much more opportunities for subsequent compiler optimization of the
generated code.
Example of the conversion:
```mlir
Before:
func.func @foo(%arg0: tensor<8x8xf32, #CSR>) -> memref<?xindex> {
%0 = sparse_tensor.pointers %arg0 {dimension = 1 : index}
: tensor<8x8xf32, #CSR> to memref<?xindex>
return %0 : memref<?xindex>
}
After:
func.func @foo(%arg0: memref<2xindex>,
%arg1: memref<3xindex>,
%arg2: memref<?xindex>,
%arg3: memref<?xindex>,
%arg4: memref<?xf32>) -> memref<?xindex> {
return %arg2 : memref<?xindex>
}
```
}];
let constructor = "mlir::createSparseTensorCodegenPass()";
let dependentDialects = [
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
let options = [
Option<"enableBufferInitialization", "enable-buffer-initialization", "bool",
"false", "Enable zero-initialization of the memory buffers">,
Option<"createSparseDeallocs", "create-sparse-deallocs", "bool",
"true", "Specify if the temporary buffers created by the sparse "
"compiler should be deallocated. For compatibility with core "
"bufferization passes. "
"This option is only used when enable-runtime-library=false. "
"See also create-deallocs for BufferizationOption.">,
];
}
def SparseBufferRewrite : Pass<"sparse-buffer-rewrite", "ModuleOp"> {
let summary = "Rewrite sparse primitives on buffers to actual code";
let description = [{
A pass that rewrites sparse primitives on buffers to the MLIR implementation
of the primitives. For example, sparse_tensor.sort operator is implemented
in this pass.
}];
let constructor = "mlir::createSparseBufferRewritePass()";
let dependentDialects = [
"arith::ArithDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
let options = [
Option<"enableBufferInitialization", "enable-buffer-initialization", "bool",
"false", "Enable zero-initialization of the memory buffers">,
];
}
def SparseVectorization : Pass<"sparse-vectorization", "ModuleOp"> {
let summary = "Vectorizes loops after sparsification";
let description = [{
A pass that converts loops after sparsification into vector loops.
The vector dialect is used as target to provide an architectural
neutral way of exploiting any platform that supports SIMD instructions.
The vector length (viz. `vl`) describes the number of packed data elements
(e.g. both vector<16xf32> and vector<16xf64> have a vector length of 16 even
though the actual bitwidths differ). A small multiple of the actual lengths
supported in hardware typically results in efficient SIMD code, since the
backend will map longer vectors to multiple vector registers, thereby
effectively unrolling an addition level within the generated for-loop.
Example of the conversion:
```mlir
Before:
%3 = memref.load %2[] : memref<f32>
%4 = scf.for %arg3 = %c0 to %c1024 step %c1 iter_args(%arg4 = %3) -> (f32) {
%6 = memref.load %0[%arg3] : memref<?xf32>
%7 = memref.load %1[%arg3] : memref<1024xf32>
%8 = arith.mulf %6, %7 : f32
%9 = arith.addf %arg4, %8 : f32
scf.yield %9 : f32
}
memref.store %4, %2[] : memref<f32>
After:
%3 = memref.load %2[] : memref<f32>
%4 = vector.insertelement %3, %cst[%c0 : index] : vector<32xf32>
%5 = scf.for %arg3 = %c0 to %c1024 step %c32 iter_args(%arg4 = %4) -> (vector<32xf32>) {
%8 = vector.load %0[%arg3] : memref<?xf32>, vector<32xf32>
%9 = vector.load %1[%arg3] : memref<1024xf32>, vector<32xf32>
%10 = arith.mulf %8, %9 : vector<32xf32>
%11 = arith.addf %arg4, %10 : vector<32xf32>
scf.yield %11 : vector<32xf32>
}
%6 = vector.reduction <add>, %5 : vector<32xf32> into f32
memref.store %6, %2[] : memref<f32>
```
}];
let constructor = "mlir::createSparseVectorizationPass()";
let dependentDialects = [
"arith::ArithDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
"vector::VectorDialect",
];
let options = [
Option<"vectorLength", "vl", "int32_t", "0",
"Set the vector length (use 0 to disable vectorization)">,
Option<"enableVLAVectorization", "enable-vla-vectorization", "bool",
"false", "Enable vector length agnostic vectorization">,
Option<"enableSIMDIndex32", "enable-simd-index32", "bool", "false",
"Enable i32 indexing into vectors (for efficient gather/scatter)">,
];
}
def SparseGPUCodegen : Pass<"sparse-gpu-codegen", "ModuleOp"> {
let summary = "Generates GPU code during sparsification";
let description = [{
Enables the sparsifier to use GPU acceleration. When the number of GPU
threads is set to zero, the pass tries to enable GPU acceleration by
means of direct library calls (like cuSPARSE).
}];
let constructor = "mlir::createSparseGPUCodegenPass()";
let dependentDialects = [
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"gpu::GPUDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
let options = [
Option<"numThreads", "num-threads", "int32_t", "1024", "Sets the number of GPU threads">,
Option<"enableRuntimeLibrary", "enable-runtime-library", "bool",
"true", "Enable runtime library for manipulating sparse tensors">,
];
}
def StorageSpecifierToLLVM : Pass<"sparse-storage-specifier-to-llvm", "ModuleOp"> {
let summary = "Lower sparse storage specifer to llvm structure";
let description = [{
This pass rewrites sparse tensor storage specifier-related operations into
LLVMDialect, and converts sparse tensor storage specifier into an llvm.struct.
Example of the conversion:
```mlir
Before:
%0 = sparse_tensor.storage_specifier.get %arg0 dim_sz at 0
: !sparse_tensor.storage_specifier<#CSR> to i64
After:
%0 = llvm.extractvalue %arg0[0, 0] : !llvm.struct<(array<2 x i64>, array<3 x i64>)>
```
}];
let constructor = "mlir::createStorageSpecifierToLLVMPass()";
let dependentDialects = [
"arith::ArithDialect",
"LLVM::LLVMDialect",
"sparse_tensor::SparseTensorDialect",
];
}
def SparsificationAndBufferization : Pass<"sparsification-and-bufferization", "ModuleOp"> {
let summary = "Mini-pipeline that combines bufferization and sparsifiation";
let description = [{
This pass forms a mini-pipeline that combines bufferization and sparsifiation.
}];
let constructor = "mlir::createSparsificationAndBufferizationPass()";
let dependentDialects = [
"affine::AffineDialect",
"arith::ArithDialect",
"bufferization::BufferizationDialect",
"gpu::GPUDialect",
"LLVM::LLVMDialect",
"linalg::LinalgDialect",
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
"vector::VectorDialect"
];
// Important optimization options are made visible to the mini-pipeline
// so that clients can set these (when not using the full pipeline).
let options = [
Option<"vectorLength", "vl", "int32_t", "0",
"Set the vector length (use 0 to disable vectorization)">,
Option<"enableVLAVectorization", "enable-vla-vectorization", "bool", "false",
"Enable vector length agnostic vectorization">,
Option<"enableSIMDIndex32", "enable-simd-index32", "bool", "false",
"Enable i32 indexing into vectors (for efficient gather/scatter)">,
Option<"enableGPULibgen", "enable-gpu-libgen", "bool", "false",
"Enable GPU acceleration by means of direct library calls">,
Option<"sparseEmitStrategy", "sparse-emit-strategy", "mlir::SparseEmitStrategy",
"mlir::SparseEmitStrategy::kFunctional",
"Emit functional code or interfaces (to debug) for sparse loops", [{llvm::cl::values(
clEnumValN(mlir::SparseEmitStrategy::kFunctional, "functional",
"Emit functional code (with scf.for/while)."),
clEnumValN(mlir::SparseEmitStrategy::kSparseIterator, "sparse-iterator",
"Emit (experimental) loops (with sparse.iterate)."),
clEnumValN(mlir::SparseEmitStrategy::kDebugInterface, "debug-interface",
"Emit non-functional but easy-to-read interfaces to debug."))}]>,
Option<"parallelization", "parallelization-strategy", "mlir::SparseParallelizationStrategy",
"mlir::SparseParallelizationStrategy::kNone",
"Set the parallelization strategy", [{llvm::cl::values(
clEnumValN(mlir::SparseParallelizationStrategy::kNone, "none",
"Turn off sparse parallelization."),
clEnumValN(mlir::SparseParallelizationStrategy::kDenseOuterLoop,
"dense-outer-loop",
"Enable dense outer loop sparse parallelization."),
clEnumValN(mlir::SparseParallelizationStrategy::kAnyStorageOuterLoop,
"any-storage-outer-loop",
"Enable sparse parallelization regardless of storage for the outer loop."),
clEnumValN(mlir::SparseParallelizationStrategy::kDenseAnyLoop,
"dense-any-loop",
"Enable dense parallelization for any loop."),
clEnumValN(mlir::SparseParallelizationStrategy::kAnyStorageAnyLoop,
"any-storage-any-loop",
"Enable sparse parallelization for any storage and loop."))}]>,
];
}
//===----------------------------------------------------------------------===//
// Sparse Iteration Transform Passes
//===----------------------------------------------------------------------===//
def SparseSpaceCollapse : Pass<"sparse-space-collapse", "func::FuncOp"> {
let summary = "sparse space collapsing pass";
let description = [{
This pass collapses consecutive sparse spaces (extracted from the same tensor)
into one multi-dimensional space. The pass is not yet stabilized.
}];
let constructor = "mlir::createSparseSpaceCollapsePass()";
let dependentDialects = [
"sparse_tensor::SparseTensorDialect",
];
}
def LowerSparseIterationToSCF : Pass<"lower-sparse-iteration-to-scf", "func::FuncOp"> {
let summary = "lower sparse_tensor.iterate/coiterate into scf loops";
let description = [{
This pass lowers `sparse_tensor.iterate` operations into `scf.for/while` operations.
The pass is not yet stabilized.
}];
let constructor = "mlir::createLowerSparseIterationToSCFPass()";
let dependentDialects = [
"memref::MemRefDialect",
"scf::SCFDialect",
"sparse_tensor::SparseTensorDialect",
];
}
#endif // MLIR_DIALECT_SPARSETENSOR_TRANSFORMS_PASSES