diff options
author | Aart Bik <ajcbik@google.com> | 2022-12-15 11:38:55 -0800 |
---|---|---|
committer | Aart Bik <ajcbik@google.com> | 2022-12-15 12:12:06 -0800 |
commit | 78ba3aa765d079e1f83bfc4bf30f55ddaed518ce (patch) | |
tree | abc22b5a4f67fb5746c33b5d03ccf44e82aeef65 | |
parent | 795868285db9e8916418cc83f74416b0775d20dc (diff) | |
download | llvm-78ba3aa765d079e1f83bfc4bf30f55ddaed518ce.zip llvm-78ba3aa765d079e1f83bfc4bf30f55ddaed518ce.tar.gz llvm-78ba3aa765d079e1f83bfc4bf30f55ddaed518ce.tar.bz2 |
[mlir][sparse] performs a tab cleanup (NFC)
Reviewed By: Peiming
Differential Revision: https://reviews.llvm.org/D140142
7 files changed, 10 insertions, 10 deletions
diff --git a/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorOps.td b/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorOps.td index 31d5eb5..30f864f 100644 --- a/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorOps.td +++ b/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorOps.td @@ -451,7 +451,7 @@ def SparseTensor_CompressOp : SparseTensor_Op<"compress", StridedMemRefRankOf<[Index],[1]>:$added, Index:$count, AnySparseTensor:$tensor, - Variadic<Index>:$indices)>, + Variadic<Index>:$indices)>, Results<(outs AnySparseTensor:$result)> { string summary = "Compressed an access pattern for insertion"; string description = [{ @@ -477,8 +477,8 @@ def SparseTensor_CompressOp : SparseTensor_Op<"compress", }]; let assemblyFormat = "$values `,` $filled `,` $added `,` $count" " `into` $tensor `[` $indices `]` attr-dict" - " `:` type($values) `,` type($filled) `,` type($added)" - " `,` type($tensor)"; + " `:` type($values) `,` type($filled) `,` type($added)" + " `,` type($tensor)"; let hasVerifier = 1; } diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir index e12e9bf..79d1e35 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir @@ -83,8 +83,8 @@ module { %v2 = arith.constant sparse< [ [0], [3], [5], [11], [13], [17], [18], [21], [31] ], [ -2147483648, -2147483647, -1000, -1, 0, - 1, 1000, 2147483646, 2147483647 - ] + 1, 1000, 2147483646, 2147483647 + ] > : tensor<32xi32> %sv1 = sparse_tensor.convert %v1 : tensor<32xf64> to tensor<?xf64, #SparseVector> diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_flatten.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_flatten.mlir index 41b5410..6917274 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_flatten.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_flatten.mlir @@ -48,7 +48,7 @@ module { // func.func @kernel_flatten(%arga: tensor<7x3x3x3x3x3x5x3xf64, #SparseTensor>, %argx: tensor<7x3xf64>) - -> tensor<7x3xf64> { + -> tensor<7x3xf64> { %0 = linalg.generic #trait_flatten ins(%arga: tensor<7x3x3x3x3x3x5x3xf64, #SparseTensor>) outs(%argx: tensor<7x3xf64>) { diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_matvec.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_matvec.mlir index 9430070..2c18c2b 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_matvec.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_matvec.mlir @@ -55,7 +55,7 @@ module { func.func @kernel_matvec(%arga: tensor<?x?xi32, #SparseMatrix>, %argb: tensor<?xi32>, %argx: tensor<?xi32>) - -> tensor<?xi32> { + -> tensor<?xi32> { %0 = linalg.generic #matvec ins(%arga, %argb: tensor<?x?xi32, #SparseMatrix>, tensor<?xi32>) outs(%argx: tensor<?xi32>) { diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_mttkrp.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_mttkrp.mlir index 0933e82..375db9a 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_mttkrp.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_mttkrp.mlir @@ -49,7 +49,7 @@ module { %argc: tensor<?x?xf64>, %argd: tensor<?x?xf64>, %arga: tensor<?x?xf64>) - -> tensor<?x?xf64> { + -> tensor<?x?xf64> { %0 = linalg.generic #mttkrp ins(%argb, %argc, %argd: tensor<?x?x?xf64, #SparseTensor>, tensor<?x?xf64>, tensor<?x?xf64>) diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_out_reduction.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_out_reduction.mlir index 36c09a0..4978434 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_out_reduction.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_out_reduction.mlir @@ -36,7 +36,7 @@ module { func.func @redsum(%arga: tensor<?x?x?xi32, #SparseTensor>, %argb: tensor<?x?x?xi32, #SparseTensor>) - -> tensor<?x?xi32, #SparseMatrix> { + -> tensor<?x?xi32, #SparseMatrix> { %c0 = arith.constant 0 : index %c1 = arith.constant 1 : index %d0 = tensor.dim %arga, %c0 : tensor<?x?x?xi32, #SparseTensor> diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_quantized_matmul.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_quantized_matmul.mlir index 7b982f7..2c1200a 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_quantized_matmul.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_quantized_matmul.mlir @@ -59,7 +59,7 @@ module { %0 = call @quantized_matmul(%input1, %sparse_input2, %output) : (tensor<5x3xi8>, tensor<3x6xi8, #DCSR>, - tensor<5x6xi32>) -> tensor<5x6xi32> + tensor<5x6xi32>) -> tensor<5x6xi32> // // Verify the output. |