aboutsummaryrefslogtreecommitdiff
path: root/llvm/lib/CodeGen/TargetLoweringObjectFileImpl.cpp
diff options
context:
space:
mode:
authorPeiming Liu <peiming@google.com>2023-02-28 01:18:12 +0000
committerPeiming Liu <peiming@google.com>2023-02-28 21:02:17 +0000
commit849529ba8a2e155cd90b0641bcf20bc566222db6 (patch)
tree6685c59910a848c846c397187575ac4c35dc849a /llvm/lib/CodeGen/TargetLoweringObjectFileImpl.cpp
parent35d17c17a6702c70e877c63e9f7c4ce3e35c1bb9 (diff)
downloadllvm-849529ba8a2e155cd90b0641bcf20bc566222db6.zip
llvm-849529ba8a2e155cd90b0641bcf20bc566222db6.tar.gz
llvm-849529ba8a2e155cd90b0641bcf20bc566222db6.tar.bz2
[mlir][sparse] fix performance bug in matmul with a sparse rhs due to suboptimal iteration graphs.
While dense tensors support random accesses, it is critical to visit them in a row-major order for better cache locality. However, we previously consider dense inputs and outputs together when computing constraints for building iteration graph, it could lead us to less efficient iteration graphs. This patch adds a new `SortMask::kIncludeDenseInput` to treat dense inputs/outputs separately when building iteration graph, thus increasing the chance for use to construct a better iteration graph. A more fine-grained approach is to treat each input separately. Note, related to: https://github.com/llvm/llvm-project/issues/51651 Reviewed By: aartbik Differential Revision: https://reviews.llvm.org/D144932
Diffstat (limited to 'llvm/lib/CodeGen/TargetLoweringObjectFileImpl.cpp')
0 files changed, 0 insertions, 0 deletions