diff options
author | Archibald Elliott <archibald.elliott@arm.com> | 2022-12-07 10:04:30 +0000 |
---|---|---|
committer | Archibald Elliott <archibald.elliott@arm.com> | 2022-12-16 14:42:27 +0000 |
commit | 82b51a14280414a53413ed62c001d2c589c649c3 (patch) | |
tree | 4f0a241f025d80eff06abdbb261267cf7dc19b10 /llvm/lib | |
parent | f86cdb4853618603b8889dfeb932fd4ef8efd010 (diff) | |
download | llvm-82b51a14280414a53413ed62c001d2c589c649c3.zip llvm-82b51a14280414a53413ed62c001d2c589c649c3.tar.gz llvm-82b51a14280414a53413ed62c001d2c589c649c3.tar.bz2 |
[AArch64] Support SLC in ACLE prefetch intrinsics
This change:
- Modifies the ACLE code to allow the new SLC value (3) for the prefetch
target.
- Introduces a new intrinsic, @llvm.aarch64.prefetch which matches the
PRFM family instructions much more closely, and can represent all
values for the PRFM immediate.
The target-independent @llvm.prefetch intrinsic does not have enough
information for us to be able to lower to it from the ACLE intrinsics
correctly.
- Lowers the acle calls to the new intrinsic on aarch64 (the ARM
lowering is unchanged).
- Implements code generation for the new intrinsic in both SelectionDAG
and GlobalISel. We specifically choose to continue to support lowering
the target-independent @llvm.prefetch intrinsic so that other
frontends can continue to use it.
Differential Revision: https://reviews.llvm.org/D139443
Diffstat (limited to 'llvm/lib')
-rw-r--r-- | llvm/lib/IR/Verifier.cpp | 11 | ||||
-rw-r--r-- | llvm/lib/Target/AArch64/AArch64ISelLowering.cpp | 58 | ||||
-rw-r--r-- | llvm/lib/Target/AArch64/AArch64ISelLowering.h | 1 | ||||
-rw-r--r-- | llvm/lib/Target/AArch64/GISel/AArch64LegalizerInfo.cpp | 18 |
4 files changed, 72 insertions, 16 deletions
diff --git a/llvm/lib/IR/Verifier.cpp b/llvm/lib/IR/Verifier.cpp index 093b30d..910d1db 100644 --- a/llvm/lib/IR/Verifier.cpp +++ b/llvm/lib/IR/Verifier.cpp @@ -5754,6 +5754,17 @@ void Verifier::visitIntrinsicCall(Intrinsic::ID ID, CallBase &Call) { &Call); break; } + case Intrinsic::aarch64_prefetch: { + Check(cast<ConstantInt>(Call.getArgOperand(1))->getZExtValue() < 2, + "write argument to llvm.aarch64.prefetch must be 0 or 1", Call); + Check(cast<ConstantInt>(Call.getArgOperand(2))->getZExtValue() < 4, + "target argument to llvm.aarch64.prefetch must be 0-3", Call); + Check(cast<ConstantInt>(Call.getArgOperand(3))->getZExtValue() < 2, + "stream argument to llvm.aarch64.prefetch must be 0 or 1", Call); + Check(cast<ConstantInt>(Call.getArgOperand(4))->getZExtValue() < 2, + "isdata argument to llvm.aarch64.prefetch must be 0 or 1", Call); + break; + } }; } diff --git a/llvm/lib/Target/AArch64/AArch64ISelLowering.cpp b/llvm/lib/Target/AArch64/AArch64ISelLowering.cpp index f6501a4..0289da5 100644 --- a/llvm/lib/Target/AArch64/AArch64ISelLowering.cpp +++ b/llvm/lib/Target/AArch64/AArch64ISelLowering.cpp @@ -1189,9 +1189,6 @@ AArch64TargetLowering::AArch64TargetLowering(const TargetMachine &TM, } } - if (Subtarget->hasSME()) - setOperationAction(ISD::INTRINSIC_VOID, MVT::Other, Custom); - if (Subtarget->hasSVE()) { for (auto VT : {MVT::nxv16i8, MVT::nxv8i16, MVT::nxv4i32, MVT::nxv2i64}) { setOperationAction(ISD::BITREVERSE, VT, Custom); @@ -1520,6 +1517,8 @@ AArch64TargetLowering::AArch64TargetLowering(const TargetMachine &TM, setOperationAction(ISD::INTRINSIC_W_CHAIN, MVT::i8, Custom); } + setOperationAction(ISD::INTRINSIC_VOID, MVT::Other, Custom); + PredictableSelectIsExpensive = Subtarget->predictableSelectIsExpensive(); IsStrictFPEnabled = true; @@ -4724,6 +4723,44 @@ static std::optional<SMEAttrs> getCalleeAttrsFromExternalFunction(SDValue V) { return std::nullopt; } +SDValue AArch64TargetLowering::LowerINTRINSIC_VOID(SDValue Op, + SelectionDAG &DAG) const { + unsigned IntNo = Op.getConstantOperandVal(1); + SDLoc DL(Op); + switch (IntNo) { + default: + return SDValue(); // Don't custom lower most intrinsics. + case Intrinsic::aarch64_prefetch: { + SDValue Chain = Op.getOperand(0); + SDValue Addr = Op.getOperand(2); + + unsigned IsWrite = cast<ConstantSDNode>(Op.getOperand(3))->getZExtValue(); + unsigned Locality = cast<ConstantSDNode>(Op.getOperand(4))->getZExtValue(); + unsigned IsStream = cast<ConstantSDNode>(Op.getOperand(5))->getZExtValue(); + unsigned IsData = cast<ConstantSDNode>(Op.getOperand(6))->getZExtValue(); + unsigned PrfOp = (IsWrite << 4) | // Load/Store bit + (!IsData << 3) | // IsDataCache bit + (Locality << 1) | // Cache level bits + (unsigned)IsStream; // Stream bit + + return DAG.getNode(AArch64ISD::PREFETCH, DL, MVT::Other, Chain, + DAG.getTargetConstant(PrfOp, DL, MVT::i32), Addr); + } + case Intrinsic::aarch64_sme_za_enable: + return DAG.getNode( + AArch64ISD::SMSTART, DL, MVT::Other, + Op->getOperand(0), // Chain + DAG.getTargetConstant((int32_t)(AArch64SVCR::SVCRZA), DL, MVT::i32), + DAG.getConstant(0, DL, MVT::i64), DAG.getConstant(1, DL, MVT::i64)); + case Intrinsic::aarch64_sme_za_disable: + return DAG.getNode( + AArch64ISD::SMSTOP, DL, MVT::Other, + Op->getOperand(0), // Chain + DAG.getTargetConstant((int32_t)(AArch64SVCR::SVCRZA), DL, MVT::i32), + DAG.getConstant(0, DL, MVT::i64), DAG.getConstant(1, DL, MVT::i64)); + } +} + SDValue AArch64TargetLowering::LowerINTRINSIC_W_CHAIN(SDValue Op, SelectionDAG &DAG) const { unsigned IntNo = Op.getConstantOperandVal(1); @@ -4754,18 +4791,6 @@ SDValue AArch64TargetLowering::LowerINTRINSIC_W_CHAIN(SDValue Op, // changed. return DAG.getMergeValues({MS.getValue(0), MS.getValue(2)}, DL); } - case Intrinsic::aarch64_sme_za_enable: - return DAG.getNode( - AArch64ISD::SMSTART, DL, MVT::Other, - Op->getOperand(0), // Chain - DAG.getTargetConstant((int32_t)(AArch64SVCR::SVCRZA), DL, MVT::i32), - DAG.getConstant(0, DL, MVT::i64), DAG.getConstant(1, DL, MVT::i64)); - case Intrinsic::aarch64_sme_za_disable: - return DAG.getNode( - AArch64ISD::SMSTOP, DL, MVT::Other, - Op->getOperand(0), // Chain - DAG.getTargetConstant((int32_t)(AArch64SVCR::SVCRZA), DL, MVT::i32), - DAG.getConstant(0, DL, MVT::i64), DAG.getConstant(1, DL, MVT::i64)); } } @@ -5872,11 +5897,12 @@ SDValue AArch64TargetLowering::LowerOperation(SDValue Op, return LowerToPredicatedOp(Op, DAG, AArch64ISD::MULHS_PRED); case ISD::MULHU: return LowerToPredicatedOp(Op, DAG, AArch64ISD::MULHU_PRED); - case ISD::INTRINSIC_VOID: case ISD::INTRINSIC_W_CHAIN: return LowerINTRINSIC_W_CHAIN(Op, DAG); case ISD::INTRINSIC_WO_CHAIN: return LowerINTRINSIC_WO_CHAIN(Op, DAG); + case ISD::INTRINSIC_VOID: + return LowerINTRINSIC_VOID(Op, DAG); case ISD::ATOMIC_STORE: if (cast<MemSDNode>(Op)->getMemoryVT() == MVT::i128) { assert(Subtarget->hasLSE2()); diff --git a/llvm/lib/Target/AArch64/AArch64ISelLowering.h b/llvm/lib/Target/AArch64/AArch64ISelLowering.h index 49d43d5..2869604 100644 --- a/llvm/lib/Target/AArch64/AArch64ISelLowering.h +++ b/llvm/lib/Target/AArch64/AArch64ISelLowering.h @@ -960,6 +960,7 @@ private: SDValue LowerINTRINSIC_W_CHAIN(SDValue Op, SelectionDAG &DAG) const; SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const; + SDValue LowerINTRINSIC_VOID(SDValue Op, SelectionDAG &DAG) const; bool isEligibleForTailCallOptimization(const CallLoweringInfo &CLI) const; diff --git a/llvm/lib/Target/AArch64/GISel/AArch64LegalizerInfo.cpp b/llvm/lib/Target/AArch64/GISel/AArch64LegalizerInfo.cpp index 48d6553..8a760a6 100644 --- a/llvm/lib/Target/AArch64/GISel/AArch64LegalizerInfo.cpp +++ b/llvm/lib/Target/AArch64/GISel/AArch64LegalizerInfo.cpp @@ -1070,6 +1070,24 @@ bool AArch64LegalizerInfo::legalizeIntrinsic(LegalizerHelper &Helper, MI.eraseFromParent(); return true; } + case Intrinsic::aarch64_prefetch: { + MachineIRBuilder MIB(MI); + auto &AddrVal = MI.getOperand(1); + + int64_t IsWrite = MI.getOperand(2).getImm(); + int64_t Target = MI.getOperand(3).getImm(); + int64_t IsStream = MI.getOperand(4).getImm(); + int64_t IsData = MI.getOperand(5).getImm(); + + unsigned PrfOp = (IsWrite << 4) | // Load/Store bit + (!IsData << 3) | // IsDataCache bit + (Target << 1) | // Cache level bits + (unsigned)IsStream; // Stream bit + + MIB.buildInstr(AArch64::G_PREFETCH).addImm(PrfOp).add(AddrVal); + MI.eraseFromParent(); + return true; + } } return true; |