35#include "llvm/IR/IntrinsicsWebAssembly.h"
42#define DEBUG_TYPE "wasm-lower"
47 auto MVTPtr = Subtarget->hasAddr64() ? MVT::i64 : MVT::i32;
61 Subtarget->hasAddr64() ? WebAssembly::SP64 : WebAssembly::SP32);
67 if (Subtarget->hasSIMD128()) {
75 if (Subtarget->hasFP16()) {
78 if (Subtarget->hasReferenceTypes()) {
81 if (Subtarget->hasExceptionHandling()) {
90 for (
auto T : {MVT::i32, MVT::i64, MVT::f32, MVT::f64}) {
94 if (Subtarget->hasSIMD128()) {
95 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
101 if (Subtarget->hasFP16()) {
105 if (Subtarget->hasReferenceTypes()) {
108 for (
auto T : {MVT::externref, MVT::funcref, MVT::Other}) {
129 for (
auto T : {MVT::f32, MVT::f64, MVT::v4f32, MVT::v2f64, MVT::v8f16}) {
130 if (!Subtarget->hasFP16() &&
T == MVT::v8f16) {
145 for (
auto Op : {ISD::FCEIL, ISD::FFLOOR, ISD::FTRUNC, ISD::FNEARBYINT,
146 ISD::FRINT, ISD::FROUNDEVEN})
153 if (
T != MVT::v8f16) {
166 for (
auto T : {MVT::i32, MVT::i64})
168 if (Subtarget->hasSIMD128())
169 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v2i64})
173 if (Subtarget->hasWideArithmetic()) {
181 if (Subtarget->hasNontrappingFPToInt())
183 for (
auto T : {MVT::i32, MVT::i64})
187 if (Subtarget->hasSIMD128()) {
221 for (
auto T : {MVT::v16i8, MVT::v8i16})
225 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v2i64})
229 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
233 if (Subtarget->hasFP16())
237 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
241 if (Subtarget->hasFP16())
245 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
253 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v2i64})
258 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
266 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v4f32, MVT::v2i64,
273 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v2i64})
278 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32})
288 for (
auto T : {MVT::v8i16, MVT::v4i32, MVT::v2i64})
293 ISD::FEXP, ISD::FEXP2, ISD::FEXP10})
294 for (
auto T : {MVT::v4f32, MVT::v2f64})
304 for (
auto T : {MVT::v2i64, MVT::v2f64})
310 if (Subtarget->hasFP16()) {
325 if (!Subtarget->hasSignExt()) {
327 auto Action = Subtarget->hasSIMD128() ?
Custom :
Expand;
328 for (
auto T : {MVT::i8, MVT::i16, MVT::i32})
344 for (
auto T : {MVT::i32, MVT::i64, MVT::f32, MVT::f64})
361 if (Subtarget->hasSIMD128()) {
362 for (
auto T : {MVT::v16i8, MVT::v8i16, MVT::v4i32, MVT::v2i64, MVT::v4f32,
365 if (
MVT(
T) != MemT) {
404 return MVT::externref;
413 return MVT::externref;
419bool WebAssemblyTargetLowering::shouldExpandPartialReductionIntrinsic(
421 if (
I->getIntrinsicID() != Intrinsic::vector_partial_reduce_add)
428 auto Op1 =
I->getOperand(1);
455WebAssemblyTargetLowering::shouldExpandAtomicRMWInIR(
AtomicRMWInst *AI)
const {
471bool WebAssemblyTargetLowering::shouldScalarizeBinop(
SDValue VecOp)
const {
491FastISel *WebAssemblyTargetLowering::createFastISel(
496MVT WebAssemblyTargetLowering::getScalarShiftAmountTy(
const DataLayout & ,
507 "32-bit shift counts ought to be enough for anyone");
512 "Unable to represent scalar shift amount type");
522 bool IsUnsigned,
bool Int64,
523 bool Float64,
unsigned LoweredOpcode) {
529 unsigned Abs = Float64 ? WebAssembly::ABS_F64 : WebAssembly::ABS_F32;
530 unsigned FConst = Float64 ? WebAssembly::CONST_F64 : WebAssembly::CONST_F32;
531 unsigned LT = Float64 ? WebAssembly::LT_F64 : WebAssembly::LT_F32;
532 unsigned GE = Float64 ? WebAssembly::GE_F64 : WebAssembly::GE_F32;
533 unsigned IConst =
Int64 ? WebAssembly::CONST_I64 : WebAssembly::CONST_I32;
534 unsigned Eqz = WebAssembly::EQZ_I32;
535 unsigned And = WebAssembly::AND_I32;
537 int64_t Substitute = IsUnsigned ? 0 : Limit;
538 double CmpVal = IsUnsigned ? -(double)Limit * 2.0 : -(double)Limit;
549 F->insert(It, FalseMBB);
550 F->insert(It, TrueMBB);
551 F->insert(It, DoneMBB);
554 DoneMBB->
splice(DoneMBB->
begin(), BB, std::next(
MI.getIterator()), BB->
end());
562 unsigned Tmp0, Tmp1, CmpReg, EqzReg, FalseReg, TrueReg;
563 Tmp0 =
MRI.createVirtualRegister(
MRI.getRegClass(InReg));
564 Tmp1 =
MRI.createVirtualRegister(
MRI.getRegClass(InReg));
565 CmpReg =
MRI.createVirtualRegister(&WebAssembly::I32RegClass);
566 EqzReg =
MRI.createVirtualRegister(&WebAssembly::I32RegClass);
567 FalseReg =
MRI.createVirtualRegister(
MRI.getRegClass(OutReg));
568 TrueReg =
MRI.createVirtualRegister(
MRI.getRegClass(OutReg));
570 MI.eraseFromParent();
584 Tmp1 =
MRI.createVirtualRegister(
MRI.getRegClass(InReg));
586 MRI.createVirtualRegister(&WebAssembly::I32RegClass);
587 Register AndReg =
MRI.createVirtualRegister(&WebAssembly::I32RegClass);
631 unsigned Eqz =
Int64 ? WebAssembly::EQZ_I64 : WebAssembly::EQZ_I32;
632 unsigned MemoryCopy =
633 Int64 ? WebAssembly::MEMORY_COPY_A64 : WebAssembly::MEMORY_COPY_A32;
644 F->insert(It, TrueMBB);
645 F->insert(It, DoneMBB);
648 DoneMBB->
splice(DoneMBB->
begin(), BB, std::next(
MI.getIterator()), BB->
end());
658 EqzReg =
MRI.createVirtualRegister(&WebAssembly::I32RegClass);
661 MI.eraseFromParent();
699 unsigned Eqz =
Int64 ? WebAssembly::EQZ_I64 : WebAssembly::EQZ_I32;
700 unsigned MemoryFill =
701 Int64 ? WebAssembly::MEMORY_FILL_A64 : WebAssembly::MEMORY_FILL_A32;
712 F->insert(It, TrueMBB);
713 F->insert(It, DoneMBB);
716 DoneMBB->
splice(DoneMBB->
begin(), BB, std::next(
MI.getIterator()), BB->
end());
726 EqzReg =
MRI.createVirtualRegister(&WebAssembly::I32RegClass);
729 MI.eraseFromParent();
751 CallResults.
getOpcode() == WebAssembly::RET_CALL_RESULTS);
755 bool IsRetCall = CallResults.
getOpcode() == WebAssembly::RET_CALL_RESULTS;
757 bool IsFuncrefCall =
false;
763 IsFuncrefCall = (TRC == &WebAssembly::FUNCREFRegClass);
768 if (IsIndirect && IsRetCall) {
769 CallOp = WebAssembly::RET_CALL_INDIRECT;
770 }
else if (IsIndirect) {
771 CallOp = WebAssembly::CALL_INDIRECT;
772 }
else if (IsRetCall) {
773 CallOp = WebAssembly::RET_CALL;
775 CallOp = WebAssembly::CALL;
804 for (
auto Def : CallResults.
defs())
828 for (
auto Use : CallParams.
uses())
844 if (IsIndirect && IsFuncrefCall) {
856 BuildMI(MF,
DL,
TII.get(WebAssembly::REF_NULL_FUNCREF), RegFuncref);
860 BuildMI(MF,
DL,
TII.get(WebAssembly::TABLE_SET_FUNCREF))
872 const TargetInstrInfo &
TII = *Subtarget->getInstrInfo();
875 switch (
MI.getOpcode()) {
878 case WebAssembly::FP_TO_SINT_I32_F32:
880 WebAssembly::I32_TRUNC_S_F32);
881 case WebAssembly::FP_TO_UINT_I32_F32:
883 WebAssembly::I32_TRUNC_U_F32);
884 case WebAssembly::FP_TO_SINT_I64_F32:
886 WebAssembly::I64_TRUNC_S_F32);
887 case WebAssembly::FP_TO_UINT_I64_F32:
889 WebAssembly::I64_TRUNC_U_F32);
890 case WebAssembly::FP_TO_SINT_I32_F64:
892 WebAssembly::I32_TRUNC_S_F64);
893 case WebAssembly::FP_TO_UINT_I32_F64:
895 WebAssembly::I32_TRUNC_U_F64);
896 case WebAssembly::FP_TO_SINT_I64_F64:
898 WebAssembly::I64_TRUNC_S_F64);
899 case WebAssembly::FP_TO_UINT_I64_F64:
901 WebAssembly::I64_TRUNC_U_F64);
902 case WebAssembly::MEMCPY_A32:
904 case WebAssembly::MEMCPY_A64:
906 case WebAssembly::MEMSET_A32:
908 case WebAssembly::MEMSET_A64:
910 case WebAssembly::CALL_RESULTS:
911 case WebAssembly::RET_CALL_RESULTS:
917WebAssemblyTargetLowering::getTargetNodeName(
unsigned Opcode)
const {
921#define HANDLE_NODETYPE(NODE) \
922 case WebAssemblyISD::NODE: \
923 return "WebAssemblyISD::" #NODE;
924#include "WebAssemblyISD.def"
925#undef HANDLE_NODETYPE
930std::pair<unsigned, const TargetRegisterClass *>
931WebAssemblyTargetLowering::getRegForInlineAsmConstraint(
935 if (Constraint.
size() == 1) {
936 switch (Constraint[0]) {
938 assert(VT != MVT::iPTR &&
"Pointer MVT not expected here");
939 if (Subtarget->hasSIMD128() && VT.
isVector()) {
941 return std::make_pair(0U, &WebAssembly::V128RegClass);
945 return std::make_pair(0U, &WebAssembly::I32RegClass);
947 return std::make_pair(0U, &WebAssembly::I64RegClass);
952 return std::make_pair(0U, &WebAssembly::F32RegClass);
954 return std::make_pair(0U, &WebAssembly::F64RegClass);
968bool WebAssemblyTargetLowering::isCheapToSpeculateCttz(
Type *Ty)
const {
973bool WebAssemblyTargetLowering::isCheapToSpeculateCtlz(
Type *Ty)
const {
978bool WebAssemblyTargetLowering::isLegalAddressingMode(
const DataLayout &
DL,
980 Type *Ty,
unsigned AS,
996bool WebAssemblyTargetLowering::allowsMisalignedMemoryAccesses(
1010bool WebAssemblyTargetLowering::isIntDivCheap(
EVT VT,
1011 AttributeList Attr)
const {
1017bool WebAssemblyTargetLowering::isVectorLoadExtDesirable(
SDValue ExtVal)
const {
1020 return (ExtT == MVT::v8i16 && MemT == MVT::v8i8) ||
1021 (ExtT == MVT::v4i32 && MemT == MVT::v4i16) ||
1022 (ExtT == MVT::v2i64 && MemT == MVT::v2i32);
1025bool WebAssemblyTargetLowering::isOffsetFoldingLegal(
1028 const GlobalValue *GV = GA->
getGlobal();
1032EVT WebAssemblyTargetLowering::getSetCCResultType(
const DataLayout &
DL,
1045bool WebAssemblyTargetLowering::getTgtMemIntrinsic(IntrinsicInfo &
Info,
1049 switch (Intrinsic) {
1050 case Intrinsic::wasm_memory_atomic_notify:
1052 Info.memVT = MVT::i32;
1053 Info.ptrVal =
I.getArgOperand(0);
1064 case Intrinsic::wasm_memory_atomic_wait32:
1066 Info.memVT = MVT::i32;
1067 Info.ptrVal =
I.getArgOperand(0);
1072 case Intrinsic::wasm_memory_atomic_wait64:
1074 Info.memVT = MVT::i64;
1075 Info.ptrVal =
I.getArgOperand(0);
1080 case Intrinsic::wasm_loadf16_f32:
1082 Info.memVT = MVT::f16;
1083 Info.ptrVal =
I.getArgOperand(0);
1088 case Intrinsic::wasm_storef16_f32:
1090 Info.memVT = MVT::f16;
1091 Info.ptrVal =
I.getArgOperand(1);
1101void WebAssemblyTargetLowering::computeKnownBitsForTargetNode(
1104 switch (
Op.getOpcode()) {
1108 unsigned IntNo =
Op.getConstantOperandVal(0);
1112 case Intrinsic::wasm_bitmask: {
1114 EVT VT =
Op.getOperand(1).getSimpleValueType();
1117 Known.
Zero |= ZeroMask;
1127 case WebAssemblyISD::I64_ADD128:
1128 if (
Op.getResNo() == 1) {
1139WebAssemblyTargetLowering::getPreferredVectorAction(
MVT VT)
const {
1145 if (EltVT == MVT::i8 || EltVT == MVT::i16 || EltVT == MVT::i32 ||
1146 EltVT == MVT::i64 || EltVT == MVT::f32 || EltVT == MVT::f64)
1153bool WebAssemblyTargetLowering::shouldSimplifyDemandedVectorElts(
1154 SDValue Op,
const TargetLoweringOpt &TLO)
const {
1207WebAssemblyTargetLowering::LowerCall(CallLoweringInfo &CLI,
1209 SelectionDAG &DAG = CLI.DAG;
1219 "WebAssembly doesn't support language-specific or target-specific "
1220 "calling conventions yet");
1221 if (CLI.IsPatchPoint)
1222 fail(
DL, DAG,
"WebAssembly doesn't support patch point yet");
1224 if (CLI.IsTailCall) {
1225 auto NoTail = [&](
const char *Msg) {
1226 if (CLI.CB && CLI.CB->isMustTailCall())
1228 CLI.IsTailCall =
false;
1231 if (!Subtarget->hasTailCall())
1232 NoTail(
"WebAssembly 'tail-call' feature not enabled");
1236 NoTail(
"WebAssembly does not support varargs tail calls");
1241 Type *RetTy =
F.getReturnType();
1246 bool TypesMatch = CallerRetTys.
size() == CalleeRetTys.
size() &&
1247 std::equal(CallerRetTys.
begin(), CallerRetTys.
end(),
1248 CalleeRetTys.
begin());
1250 NoTail(
"WebAssembly tail call requires caller and callee return types to "
1255 for (
auto &Arg : CLI.CB->args()) {
1256 Value *Val = Arg.get();
1261 Src =
GEP->getPointerOperand();
1268 "WebAssembly does not support tail calling with stack arguments");
1275 SmallVectorImpl<ISD::InputArg> &
Ins = CLI.Ins;
1276 SmallVectorImpl<ISD::OutputArg> &Outs = CLI.Outs;
1277 SmallVectorImpl<SDValue> &OutVals = CLI.OutVals;
1283 Outs[0].Flags.isSRet()) {
1288 bool HasSwiftSelfArg =
false;
1289 bool HasSwiftErrorArg =
false;
1290 unsigned NumFixedArgs = 0;
1291 for (
unsigned I = 0;
I < Outs.
size(); ++
I) {
1292 const ISD::OutputArg &Out = Outs[
I];
1297 fail(
DL, DAG,
"WebAssembly hasn't implemented nest arguments");
1299 fail(
DL, DAG,
"WebAssembly hasn't implemented inalloca arguments");
1301 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs arguments");
1303 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs last arguments");
1312 Chain = DAG.
getMemcpy(Chain,
DL, FINode, OutVal, SizeNode,
1315 nullptr, std::nullopt, MachinePointerInfo(),
1316 MachinePointerInfo());
1323 bool IsVarArg = CLI.IsVarArg;
1332 if (!HasSwiftSelfArg) {
1334 ISD::ArgFlagsTy
Flags;
1335 Flags.setSwiftSelf();
1336 ISD::OutputArg Arg(Flags, PtrVT, EVT(PtrVT), PtrTy, 0, 0);
1337 CLI.Outs.push_back(Arg);
1339 CLI.OutVals.push_back(ArgVal);
1341 if (!HasSwiftErrorArg) {
1343 ISD::ArgFlagsTy
Flags;
1344 Flags.setSwiftError();
1345 ISD::OutputArg Arg(Flags, PtrVT, EVT(PtrVT), PtrTy, 0, 0);
1346 CLI.Outs.push_back(Arg);
1348 CLI.OutVals.push_back(ArgVal);
1354 CCState CCInfo(CallConv, IsVarArg, MF, ArgLocs, *DAG.
getContext());
1359 for (
unsigned I = NumFixedArgs;
I < Outs.
size(); ++
I) {
1360 const ISD::OutputArg &Out = Outs[
I];
1363 assert(VT != MVT::iPTR &&
"Legalized args should be concrete");
1368 CCInfo.AllocateStack(Layout.getTypeAllocSize(Ty), Alignment);
1375 unsigned NumBytes = CCInfo.getAlignedCallFrameSize();
1378 if (IsVarArg && NumBytes) {
1381 MaybeAlign StackAlign = Layout.getStackAlignment();
1382 assert(StackAlign &&
"data layout string is missing stack alignment");
1388 assert(ArgLocs[ValNo].getValNo() == ValNo &&
1389 "ArgLocs should remain in order and only hold varargs args");
1390 unsigned Offset = ArgLocs[ValNo++].getLocMemOffset();
1398 if (!Chains.
empty())
1400 }
else if (IsVarArg) {
1418 Ops.push_back(Chain);
1419 Ops.push_back(Callee);
1424 IsVarArg ? OutVals.
begin() + NumFixedArgs : OutVals.
end());
1427 Ops.push_back(FINode);
1430 for (
const auto &In : Ins) {
1431 assert(!
In.Flags.isByVal() &&
"byval is not valid for return values");
1432 assert(!
In.Flags.isNest() &&
"nest is not valid for return values");
1433 if (
In.Flags.isInAlloca())
1434 fail(
DL, DAG,
"WebAssembly hasn't implemented inalloca return values");
1435 if (
In.Flags.isInConsecutiveRegs())
1436 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs return values");
1437 if (
In.Flags.isInConsecutiveRegsLast())
1439 "WebAssembly hasn't implemented cons regs last return values");
1448 CLI.CB->getCalledOperand()->getType())) {
1463 WebAssemblyISD::TABLE_SET,
DL, DAG.
getVTList(MVT::Other), TableSetOps,
1468 CLI.CB->getCalledOperand()->getPointerAlignment(DAG.
getDataLayout()),
1474 if (CLI.IsTailCall) {
1476 SDVTList NodeTys = DAG.
getVTList(MVT::Other, MVT::Glue);
1477 return DAG.
getNode(WebAssemblyISD::RET_CALL,
DL, NodeTys,
Ops);
1481 SDVTList InTyList = DAG.
getVTList(InTys);
1484 for (
size_t I = 0;
I <
Ins.size(); ++
I)
1491bool WebAssemblyTargetLowering::CanLowerReturn(
1494 const Type *RetTy)
const {
1499SDValue WebAssemblyTargetLowering::LowerReturn(
1505 "MVP WebAssembly can only return up to one value");
1507 fail(
DL, DAG,
"WebAssembly doesn't support non-C calling conventions");
1510 RetOps.append(OutVals.
begin(), OutVals.
end());
1511 Chain = DAG.
getNode(WebAssemblyISD::RETURN,
DL, MVT::Other, RetOps);
1514 for (
const ISD::OutputArg &Out : Outs) {
1519 fail(
DL, DAG,
"WebAssembly hasn't implemented inalloca results");
1521 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs results");
1523 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs last results");
1529SDValue WebAssemblyTargetLowering::LowerFormalArguments(
1534 fail(
DL, DAG,
"WebAssembly doesn't support non-C calling conventions");
1537 auto *MFI = MF.
getInfo<WebAssemblyFunctionInfo>();
1543 bool HasSwiftErrorArg =
false;
1544 bool HasSwiftSelfArg =
false;
1545 for (
const ISD::InputArg &In : Ins) {
1546 HasSwiftSelfArg |=
In.Flags.isSwiftSelf();
1547 HasSwiftErrorArg |=
In.Flags.isSwiftError();
1548 if (
In.Flags.isInAlloca())
1549 fail(
DL, DAG,
"WebAssembly hasn't implemented inalloca arguments");
1550 if (
In.Flags.isNest())
1551 fail(
DL, DAG,
"WebAssembly hasn't implemented nest arguments");
1552 if (
In.Flags.isInConsecutiveRegs())
1553 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs arguments");
1554 if (
In.Flags.isInConsecutiveRegsLast())
1555 fail(
DL, DAG,
"WebAssembly hasn't implemented cons regs last arguments");
1564 MFI->addParam(
In.VT);
1573 if (!HasSwiftSelfArg) {
1574 MFI->addParam(PtrVT);
1576 if (!HasSwiftErrorArg) {
1577 MFI->addParam(PtrVT);
1586 MFI->setVarargBufferVreg(VarargVreg);
1588 Chain,
DL, VarargVreg,
1589 DAG.
getNode(WebAssemblyISD::ARGUMENT,
DL, PtrVT,
1591 MFI->addParam(PtrVT);
1603 assert(MFI->getParams().size() == Params.
size() &&
1604 std::equal(MFI->getParams().begin(), MFI->getParams().end(),
1610void WebAssemblyTargetLowering::ReplaceNodeResults(
1612 switch (
N->getOpcode()) {
1626 Results.push_back(Replace128Op(
N, DAG));
1630 "ReplaceNodeResults not implemented for this op for WebAssembly!");
1641 switch (
Op.getOpcode()) {
1646 return LowerFrameIndex(
Op, DAG);
1648 return LowerGlobalAddress(
Op, DAG);
1650 return LowerGlobalTLSAddress(
Op, DAG);
1652 return LowerExternalSymbol(
Op, DAG);
1654 return LowerJumpTable(
Op, DAG);
1656 return LowerBR_JT(
Op, DAG);
1658 return LowerVASTART(
Op, DAG);
1661 fail(
DL, DAG,
"WebAssembly hasn't implemented computed gotos");
1664 return LowerRETURNADDR(
Op, DAG);
1666 return LowerFRAMEADDR(
Op, DAG);
1668 return LowerCopyToReg(
Op, DAG);
1671 return LowerAccessVectorElement(
Op, DAG);
1675 return LowerIntrinsic(
Op, DAG);
1677 return LowerSIGN_EXTEND_INREG(
Op, DAG);
1680 return LowerEXTEND_VECTOR_INREG(
Op, DAG);
1682 return LowerBUILD_VECTOR(
Op, DAG);
1684 return LowerVECTOR_SHUFFLE(
Op, DAG);
1686 return LowerSETCC(
Op, DAG);
1690 return LowerShift(
Op, DAG);
1693 return LowerFP_TO_INT_SAT(
Op, DAG);
1695 return LowerLoad(
Op, DAG);
1697 return LowerStore(
Op, DAG);
1706 return LowerMUL_LOHI(
Op, DAG);
1708 return LowerUADDO(
Op, DAG);
1723 return std::nullopt;
1742 SDVTList Tys = DAG.
getVTList(MVT::Other);
1754 SDVTList Tys = DAG.
getVTList(MVT::Other);
1756 return DAG.
getNode(WebAssemblyISD::LOCAL_SET,
DL, Tys,
Ops);
1761 "Encountered an unlowerable store to the wasm_var address space",
1777 "unexpected offset when loading from webassembly global",
false);
1788 "unexpected offset when loading from webassembly local",
false);
1795 assert(
Result->getNumValues() == 2 &&
"Loads must carry a chain!");
1801 "Encountered an unlowerable load from the wasm_var address space",
1809 assert(Subtarget->hasWideArithmetic());
1810 assert(
Op.getValueType() == MVT::i64);
1813 switch (
Op.getOpcode()) {
1815 Opcode = WebAssemblyISD::I64_MUL_WIDE_U;
1818 Opcode = WebAssemblyISD::I64_MUL_WIDE_S;
1839 assert(Subtarget->hasWideArithmetic());
1840 assert(
Op.getValueType() == MVT::i64);
1847 DAG.
getNode(WebAssemblyISD::I64_ADD128,
DL,
1857 assert(Subtarget->hasWideArithmetic());
1858 assert(
N->getValueType(0) == MVT::i128);
1861 switch (
N->getOpcode()) {
1863 Opcode = WebAssemblyISD::I64_ADD128;
1866 Opcode = WebAssemblyISD::I64_SUB128;
1881 LHS_0, LHS_1, RHS_0, RHS_1);
1898 EVT VT = Src.getValueType();
1900 : WebAssembly::COPY_I64,
1903 return Op.getNode()->getNumValues() == 1
1922 if (!Subtarget->getTargetTriple().isOSEmscripten()) {
1924 "Non-Emscripten WebAssembly hasn't implemented "
1925 "__builtin_return_address");
1929 unsigned Depth =
Op.getConstantOperandVal(0);
1931 return makeLibCall(DAG, RTLIB::RETURN_ADDRESS,
Op.getValueType(),
1932 {DAG.getConstant(Depth, DL, MVT::i32)}, CallOptions,
DL)
1941 if (
Op.getConstantOperandVal(0) > 0)
1945 EVT VT =
Op.getValueType();
1952WebAssemblyTargetLowering::LowerGlobalTLSAddress(
SDValue Op,
1958 if (!MF.
getSubtarget<WebAssemblySubtarget>().hasBulkMemory())
1962 const GlobalValue *GV = GA->
getGlobal();
1967 auto model = Subtarget->getTargetTriple().isOSEmscripten()
1982 auto GlobalGet = PtrVT == MVT::i64 ? WebAssembly::GLOBAL_GET_I64
1983 : WebAssembly::GLOBAL_GET_I32;
1994 DAG.
getNode(WebAssemblyISD::WrapperREL,
DL, PtrVT, TLSOffset);
2001 EVT VT =
Op.getValueType();
2002 return DAG.
getNode(WebAssemblyISD::Wrapper,
DL, VT,
2012 EVT VT =
Op.getValueType();
2014 "Unexpected target flags on generic GlobalAddressSDNode");
2016 fail(
DL, DAG,
"Invalid address space for WebAssembly target");
2019 const GlobalValue *GV = GA->
getGlobal();
2027 const char *BaseName;
2036 DAG.
getNode(WebAssemblyISD::Wrapper,
DL, PtrVT,
2040 WebAssemblyISD::WrapperREL,
DL, VT,
2049 return DAG.
getNode(WebAssemblyISD::Wrapper,
DL, VT,
2055WebAssemblyTargetLowering::LowerExternalSymbol(
SDValue Op,
2059 EVT VT =
Op.getValueType();
2060 assert(ES->getTargetFlags() == 0 &&
2061 "Unexpected target flags on generic ExternalSymbolSDNode");
2062 return DAG.
getNode(WebAssemblyISD::Wrapper,
DL, VT,
2073 JT->getTargetFlags());
2082 assert(
JT->getTargetFlags() == 0 &&
"WebAssembly doesn't set target flags");
2085 Ops.push_back(Chain);
2086 Ops.push_back(Index);
2092 for (
auto *
MBB : MBBs)
2099 return DAG.
getNode(WebAssemblyISD::BR_TABLE,
DL, MVT::Other,
Ops);
2111 MFI->getVarargBufferVreg(), PtrVT);
2112 return DAG.
getStore(
Op.getOperand(0),
DL, ArgN,
Op.getOperand(1),
2113 MachinePointerInfo(SV));
2120 if (
N->getConstantOperandVal(0) != Intrinsic::vector_partial_reduce_add)
2123 assert(
N->getValueType(0) == MVT::v4i32 &&
"can only support v4i32");
2132 "expected widening mul or add");
2134 "expected binop to use the same extend for both operands");
2140 IsSigned ? WebAssemblyISD::EXTEND_LOW_S : WebAssemblyISD::EXTEND_LOW_U;
2141 unsigned HighOpc = IsSigned ? WebAssemblyISD::EXTEND_HIGH_S
2142 : WebAssemblyISD::EXTEND_HIGH_U;
2148 auto AssignInputs = [&](
MVT VT) {
2149 LowLHS = DAG.
getNode(LowOpc,
DL, VT, ExtendInLHS);
2150 LowRHS = DAG.
getNode(LowOpc,
DL, VT, ExtendInRHS);
2151 HighLHS = DAG.
getNode(HighOpc,
DL, VT, ExtendInLHS);
2152 HighRHS = DAG.
getNode(HighOpc,
DL, VT, ExtendInRHS);
2159 ExtendInLHS, ExtendInRHS);
2164 MVT VT = MVT::v4i32;
2172 "expected v16i8 input types");
2173 AssignInputs(MVT::v8i16);
2178 DAG.
getNode(WebAssemblyISD::DOT,
DL, MVT::v4i32, LowLHS, LowRHS);
2180 DAG.
getNode(WebAssemblyISD::DOT,
DL, MVT::v4i32, HighLHS, HighRHS);
2189 MVT::v4i32, MulLow);
2191 MVT::v4i32, MulHigh);
2199 unsigned PairwiseOpc = IsSigned ? WebAssemblyISD::EXT_ADD_PAIRWISE_S
2200 : WebAssemblyISD::EXT_ADD_PAIRWISE_U;
2208 "expected v16i8 input types");
2210 DAG.
getNode(PairwiseOpc,
DL, MVT::v4i32,
2211 DAG.
getNode(PairwiseOpc,
DL, MVT::v8i16, ExtendIn));
2220 switch (
Op.getOpcode()) {
2223 IntNo =
Op.getConstantOperandVal(1);
2226 IntNo =
Op.getConstantOperandVal(0);
2237 case Intrinsic::wasm_lsda: {
2246 DAG.
getNode(WebAssemblyISD::Wrapper,
DL, PtrVT,
2249 DAG.
getNode(WebAssemblyISD::WrapperREL,
DL, PtrVT, Node);
2253 return DAG.
getNode(WebAssemblyISD::Wrapper,
DL, PtrVT, Node);
2256 case Intrinsic::wasm_shuffle: {
2262 while (
OpIdx < 18) {
2271 return DAG.
getNode(WebAssemblyISD::SHUFFLE,
DL,
Op.getValueType(),
Ops);
2274 case Intrinsic::thread_pointer: {
2276 auto GlobalGet = PtrVT == MVT::i64 ? WebAssembly::GLOBAL_GET_I64
2277 : WebAssembly::GLOBAL_GET_I32;
2288WebAssemblyTargetLowering::LowerSIGN_EXTEND_INREG(
SDValue Op,
2298 assert(!Subtarget->hasSignExt() && Subtarget->hasSIMD128());
2302 const SDValue &Extract =
Op.getOperand(0);
2306 MVT ExtractedLaneT =
2310 if (ExtractedVecT == VecT)
2317 unsigned IndexVal =
Index->getAsZExtVal();
2335 assert((UserOpc == WebAssemblyISD::EXTEND_LOW_U ||
2336 UserOpc == WebAssemblyISD::EXTEND_LOW_S) &&
2337 "expected extend_low");
2342 size_t FirstIdx = Mask.size() / 2;
2343 for (
size_t i = 0; i < Mask.size() / 2; ++i) {
2344 if (Mask[i] !=
static_cast<int>(FirstIdx + i)) {
2350 unsigned Opc = UserOpc == WebAssemblyISD::EXTEND_LOW_S
2351 ? WebAssemblyISD::EXTEND_HIGH_S
2352 : WebAssemblyISD::EXTEND_HIGH_U;
2353 return DAG.
getNode(
Opc,
DL, VT, Shuffle->getOperand(0));
2357WebAssemblyTargetLowering::LowerEXTEND_VECTOR_INREG(
SDValue Op,
2360 EVT VT =
Op.getValueType();
2362 EVT SrcVT = Src.getValueType();
2369 "Unexpected extension factor.");
2372 if (Scale != 2 && Scale != 4 && Scale != 8)
2376 switch (
Op.getOpcode()) {
2378 Ext = WebAssemblyISD::EXTEND_LOW_U;
2381 Ext = WebAssemblyISD::EXTEND_LOW_S;
2392 while (Scale != 1) {
2395 .widenIntegerVectorElementType(*DAG.
getContext())
2396 .getHalfNumVectorElementsVT(*DAG.
getContext()),
2406 if (
Op.getValueType() != MVT::v2f64)
2410 unsigned &Index) ->
bool {
2411 switch (
Op.getOpcode()) {
2413 Opcode = WebAssemblyISD::CONVERT_LOW_S;
2416 Opcode = WebAssemblyISD::CONVERT_LOW_U;
2418 case ISD::FP_EXTEND:
2419 Opcode = WebAssemblyISD::PROMOTE_LOW;
2425 auto ExtractVector =
Op.getOperand(0);
2432 SrcVec = ExtractVector.getOperand(0);
2433 Index = ExtractVector.getConstantOperandVal(1);
2437 unsigned LHSOpcode, RHSOpcode, LHSIndex, RHSIndex;
2439 if (!GetConvertedLane(
Op.getOperand(0), LHSOpcode, LHSSrcVec, LHSIndex) ||
2440 !GetConvertedLane(
Op.getOperand(1), RHSOpcode, RHSSrcVec, RHSIndex))
2443 if (LHSOpcode != RHSOpcode)
2447 switch (LHSOpcode) {
2448 case WebAssemblyISD::CONVERT_LOW_S:
2449 case WebAssemblyISD::CONVERT_LOW_U:
2450 ExpectedSrcVT = MVT::v4i32;
2452 case WebAssemblyISD::PROMOTE_LOW:
2453 ExpectedSrcVT = MVT::v4f32;
2459 auto Src = LHSSrcVec;
2460 if (LHSIndex != 0 || RHSIndex != 1 || LHSSrcVec != RHSSrcVec) {
2463 ExpectedSrcVT,
DL, LHSSrcVec, RHSSrcVec,
2464 {
static_cast<int>(LHSIndex),
static_cast<int>(RHSIndex) + 4, -1, -1});
2466 return DAG.
getNode(LHSOpcode,
DL, MVT::v2f64, Src);
2471 MVT VT =
Op.getSimpleValueType();
2472 if (VT == MVT::v8f16) {
2487 const EVT VecT =
Op.getValueType();
2488 const EVT LaneT =
Op.getOperand(0).getValueType();
2490 bool CanSwizzle = VecT == MVT::v16i8;
2511 auto GetSwizzleSrcs = [](
size_t I,
const SDValue &Lane) {
2515 const SDValue &SwizzleSrc = Lane->getOperand(0);
2516 const SDValue &IndexExt = Lane->getOperand(1);
2526 Index->getConstantOperandVal(1) !=
I)
2528 return std::make_pair(SwizzleSrc, SwizzleIndices);
2535 auto GetShuffleSrc = [&](
const SDValue &Lane) {
2540 if (Lane->getOperand(0).getValueType().getVectorNumElements() >
2543 return Lane->getOperand(0);
2546 using ValueEntry = std::pair<SDValue, size_t>;
2549 using SwizzleEntry = std::pair<std::pair<SDValue, SDValue>,
size_t>;
2552 using ShuffleEntry = std::pair<SDValue, size_t>;
2555 auto AddCount = [](
auto &Counts,
const auto &Val) {
2558 if (CountIt == Counts.end()) {
2559 Counts.emplace_back(Val, 1);
2565 auto GetMostCommon = [](
auto &Counts) {
2567 assert(CommonIt != Counts.end() &&
"Unexpected all-undef build_vector");
2571 size_t NumConstantLanes = 0;
2574 for (
size_t I = 0;
I < Lanes; ++
I) {
2579 AddCount(SplatValueCounts, Lane);
2583 if (
auto ShuffleSrc = GetShuffleSrc(Lane))
2584 AddCount(ShuffleCounts, ShuffleSrc);
2586 auto SwizzleSrcs = GetSwizzleSrcs(
I, Lane);
2587 if (SwizzleSrcs.first)
2588 AddCount(SwizzleCounts, SwizzleSrcs);
2593 size_t NumSplatLanes;
2594 std::tie(SplatValue, NumSplatLanes) = GetMostCommon(SplatValueCounts);
2598 size_t NumSwizzleLanes = 0;
2599 if (SwizzleCounts.
size())
2600 std::forward_as_tuple(std::tie(SwizzleSrc, SwizzleIndices),
2601 NumSwizzleLanes) = GetMostCommon(SwizzleCounts);
2605 SDValue ShuffleSrc1, ShuffleSrc2;
2606 size_t NumShuffleLanes = 0;
2607 if (ShuffleCounts.
size()) {
2608 std::tie(ShuffleSrc1, NumShuffleLanes) = GetMostCommon(ShuffleCounts);
2610 [&](
const auto &Pair) {
return Pair.first == ShuffleSrc1; });
2612 if (ShuffleCounts.
size()) {
2613 size_t AdditionalShuffleLanes;
2614 std::tie(ShuffleSrc2, AdditionalShuffleLanes) =
2615 GetMostCommon(ShuffleCounts);
2616 NumShuffleLanes += AdditionalShuffleLanes;
2621 std::function<bool(
size_t,
const SDValue &)> IsLaneConstructed;
2624 if (NumSwizzleLanes >= NumShuffleLanes &&
2625 NumSwizzleLanes >= NumConstantLanes && NumSwizzleLanes >= NumSplatLanes) {
2628 auto Swizzled = std::make_pair(SwizzleSrc, SwizzleIndices);
2629 IsLaneConstructed = [&, Swizzled](
size_t I,
const SDValue &Lane) {
2630 return Swizzled == GetSwizzleSrcs(
I, Lane);
2632 }
else if (NumShuffleLanes >= NumConstantLanes &&
2633 NumShuffleLanes >= NumSplatLanes) {
2643 assert(LaneSize > DestLaneSize);
2644 Scale1 = LaneSize / DestLaneSize;
2650 assert(LaneSize > DestLaneSize);
2651 Scale2 = LaneSize / DestLaneSize;
2656 assert(DestLaneCount <= 16);
2657 for (
size_t I = 0;
I < DestLaneCount; ++
I) {
2659 SDValue Src = GetShuffleSrc(Lane);
2660 if (Src == ShuffleSrc1) {
2662 }
else if (Src && Src == ShuffleSrc2) {
2668 ArrayRef<int> MaskRef(Mask, DestLaneCount);
2670 IsLaneConstructed = [&](size_t,
const SDValue &Lane) {
2671 auto Src = GetShuffleSrc(Lane);
2672 return Src == ShuffleSrc1 || (Src && Src == ShuffleSrc2);
2674 }
else if (NumConstantLanes >= NumSplatLanes) {
2676 for (
const SDValue &Lane :
Op->op_values()) {
2685 int64_t Val =
Const ?
Const->getSExtValue() : 0;
2686 uint64_t LaneBits = 128 / Lanes;
2687 assert((LaneBits == 64 || Val >= -(1ll << (LaneBits - 1))) &&
2688 "Unexpected out of bounds negative value");
2689 if (Const && LaneBits != 64 && Val > (1ll << (LaneBits - 1)) - 1) {
2690 uint64_t
Mask = (1ll << LaneBits) - 1;
2691 auto NewVal = (((uint64_t)Val &
Mask) - (1ll << LaneBits)) & Mask;
2708 if (NumSplatLanes == 1 &&
Op->getOperand(0) == SplatValue &&
2709 (DestLaneSize == 32 || DestLaneSize == 64)) {
2716 IsLaneConstructed = [&SplatValue](
size_t _,
const SDValue &Lane) {
2717 return Lane == SplatValue;
2722 assert(IsLaneConstructed);
2725 for (
size_t I = 0;
I < Lanes; ++
I) {
2727 if (!Lane.
isUndef() && !IsLaneConstructed(
I, Lane))
2736WebAssemblyTargetLowering::LowerVECTOR_SHUFFLE(
SDValue Op,
2740 MVT VecType =
Op.getOperand(0).getSimpleValueType();
2751 for (
int M : Mask) {
2752 for (
size_t J = 0; J < LaneBytes; ++J) {
2756 uint64_t ByteIndex =
M == -1 ? J : (uint64_t)M * LaneBytes + J;
2761 return DAG.
getNode(WebAssemblyISD::SHUFFLE,
DL,
Op.getValueType(),
Ops);
2769 assert(
Op->getOperand(0)->getSimpleValueType(0) == MVT::v2i64);
2774 auto MakeLane = [&](
unsigned I) {
2780 {MakeLane(0), MakeLane(1)});
2784WebAssemblyTargetLowering::LowerAccessVectorElement(
SDValue Op,
2801 EVT LaneT =
Op.getSimpleValueType().getVectorElementType();
2803 if (LaneT.
bitsGE(MVT::i32))
2807 size_t NumLanes =
Op.getSimpleValueType().getVectorNumElements();
2809 unsigned ShiftOpcode =
Op.getOpcode();
2815 for (
size_t i = 0; i < NumLanes; ++i) {
2818 SDValue ShiftedValue = ShiftedElements[i];
2823 DAG.
getNode(ShiftOpcode,
DL, MVT::i32, ShiftedValue, MaskedShiftValue));
2833 assert(
Op.getSimpleValueType().isVector());
2835 uint64_t LaneBits =
Op.getValueType().getScalarSizeInBits();
2836 auto ShiftVal =
Op.getOperand(1);
2839 auto SkipImpliedMask = [](
SDValue MaskOp, uint64_t MaskBits) {
2850 MaskVal == MaskBits)
2857 if (ConstantRHS && ConstantRHS->getAPIntValue() == MaskBits)
2865 ShiftVal = SkipImpliedMask(ShiftVal, LaneBits - 1);
2871 ShiftVal = SkipImpliedMask(ShiftVal, LaneBits - 1);
2876 switch (
Op.getOpcode()) {
2878 Opcode = WebAssemblyISD::VEC_SHL;
2881 Opcode = WebAssemblyISD::VEC_SHR_S;
2884 Opcode = WebAssemblyISD::VEC_SHR_U;
2890 return DAG.
getNode(Opcode,
DL,
Op.getValueType(),
Op.getOperand(0), ShiftVal);
2895 EVT ResT =
Op.getValueType();
2898 if ((ResT == MVT::i32 || ResT == MVT::i64) &&
2899 (SatVT == MVT::i32 || SatVT == MVT::i64))
2902 if (ResT == MVT::v4i32 && SatVT == MVT::i32)
2905 if (ResT == MVT::v8i16 && SatVT == MVT::i16)
2916 auto &DAG = DCI.
DAG;
2923 SDValue Bitcast =
N->getOperand(0);
2924 if (Bitcast.getOpcode() != ISD::BITCAST)
2926 if (!
N->getOperand(1).isUndef())
2928 SDValue CastOp = Bitcast.getOperand(0);
2930 EVT DstType = Bitcast.getValueType();
2931 if (!SrcType.is128BitVector() ||
2932 SrcType.getVectorNumElements() != DstType.getVectorNumElements())
2935 SrcType,
SDLoc(
N), CastOp, DAG.
getUNDEF(SrcType), Shuffle->getMask());
2945 auto &DAG = DCI.
DAG;
2949 EVT InVT =
N->getOperand(0)->getValueType(0);
2950 EVT ResVT =
N->getValueType(0);
2952 if (ResVT == MVT::v4f32 && (InVT == MVT::v4i16 || InVT == MVT::v4i8))
2954 else if (ResVT == MVT::v2f64 && (InVT == MVT::v2i16 || InVT == MVT::v2i8))
2968 auto &DAG = DCI.
DAG;
2972 EVT VT =
N->getValueType(0);
2986 auto &DAG = DCI.
DAG;
2992 auto Extract =
N->getOperand(0);
2997 if (IndexNode ==
nullptr)
2999 auto Index = IndexNode->getZExtValue();
3003 EVT ResVT =
N->getValueType(0);
3004 if (ResVT == MVT::v8i16) {
3006 Source.getValueType() != MVT::v16i8 || (Index != 0 && Index != 8))
3008 }
else if (ResVT == MVT::v4i32) {
3010 Source.getValueType() != MVT::v8i16 || (Index != 0 && Index != 4))
3012 }
else if (ResVT == MVT::v2i64) {
3014 Source.getValueType() != MVT::v4i32 || (Index != 0 && Index != 2))
3021 bool IsLow = Index == 0;
3023 unsigned Op = IsSext ? (IsLow ? WebAssemblyISD::EXTEND_LOW_S
3024 : WebAssemblyISD::EXTEND_HIGH_S)
3025 : (IsLow ? WebAssemblyISD::EXTEND_LOW_U
3026 : WebAssemblyISD::EXTEND_HIGH_U);
3033 auto &DAG = DCI.
DAG;
3035 auto GetWasmConversionOp = [](
unsigned Op) {
3038 return WebAssemblyISD::TRUNC_SAT_ZERO_S;
3040 return WebAssemblyISD::TRUNC_SAT_ZERO_U;
3042 return WebAssemblyISD::DEMOTE_ZERO;
3047 auto IsZeroSplat = [](
SDValue SplatVal) {
3049 APInt SplatValue, SplatUndef;
3050 unsigned SplatBitSize;
3055 Splat->isConstantSplat(SplatValue, SplatUndef, SplatBitSize,
3073 EVT ExpectedConversionType;
3076 switch (ConversionOp) {
3080 ExpectedConversionType = MVT::v2i32;
3084 ExpectedConversionType = MVT::v2f32;
3090 if (
N->getValueType(0) != ResVT)
3093 if (
Conversion.getValueType() != ExpectedConversionType)
3097 if (Source.getValueType() != MVT::v2f64)
3100 if (!IsZeroSplat(
N->getOperand(1)) ||
3101 N->getOperand(1).getValueType() != ExpectedConversionType)
3104 unsigned Op = GetWasmConversionOp(ConversionOp);
3120 auto ConversionOp =
N->getOpcode();
3121 switch (ConversionOp) {
3133 if (
N->getValueType(0) != ResVT)
3136 auto Concat =
N->getOperand(0);
3137 if (
Concat.getValueType() != MVT::v4f64)
3140 auto Source =
Concat.getOperand(0);
3141 if (Source.getValueType() != MVT::v2f64)
3144 if (!IsZeroSplat(
Concat.getOperand(1)) ||
3145 Concat.getOperand(1).getValueType() != MVT::v2f64)
3148 unsigned Op = GetWasmConversionOp(ConversionOp);
3154 const SDLoc &
DL,
unsigned VectorWidth) {
3162 unsigned ElemsPerChunk = VectorWidth / ElVT.
getSizeInBits();
3167 IdxVal &= ~(ElemsPerChunk - 1);
3172 Vec->
ops().slice(IdxVal, ElemsPerChunk));
3184 EVT SrcVT = In.getValueType();
3202 EVT InVT = MVT::i16, OutVT = MVT::i8;
3207 unsigned SubSizeInBits = SrcSizeInBits / 2;
3209 OutVT =
EVT::getVectorVT(Ctx, OutVT, SubSizeInBits / OutVT.getSizeInBits());
3235 auto &DAG = DCI.
DAG;
3238 EVT InVT = In.getValueType();
3242 EVT OutVT =
N->getValueType(0);
3249 if (!((InSVT == MVT::i16 || InSVT == MVT::i32 || InSVT == MVT::i64) &&
3250 (OutSVT == MVT::i8 || OutSVT == MVT::i16) && OutVT.
is128BitVector()))
3263 auto &DAG = DCI.
DAG;
3266 EVT VT =
N->getValueType(0);
3267 EVT SrcVT = Src.getValueType();
3278 if (NumElts == 2 || NumElts == 4 || NumElts == 8 || NumElts == 16) {
3281 {DAG.getConstant(Intrinsic::wasm_bitmask, DL, MVT::i32),
3282 DAG.getSExtOrTrunc(N->getOperand(0), DL,
3283 SrcVT.changeVectorElementType(Width))}),
3288 if (NumElts == 32 || NumElts == 64) {
3314 MVT ReturnType = VectorsToShuffle.
size() == 2 ? MVT::i32 : MVT::i64;
3317 for (
SDValue V : VectorsToShuffle) {
3318 ReturningInteger = DAG.
getNode(
3327 return ReturningInteger;
3345 EVT LT =
LHS.getValueType();
3346 if (LT.getScalarSizeInBits() > 128 / LT.getVectorNumElements())
3349 auto CombineSetCC = [&
N, &DAG](Intrinsic::WASMIntrinsics InPre,
3351 Intrinsic::WASMIntrinsics InPost) {
3352 if (
N->getConstantOperandVal(0) != InPre)
3363 {DAG.getConstant(InPost, DL, MVT::i32), LHS}),
3366 Ret = DAG.
getNOT(
DL, Ret, MVT::i1);
3371 Intrinsic::wasm_alltrue))
3374 Intrinsic::wasm_anytrue))
3377 Intrinsic::wasm_anytrue))
3380 Intrinsic::wasm_alltrue))
3386template <
int MatchRHS,
ISD::CondCode MatchCond,
bool RequiresNegate,
3401 {DAG.getConstant(Intrin, DL, MVT::i32),
3402 DAG.getSExtOrTrunc(LHS->getOperand(0), DL, VecVT)}),
3405 Ret = DAG.
getNOT(
DL, Ret, MVT::i1);
3418 EVT VT =
N->getValueType(0);
3419 EVT OpVT =
X.getValueType();
3423 Attribute::NoImplicitFloat))
3429 !Subtarget->
hasSIMD128() || !isIntEqualitySetCC(CC))
3433 auto IsVectorBitCastCheap = [](
SDValue X) {
3438 if (!IsVectorBitCastCheap(
X) || !IsVectorBitCastCheap(
Y))
3448 : Intrinsic::wasm_anytrue,
3462 EVT VT =
N->getValueType(0);
3470 if (
LHS->getOpcode() != ISD::BITCAST)
3473 EVT FromVT =
LHS->getOperand(0).getValueType();
3478 if (NumElts != 2 && NumElts != 4 && NumElts != 8 && NumElts != 16)
3485 auto &DAG = DCI.
DAG;
3514 EVT VT =
N->getValueType(0);
3515 if (VT != MVT::v8i32 && VT != MVT::v16i32)
3521 if (
LHS.getOpcode() !=
RHS.getOpcode())
3528 if (
LHS->getOperand(0).getValueType() !=
RHS->getOperand(0).getValueType())
3531 EVT FromVT =
LHS->getOperand(0).getValueType();
3533 if (EltTy != MVT::i8)
3561 unsigned ExtendLowOpc =
3562 IsSigned ? WebAssemblyISD::EXTEND_LOW_S : WebAssemblyISD::EXTEND_LOW_U;
3563 unsigned ExtendHighOpc =
3564 IsSigned ? WebAssemblyISD::EXTEND_HIGH_S : WebAssemblyISD::EXTEND_HIGH_U;
3566 auto GetExtendLow = [&DAG, &
DL, &ExtendLowOpc](
EVT VT,
SDValue Op) {
3573 if (NumElts == 16) {
3574 SDValue LowLHS = GetExtendLow(MVT::v8i16, ExtendInLHS);
3575 SDValue LowRHS = GetExtendLow(MVT::v8i16, ExtendInRHS);
3581 GetExtendLow(MVT::v4i32, MulLow),
3583 GetExtendLow(MVT::v4i32, MulHigh),
3592 SDValue Lo = GetExtendLow(MVT::v4i32, MulLow);
3602 EVT VT =
N->getValueType(0);
3619 EVT MulVT = MVT::v8i16;
3621 if (VT == MVT::v8i8) {
3627 DAG.
getNode(WebAssemblyISD::EXTEND_LOW_U,
DL, MulVT, PromotedLHS);
3629 DAG.
getNode(WebAssemblyISD::EXTEND_LOW_U,
DL, MulVT, PromotedRHS);
3634 MVT::v16i8,
DL, MulLow, DAG.
getUNDEF(MVT::v16i8),
3635 {0, 2, 4, 6, 8, 10, 12, 14, -1, -1, -1, -1, -1, -1, -1, -1});
3638 assert(VT == MVT::v16i8 &&
"Expected v16i8");
3642 DAG.
getNode(WebAssemblyISD::EXTEND_HIGH_U,
DL, MulVT,
LHS);
3644 DAG.
getNode(WebAssemblyISD::EXTEND_HIGH_U,
DL, MulVT,
RHS);
3653 VT,
DL, MulLow, MulHigh,
3654 {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30});
3659WebAssemblyTargetLowering::PerformDAGCombine(
SDNode *
N,
3660 DAGCombinerInfo &DCI)
const {
3661 switch (
N->getOpcode()) {
3688 return AnyAllCombine;
unsigned const MachineRegisterInfo * MRI
static SDValue performMulCombine(SDNode *N, SelectionDAG &DAG, TargetLowering::DAGCombinerInfo &DCI, const AArch64Subtarget *Subtarget)
static SDValue performTruncateCombine(SDNode *N, SelectionDAG &DAG, TargetLowering::DAGCombinerInfo &DCI)
static SDValue performSETCCCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI, SelectionDAG &DAG)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis false
Function Alias Analysis Results
static void fail(const SDLoc &DL, SelectionDAG &DAG, const Twine &Msg, SDValue Val={})
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
Analysis containing CSE Info
const HexagonInstrInfo * TII
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
Register const TargetRegisterInfo * TRI
Promote Memory to Register
MachineInstr unsigned OpIdx
static SDValue performVECTOR_SHUFFLECombine(SDNode *N, SelectionDAG &DAG, const RISCVSubtarget &Subtarget, const RISCVTargetLowering &TLI)
static SDValue combineVectorSizedSetCCEquality(EVT VT, SDValue X, SDValue Y, ISD::CondCode CC, const SDLoc &DL, SelectionDAG &DAG, const RISCVSubtarget &Subtarget)
Try to map an integer comparison with size > XLEN to vector instructions before type legalization spl...
const SmallVectorImpl< MachineOperand > & Cond
Contains matchers for matching SelectionDAG nodes and values.
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static MachineBasicBlock * LowerFPToInt(MachineInstr &MI, DebugLoc DL, MachineBasicBlock *BB, const TargetInstrInfo &TII, bool IsUnsigned, bool Int64, bool Float64, unsigned LoweredOpcode)
static bool callingConvSupported(CallingConv::ID CallConv)
static SDValue TryWideExtMulCombine(SDNode *N, SelectionDAG &DAG)
static MachineBasicBlock * LowerMemcpy(MachineInstr &MI, DebugLoc DL, MachineBasicBlock *BB, const TargetInstrInfo &TII, bool Int64)
static std::optional< unsigned > IsWebAssemblyLocal(SDValue Op, SelectionDAG &DAG)
static SDValue performVectorExtendCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI)
static SDValue performVectorNonNegToFPCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI)
static SDValue unrollVectorShift(SDValue Op, SelectionDAG &DAG)
static SDValue performAnyAllCombine(SDNode *N, SelectionDAG &DAG)
static MachineBasicBlock * LowerCallResults(MachineInstr &CallResults, DebugLoc DL, MachineBasicBlock *BB, const WebAssemblySubtarget *Subtarget, const TargetInstrInfo &TII)
static SDValue TryMatchTrue(SDNode *N, EVT VecVT, SelectionDAG &DAG)
static SDValue GetExtendHigh(SDValue Op, unsigned UserOpc, EVT VT, SelectionDAG &DAG)
static SDValue performVectorTruncZeroCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI)
static bool IsWebAssemblyGlobal(SDValue Op)
static MachineBasicBlock * LowerMemset(MachineInstr &MI, DebugLoc DL, MachineBasicBlock *BB, const TargetInstrInfo &TII, bool Int64)
static SDValue performVectorExtendToFPCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI)
Convert ({u,s}itofp vec) --> ({u,s}itofp ({s,z}ext vec)) so it doesn't get split up into scalar instr...
static SDValue LowerConvertLow(SDValue Op, SelectionDAG &DAG)
SDValue performLowerPartialReduction(SDNode *N, SelectionDAG &DAG)
static SDValue extractSubVector(SDValue Vec, unsigned IdxVal, SelectionDAG &DAG, const SDLoc &DL, unsigned VectorWidth)
static SDValue performBitcastCombine(SDNode *N, TargetLowering::DAGCombinerInfo &DCI)
static SDValue truncateVectorWithNARROW(EVT DstVT, SDValue In, const SDLoc &DL, SelectionDAG &DAG)
This file defines the interfaces that WebAssembly uses to lower LLVM code into a selection DAG.
This file provides WebAssembly-specific target descriptions.
This file declares WebAssembly-specific per-machine-function information.
This file declares the WebAssembly-specific subclass of TargetSubtarget.
This file declares the WebAssembly-specific subclass of TargetMachine.
This file contains the declaration of the WebAssembly-specific type parsing utility functions.
This file contains the declaration of the WebAssembly-specific utility functions.
static constexpr int Concat[]
Class for arbitrary precision integers.
void setBitsFrom(unsigned loBit)
Set the top bits starting from loBit.
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
static APInt getHighBitsSet(unsigned numBits, unsigned hiBitsSet)
Constructs an APInt value that has the top hiBitsSet bits set.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
an instruction that atomically reads a memory location, combines it with another value,...
BinOp getOperation() const
LLVM Basic Block Representation.
static CCValAssign getMem(unsigned ValNo, MVT ValVT, int64_t Offset, MVT LocVT, LocInfo HTP, bool IsCustom=false)
This class represents a function call, abstracting a target machine's calling convention.
uint64_t getNumOperands() const
A parsed version of the target data layout string in and methods for querying it.
Diagnostic information for unsupported feature in backend.
This is a fast-path instruction selection class that generates poor code and doesn't support illegal ...
FunctionLoweringInfo - This contains information that is global to a function that is used when lower...
FunctionType * getFunctionType() const
Returns the FunctionType for me.
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
int64_t getOffset() const
LLVM_ABI unsigned getAddressSpace() const
unsigned getTargetFlags() const
const GlobalValue * getGlobal() const
ThreadLocalMode getThreadLocalMode() const
Type * getValueType() const
A wrapper class for inspecting calls to intrinsic functions.
This is an important class for using LLVM in a threaded context.
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
const SDValue & getBasePtr() const
const SDValue & getOffset() const
Describe properties that are true of each instruction in the target description file.
bool is128BitVector() const
Return true if this is a 128-bit vector type.
@ INVALID_SIMPLE_VALUE_TYPE
static auto integer_fixedlen_vector_valuetypes()
MVT changeVectorElementType(MVT EltVT) const
Return a VT for a vector type whose attributes match ourselves with the exception of the element type...
unsigned getVectorNumElements() const
bool isVector() const
Return true if this is a vector value type.
bool isInteger() const
Return true if this is an integer or a vector integer type.
static auto integer_valuetypes()
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
static auto fixedlen_vector_valuetypes()
bool isFixedLengthVector() const
static MVT getVectorVT(MVT VT, unsigned NumElements)
MVT getVectorElementType() const
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
static MVT getIntegerVT(unsigned BitWidth)
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
iterator insertAfter(iterator I, MachineInstr *MI)
Insert MI into the instruction list after I.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
void setFrameAddressIsTaken(bool T)
unsigned getFunctionNumber() const
getFunctionNumber - Return a unique ID for the current function.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
const char * createExternalSymbolName(StringRef Name)
Allocate a string and populate it with the given external symbol name.
MCContext & getContext() const
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addFPImm(const ConstantFP *Val) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
MachineInstr * getInstr() const
If conversion operators fail, use this method to get the MachineInstr explicitly.
Representation of each machine instruction.
mop_range defs()
Returns all explicit operands that are register definitions.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
LLVM_ABI void addOperand(MachineFunction &MF, const MachineOperand &Op)
Add the specified operand to the instruction.
mop_range uses()
Returns all operands which may be register uses.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
const MachineOperand & getOperand(unsigned i) const
const std::vector< MachineJumpTableEntry > & getJumpTables() const
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
MachineOperand class - Representation of each machine instruction operand.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
void setIsKill(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
void addLiveIn(MCRegister Reg, Register vreg=Register())
addLiveIn - Add the specified register as a live-in.
unsigned getAddressSpace() const
Return the address space for the associated pointer.
MachineMemOperand * getMemOperand() const
Return a MachineMemOperand object describing the memory reference performed by operation.
const SDValue & getChain() const
EVT getMemoryVT() const
Return the type of the in-memory value.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
Wrapper class representing virtual and physical registers.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
ArrayRef< SDUse > ops() const
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
uint64_t getAsZExtVal() const
Helper method returns the zero-extended integer value of a ConstantSDNode.
const SDValue & getOperand(unsigned Num) const
uint64_t getConstantOperandVal(unsigned Num) const
Helper method returns the integer value of a ConstantSDNode operand.
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
const SDValue & getOperand(unsigned i) const
MVT getSimpleValueType() const
Return the simple ValueType of the referenced return value.
unsigned getOpcode() const
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
SDValue getTargetGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, unsigned TargetFlags=0)
SDValue getCopyToReg(SDValue Chain, const SDLoc &dl, Register Reg, SDValue N)
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI SDValue getShiftAmountConstant(uint64_t Val, EVT VT, const SDLoc &DL)
LLVM_ABI SDValue getSplatValue(SDValue V, bool LegalTypes=false)
If V is a splat vector, return its scalar source operand by extracting that element from the source v...
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI void ExtractVectorElements(SDValue Op, SmallVectorImpl< SDValue > &Args, unsigned Start=0, unsigned Count=0, EVT EltVT=EVT())
Append the extracted elements from Start to Count out of the vector Op in Args.
SDValue getSetCC(const SDLoc &DL, EVT VT, SDValue LHS, SDValue RHS, ISD::CondCode Cond, SDValue Chain=SDValue(), bool IsSignaling=false)
Helper function to make it easier to build SetCC's if you just have an ISD::CondCode instead of an SD...
LLVM_ABI SDValue UnrollVectorOp(SDNode *N, unsigned ResNE=0)
Utility function used by legalize and lowering to "unroll" a vector operation by splitting out the sc...
LLVM_ABI SDValue getConstantFP(double Val, const SDLoc &DL, EVT VT, bool isTarget=false)
Create a ConstantFPSDNode wrapping a constant value.
LLVM_ABI SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef< SDValue > Ops, EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MOLoad|MachineMemOperand::MOStore, LocationSize Size=LocationSize::precise(0), const AAMDNodes &AAInfo=AAMDNodes())
Creates a MemIntrinsicNode that may produce a result and takes a list of operands.
LLVM_ABI SDValue getMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
LLVM_ABI SDValue getNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a bitwise NOT operation as (XOR Val, -1).
SDValue getTargetJumpTable(int JTI, EVT VT, unsigned TargetFlags=0)
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
LLVM_ABI SDValue getBitcast(EVT VT, SDValue V)
Return a bitcast using the SDLoc of the value operand, and casting to the provided type.
SDValue getCopyFromReg(SDValue Chain, const SDLoc &dl, Register Reg, EVT VT)
const DataLayout & getDataLayout() const
SDValue getTargetFrameIndex(int FI, EVT VT)
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
Helper function to build ISD::STORE nodes.
LLVM_ABI bool SignBitIsZero(SDValue Op, unsigned Depth=0) const
Return true if the sign bit of Op is known to be zero.
LLVM_ABI SDValue getBasicBlock(MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
LLVM_ABI SDValue getAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either any-extending or truncat...
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
MachineFunction & getMachineFunction() const
SDValue getSplatBuildVector(EVT VT, const SDLoc &DL, SDValue Op)
Return a splat ISD::BUILD_VECTOR node, consisting of Op splatted to all elements.
LLVM_ABI SDValue getFrameIndex(int FI, EVT VT, bool isTarget=false)
LLVM_ABI SDValue getZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either zero-extending or trunca...
LLVMContext * getContext() const
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
LLVM_ABI SDValue getMCSymbol(MCSymbol *Sym, EVT VT)
SDValue getEntryNode() const
Return the token chain corresponding to the entry of the function.
LLVM_ABI SDValue getVectorShuffle(EVT VT, const SDLoc &dl, SDValue N1, SDValue N2, ArrayRef< int > Mask)
Return an ISD::VECTOR_SHUFFLE node.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
const SDValue & getBasePtr() const
const SDValue & getOffset() const
const SDValue & getValue() const
StringRef - Represent a constant reference to a string, i.e.
constexpr size_t size() const
size - Get the string size.
TargetInstrInfo - Interface to description of machine instruction set.
Provides information about what library functions are available for the current target.
int InstructionOpcodeToISD(unsigned Opcode) const
Get the ISD node that corresponds to the Instruction class opcode.
void setBooleanVectorContents(BooleanContent Ty)
Specify how the target extends the result of a vector boolean value from a vector of i1 to a wider ty...
void setOperationAction(unsigned Op, MVT VT, LegalizeAction Action)
Indicate that the specified operation does not work with the specified type and indicate what to do a...
virtual const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent=false) const
Return the register class that should be used for the specified value type.
const TargetMachine & getTargetMachine() const
unsigned MaxLoadsPerMemcmp
Specify maximum number of load instructions per memcmp call.
LegalizeTypeAction
This enum indicates whether a types are legal for a target, and if not, what action should be used to...
void setMaxAtomicSizeInBitsSupported(unsigned SizeInBits)
Set the maximum atomic operation size supported by the backend.
virtual TargetLoweringBase::LegalizeTypeAction getPreferredVectorAction(MVT VT) const
Return the preferred vector type legalization action.
void setBooleanContents(BooleanContent Ty)
Specify how the target extends the result of integer and floating point boolean values from i1 to a w...
void computeRegisterProperties(const TargetRegisterInfo *TRI)
Once all of the register classes are added, this allows us to compute derived properties we expose.
void addRegisterClass(MVT VT, const TargetRegisterClass *RC)
Add the specified register class as an available regclass for the specified value type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
void setMinimumJumpTableEntries(unsigned Val)
Indicate the minimum number of blocks to generate jump tables.
void setTruncStoreAction(MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified truncating store does not work with the specified type and indicate what ...
@ ZeroOrOneBooleanContent
@ ZeroOrNegativeOneBooleanContent
unsigned MaxLoadsPerMemcmpOptSize
Likewise for functions with the OptSize attribute.
virtual bool isBinOp(unsigned Opcode) const
Return true if the node is a math/logic binary operator.
void setStackPointerRegisterToSaveRestore(Register R)
If set to a physical register, this specifies the register that llvm.savestack/llvm....
AtomicExpansionKind
Enum that specifies what an atomic load/AtomicRMWInst is expanded to, if at all.
void setCondCodeAction(ArrayRef< ISD::CondCode > CCs, MVT VT, LegalizeAction Action)
Indicate that the specified condition code is or isn't supported on the target and indicate what to d...
void setTargetDAGCombine(ArrayRef< ISD::NodeType > NTs)
Targets should invoke this method for each target independent node that they want to provide a custom...
void setLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified load with extension does not work with the specified type and indicate wh...
virtual MVT getPointerMemTy(const DataLayout &DL, uint32_t AS=0) const
Return the in-memory pointer type for the given address space, defaults to the pointer type from the ...
void setSchedulingPreference(Sched::Preference Pref)
Specify the target scheduling preference.
bool isOperationLegalOrCustomOrPromote(unsigned Op, EVT VT, bool LegalOnly=false) const
Return true if the specified operation is legal on this target or can be made legal with custom lower...
std::pair< SDValue, SDValue > makeLibCall(SelectionDAG &DAG, RTLIB::Libcall LC, EVT RetVT, ArrayRef< SDValue > Ops, MakeLibCallOptions CallOptions, const SDLoc &dl, SDValue Chain=SDValue()) const
Returns a pair of (return value, chain).
bool isPositionIndependent() const
virtual std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const
Given a physical register constraint (e.g.
TargetLowering(const TargetLowering &)=delete
virtual bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const
Return true if folding a constant offset with the given GlobalAddress is legal.
Primary interface to the complete machine description for the target machine.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
The instances of the Type class are immutable: once they are created, they are never changed.
bool isFunctionTy() const
True if this is an instance of FunctionType.
static LLVM_ABI Type * getDoubleTy(LLVMContext &C)
static LLVM_ABI Type * getFloatTy(LLVMContext &C)
A Use represents the edge between a Value definition and its users.
LLVM_ABI const Value * stripPointerCastsAndAliases() const
Strip off pointer casts, all-zero GEPs, address space casts, and aliases.
static std::optional< unsigned > getLocalForStackObject(MachineFunction &MF, int FrameIndex)
bool hasCallIndirectOverlong() const
bool hasReferenceTypes() const
WebAssemblyTargetLowering(const TargetMachine &TM, const WebAssemblySubtarget &STI)
MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const override
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
MVT getPointerMemTy(const DataLayout &DL, uint32_t AS=0) const override
Return the in-memory pointer type for the given address space, defaults to the pointer type from the ...
self_iterator getIterator()
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Swift
Calling convention for Swift.
@ PreserveMost
Used for runtime calls that preserves most registers.
@ CXX_FAST_TLS
Used for access functions.
@ WASM_EmscriptenInvoke
For emscripten __invoke_* functions.
@ Cold
Attempts to make code in the caller as efficient as possible under the assumption that the call is no...
@ PreserveAll
Used for runtime calls that preserves (almost) all registers.
@ Fast
Attempts to make calls as fast as possible (e.g.
@ C
The default llvm calling convention, compatible with C.
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
@ BSWAP
Byte Swap and Counting operators.
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
@ ADD
Simple integer binary arithmetic operators.
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
@ INTRINSIC_VOID
OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...) This node represents a target intrin...
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
@ CONCAT_VECTORS
CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of vector type with the same length ...
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
@ BUILD_PAIR
BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
@ BUILTIN_OP_END
BUILTIN_OP_END - This must be the last enum value in this list.
@ SIGN_EXTEND
Conversion operators.
@ SCALAR_TO_VECTOR
SCALAR_TO_VECTOR(VAL) - This represents the operation of loading a scalar value into element 0 of the...
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
@ EXTRACT_ELEMENT
EXTRACT_ELEMENT - This is used to get the lower or upper (determined by a Constant,...
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
@ SHL
Shift and rotation operations.
@ VECTOR_SHUFFLE
VECTOR_SHUFFLE(VEC1, VEC2) - Returns a vector, of the same type as VEC1/VEC2.
@ EXTRACT_SUBVECTOR
EXTRACT_SUBVECTOR(VECTOR, IDX) - Returns a subvector from VECTOR.
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
@ CopyToReg
CopyToReg - This node has three operands: a chain, a register number to set to this value,...
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
@ FRAMEADDR
FRAMEADDR, RETURNADDR - These nodes represent llvm.frameaddress and llvm.returnaddress on the DAG.
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
@ TargetConstant
TargetConstant* - Like Constant*, but the DAG does not do any folding, simplification,...
@ AND
Bitwise operators - logical and, logical or, logical xor.
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
@ INTRINSIC_W_CHAIN
RESULT,OUTCHAIN = INTRINSIC_W_CHAIN(INCHAIN, INTRINSICID, arg1, ...) This node represents a target in...
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
bool isExtOpcode(unsigned Opcode)
LLVM_ABI bool isConstantSplatVector(const SDNode *N, APInt &SplatValue)
Node predicates.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
This namespace contains an enum with a value for every intrinsic/builtin function known by LLVM.
OperandFlags
These are flags set on operands, but should be considered private, all access should go through the M...
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
bool sd_match(SDNode *N, const SelectionDAG *DAG, Pattern &&P)
CondCode_match m_SpecificCondCode(ISD::CondCode CC)
Match a conditional code SDNode with a specific ISD::CondCode.
CondCode_match m_CondCode()
Match any conditional code SDNode.
TernaryOpc_match< T0_P, T1_P, T2_P, true, false > m_c_SetCC(const T0_P &LHS, const T1_P &RHS, const T2_P &CC)
MCSymbolWasm * getOrCreateFunctionTableSymbol(MCContext &Ctx, const WebAssemblySubtarget *Subtarget)
Returns the __indirect_function_table, for use in call_indirect and in function bitcasts.
@ WASM_ADDRESS_SPACE_EXTERNREF
@ WASM_ADDRESS_SPACE_FUNCREF
bool isWebAssemblyFuncrefType(const Type *Ty)
Return true if this is a WebAssembly Funcref Type.
bool isWebAssemblyTableType(const Type *Ty)
Return true if the table represents a WebAssembly table type.
MCSymbolWasm * getOrCreateFuncrefCallTableSymbol(MCContext &Ctx, const WebAssemblySubtarget *Subtarget)
Returns the __funcref_call_table, for use in funcref calls when lowered to table.set + call_indirect.
FastISel * createFastISel(FunctionLoweringInfo &funcInfo, const TargetLibraryInfo *libInfo)
bool isValidAddressSpace(unsigned AS)
bool canLowerReturn(size_t ResultSize, const WebAssemblySubtarget *Subtarget)
Returns true if the function's return value(s) can be lowered directly, i.e., not indirectly via a po...
bool isWasmVarAddressSpace(unsigned AS)
NodeAddr< NodeBase * > Node
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
unsigned Log2_32_Ceil(uint32_t Value)
Return the ceil log base 2 of the specified value, 32 if the value is zero.
FunctionAddr VTableAddr Value
void computeSignatureVTs(const FunctionType *Ty, const Function *TargetFunc, const Function &ContextFunc, const TargetMachine &TM, SmallVectorImpl< MVT > &Params, SmallVectorImpl< MVT > &Results)
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
LLVM_ABI bool isNullConstant(SDValue V)
Returns true if V is a constant integer zero.
LLVM_ABI SDValue peekThroughBitcasts(SDValue V)
Return the non-bitcasted source operand of V if it exists.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
DWARFExpression::Operation Op
auto max_element(R &&Range)
Provide wrappers to std::max_element which take ranges instead of having to pass begin/end explicitly...
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
auto find_if(R &&Range, UnaryPredicate P)
Provide wrappers to std::find_if which take ranges instead of having to pass begin/end explicitly.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
void computeLegalValueVTs(const WebAssemblyTargetLowering &TLI, LLVMContext &Ctx, const DataLayout &DL, Type *Ty, SmallVectorImpl< MVT > &ValueVTs)
constexpr uint64_t NextPowerOf2(uint64_t A)
Returns the next power of two (in 64-bits) that is strictly greater than A.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
EVT changeVectorElementTypeToInteger() const
Return a vector with the same number of elements as this vector, but with the element type converted ...
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
ElementCount getVectorElementCount() const
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
bool isByteSized() const
Return true if the bit size is a multiple of 8.
uint64_t getScalarSizeInBits() const
static LLVM_ABI EVT getEVT(Type *Ty, bool HandleUnknown=false)
Return the value type corresponding to the specified type.
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
bool is128BitVector() const
Return true if this is a 128-bit vector type.
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
uint64_t getFixedSizeInBits() const
Return the size of the specified fixed width value type in bits.
bool isFixedLengthVector() const
bool isVector() const
Return true if this is a vector value type.
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
bool bitsGE(EVT VT) const
Return true if this has no less bits than VT.
bool is256BitVector() const
Return true if this is a 256-bit vector type.
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
EVT getVectorElementType() const
Given a vector type, return the type of each element.
bool isScalarInteger() const
Return true if this is an integer, but not a vector.
EVT changeVectorElementType(EVT EltVT) const
Return a VT for a vector type whose attributes match ourselves with the exception of the element type...
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
bool isInConsecutiveRegs() const
Align getNonZeroOrigAlign() const
bool isSwiftError() const
unsigned getByValSize() const
bool isInConsecutiveRegsLast() const
Align getNonZeroByValAlign() const
unsigned getBitWidth() const
Get the bit width of this value.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
These are IR-level optimization flags that may be propagated to SDNodes.
bool isBeforeLegalize() const
This structure is used to pass arguments to makeLibCall function.