LLVM 19.0.0git
LegalizeVectorOps.cpp
Go to the documentation of this file.
1//===- LegalizeVectorOps.cpp - Implement SelectionDAG::LegalizeVectors ----===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the SelectionDAG::LegalizeVectors method.
10//
11// The vector legalizer looks for vector operations which might need to be
12// scalarized and legalizes them. This is a separate step from Legalize because
13// scalarizing can introduce illegal types. For example, suppose we have an
14// ISD::SDIV of type v2i64 on x86-32. The type is legal (for example, addition
15// on a v2i64 is legal), but ISD::SDIV isn't legal, so we have to unroll the
16// operation, which introduces nodes with the illegal type i64 which must be
17// expanded. Similarly, suppose we have an ISD::SRA of type v16i8 on PowerPC;
18// the operation must be unrolled, which introduces nodes with the illegal
19// type i8 which must be promoted.
20//
21// This does not legalize vector manipulations like ISD::BUILD_VECTOR,
22// or operations that happen to take a vector which are custom-lowered;
23// the legalization for such operations never produces nodes
24// with illegal types, so it's okay to put off legalizing them until
25// SelectionDAG::Legalize runs.
26//
27//===----------------------------------------------------------------------===//
28
29#include "llvm/ADT/DenseMap.h"
39#include "llvm/IR/DataLayout.h"
42#include "llvm/Support/Debug.h"
44#include <cassert>
45#include <cstdint>
46#include <iterator>
47#include <utility>
48
49using namespace llvm;
50
51#define DEBUG_TYPE "legalizevectorops"
52
53namespace {
54
55class VectorLegalizer {
56 SelectionDAG& DAG;
57 const TargetLowering &TLI;
58 bool Changed = false; // Keep track of whether anything changed
59
60 /// For nodes that are of legal width, and that have more than one use, this
61 /// map indicates what regularized operand to use. This allows us to avoid
62 /// legalizing the same thing more than once.
64
65 /// Adds a node to the translation cache.
66 void AddLegalizedOperand(SDValue From, SDValue To) {
67 LegalizedNodes.insert(std::make_pair(From, To));
68 // If someone requests legalization of the new node, return itself.
69 if (From != To)
70 LegalizedNodes.insert(std::make_pair(To, To));
71 }
72
73 /// Legalizes the given node.
74 SDValue LegalizeOp(SDValue Op);
75
76 /// Assuming the node is legal, "legalize" the results.
77 SDValue TranslateLegalizeResults(SDValue Op, SDNode *Result);
78
79 /// Make sure Results are legal and update the translation cache.
80 SDValue RecursivelyLegalizeResults(SDValue Op,
82
83 /// Wrapper to interface LowerOperation with a vector of Results.
84 /// Returns false if the target wants to use default expansion. Otherwise
85 /// returns true. If return is true and the Results are empty, then the
86 /// target wants to keep the input node as is.
87 bool LowerOperationWrapper(SDNode *N, SmallVectorImpl<SDValue> &Results);
88
89 /// Implements unrolling a VSETCC.
90 SDValue UnrollVSETCC(SDNode *Node);
91
92 /// Implement expand-based legalization of vector operations.
93 ///
94 /// This is just a high-level routine to dispatch to specific code paths for
95 /// operations to legalize them.
97
98 /// Implements expansion for FP_TO_UINT; falls back to UnrollVectorOp if
99 /// FP_TO_SINT isn't legal.
100 void ExpandFP_TO_UINT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
101
102 /// Implements expansion for UINT_TO_FLOAT; falls back to UnrollVectorOp if
103 /// SINT_TO_FLOAT and SHR on vectors isn't legal.
104 void ExpandUINT_TO_FLOAT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
105
106 /// Implement expansion for SIGN_EXTEND_INREG using SRL and SRA.
107 SDValue ExpandSEXTINREG(SDNode *Node);
108
109 /// Implement expansion for ANY_EXTEND_VECTOR_INREG.
110 ///
111 /// Shuffles the low lanes of the operand into place and bitcasts to the proper
112 /// type. The contents of the bits in the extended part of each element are
113 /// undef.
114 SDValue ExpandANY_EXTEND_VECTOR_INREG(SDNode *Node);
115
116 /// Implement expansion for SIGN_EXTEND_VECTOR_INREG.
117 ///
118 /// Shuffles the low lanes of the operand into place, bitcasts to the proper
119 /// type, then shifts left and arithmetic shifts right to introduce a sign
120 /// extension.
121 SDValue ExpandSIGN_EXTEND_VECTOR_INREG(SDNode *Node);
122
123 /// Implement expansion for ZERO_EXTEND_VECTOR_INREG.
124 ///
125 /// Shuffles the low lanes of the operand into place and blends zeros into
126 /// the remaining lanes, finally bitcasting to the proper type.
127 SDValue ExpandZERO_EXTEND_VECTOR_INREG(SDNode *Node);
128
129 /// Expand bswap of vectors into a shuffle if legal.
130 SDValue ExpandBSWAP(SDNode *Node);
131
132 /// Implement vselect in terms of XOR, AND, OR when blend is not
133 /// supported by the target.
134 SDValue ExpandVSELECT(SDNode *Node);
135 SDValue ExpandVP_SELECT(SDNode *Node);
136 SDValue ExpandVP_MERGE(SDNode *Node);
137 SDValue ExpandVP_REM(SDNode *Node);
138 SDValue ExpandSELECT(SDNode *Node);
139 std::pair<SDValue, SDValue> ExpandLoad(SDNode *N);
140 SDValue ExpandStore(SDNode *N);
141 SDValue ExpandFNEG(SDNode *Node);
142 void ExpandFSUB(SDNode *Node, SmallVectorImpl<SDValue> &Results);
143 void ExpandSETCC(SDNode *Node, SmallVectorImpl<SDValue> &Results);
144 void ExpandBITREVERSE(SDNode *Node, SmallVectorImpl<SDValue> &Results);
145 void ExpandUADDSUBO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
146 void ExpandSADDSUBO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
147 void ExpandMULO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
148 void ExpandFixedPointDiv(SDNode *Node, SmallVectorImpl<SDValue> &Results);
149 void ExpandStrictFPOp(SDNode *Node, SmallVectorImpl<SDValue> &Results);
150 void ExpandREM(SDNode *Node, SmallVectorImpl<SDValue> &Results);
151
152 bool tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall LC,
154 bool tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall Call_F32,
155 RTLIB::Libcall Call_F64, RTLIB::Libcall Call_F80,
156 RTLIB::Libcall Call_F128,
157 RTLIB::Libcall Call_PPCF128,
159
160 void UnrollStrictFPOp(SDNode *Node, SmallVectorImpl<SDValue> &Results);
161
162 /// Implements vector promotion.
163 ///
164 /// This is essentially just bitcasting the operands to a different type and
165 /// bitcasting the result back to the original type.
167
168 /// Implements [SU]INT_TO_FP vector promotion.
169 ///
170 /// This is a [zs]ext of the input operand to a larger integer type.
171 void PromoteINT_TO_FP(SDNode *Node, SmallVectorImpl<SDValue> &Results);
172
173 /// Implements FP_TO_[SU]INT vector promotion of the result type.
174 ///
175 /// It is promoted to a larger integer type. The result is then
176 /// truncated back to the original type.
177 void PromoteFP_TO_INT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
178
179 /// Implements vector setcc operation promotion.
180 ///
181 /// All vector operands are promoted to a vector type with larger element
182 /// type.
183 void PromoteSETCC(SDNode *Node, SmallVectorImpl<SDValue> &Results);
184
185 void PromoteSTRICT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
186
187public:
188 VectorLegalizer(SelectionDAG& dag) :
189 DAG(dag), TLI(dag.getTargetLoweringInfo()) {}
190
191 /// Begin legalizer the vector operations in the DAG.
192 bool Run();
193};
194
195} // end anonymous namespace
196
197bool VectorLegalizer::Run() {
198 // Before we start legalizing vector nodes, check if there are any vectors.
199 bool HasVectors = false;
200 for (SelectionDAG::allnodes_iterator I = DAG.allnodes_begin(),
201 E = std::prev(DAG.allnodes_end()); I != std::next(E); ++I) {
202 // Check if the values of the nodes contain vectors. We don't need to check
203 // the operands because we are going to check their values at some point.
204 HasVectors = llvm::any_of(I->values(), [](EVT T) { return T.isVector(); });
205
206 // If we found a vector node we can start the legalization.
207 if (HasVectors)
208 break;
209 }
210
211 // If this basic block has no vectors then no need to legalize vectors.
212 if (!HasVectors)
213 return false;
214
215 // The legalize process is inherently a bottom-up recursive process (users
216 // legalize their uses before themselves). Given infinite stack space, we
217 // could just start legalizing on the root and traverse the whole graph. In
218 // practice however, this causes us to run out of stack space on large basic
219 // blocks. To avoid this problem, compute an ordering of the nodes where each
220 // node is only legalized after all of its operands are legalized.
221 DAG.AssignTopologicalOrder();
222 for (SelectionDAG::allnodes_iterator I = DAG.allnodes_begin(),
223 E = std::prev(DAG.allnodes_end()); I != std::next(E); ++I)
224 LegalizeOp(SDValue(&*I, 0));
225
226 // Finally, it's possible the root changed. Get the new root.
227 SDValue OldRoot = DAG.getRoot();
228 assert(LegalizedNodes.count(OldRoot) && "Root didn't get legalized?");
229 DAG.setRoot(LegalizedNodes[OldRoot]);
230
231 LegalizedNodes.clear();
232
233 // Remove dead nodes now.
234 DAG.RemoveDeadNodes();
235
236 return Changed;
237}
238
239SDValue VectorLegalizer::TranslateLegalizeResults(SDValue Op, SDNode *Result) {
240 assert(Op->getNumValues() == Result->getNumValues() &&
241 "Unexpected number of results");
242 // Generic legalization: just pass the operand through.
243 for (unsigned i = 0, e = Op->getNumValues(); i != e; ++i)
244 AddLegalizedOperand(Op.getValue(i), SDValue(Result, i));
245 return SDValue(Result, Op.getResNo());
246}
247
249VectorLegalizer::RecursivelyLegalizeResults(SDValue Op,
251 assert(Results.size() == Op->getNumValues() &&
252 "Unexpected number of results");
253 // Make sure that the generated code is itself legal.
254 for (unsigned i = 0, e = Results.size(); i != e; ++i) {
255 Results[i] = LegalizeOp(Results[i]);
256 AddLegalizedOperand(Op.getValue(i), Results[i]);
257 }
258
259 return Results[Op.getResNo()];
260}
261
262SDValue VectorLegalizer::LegalizeOp(SDValue Op) {
263 // Note that LegalizeOp may be reentered even from single-use nodes, which
264 // means that we always must cache transformed nodes.
265 DenseMap<SDValue, SDValue>::iterator I = LegalizedNodes.find(Op);
266 if (I != LegalizedNodes.end()) return I->second;
267
268 // Legalize the operands
270 for (const SDValue &Oper : Op->op_values())
271 Ops.push_back(LegalizeOp(Oper));
272
273 SDNode *Node = DAG.UpdateNodeOperands(Op.getNode(), Ops);
274
275 bool HasVectorValueOrOp =
276 llvm::any_of(Node->values(), [](EVT T) { return T.isVector(); }) ||
277 llvm::any_of(Node->op_values(),
278 [](SDValue O) { return O.getValueType().isVector(); });
279 if (!HasVectorValueOrOp)
280 return TranslateLegalizeResults(Op, Node);
281
282 TargetLowering::LegalizeAction Action = TargetLowering::Legal;
283 EVT ValVT;
284 switch (Op.getOpcode()) {
285 default:
286 return TranslateLegalizeResults(Op, Node);
287 case ISD::LOAD: {
288 LoadSDNode *LD = cast<LoadSDNode>(Node);
289 ISD::LoadExtType ExtType = LD->getExtensionType();
290 EVT LoadedVT = LD->getMemoryVT();
291 if (LoadedVT.isVector() && ExtType != ISD::NON_EXTLOAD)
292 Action = TLI.getLoadExtAction(ExtType, LD->getValueType(0), LoadedVT);
293 break;
294 }
295 case ISD::STORE: {
296 StoreSDNode *ST = cast<StoreSDNode>(Node);
297 EVT StVT = ST->getMemoryVT();
298 MVT ValVT = ST->getValue().getSimpleValueType();
299 if (StVT.isVector() && ST->isTruncatingStore())
300 Action = TLI.getTruncStoreAction(ValVT, StVT);
301 break;
302 }
304 Action = TLI.getOperationAction(Node->getOpcode(), Node->getValueType(0));
305 // This operation lies about being legal: when it claims to be legal,
306 // it should actually be expanded.
307 if (Action == TargetLowering::Legal)
308 Action = TargetLowering::Expand;
309 break;
310#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
311 case ISD::STRICT_##DAGN:
312#include "llvm/IR/ConstrainedOps.def"
313 ValVT = Node->getValueType(0);
314 if (Op.getOpcode() == ISD::STRICT_SINT_TO_FP ||
315 Op.getOpcode() == ISD::STRICT_UINT_TO_FP)
316 ValVT = Node->getOperand(1).getValueType();
317 if (Op.getOpcode() == ISD::STRICT_FSETCC ||
318 Op.getOpcode() == ISD::STRICT_FSETCCS) {
319 MVT OpVT = Node->getOperand(1).getSimpleValueType();
320 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(3))->get();
321 Action = TLI.getCondCodeAction(CCCode, OpVT);
322 if (Action == TargetLowering::Legal)
323 Action = TLI.getOperationAction(Node->getOpcode(), OpVT);
324 } else {
325 Action = TLI.getOperationAction(Node->getOpcode(), ValVT);
326 }
327 // If we're asked to expand a strict vector floating-point operation,
328 // by default we're going to simply unroll it. That is usually the
329 // best approach, except in the case where the resulting strict (scalar)
330 // operations would themselves use the fallback mutation to non-strict.
331 // In that specific case, just do the fallback on the vector op.
332 if (Action == TargetLowering::Expand && !TLI.isStrictFPEnabled() &&
333 TLI.getStrictFPOperationAction(Node->getOpcode(), ValVT) ==
334 TargetLowering::Legal) {
335 EVT EltVT = ValVT.getVectorElementType();
336 if (TLI.getOperationAction(Node->getOpcode(), EltVT)
337 == TargetLowering::Expand &&
338 TLI.getStrictFPOperationAction(Node->getOpcode(), EltVT)
339 == TargetLowering::Legal)
340 Action = TargetLowering::Legal;
341 }
342 break;
343 case ISD::ADD:
344 case ISD::SUB:
345 case ISD::MUL:
346 case ISD::MULHS:
347 case ISD::MULHU:
348 case ISD::SDIV:
349 case ISD::UDIV:
350 case ISD::SREM:
351 case ISD::UREM:
352 case ISD::SDIVREM:
353 case ISD::UDIVREM:
354 case ISD::FADD:
355 case ISD::FSUB:
356 case ISD::FMUL:
357 case ISD::FDIV:
358 case ISD::FREM:
359 case ISD::AND:
360 case ISD::OR:
361 case ISD::XOR:
362 case ISD::SHL:
363 case ISD::SRA:
364 case ISD::SRL:
365 case ISD::FSHL:
366 case ISD::FSHR:
367 case ISD::ROTL:
368 case ISD::ROTR:
369 case ISD::ABS:
370 case ISD::BSWAP:
371 case ISD::BITREVERSE:
372 case ISD::CTLZ:
373 case ISD::CTTZ:
376 case ISD::CTPOP:
377 case ISD::SELECT:
378 case ISD::VSELECT:
379 case ISD::SELECT_CC:
380 case ISD::ZERO_EXTEND:
381 case ISD::ANY_EXTEND:
382 case ISD::TRUNCATE:
383 case ISD::SIGN_EXTEND:
384 case ISD::FP_TO_SINT:
385 case ISD::FP_TO_UINT:
386 case ISD::FNEG:
387 case ISD::FABS:
388 case ISD::FMINNUM:
389 case ISD::FMAXNUM:
392 case ISD::FMINIMUM:
393 case ISD::FMAXIMUM:
394 case ISD::FCOPYSIGN:
395 case ISD::FSQRT:
396 case ISD::FSIN:
397 case ISD::FCOS:
398 case ISD::FLDEXP:
399 case ISD::FPOWI:
400 case ISD::FPOW:
401 case ISD::FLOG:
402 case ISD::FLOG2:
403 case ISD::FLOG10:
404 case ISD::FEXP:
405 case ISD::FEXP2:
406 case ISD::FEXP10:
407 case ISD::FCEIL:
408 case ISD::FTRUNC:
409 case ISD::FRINT:
410 case ISD::FNEARBYINT:
411 case ISD::FROUND:
412 case ISD::FROUNDEVEN:
413 case ISD::FFLOOR:
414 case ISD::FP_ROUND:
415 case ISD::FP_EXTEND:
417 case ISD::FMA:
422 case ISD::SMIN:
423 case ISD::SMAX:
424 case ISD::UMIN:
425 case ISD::UMAX:
426 case ISD::SMUL_LOHI:
427 case ISD::UMUL_LOHI:
428 case ISD::SADDO:
429 case ISD::UADDO:
430 case ISD::SSUBO:
431 case ISD::USUBO:
432 case ISD::SMULO:
433 case ISD::UMULO:
435 case ISD::FFREXP:
436 case ISD::SADDSAT:
437 case ISD::UADDSAT:
438 case ISD::SSUBSAT:
439 case ISD::USUBSAT:
440 case ISD::SSHLSAT:
441 case ISD::USHLSAT:
444 case ISD::MGATHER:
445 Action = TLI.getOperationAction(Node->getOpcode(), Node->getValueType(0));
446 break;
447 case ISD::SMULFIX:
448 case ISD::SMULFIXSAT:
449 case ISD::UMULFIX:
450 case ISD::UMULFIXSAT:
451 case ISD::SDIVFIX:
452 case ISD::SDIVFIXSAT:
453 case ISD::UDIVFIX:
454 case ISD::UDIVFIXSAT: {
455 unsigned Scale = Node->getConstantOperandVal(2);
456 Action = TLI.getFixedPointOperationAction(Node->getOpcode(),
457 Node->getValueType(0), Scale);
458 break;
459 }
460 case ISD::LRINT:
461 case ISD::LLRINT:
462 case ISD::SINT_TO_FP:
463 case ISD::UINT_TO_FP:
479 Action = TLI.getOperationAction(Node->getOpcode(),
480 Node->getOperand(0).getValueType());
481 break;
484 Action = TLI.getOperationAction(Node->getOpcode(),
485 Node->getOperand(1).getValueType());
486 break;
487 case ISD::SETCC: {
488 MVT OpVT = Node->getOperand(0).getSimpleValueType();
489 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(2))->get();
490 Action = TLI.getCondCodeAction(CCCode, OpVT);
491 if (Action == TargetLowering::Legal)
492 Action = TLI.getOperationAction(Node->getOpcode(), OpVT);
493 break;
494 }
495
496#define BEGIN_REGISTER_VP_SDNODE(VPID, LEGALPOS, ...) \
497 case ISD::VPID: { \
498 EVT LegalizeVT = LEGALPOS < 0 ? Node->getValueType(-(1 + LEGALPOS)) \
499 : Node->getOperand(LEGALPOS).getValueType(); \
500 if (ISD::VPID == ISD::VP_SETCC) { \
501 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(2))->get(); \
502 Action = TLI.getCondCodeAction(CCCode, LegalizeVT.getSimpleVT()); \
503 if (Action != TargetLowering::Legal) \
504 break; \
505 } \
506 /* Defer non-vector results to LegalizeDAG. */ \
507 if (!Node->getValueType(0).isVector()) { \
508 Action = TargetLowering::Legal; \
509 break; \
510 } \
511 Action = TLI.getOperationAction(Node->getOpcode(), LegalizeVT); \
512 } break;
513#include "llvm/IR/VPIntrinsics.def"
514 }
515
516 LLVM_DEBUG(dbgs() << "\nLegalizing vector op: "; Node->dump(&DAG));
517
518 SmallVector<SDValue, 8> ResultVals;
519 switch (Action) {
520 default: llvm_unreachable("This action is not supported yet!");
521 case TargetLowering::Promote:
522 assert((Op.getOpcode() != ISD::LOAD && Op.getOpcode() != ISD::STORE) &&
523 "This action is not supported yet!");
524 LLVM_DEBUG(dbgs() << "Promoting\n");
525 Promote(Node, ResultVals);
526 assert(!ResultVals.empty() && "No results for promotion?");
527 break;
528 case TargetLowering::Legal:
529 LLVM_DEBUG(dbgs() << "Legal node: nothing to do\n");
530 break;
531 case TargetLowering::Custom:
532 LLVM_DEBUG(dbgs() << "Trying custom legalization\n");
533 if (LowerOperationWrapper(Node, ResultVals))
534 break;
535 LLVM_DEBUG(dbgs() << "Could not custom legalize node\n");
536 [[fallthrough]];
537 case TargetLowering::Expand:
538 LLVM_DEBUG(dbgs() << "Expanding\n");
539 Expand(Node, ResultVals);
540 break;
541 }
542
543 if (ResultVals.empty())
544 return TranslateLegalizeResults(Op, Node);
545
546 Changed = true;
547 return RecursivelyLegalizeResults(Op, ResultVals);
548}
549
550// FIXME: This is very similar to TargetLowering::LowerOperationWrapper. Can we
551// merge them somehow?
552bool VectorLegalizer::LowerOperationWrapper(SDNode *Node,
554 SDValue Res = TLI.LowerOperation(SDValue(Node, 0), DAG);
555
556 if (!Res.getNode())
557 return false;
558
559 if (Res == SDValue(Node, 0))
560 return true;
561
562 // If the original node has one result, take the return value from
563 // LowerOperation as is. It might not be result number 0.
564 if (Node->getNumValues() == 1) {
565 Results.push_back(Res);
566 return true;
567 }
568
569 // If the original node has multiple results, then the return node should
570 // have the same number of results.
571 assert((Node->getNumValues() == Res->getNumValues()) &&
572 "Lowering returned the wrong number of results!");
573
574 // Places new result values base on N result number.
575 for (unsigned I = 0, E = Node->getNumValues(); I != E; ++I)
576 Results.push_back(Res.getValue(I));
577
578 return true;
579}
580
581void VectorLegalizer::PromoteSETCC(SDNode *Node,
583 MVT VecVT = Node->getOperand(0).getSimpleValueType();
584 MVT NewVecVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VecVT);
585
586 unsigned ExtOp = VecVT.isFloatingPoint() ? ISD::FP_EXTEND : ISD::ANY_EXTEND;
587
588 SDLoc DL(Node);
589 SmallVector<SDValue, 5> Operands(Node->getNumOperands());
590
591 Operands[0] = DAG.getNode(ExtOp, DL, NewVecVT, Node->getOperand(0));
592 Operands[1] = DAG.getNode(ExtOp, DL, NewVecVT, Node->getOperand(1));
593 Operands[2] = Node->getOperand(2);
594
595 if (Node->getOpcode() == ISD::VP_SETCC) {
596 Operands[3] = Node->getOperand(3); // mask
597 Operands[4] = Node->getOperand(4); // evl
598 }
599
600 SDValue Res = DAG.getNode(Node->getOpcode(), DL, Node->getSimpleValueType(0),
601 Operands, Node->getFlags());
602
603 Results.push_back(Res);
604}
605
606void VectorLegalizer::PromoteSTRICT(SDNode *Node,
608 MVT VecVT = Node->getOperand(1).getSimpleValueType();
609 MVT NewVecVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VecVT);
610
611 assert(VecVT.isFloatingPoint());
612
613 SDLoc DL(Node);
614 SmallVector<SDValue, 5> Operands(Node->getNumOperands());
616
617 for (unsigned j = 1; j != Node->getNumOperands(); ++j)
618 if (Node->getOperand(j).getValueType().isVector() &&
619 !(ISD::isVPOpcode(Node->getOpcode()) &&
620 ISD::getVPMaskIdx(Node->getOpcode()) == j)) // Skip mask operand.
621 {
622 // promote the vector operand.
623 SDValue Ext =
624 DAG.getNode(ISD::STRICT_FP_EXTEND, DL, {NewVecVT, MVT::Other},
625 {Node->getOperand(0), Node->getOperand(j)});
626 Operands[j] = Ext.getValue(0);
627 Chains.push_back(Ext.getValue(1));
628 } else
629 Operands[j] = Node->getOperand(j); // Skip no vector operand.
630
631 SDVTList VTs = DAG.getVTList(NewVecVT, Node->getValueType(1));
632
633 Operands[0] = DAG.getNode(ISD::TokenFactor, DL, MVT::Other, Chains);
634
635 SDValue Res =
636 DAG.getNode(Node->getOpcode(), DL, VTs, Operands, Node->getFlags());
637
638 SDValue Round =
639 DAG.getNode(ISD::STRICT_FP_ROUND, DL, {VecVT, MVT::Other},
640 {Res.getValue(1), Res.getValue(0),
641 DAG.getIntPtrConstant(0, DL, /*isTarget=*/true)});
642
643 Results.push_back(Round.getValue(0));
644 Results.push_back(Round.getValue(1));
645}
646
647void VectorLegalizer::Promote(SDNode *Node, SmallVectorImpl<SDValue> &Results) {
648 // For a few operations there is a specific concept for promotion based on
649 // the operand's type.
650 switch (Node->getOpcode()) {
651 case ISD::SINT_TO_FP:
652 case ISD::UINT_TO_FP:
655 // "Promote" the operation by extending the operand.
656 PromoteINT_TO_FP(Node, Results);
657 return;
658 case ISD::FP_TO_UINT:
659 case ISD::FP_TO_SINT:
662 // Promote the operation by extending the operand.
663 PromoteFP_TO_INT(Node, Results);
664 return;
665 case ISD::VP_SETCC:
666 case ISD::SETCC:
667 // Promote the operation by extending the operand.
668 PromoteSETCC(Node, Results);
669 return;
670 case ISD::STRICT_FADD:
671 case ISD::STRICT_FSUB:
672 case ISD::STRICT_FMUL:
673 case ISD::STRICT_FDIV:
675 case ISD::STRICT_FMA:
676 PromoteSTRICT(Node, Results);
677 return;
678 case ISD::FP_ROUND:
679 case ISD::FP_EXTEND:
680 // These operations are used to do promotion so they can't be promoted
681 // themselves.
682 llvm_unreachable("Don't know how to promote this operation!");
683 }
684
685 // There are currently two cases of vector promotion:
686 // 1) Bitcasting a vector of integers to a different type to a vector of the
687 // same overall length. For example, x86 promotes ISD::AND v2i32 to v1i64.
688 // 2) Extending a vector of floats to a vector of the same number of larger
689 // floats. For example, AArch64 promotes ISD::FADD on v4f16 to v4f32.
690 assert(Node->getNumValues() == 1 &&
691 "Can't promote a vector with multiple results!");
692 MVT VT = Node->getSimpleValueType(0);
693 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
694 SDLoc dl(Node);
695 SmallVector<SDValue, 4> Operands(Node->getNumOperands());
696
697 for (unsigned j = 0; j != Node->getNumOperands(); ++j) {
698 // Do not promote the mask operand of a VP OP.
699 bool SkipPromote = ISD::isVPOpcode(Node->getOpcode()) &&
700 ISD::getVPMaskIdx(Node->getOpcode()) == j;
701 if (Node->getOperand(j).getValueType().isVector() && !SkipPromote)
702 if (Node->getOperand(j)
703 .getValueType()
704 .getVectorElementType()
705 .isFloatingPoint() &&
707 Operands[j] = DAG.getNode(ISD::FP_EXTEND, dl, NVT, Node->getOperand(j));
708 else
709 Operands[j] = DAG.getNode(ISD::BITCAST, dl, NVT, Node->getOperand(j));
710 else
711 Operands[j] = Node->getOperand(j);
712 }
713
714 SDValue Res =
715 DAG.getNode(Node->getOpcode(), dl, NVT, Operands, Node->getFlags());
716
717 if ((VT.isFloatingPoint() && NVT.isFloatingPoint()) ||
720 Res = DAG.getNode(ISD::FP_ROUND, dl, VT, Res,
721 DAG.getIntPtrConstant(0, dl, /*isTarget=*/true));
722 else
723 Res = DAG.getNode(ISD::BITCAST, dl, VT, Res);
724
725 Results.push_back(Res);
726}
727
728void VectorLegalizer::PromoteINT_TO_FP(SDNode *Node,
730 // INT_TO_FP operations may require the input operand be promoted even
731 // when the type is otherwise legal.
732 bool IsStrict = Node->isStrictFPOpcode();
733 MVT VT = Node->getOperand(IsStrict ? 1 : 0).getSimpleValueType();
734 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
736 "Vectors have different number of elements!");
737
738 SDLoc dl(Node);
739 SmallVector<SDValue, 4> Operands(Node->getNumOperands());
740
741 unsigned Opc = (Node->getOpcode() == ISD::UINT_TO_FP ||
742 Node->getOpcode() == ISD::STRICT_UINT_TO_FP)
745 for (unsigned j = 0; j != Node->getNumOperands(); ++j) {
746 if (Node->getOperand(j).getValueType().isVector())
747 Operands[j] = DAG.getNode(Opc, dl, NVT, Node->getOperand(j));
748 else
749 Operands[j] = Node->getOperand(j);
750 }
751
752 if (IsStrict) {
753 SDValue Res = DAG.getNode(Node->getOpcode(), dl,
754 {Node->getValueType(0), MVT::Other}, Operands);
755 Results.push_back(Res);
756 Results.push_back(Res.getValue(1));
757 return;
758 }
759
760 SDValue Res =
761 DAG.getNode(Node->getOpcode(), dl, Node->getValueType(0), Operands);
762 Results.push_back(Res);
763}
764
765// For FP_TO_INT we promote the result type to a vector type with wider
766// elements and then truncate the result. This is different from the default
767// PromoteVector which uses bitcast to promote thus assumning that the
768// promoted vector type has the same overall size.
769void VectorLegalizer::PromoteFP_TO_INT(SDNode *Node,
771 MVT VT = Node->getSimpleValueType(0);
772 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
773 bool IsStrict = Node->isStrictFPOpcode();
775 "Vectors have different number of elements!");
776
777 unsigned NewOpc = Node->getOpcode();
778 // Change FP_TO_UINT to FP_TO_SINT if possible.
779 // TODO: Should we only do this if FP_TO_UINT itself isn't legal?
780 if (NewOpc == ISD::FP_TO_UINT &&
781 TLI.isOperationLegalOrCustom(ISD::FP_TO_SINT, NVT))
782 NewOpc = ISD::FP_TO_SINT;
783
784 if (NewOpc == ISD::STRICT_FP_TO_UINT &&
785 TLI.isOperationLegalOrCustom(ISD::STRICT_FP_TO_SINT, NVT))
786 NewOpc = ISD::STRICT_FP_TO_SINT;
787
788 SDLoc dl(Node);
789 SDValue Promoted, Chain;
790 if (IsStrict) {
791 Promoted = DAG.getNode(NewOpc, dl, {NVT, MVT::Other},
792 {Node->getOperand(0), Node->getOperand(1)});
793 Chain = Promoted.getValue(1);
794 } else
795 Promoted = DAG.getNode(NewOpc, dl, NVT, Node->getOperand(0));
796
797 // Assert that the converted value fits in the original type. If it doesn't
798 // (eg: because the value being converted is too big), then the result of the
799 // original operation was undefined anyway, so the assert is still correct.
800 if (Node->getOpcode() == ISD::FP_TO_UINT ||
801 Node->getOpcode() == ISD::STRICT_FP_TO_UINT)
802 NewOpc = ISD::AssertZext;
803 else
804 NewOpc = ISD::AssertSext;
805
806 Promoted = DAG.getNode(NewOpc, dl, NVT, Promoted,
807 DAG.getValueType(VT.getScalarType()));
808 Promoted = DAG.getNode(ISD::TRUNCATE, dl, VT, Promoted);
809 Results.push_back(Promoted);
810 if (IsStrict)
811 Results.push_back(Chain);
812}
813
814std::pair<SDValue, SDValue> VectorLegalizer::ExpandLoad(SDNode *N) {
815 LoadSDNode *LD = cast<LoadSDNode>(N);
816 return TLI.scalarizeVectorLoad(LD, DAG);
817}
818
819SDValue VectorLegalizer::ExpandStore(SDNode *N) {
820 StoreSDNode *ST = cast<StoreSDNode>(N);
821 SDValue TF = TLI.scalarizeVectorStore(ST, DAG);
822 return TF;
823}
824
825void VectorLegalizer::Expand(SDNode *Node, SmallVectorImpl<SDValue> &Results) {
826 switch (Node->getOpcode()) {
827 case ISD::LOAD: {
828 std::pair<SDValue, SDValue> Tmp = ExpandLoad(Node);
829 Results.push_back(Tmp.first);
830 Results.push_back(Tmp.second);
831 return;
832 }
833 case ISD::STORE:
834 Results.push_back(ExpandStore(Node));
835 return;
837 for (unsigned i = 0, e = Node->getNumValues(); i != e; ++i)
838 Results.push_back(Node->getOperand(i));
839 return;
841 Results.push_back(ExpandSEXTINREG(Node));
842 return;
844 Results.push_back(ExpandANY_EXTEND_VECTOR_INREG(Node));
845 return;
847 Results.push_back(ExpandSIGN_EXTEND_VECTOR_INREG(Node));
848 return;
850 Results.push_back(ExpandZERO_EXTEND_VECTOR_INREG(Node));
851 return;
852 case ISD::BSWAP:
853 Results.push_back(ExpandBSWAP(Node));
854 return;
855 case ISD::VP_BSWAP:
856 Results.push_back(TLI.expandVPBSWAP(Node, DAG));
857 return;
858 case ISD::VSELECT:
859 Results.push_back(ExpandVSELECT(Node));
860 return;
861 case ISD::VP_SELECT:
862 Results.push_back(ExpandVP_SELECT(Node));
863 return;
864 case ISD::VP_SREM:
865 case ISD::VP_UREM:
866 if (SDValue Expanded = ExpandVP_REM(Node)) {
867 Results.push_back(Expanded);
868 return;
869 }
870 break;
871 case ISD::SELECT:
872 Results.push_back(ExpandSELECT(Node));
873 return;
874 case ISD::SELECT_CC: {
875 if (Node->getValueType(0).isScalableVector()) {
876 EVT CondVT = TLI.getSetCCResultType(
877 DAG.getDataLayout(), *DAG.getContext(), Node->getValueType(0));
878 SDValue SetCC =
879 DAG.getNode(ISD::SETCC, SDLoc(Node), CondVT, Node->getOperand(0),
880 Node->getOperand(1), Node->getOperand(4));
881 Results.push_back(DAG.getSelect(SDLoc(Node), Node->getValueType(0), SetCC,
882 Node->getOperand(2),
883 Node->getOperand(3)));
884 return;
885 }
886 break;
887 }
888 case ISD::FP_TO_UINT:
889 ExpandFP_TO_UINT(Node, Results);
890 return;
891 case ISD::UINT_TO_FP:
892 ExpandUINT_TO_FLOAT(Node, Results);
893 return;
894 case ISD::FNEG:
895 Results.push_back(ExpandFNEG(Node));
896 return;
897 case ISD::FSUB:
898 ExpandFSUB(Node, Results);
899 return;
900 case ISD::SETCC:
901 case ISD::VP_SETCC:
902 ExpandSETCC(Node, Results);
903 return;
904 case ISD::ABS:
905 if (SDValue Expanded = TLI.expandABS(Node, DAG)) {
906 Results.push_back(Expanded);
907 return;
908 }
909 break;
910 case ISD::ABDS:
911 case ISD::ABDU:
912 if (SDValue Expanded = TLI.expandABD(Node, DAG)) {
913 Results.push_back(Expanded);
914 return;
915 }
916 break;
917 case ISD::BITREVERSE:
918 ExpandBITREVERSE(Node, Results);
919 return;
920 case ISD::VP_BITREVERSE:
921 if (SDValue Expanded = TLI.expandVPBITREVERSE(Node, DAG)) {
922 Results.push_back(Expanded);
923 return;
924 }
925 break;
926 case ISD::CTPOP:
927 if (SDValue Expanded = TLI.expandCTPOP(Node, DAG)) {
928 Results.push_back(Expanded);
929 return;
930 }
931 break;
932 case ISD::VP_CTPOP:
933 if (SDValue Expanded = TLI.expandVPCTPOP(Node, DAG)) {
934 Results.push_back(Expanded);
935 return;
936 }
937 break;
938 case ISD::CTLZ:
940 if (SDValue Expanded = TLI.expandCTLZ(Node, DAG)) {
941 Results.push_back(Expanded);
942 return;
943 }
944 break;
945 case ISD::VP_CTLZ:
946 case ISD::VP_CTLZ_ZERO_UNDEF:
947 if (SDValue Expanded = TLI.expandVPCTLZ(Node, DAG)) {
948 Results.push_back(Expanded);
949 return;
950 }
951 break;
952 case ISD::CTTZ:
954 if (SDValue Expanded = TLI.expandCTTZ(Node, DAG)) {
955 Results.push_back(Expanded);
956 return;
957 }
958 break;
959 case ISD::VP_CTTZ:
960 case ISD::VP_CTTZ_ZERO_UNDEF:
961 if (SDValue Expanded = TLI.expandVPCTTZ(Node, DAG)) {
962 Results.push_back(Expanded);
963 return;
964 }
965 break;
966 case ISD::FSHL:
967 case ISD::VP_FSHL:
968 case ISD::FSHR:
969 case ISD::VP_FSHR:
970 if (SDValue Expanded = TLI.expandFunnelShift(Node, DAG)) {
971 Results.push_back(Expanded);
972 return;
973 }
974 break;
975 case ISD::ROTL:
976 case ISD::ROTR:
977 if (SDValue Expanded = TLI.expandROT(Node, false /*AllowVectorOps*/, DAG)) {
978 Results.push_back(Expanded);
979 return;
980 }
981 break;
982 case ISD::FMINNUM:
983 case ISD::FMAXNUM:
984 if (SDValue Expanded = TLI.expandFMINNUM_FMAXNUM(Node, DAG)) {
985 Results.push_back(Expanded);
986 return;
987 }
988 break;
989 case ISD::FMINIMUM:
990 case ISD::FMAXIMUM:
991 if (SDValue Expanded = TLI.expandFMINIMUM_FMAXIMUM(Node, DAG)) {
992 Results.push_back(Expanded);
993 return;
994 }
995 break;
996 case ISD::SMIN:
997 case ISD::SMAX:
998 case ISD::UMIN:
999 case ISD::UMAX:
1000 if (SDValue Expanded = TLI.expandIntMINMAX(Node, DAG)) {
1001 Results.push_back(Expanded);
1002 return;
1003 }
1004 break;
1005 case ISD::UADDO:
1006 case ISD::USUBO:
1007 ExpandUADDSUBO(Node, Results);
1008 return;
1009 case ISD::SADDO:
1010 case ISD::SSUBO:
1011 ExpandSADDSUBO(Node, Results);
1012 return;
1013 case ISD::UMULO:
1014 case ISD::SMULO:
1015 ExpandMULO(Node, Results);
1016 return;
1017 case ISD::USUBSAT:
1018 case ISD::SSUBSAT:
1019 case ISD::UADDSAT:
1020 case ISD::SADDSAT:
1021 if (SDValue Expanded = TLI.expandAddSubSat(Node, DAG)) {
1022 Results.push_back(Expanded);
1023 return;
1024 }
1025 break;
1026 case ISD::USHLSAT:
1027 case ISD::SSHLSAT:
1028 if (SDValue Expanded = TLI.expandShlSat(Node, DAG)) {
1029 Results.push_back(Expanded);
1030 return;
1031 }
1032 break;
1035 // Expand the fpsosisat if it is scalable to prevent it from unrolling below.
1036 if (Node->getValueType(0).isScalableVector()) {
1037 if (SDValue Expanded = TLI.expandFP_TO_INT_SAT(Node, DAG)) {
1038 Results.push_back(Expanded);
1039 return;
1040 }
1041 }
1042 break;
1043 case ISD::SMULFIX:
1044 case ISD::UMULFIX:
1045 if (SDValue Expanded = TLI.expandFixedPointMul(Node, DAG)) {
1046 Results.push_back(Expanded);
1047 return;
1048 }
1049 break;
1050 case ISD::SMULFIXSAT:
1051 case ISD::UMULFIXSAT:
1052 // FIXME: We do not expand SMULFIXSAT/UMULFIXSAT here yet, not sure exactly
1053 // why. Maybe it results in worse codegen compared to the unroll for some
1054 // targets? This should probably be investigated. And if we still prefer to
1055 // unroll an explanation could be helpful.
1056 break;
1057 case ISD::SDIVFIX:
1058 case ISD::UDIVFIX:
1059 ExpandFixedPointDiv(Node, Results);
1060 return;
1061 case ISD::SDIVFIXSAT:
1062 case ISD::UDIVFIXSAT:
1063 break;
1064#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
1065 case ISD::STRICT_##DAGN:
1066#include "llvm/IR/ConstrainedOps.def"
1067 ExpandStrictFPOp(Node, Results);
1068 return;
1069 case ISD::VECREDUCE_ADD:
1070 case ISD::VECREDUCE_MUL:
1071 case ISD::VECREDUCE_AND:
1072 case ISD::VECREDUCE_OR:
1073 case ISD::VECREDUCE_XOR:
1084 Results.push_back(TLI.expandVecReduce(Node, DAG));
1085 return;
1088 Results.push_back(TLI.expandVecReduceSeq(Node, DAG));
1089 return;
1090 case ISD::SREM:
1091 case ISD::UREM:
1092 ExpandREM(Node, Results);
1093 return;
1094 case ISD::VP_MERGE:
1095 Results.push_back(ExpandVP_MERGE(Node));
1096 return;
1097 case ISD::FREM:
1098 if (tryExpandVecMathCall(Node, RTLIB::REM_F32, RTLIB::REM_F64,
1099 RTLIB::REM_F80, RTLIB::REM_F128,
1100 RTLIB::REM_PPCF128, Results))
1101 return;
1102
1103 break;
1104 }
1105
1106 SDValue Unrolled = DAG.UnrollVectorOp(Node);
1107 if (Node->getNumValues() == 1) {
1108 Results.push_back(Unrolled);
1109 } else {
1110 assert(Node->getNumValues() == Unrolled->getNumValues() &&
1111 "VectorLegalizer Expand returned wrong number of results!");
1112 for (unsigned I = 0, E = Unrolled->getNumValues(); I != E; ++I)
1113 Results.push_back(Unrolled.getValue(I));
1114 }
1115}
1116
1117SDValue VectorLegalizer::ExpandSELECT(SDNode *Node) {
1118 // Lower a select instruction where the condition is a scalar and the
1119 // operands are vectors. Lower this select to VSELECT and implement it
1120 // using XOR AND OR. The selector bit is broadcasted.
1121 EVT VT = Node->getValueType(0);
1122 SDLoc DL(Node);
1123
1124 SDValue Mask = Node->getOperand(0);
1125 SDValue Op1 = Node->getOperand(1);
1126 SDValue Op2 = Node->getOperand(2);
1127
1128 assert(VT.isVector() && !Mask.getValueType().isVector()
1129 && Op1.getValueType() == Op2.getValueType() && "Invalid type");
1130
1131 // If we can't even use the basic vector operations of
1132 // AND,OR,XOR, we will have to scalarize the op.
1133 // Notice that the operation may be 'promoted' which means that it is
1134 // 'bitcasted' to another type which is handled.
1135 // Also, we need to be able to construct a splat vector using either
1136 // BUILD_VECTOR or SPLAT_VECTOR.
1137 // FIXME: Should we also permit fixed-length SPLAT_VECTOR as a fallback to
1138 // BUILD_VECTOR?
1139 if (TLI.getOperationAction(ISD::AND, VT) == TargetLowering::Expand ||
1140 TLI.getOperationAction(ISD::XOR, VT) == TargetLowering::Expand ||
1141 TLI.getOperationAction(ISD::OR, VT) == TargetLowering::Expand ||
1142 TLI.getOperationAction(VT.isFixedLengthVector() ? ISD::BUILD_VECTOR
1144 VT) == TargetLowering::Expand)
1145 return DAG.UnrollVectorOp(Node);
1146
1147 // Generate a mask operand.
1149
1150 // What is the size of each element in the vector mask.
1151 EVT BitTy = MaskTy.getScalarType();
1152
1153 Mask = DAG.getSelect(DL, BitTy, Mask, DAG.getAllOnesConstant(DL, BitTy),
1154 DAG.getConstant(0, DL, BitTy));
1155
1156 // Broadcast the mask so that the entire vector is all one or all zero.
1157 Mask = DAG.getSplat(MaskTy, DL, Mask);
1158
1159 // Bitcast the operands to be the same type as the mask.
1160 // This is needed when we select between FP types because
1161 // the mask is a vector of integers.
1162 Op1 = DAG.getNode(ISD::BITCAST, DL, MaskTy, Op1);
1163 Op2 = DAG.getNode(ISD::BITCAST, DL, MaskTy, Op2);
1164
1165 SDValue NotMask = DAG.getNOT(DL, Mask, MaskTy);
1166
1167 Op1 = DAG.getNode(ISD::AND, DL, MaskTy, Op1, Mask);
1168 Op2 = DAG.getNode(ISD::AND, DL, MaskTy, Op2, NotMask);
1169 SDValue Val = DAG.getNode(ISD::OR, DL, MaskTy, Op1, Op2);
1170 return DAG.getNode(ISD::BITCAST, DL, Node->getValueType(0), Val);
1171}
1172
1173SDValue VectorLegalizer::ExpandSEXTINREG(SDNode *Node) {
1174 EVT VT = Node->getValueType(0);
1175
1176 // Make sure that the SRA and SHL instructions are available.
1177 if (TLI.getOperationAction(ISD::SRA, VT) == TargetLowering::Expand ||
1178 TLI.getOperationAction(ISD::SHL, VT) == TargetLowering::Expand)
1179 return DAG.UnrollVectorOp(Node);
1180
1181 SDLoc DL(Node);
1182 EVT OrigTy = cast<VTSDNode>(Node->getOperand(1))->getVT();
1183
1184 unsigned BW = VT.getScalarSizeInBits();
1185 unsigned OrigBW = OrigTy.getScalarSizeInBits();
1186 SDValue ShiftSz = DAG.getConstant(BW - OrigBW, DL, VT);
1187
1188 SDValue Op = DAG.getNode(ISD::SHL, DL, VT, Node->getOperand(0), ShiftSz);
1189 return DAG.getNode(ISD::SRA, DL, VT, Op, ShiftSz);
1190}
1191
1192// Generically expand a vector anyext in register to a shuffle of the relevant
1193// lanes into the appropriate locations, with other lanes left undef.
1194SDValue VectorLegalizer::ExpandANY_EXTEND_VECTOR_INREG(SDNode *Node) {
1195 SDLoc DL(Node);
1196 EVT VT = Node->getValueType(0);
1197 int NumElements = VT.getVectorNumElements();
1198 SDValue Src = Node->getOperand(0);
1199 EVT SrcVT = Src.getValueType();
1200 int NumSrcElements = SrcVT.getVectorNumElements();
1201
1202 // *_EXTEND_VECTOR_INREG SrcVT can be smaller than VT - so insert the vector
1203 // into a larger vector type.
1204 if (SrcVT.bitsLE(VT)) {
1205 assert((VT.getSizeInBits() % SrcVT.getScalarSizeInBits()) == 0 &&
1206 "ANY_EXTEND_VECTOR_INREG vector size mismatch");
1207 NumSrcElements = VT.getSizeInBits() / SrcVT.getScalarSizeInBits();
1208 SrcVT = EVT::getVectorVT(*DAG.getContext(), SrcVT.getScalarType(),
1209 NumSrcElements);
1210 Src = DAG.getNode(ISD::INSERT_SUBVECTOR, DL, SrcVT, DAG.getUNDEF(SrcVT),
1211 Src, DAG.getVectorIdxConstant(0, DL));
1212 }
1213
1214 // Build a base mask of undef shuffles.
1215 SmallVector<int, 16> ShuffleMask;
1216 ShuffleMask.resize(NumSrcElements, -1);
1217
1218 // Place the extended lanes into the correct locations.
1219 int ExtLaneScale = NumSrcElements / NumElements;
1220 int EndianOffset = DAG.getDataLayout().isBigEndian() ? ExtLaneScale - 1 : 0;
1221 for (int i = 0; i < NumElements; ++i)
1222 ShuffleMask[i * ExtLaneScale + EndianOffset] = i;
1223
1224 return DAG.getNode(
1225 ISD::BITCAST, DL, VT,
1226 DAG.getVectorShuffle(SrcVT, DL, Src, DAG.getUNDEF(SrcVT), ShuffleMask));
1227}
1228
1229SDValue VectorLegalizer::ExpandSIGN_EXTEND_VECTOR_INREG(SDNode *Node) {
1230 SDLoc DL(Node);
1231 EVT VT = Node->getValueType(0);
1232 SDValue Src = Node->getOperand(0);
1233 EVT SrcVT = Src.getValueType();
1234
1235 // First build an any-extend node which can be legalized above when we
1236 // recurse through it.
1237 SDValue Op = DAG.getNode(ISD::ANY_EXTEND_VECTOR_INREG, DL, VT, Src);
1238
1239 // Now we need sign extend. Do this by shifting the elements. Even if these
1240 // aren't legal operations, they have a better chance of being legalized
1241 // without full scalarization than the sign extension does.
1242 unsigned EltWidth = VT.getScalarSizeInBits();
1243 unsigned SrcEltWidth = SrcVT.getScalarSizeInBits();
1244 SDValue ShiftAmount = DAG.getConstant(EltWidth - SrcEltWidth, DL, VT);
1245 return DAG.getNode(ISD::SRA, DL, VT,
1246 DAG.getNode(ISD::SHL, DL, VT, Op, ShiftAmount),
1247 ShiftAmount);
1248}
1249
1250// Generically expand a vector zext in register to a shuffle of the relevant
1251// lanes into the appropriate locations, a blend of zero into the high bits,
1252// and a bitcast to the wider element type.
1253SDValue VectorLegalizer::ExpandZERO_EXTEND_VECTOR_INREG(SDNode *Node) {
1254 SDLoc DL(Node);
1255 EVT VT = Node->getValueType(0);
1256 int NumElements = VT.getVectorNumElements();
1257 SDValue Src = Node->getOperand(0);
1258 EVT SrcVT = Src.getValueType();
1259 int NumSrcElements = SrcVT.getVectorNumElements();
1260
1261 // *_EXTEND_VECTOR_INREG SrcVT can be smaller than VT - so insert the vector
1262 // into a larger vector type.
1263 if (SrcVT.bitsLE(VT)) {
1264 assert((VT.getSizeInBits() % SrcVT.getScalarSizeInBits()) == 0 &&
1265 "ZERO_EXTEND_VECTOR_INREG vector size mismatch");
1266 NumSrcElements = VT.getSizeInBits() / SrcVT.getScalarSizeInBits();
1267 SrcVT = EVT::getVectorVT(*DAG.getContext(), SrcVT.getScalarType(),
1268 NumSrcElements);
1269 Src = DAG.getNode(ISD::INSERT_SUBVECTOR, DL, SrcVT, DAG.getUNDEF(SrcVT),
1270 Src, DAG.getVectorIdxConstant(0, DL));
1271 }
1272
1273 // Build up a zero vector to blend into this one.
1274 SDValue Zero = DAG.getConstant(0, DL, SrcVT);
1275
1276 // Shuffle the incoming lanes into the correct position, and pull all other
1277 // lanes from the zero vector.
1278 auto ShuffleMask = llvm::to_vector<16>(llvm::seq<int>(0, NumSrcElements));
1279
1280 int ExtLaneScale = NumSrcElements / NumElements;
1281 int EndianOffset = DAG.getDataLayout().isBigEndian() ? ExtLaneScale - 1 : 0;
1282 for (int i = 0; i < NumElements; ++i)
1283 ShuffleMask[i * ExtLaneScale + EndianOffset] = NumSrcElements + i;
1284
1285 return DAG.getNode(ISD::BITCAST, DL, VT,
1286 DAG.getVectorShuffle(SrcVT, DL, Zero, Src, ShuffleMask));
1287}
1288
1289static void createBSWAPShuffleMask(EVT VT, SmallVectorImpl<int> &ShuffleMask) {
1290 int ScalarSizeInBytes = VT.getScalarSizeInBits() / 8;
1291 for (int I = 0, E = VT.getVectorNumElements(); I != E; ++I)
1292 for (int J = ScalarSizeInBytes - 1; J >= 0; --J)
1293 ShuffleMask.push_back((I * ScalarSizeInBytes) + J);
1294}
1295
1296SDValue VectorLegalizer::ExpandBSWAP(SDNode *Node) {
1297 EVT VT = Node->getValueType(0);
1298
1299 // Scalable vectors can't use shuffle expansion.
1300 if (VT.isScalableVector())
1301 return TLI.expandBSWAP(Node, DAG);
1302
1303 // Generate a byte wise shuffle mask for the BSWAP.
1304 SmallVector<int, 16> ShuffleMask;
1305 createBSWAPShuffleMask(VT, ShuffleMask);
1306 EVT ByteVT = EVT::getVectorVT(*DAG.getContext(), MVT::i8, ShuffleMask.size());
1307
1308 // Only emit a shuffle if the mask is legal.
1309 if (TLI.isShuffleMaskLegal(ShuffleMask, ByteVT)) {
1310 SDLoc DL(Node);
1311 SDValue Op = DAG.getNode(ISD::BITCAST, DL, ByteVT, Node->getOperand(0));
1312 Op = DAG.getVectorShuffle(ByteVT, DL, Op, DAG.getUNDEF(ByteVT), ShuffleMask);
1313 return DAG.getNode(ISD::BITCAST, DL, VT, Op);
1314 }
1315
1316 // If we have the appropriate vector bit operations, it is better to use them
1317 // than unrolling and expanding each component.
1318 if (TLI.isOperationLegalOrCustom(ISD::SHL, VT) &&
1319 TLI.isOperationLegalOrCustom(ISD::SRL, VT) &&
1320 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, VT) &&
1321 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, VT))
1322 return TLI.expandBSWAP(Node, DAG);
1323
1324 // Otherwise unroll.
1325 return DAG.UnrollVectorOp(Node);
1326}
1327
1328void VectorLegalizer::ExpandBITREVERSE(SDNode *Node,
1330 EVT VT = Node->getValueType(0);
1331
1332 // We can't unroll or use shuffles for scalable vectors.
1333 if (VT.isScalableVector()) {
1334 Results.push_back(TLI.expandBITREVERSE(Node, DAG));
1335 return;
1336 }
1337
1338 // If we have the scalar operation, it's probably cheaper to unroll it.
1339 if (TLI.isOperationLegalOrCustom(ISD::BITREVERSE, VT.getScalarType())) {
1340 SDValue Tmp = DAG.UnrollVectorOp(Node);
1341 Results.push_back(Tmp);
1342 return;
1343 }
1344
1345 // If the vector element width is a whole number of bytes, test if its legal
1346 // to BSWAP shuffle the bytes and then perform the BITREVERSE on the byte
1347 // vector. This greatly reduces the number of bit shifts necessary.
1348 unsigned ScalarSizeInBits = VT.getScalarSizeInBits();
1349 if (ScalarSizeInBits > 8 && (ScalarSizeInBits % 8) == 0) {
1350 SmallVector<int, 16> BSWAPMask;
1351 createBSWAPShuffleMask(VT, BSWAPMask);
1352
1353 EVT ByteVT = EVT::getVectorVT(*DAG.getContext(), MVT::i8, BSWAPMask.size());
1354 if (TLI.isShuffleMaskLegal(BSWAPMask, ByteVT) &&
1355 (TLI.isOperationLegalOrCustom(ISD::BITREVERSE, ByteVT) ||
1356 (TLI.isOperationLegalOrCustom(ISD::SHL, ByteVT) &&
1357 TLI.isOperationLegalOrCustom(ISD::SRL, ByteVT) &&
1358 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, ByteVT) &&
1359 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, ByteVT)))) {
1360 SDLoc DL(Node);
1361 SDValue Op = DAG.getNode(ISD::BITCAST, DL, ByteVT, Node->getOperand(0));
1362 Op = DAG.getVectorShuffle(ByteVT, DL, Op, DAG.getUNDEF(ByteVT),
1363 BSWAPMask);
1364 Op = DAG.getNode(ISD::BITREVERSE, DL, ByteVT, Op);
1365 Op = DAG.getNode(ISD::BITCAST, DL, VT, Op);
1366 Results.push_back(Op);
1367 return;
1368 }
1369 }
1370
1371 // If we have the appropriate vector bit operations, it is better to use them
1372 // than unrolling and expanding each component.
1373 if (TLI.isOperationLegalOrCustom(ISD::SHL, VT) &&
1374 TLI.isOperationLegalOrCustom(ISD::SRL, VT) &&
1375 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, VT) &&
1376 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, VT)) {
1377 Results.push_back(TLI.expandBITREVERSE(Node, DAG));
1378 return;
1379 }
1380
1381 // Otherwise unroll.
1382 SDValue Tmp = DAG.UnrollVectorOp(Node);
1383 Results.push_back(Tmp);
1384}
1385
1386SDValue VectorLegalizer::ExpandVSELECT(SDNode *Node) {
1387 // Implement VSELECT in terms of XOR, AND, OR
1388 // on platforms which do not support blend natively.
1389 SDLoc DL(Node);
1390
1391 SDValue Mask = Node->getOperand(0);
1392 SDValue Op1 = Node->getOperand(1);
1393 SDValue Op2 = Node->getOperand(2);
1394
1395 EVT VT = Mask.getValueType();
1396
1397 // If we can't even use the basic vector operations of
1398 // AND,OR,XOR, we will have to scalarize the op.
1399 // Notice that the operation may be 'promoted' which means that it is
1400 // 'bitcasted' to another type which is handled.
1401 if (TLI.getOperationAction(ISD::AND, VT) == TargetLowering::Expand ||
1402 TLI.getOperationAction(ISD::XOR, VT) == TargetLowering::Expand ||
1403 TLI.getOperationAction(ISD::OR, VT) == TargetLowering::Expand)
1404 return DAG.UnrollVectorOp(Node);
1405
1406 // This operation also isn't safe with AND, OR, XOR when the boolean type is
1407 // 0/1 and the select operands aren't also booleans, as we need an all-ones
1408 // vector constant to mask with.
1409 // FIXME: Sign extend 1 to all ones if that's legal on the target.
1410 auto BoolContents = TLI.getBooleanContents(Op1.getValueType());
1411 if (BoolContents != TargetLowering::ZeroOrNegativeOneBooleanContent &&
1412 !(BoolContents == TargetLowering::ZeroOrOneBooleanContent &&
1413 Op1.getValueType().getVectorElementType() == MVT::i1))
1414 return DAG.UnrollVectorOp(Node);
1415
1416 // If the mask and the type are different sizes, unroll the vector op. This
1417 // can occur when getSetCCResultType returns something that is different in
1418 // size from the operand types. For example, v4i8 = select v4i32, v4i8, v4i8.
1419 if (VT.getSizeInBits() != Op1.getValueSizeInBits())
1420 return DAG.UnrollVectorOp(Node);
1421
1422 // Bitcast the operands to be the same type as the mask.
1423 // This is needed when we select between FP types because
1424 // the mask is a vector of integers.
1425 Op1 = DAG.getNode(ISD::BITCAST, DL, VT, Op1);
1426 Op2 = DAG.getNode(ISD::BITCAST, DL, VT, Op2);
1427
1428 SDValue NotMask = DAG.getNOT(DL, Mask, VT);
1429
1430 Op1 = DAG.getNode(ISD::AND, DL, VT, Op1, Mask);
1431 Op2 = DAG.getNode(ISD::AND, DL, VT, Op2, NotMask);
1432 SDValue Val = DAG.getNode(ISD::OR, DL, VT, Op1, Op2);
1433 return DAG.getNode(ISD::BITCAST, DL, Node->getValueType(0), Val);
1434}
1435
1436SDValue VectorLegalizer::ExpandVP_SELECT(SDNode *Node) {
1437 // Implement VP_SELECT in terms of VP_XOR, VP_AND and VP_OR on platforms which
1438 // do not support it natively.
1439 SDLoc DL(Node);
1440
1441 SDValue Mask = Node->getOperand(0);
1442 SDValue Op1 = Node->getOperand(1);
1443 SDValue Op2 = Node->getOperand(2);
1444 SDValue EVL = Node->getOperand(3);
1445
1446 EVT VT = Mask.getValueType();
1447
1448 // If we can't even use the basic vector operations of
1449 // VP_AND,VP_OR,VP_XOR, we will have to scalarize the op.
1450 if (TLI.getOperationAction(ISD::VP_AND, VT) == TargetLowering::Expand ||
1451 TLI.getOperationAction(ISD::VP_XOR, VT) == TargetLowering::Expand ||
1452 TLI.getOperationAction(ISD::VP_OR, VT) == TargetLowering::Expand)
1453 return DAG.UnrollVectorOp(Node);
1454
1455 // This operation also isn't safe when the operands aren't also booleans.
1456 if (Op1.getValueType().getVectorElementType() != MVT::i1)
1457 return DAG.UnrollVectorOp(Node);
1458
1459 SDValue Ones = DAG.getAllOnesConstant(DL, VT);
1460 SDValue NotMask = DAG.getNode(ISD::VP_XOR, DL, VT, Mask, Ones, Ones, EVL);
1461
1462 Op1 = DAG.getNode(ISD::VP_AND, DL, VT, Op1, Mask, Ones, EVL);
1463 Op2 = DAG.getNode(ISD::VP_AND, DL, VT, Op2, NotMask, Ones, EVL);
1464 return DAG.getNode(ISD::VP_OR, DL, VT, Op1, Op2, Ones, EVL);
1465}
1466
1467SDValue VectorLegalizer::ExpandVP_MERGE(SDNode *Node) {
1468 // Implement VP_MERGE in terms of VSELECT. Construct a mask where vector
1469 // indices less than the EVL/pivot are true. Combine that with the original
1470 // mask for a full-length mask. Use a full-length VSELECT to select between
1471 // the true and false values.
1472 SDLoc DL(Node);
1473
1474 SDValue Mask = Node->getOperand(0);
1475 SDValue Op1 = Node->getOperand(1);
1476 SDValue Op2 = Node->getOperand(2);
1477 SDValue EVL = Node->getOperand(3);
1478
1479 EVT MaskVT = Mask.getValueType();
1480 bool IsFixedLen = MaskVT.isFixedLengthVector();
1481
1482 EVT EVLVecVT = EVT::getVectorVT(*DAG.getContext(), EVL.getValueType(),
1483 MaskVT.getVectorElementCount());
1484
1485 // If we can't construct the EVL mask efficiently, it's better to unroll.
1486 if ((IsFixedLen &&
1487 !TLI.isOperationLegalOrCustom(ISD::BUILD_VECTOR, EVLVecVT)) ||
1488 (!IsFixedLen &&
1489 (!TLI.isOperationLegalOrCustom(ISD::STEP_VECTOR, EVLVecVT) ||
1490 !TLI.isOperationLegalOrCustom(ISD::SPLAT_VECTOR, EVLVecVT))))
1491 return DAG.UnrollVectorOp(Node);
1492
1493 // If using a SETCC would result in a different type than the mask type,
1494 // unroll.
1495 if (TLI.getSetCCResultType(DAG.getDataLayout(), *DAG.getContext(),
1496 EVLVecVT) != MaskVT)
1497 return DAG.UnrollVectorOp(Node);
1498
1499 SDValue StepVec = DAG.getStepVector(DL, EVLVecVT);
1500 SDValue SplatEVL = DAG.getSplat(EVLVecVT, DL, EVL);
1501 SDValue EVLMask =
1502 DAG.getSetCC(DL, MaskVT, StepVec, SplatEVL, ISD::CondCode::SETULT);
1503
1504 SDValue FullMask = DAG.getNode(ISD::AND, DL, MaskVT, Mask, EVLMask);
1505 return DAG.getSelect(DL, Node->getValueType(0), FullMask, Op1, Op2);
1506}
1507
1508SDValue VectorLegalizer::ExpandVP_REM(SDNode *Node) {
1509 // Implement VP_SREM/UREM in terms of VP_SDIV/VP_UDIV, VP_MUL, VP_SUB.
1510 EVT VT = Node->getValueType(0);
1511
1512 unsigned DivOpc = Node->getOpcode() == ISD::VP_SREM ? ISD::VP_SDIV : ISD::VP_UDIV;
1513
1514 if (!TLI.isOperationLegalOrCustom(DivOpc, VT) ||
1515 !TLI.isOperationLegalOrCustom(ISD::VP_MUL, VT) ||
1516 !TLI.isOperationLegalOrCustom(ISD::VP_SUB, VT))
1517 return SDValue();
1518
1519 SDLoc DL(Node);
1520
1521 SDValue Dividend = Node->getOperand(0);
1522 SDValue Divisor = Node->getOperand(1);
1523 SDValue Mask = Node->getOperand(2);
1524 SDValue EVL = Node->getOperand(3);
1525
1526 // X % Y -> X-X/Y*Y
1527 SDValue Div = DAG.getNode(DivOpc, DL, VT, Dividend, Divisor, Mask, EVL);
1528 SDValue Mul = DAG.getNode(ISD::VP_MUL, DL, VT, Divisor, Div, Mask, EVL);
1529 return DAG.getNode(ISD::VP_SUB, DL, VT, Dividend, Mul, Mask, EVL);
1530}
1531
1532void VectorLegalizer::ExpandFP_TO_UINT(SDNode *Node,
1534 // Attempt to expand using TargetLowering.
1535 SDValue Result, Chain;
1536 if (TLI.expandFP_TO_UINT(Node, Result, Chain, DAG)) {
1537 Results.push_back(Result);
1538 if (Node->isStrictFPOpcode())
1539 Results.push_back(Chain);
1540 return;
1541 }
1542
1543 // Otherwise go ahead and unroll.
1544 if (Node->isStrictFPOpcode()) {
1545 UnrollStrictFPOp(Node, Results);
1546 return;
1547 }
1548
1549 Results.push_back(DAG.UnrollVectorOp(Node));
1550}
1551
1552void VectorLegalizer::ExpandUINT_TO_FLOAT(SDNode *Node,
1554 bool IsStrict = Node->isStrictFPOpcode();
1555 unsigned OpNo = IsStrict ? 1 : 0;
1556 SDValue Src = Node->getOperand(OpNo);
1557 EVT VT = Src.getValueType();
1558 SDLoc DL(Node);
1559
1560 // Attempt to expand using TargetLowering.
1562 SDValue Chain;
1563 if (TLI.expandUINT_TO_FP(Node, Result, Chain, DAG)) {
1564 Results.push_back(Result);
1565 if (IsStrict)
1566 Results.push_back(Chain);
1567 return;
1568 }
1569
1570 // Make sure that the SINT_TO_FP and SRL instructions are available.
1571 if (((!IsStrict && TLI.getOperationAction(ISD::SINT_TO_FP, VT) ==
1572 TargetLowering::Expand) ||
1573 (IsStrict && TLI.getOperationAction(ISD::STRICT_SINT_TO_FP, VT) ==
1574 TargetLowering::Expand)) ||
1575 TLI.getOperationAction(ISD::SRL, VT) == TargetLowering::Expand) {
1576 if (IsStrict) {
1577 UnrollStrictFPOp(Node, Results);
1578 return;
1579 }
1580
1581 Results.push_back(DAG.UnrollVectorOp(Node));
1582 return;
1583 }
1584
1585 unsigned BW = VT.getScalarSizeInBits();
1586 assert((BW == 64 || BW == 32) &&
1587 "Elements in vector-UINT_TO_FP must be 32 or 64 bits wide");
1588
1589 SDValue HalfWord = DAG.getConstant(BW / 2, DL, VT);
1590
1591 // Constants to clear the upper part of the word.
1592 // Notice that we can also use SHL+SHR, but using a constant is slightly
1593 // faster on x86.
1594 uint64_t HWMask = (BW == 64) ? 0x00000000FFFFFFFF : 0x0000FFFF;
1595 SDValue HalfWordMask = DAG.getConstant(HWMask, DL, VT);
1596
1597 // Two to the power of half-word-size.
1598 SDValue TWOHW =
1599 DAG.getConstantFP(1ULL << (BW / 2), DL, Node->getValueType(0));
1600
1601 // Clear upper part of LO, lower HI
1602 SDValue HI = DAG.getNode(ISD::SRL, DL, VT, Src, HalfWord);
1603 SDValue LO = DAG.getNode(ISD::AND, DL, VT, Src, HalfWordMask);
1604
1605 if (IsStrict) {
1606 // Convert hi and lo to floats
1607 // Convert the hi part back to the upper values
1608 // TODO: Can any fast-math-flags be set on these nodes?
1610 {Node->getValueType(0), MVT::Other},
1611 {Node->getOperand(0), HI});
1612 fHI = DAG.getNode(ISD::STRICT_FMUL, DL, {Node->getValueType(0), MVT::Other},
1613 {fHI.getValue(1), fHI, TWOHW});
1615 {Node->getValueType(0), MVT::Other},
1616 {Node->getOperand(0), LO});
1617
1618 SDValue TF = DAG.getNode(ISD::TokenFactor, DL, MVT::Other, fHI.getValue(1),
1619 fLO.getValue(1));
1620
1621 // Add the two halves
1622 SDValue Result =
1623 DAG.getNode(ISD::STRICT_FADD, DL, {Node->getValueType(0), MVT::Other},
1624 {TF, fHI, fLO});
1625
1626 Results.push_back(Result);
1627 Results.push_back(Result.getValue(1));
1628 return;
1629 }
1630
1631 // Convert hi and lo to floats
1632 // Convert the hi part back to the upper values
1633 // TODO: Can any fast-math-flags be set on these nodes?
1634 SDValue fHI = DAG.getNode(ISD::SINT_TO_FP, DL, Node->getValueType(0), HI);
1635 fHI = DAG.getNode(ISD::FMUL, DL, Node->getValueType(0), fHI, TWOHW);
1636 SDValue fLO = DAG.getNode(ISD::SINT_TO_FP, DL, Node->getValueType(0), LO);
1637
1638 // Add the two halves
1639 Results.push_back(
1640 DAG.getNode(ISD::FADD, DL, Node->getValueType(0), fHI, fLO));
1641}
1642
1643SDValue VectorLegalizer::ExpandFNEG(SDNode *Node) {
1644 if (TLI.isOperationLegalOrCustom(ISD::FSUB, Node->getValueType(0))) {
1645 SDLoc DL(Node);
1646 SDValue Zero = DAG.getConstantFP(-0.0, DL, Node->getValueType(0));
1647 // TODO: If FNEG had fast-math-flags, they'd get propagated to this FSUB.
1648 return DAG.getNode(ISD::FSUB, DL, Node->getValueType(0), Zero,
1649 Node->getOperand(0));
1650 }
1651 return DAG.UnrollVectorOp(Node);
1652}
1653
1654void VectorLegalizer::ExpandFSUB(SDNode *Node,
1656 // For floating-point values, (a-b) is the same as a+(-b). If FNEG is legal,
1657 // we can defer this to operation legalization where it will be lowered as
1658 // a+(-b).
1659 EVT VT = Node->getValueType(0);
1660 if (TLI.isOperationLegalOrCustom(ISD::FNEG, VT) &&
1661 TLI.isOperationLegalOrCustom(ISD::FADD, VT))
1662 return; // Defer to LegalizeDAG
1663
1664 SDValue Tmp = DAG.UnrollVectorOp(Node);
1665 Results.push_back(Tmp);
1666}
1667
1668void VectorLegalizer::ExpandSETCC(SDNode *Node,
1670 bool NeedInvert = false;
1671 bool IsVP = Node->getOpcode() == ISD::VP_SETCC;
1672 bool IsStrict = Node->getOpcode() == ISD::STRICT_FSETCC ||
1673 Node->getOpcode() == ISD::STRICT_FSETCCS;
1674 bool IsSignaling = Node->getOpcode() == ISD::STRICT_FSETCCS;
1675 unsigned Offset = IsStrict ? 1 : 0;
1676
1677 SDValue Chain = IsStrict ? Node->getOperand(0) : SDValue();
1678 SDValue LHS = Node->getOperand(0 + Offset);
1679 SDValue RHS = Node->getOperand(1 + Offset);
1680 SDValue CC = Node->getOperand(2 + Offset);
1681
1682 MVT OpVT = LHS.getSimpleValueType();
1683 ISD::CondCode CCCode = cast<CondCodeSDNode>(CC)->get();
1684
1685 if (TLI.getCondCodeAction(CCCode, OpVT) != TargetLowering::Expand) {
1686 if (IsStrict) {
1687 UnrollStrictFPOp(Node, Results);
1688 return;
1689 }
1690 Results.push_back(UnrollVSETCC(Node));
1691 return;
1692 }
1693
1694 SDValue Mask, EVL;
1695 if (IsVP) {
1696 Mask = Node->getOperand(3 + Offset);
1697 EVL = Node->getOperand(4 + Offset);
1698 }
1699
1700 SDLoc dl(Node);
1701 bool Legalized =
1702 TLI.LegalizeSetCCCondCode(DAG, Node->getValueType(0), LHS, RHS, CC, Mask,
1703 EVL, NeedInvert, dl, Chain, IsSignaling);
1704
1705 if (Legalized) {
1706 // If we expanded the SETCC by swapping LHS and RHS, or by inverting the
1707 // condition code, create a new SETCC node.
1708 if (CC.getNode()) {
1709 if (IsStrict) {
1710 LHS = DAG.getNode(Node->getOpcode(), dl, Node->getVTList(),
1711 {Chain, LHS, RHS, CC}, Node->getFlags());
1712 Chain = LHS.getValue(1);
1713 } else if (IsVP) {
1714 LHS = DAG.getNode(ISD::VP_SETCC, dl, Node->getValueType(0),
1715 {LHS, RHS, CC, Mask, EVL}, Node->getFlags());
1716 } else {
1717 LHS = DAG.getNode(ISD::SETCC, dl, Node->getValueType(0), LHS, RHS, CC,
1718 Node->getFlags());
1719 }
1720 }
1721
1722 // If we expanded the SETCC by inverting the condition code, then wrap
1723 // the existing SETCC in a NOT to restore the intended condition.
1724 if (NeedInvert) {
1725 if (!IsVP)
1726 LHS = DAG.getLogicalNOT(dl, LHS, LHS->getValueType(0));
1727 else
1728 LHS = DAG.getVPLogicalNOT(dl, LHS, Mask, EVL, LHS->getValueType(0));
1729 }
1730 } else {
1731 assert(!IsStrict && "Don't know how to expand for strict nodes.");
1732
1733 // Otherwise, SETCC for the given comparison type must be completely
1734 // illegal; expand it into a SELECT_CC.
1735 EVT VT = Node->getValueType(0);
1736 LHS =
1737 DAG.getNode(ISD::SELECT_CC, dl, VT, LHS, RHS,
1738 DAG.getBoolConstant(true, dl, VT, LHS.getValueType()),
1739 DAG.getBoolConstant(false, dl, VT, LHS.getValueType()), CC);
1740 LHS->setFlags(Node->getFlags());
1741 }
1742
1743 Results.push_back(LHS);
1744 if (IsStrict)
1745 Results.push_back(Chain);
1746}
1747
1748void VectorLegalizer::ExpandUADDSUBO(SDNode *Node,
1750 SDValue Result, Overflow;
1751 TLI.expandUADDSUBO(Node, Result, Overflow, DAG);
1752 Results.push_back(Result);
1753 Results.push_back(Overflow);
1754}
1755
1756void VectorLegalizer::ExpandSADDSUBO(SDNode *Node,
1758 SDValue Result, Overflow;
1759 TLI.expandSADDSUBO(Node, Result, Overflow, DAG);
1760 Results.push_back(Result);
1761 Results.push_back(Overflow);
1762}
1763
1764void VectorLegalizer::ExpandMULO(SDNode *Node,
1766 SDValue Result, Overflow;
1767 if (!TLI.expandMULO(Node, Result, Overflow, DAG))
1768 std::tie(Result, Overflow) = DAG.UnrollVectorOverflowOp(Node);
1769
1770 Results.push_back(Result);
1771 Results.push_back(Overflow);
1772}
1773
1774void VectorLegalizer::ExpandFixedPointDiv(SDNode *Node,
1776 SDNode *N = Node;
1777 if (SDValue Expanded = TLI.expandFixedPointDiv(N->getOpcode(), SDLoc(N),
1778 N->getOperand(0), N->getOperand(1), N->getConstantOperandVal(2), DAG))
1779 Results.push_back(Expanded);
1780}
1781
1782void VectorLegalizer::ExpandStrictFPOp(SDNode *Node,
1784 if (Node->getOpcode() == ISD::STRICT_UINT_TO_FP) {
1785 ExpandUINT_TO_FLOAT(Node, Results);
1786 return;
1787 }
1788 if (Node->getOpcode() == ISD::STRICT_FP_TO_UINT) {
1789 ExpandFP_TO_UINT(Node, Results);
1790 return;
1791 }
1792
1793 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1794 Node->getOpcode() == ISD::STRICT_FSETCCS) {
1795 ExpandSETCC(Node, Results);
1796 return;
1797 }
1798
1799 UnrollStrictFPOp(Node, Results);
1800}
1801
1802void VectorLegalizer::ExpandREM(SDNode *Node,
1804 assert((Node->getOpcode() == ISD::SREM || Node->getOpcode() == ISD::UREM) &&
1805 "Expected REM node");
1806
1808 if (!TLI.expandREM(Node, Result, DAG))
1809 Result = DAG.UnrollVectorOp(Node);
1810 Results.push_back(Result);
1811}
1812
1813// Try to expand libm nodes into vector math routine calls. Callers provide the
1814// LibFunc equivalent of the passed in Node, which is used to lookup mappings
1815// within TargetLibraryInfo. The only mappings considered are those where the
1816// result and all operands are the same vector type. While predicated nodes are
1817// not supported, we will emit calls to masked routines by passing in an all
1818// true mask.
1819bool VectorLegalizer::tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall LC,
1821 // Chain must be propagated but currently strict fp operations are down
1822 // converted to their none strict counterpart.
1823 assert(!Node->isStrictFPOpcode() && "Unexpected strict fp operation!");
1824
1825 const char *LCName = TLI.getLibcallName(LC);
1826 if (!LCName)
1827 return false;
1828 LLVM_DEBUG(dbgs() << "Looking for vector variant of " << LCName << "\n");
1829
1830 EVT VT = Node->getValueType(0);
1832
1833 // Lookup a vector function equivalent to the specified libcall. Prefer
1834 // unmasked variants but we will generate a mask if need be.
1835 const TargetLibraryInfo &TLibInfo = DAG.getLibInfo();
1836 const VecDesc *VD = TLibInfo.getVectorMappingInfo(LCName, VL, false);
1837 if (!VD)
1838 VD = TLibInfo.getVectorMappingInfo(LCName, VL, /*Masked=*/true);
1839 if (!VD)
1840 return false;
1841
1842 LLVMContext *Ctx = DAG.getContext();
1843 Type *Ty = VT.getTypeForEVT(*Ctx);
1844 Type *ScalarTy = Ty->getScalarType();
1845
1846 // Construct a scalar function type based on Node's operands.
1848 for (unsigned i = 0; i < Node->getNumOperands(); ++i) {
1849 assert(Node->getOperand(i).getValueType() == VT &&
1850 "Expected matching vector types!");
1851 ArgTys.push_back(ScalarTy);
1852 }
1853 FunctionType *ScalarFTy = FunctionType::get(ScalarTy, ArgTys, false);
1854
1855 // Generate call information for the vector function.
1856 const std::string MangledName = VD->getVectorFunctionABIVariantString();
1857 auto OptVFInfo = VFABI::tryDemangleForVFABI(MangledName, ScalarFTy);
1858 if (!OptVFInfo)
1859 return false;
1860
1861 LLVM_DEBUG(dbgs() << "Found vector variant " << VD->getVectorFnName()
1862 << "\n");
1863
1864 // Sanity check just in case OptVFInfo has unexpected parameters.
1865 if (OptVFInfo->Shape.Parameters.size() !=
1866 Node->getNumOperands() + VD->isMasked())
1867 return false;
1868
1869 // Collect vector call operands.
1870
1871 SDLoc DL(Node);
1874 Entry.IsSExt = false;
1875 Entry.IsZExt = false;
1876
1877 unsigned OpNum = 0;
1878 for (auto &VFParam : OptVFInfo->Shape.Parameters) {
1879 if (VFParam.ParamKind == VFParamKind::GlobalPredicate) {
1880 EVT MaskVT = TLI.getSetCCResultType(DAG.getDataLayout(), *Ctx, VT);
1881 Entry.Node = DAG.getBoolConstant(true, DL, MaskVT, VT);
1882 Entry.Ty = MaskVT.getTypeForEVT(*Ctx);
1883 Args.push_back(Entry);
1884 continue;
1885 }
1886
1887 // Only vector operands are supported.
1888 if (VFParam.ParamKind != VFParamKind::Vector)
1889 return false;
1890
1891 Entry.Node = Node->getOperand(OpNum++);
1892 Entry.Ty = Ty;
1893 Args.push_back(Entry);
1894 }
1895
1896 // Emit a call to the vector function.
1897 SDValue Callee = DAG.getExternalSymbol(VD->getVectorFnName().data(),
1898 TLI.getPointerTy(DAG.getDataLayout()));
1900 CLI.setDebugLoc(DL)
1901 .setChain(DAG.getEntryNode())
1902 .setLibCallee(CallingConv::C, Ty, Callee, std::move(Args));
1903
1904 std::pair<SDValue, SDValue> CallResult = TLI.LowerCallTo(CLI);
1905 Results.push_back(CallResult.first);
1906 return true;
1907}
1908
1909/// Try to expand the node to a vector libcall based on the result type.
1910bool VectorLegalizer::tryExpandVecMathCall(
1911 SDNode *Node, RTLIB::Libcall Call_F32, RTLIB::Libcall Call_F64,
1912 RTLIB::Libcall Call_F80, RTLIB::Libcall Call_F128,
1915 Node->getValueType(0).getVectorElementType(), Call_F32, Call_F64,
1916 Call_F80, Call_F128, Call_PPCF128);
1917
1918 if (LC == RTLIB::UNKNOWN_LIBCALL)
1919 return false;
1920
1921 return tryExpandVecMathCall(Node, LC, Results);
1922}
1923
1924void VectorLegalizer::UnrollStrictFPOp(SDNode *Node,
1926 EVT VT = Node->getValueType(0);
1927 EVT EltVT = VT.getVectorElementType();
1928 unsigned NumElems = VT.getVectorNumElements();
1929 unsigned NumOpers = Node->getNumOperands();
1930 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
1931
1932 EVT TmpEltVT = EltVT;
1933 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1934 Node->getOpcode() == ISD::STRICT_FSETCCS)
1935 TmpEltVT = TLI.getSetCCResultType(DAG.getDataLayout(),
1936 *DAG.getContext(), TmpEltVT);
1937
1938 EVT ValueVTs[] = {TmpEltVT, MVT::Other};
1939 SDValue Chain = Node->getOperand(0);
1940 SDLoc dl(Node);
1941
1942 SmallVector<SDValue, 32> OpValues;
1943 SmallVector<SDValue, 32> OpChains;
1944 for (unsigned i = 0; i < NumElems; ++i) {
1946 SDValue Idx = DAG.getVectorIdxConstant(i, dl);
1947
1948 // The Chain is the first operand.
1949 Opers.push_back(Chain);
1950
1951 // Now process the remaining operands.
1952 for (unsigned j = 1; j < NumOpers; ++j) {
1953 SDValue Oper = Node->getOperand(j);
1954 EVT OperVT = Oper.getValueType();
1955
1956 if (OperVT.isVector())
1957 Oper = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl,
1958 OperVT.getVectorElementType(), Oper, Idx);
1959
1960 Opers.push_back(Oper);
1961 }
1962
1963 SDValue ScalarOp = DAG.getNode(Node->getOpcode(), dl, ValueVTs, Opers);
1964 SDValue ScalarResult = ScalarOp.getValue(0);
1965 SDValue ScalarChain = ScalarOp.getValue(1);
1966
1967 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1968 Node->getOpcode() == ISD::STRICT_FSETCCS)
1969 ScalarResult = DAG.getSelect(dl, EltVT, ScalarResult,
1970 DAG.getAllOnesConstant(dl, EltVT),
1971 DAG.getConstant(0, dl, EltVT));
1972
1973 OpValues.push_back(ScalarResult);
1974 OpChains.push_back(ScalarChain);
1975 }
1976
1977 SDValue Result = DAG.getBuildVector(VT, dl, OpValues);
1978 SDValue NewChain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OpChains);
1979
1980 Results.push_back(Result);
1981 Results.push_back(NewChain);
1982}
1983
1984SDValue VectorLegalizer::UnrollVSETCC(SDNode *Node) {
1985 EVT VT = Node->getValueType(0);
1986 unsigned NumElems = VT.getVectorNumElements();
1987 EVT EltVT = VT.getVectorElementType();
1988 SDValue LHS = Node->getOperand(0);
1989 SDValue RHS = Node->getOperand(1);
1990 SDValue CC = Node->getOperand(2);
1991 EVT TmpEltVT = LHS.getValueType().getVectorElementType();
1992 SDLoc dl(Node);
1993 SmallVector<SDValue, 8> Ops(NumElems);
1994 for (unsigned i = 0; i < NumElems; ++i) {
1995 SDValue LHSElem = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl, TmpEltVT, LHS,
1996 DAG.getVectorIdxConstant(i, dl));
1997 SDValue RHSElem = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl, TmpEltVT, RHS,
1998 DAG.getVectorIdxConstant(i, dl));
1999 Ops[i] = DAG.getNode(ISD::SETCC, dl,
2000 TLI.getSetCCResultType(DAG.getDataLayout(),
2001 *DAG.getContext(), TmpEltVT),
2002 LHSElem, RHSElem, CC);
2003 Ops[i] = DAG.getSelect(dl, EltVT, Ops[i], DAG.getAllOnesConstant(dl, EltVT),
2004 DAG.getConstant(0, dl, EltVT));
2005 }
2006 return DAG.getBuildVector(VT, dl, Ops);
2007}
2008
2010 return VectorLegalizer(*this).Run();
2011}
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
BlockVerifier::State From
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
#define LLVM_DEBUG(X)
Definition: Debug.h:101
This file defines the DenseMap class.
static void createBSWAPShuffleMask(EVT VT, SmallVectorImpl< int > &ShuffleMask)
#define I(x, y, z)
Definition: MD5.cpp:58
mir Rename Register Operands
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
This file defines the SmallVector class.
This file describes how to lower LLVM code to machine code.
Value * RHS
Value * LHS
BinaryOperator * Mul
DEMANGLE_DUMP_METHOD void dump() const
This class represents an Operation in the Expression.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition: DenseMap.h:220
size_t size() const
Definition: Function.h:808
This is an important class for using LLVM in a threaded context.
Definition: LLVMContext.h:67
This class is used to represent ISD::LOAD nodes.
Machine Value Type.
unsigned getVectorNumElements() const
bool isVector() const
Return true if this is a vector value type.
MVT getVectorElementType() const
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
MVT getScalarType() const
If this is a vector, return the element type, otherwise return this.
MutableArrayRef - Represent a mutable reference to an array (0 or more elements consecutively in memo...
Definition: ArrayRef.h:307
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
TypeSize getValueSizeInBits() const
Returns the size of the value in bits.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
Definition: SelectionDAG.h:225
bool LegalizeVectors()
This transforms the SelectionDAG into a SelectionDAG that only uses vector math operations supported ...
const TargetLowering & getTargetLoweringInfo() const
Definition: SelectionDAG.h:478
ilist< SDNode >::iterator allnodes_iterator
Definition: SelectionDAG.h:534
bool empty() const
Definition: SmallVector.h:94
size_t size() const
Definition: SmallVector.h:91
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Definition: SmallVector.h:586
void resize(size_type N)
Definition: SmallVector.h:651
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
This class is used to represent ISD::STORE nodes.
constexpr const char * data() const
data - Get a pointer to the start of the string (which may not be null terminated).
Definition: StringRef.h:131
Provides information about what library functions are available for the current target.
const VecDesc * getVectorMappingInfo(StringRef F, const ElementCount &VF, bool Masked) const
LegalizeAction
This enum indicates whether operations are valid for a target, and if not, what action should be used...
virtual EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const
Return the ValueType of the result of SETCC operations.
std::vector< ArgListEntry > ArgListTy
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
The instances of the Type class are immutable: once they are created, they are never changed.
Definition: Type.h:45
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition: Type.h:348
Provides info so a possible vectorization of a function can be computed.
bool isMasked() const
std::string getVectorFunctionABIVariantString() const
Returns a vector function ABI variant string on the form: ZGV<isa><mask><vlen><vparams><scalarname>(<...
StringRef getVectorFnName() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
Definition: BitmaskEnum.h:121
@ C
The default llvm calling convention, compatible with C.
Definition: CallingConv.h:34
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
Definition: ISDOpcodes.h:751
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
Definition: ISDOpcodes.h:237
@ CTLZ_ZERO_UNDEF
Definition: ISDOpcodes.h:724
@ STRICT_FSETCC
STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used for floating-point operands only.
Definition: ISDOpcodes.h:477
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
Definition: ISDOpcodes.h:1346
@ VECREDUCE_SMIN
Definition: ISDOpcodes.h:1377
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
Definition: ISDOpcodes.h:251
@ INSERT_SUBVECTOR
INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2 inserted into VECTOR1.
Definition: ISDOpcodes.h:560
@ BSWAP
Byte Swap and Counting operators.
Definition: ISDOpcodes.h:715
@ SMULFIX
RESULT = [US]MULFIX(LHS, RHS, SCALE) - Perform fixed point multiplication on 2 integers with the same...
Definition: ISDOpcodes.h:368
@ FMAXNUM_IEEE
Definition: ISDOpcodes.h:986
@ ADD
Simple integer binary arithmetic operators.
Definition: ISDOpcodes.h:240
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
Definition: ISDOpcodes.h:1038
@ SMULFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition: ISDOpcodes.h:374
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
Definition: ISDOpcodes.h:784
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
Definition: ISDOpcodes.h:484
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
Definition: ISDOpcodes.h:791
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
Definition: ISDOpcodes.h:1362
@ FADD
Simple binary floating point operators.
Definition: ISDOpcodes.h:391
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
Definition: ISDOpcodes.h:1366
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
Definition: ISDOpcodes.h:689
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
Definition: ISDOpcodes.h:821
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
Definition: ISDOpcodes.h:256
@ VECREDUCE_SMAX
Definition: ISDOpcodes.h:1376
@ STRICT_FSETCCS
Definition: ISDOpcodes.h:478
@ FPTRUNC_ROUND
Definition: ISDOpcodes.h:481
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
Definition: ISDOpcodes.h:904
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
Definition: ISDOpcodes.h:940
@ SDIVFIX
RESULT = [US]DIVFIX(LHS, RHS, SCALE) - Perform fixed point division on 2 integers with the same width...
Definition: ISDOpcodes.h:381
@ STRICT_FSQRT
Constrained versions of libm-equivalent floating point intrinsics.
Definition: ISDOpcodes.h:412
@ SIGN_EXTEND
Conversion operators.
Definition: ISDOpcodes.h:775
@ STRICT_UINT_TO_FP
Definition: ISDOpcodes.h:451
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
Definition: ISDOpcodes.h:1359
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
Definition: ISDOpcodes.h:723
@ VECREDUCE_FMIN
Definition: ISDOpcodes.h:1363
@ FNEG
Perform various unary floating-point operations inspired by libm.
Definition: ISDOpcodes.h:931
@ SSUBO
Same for subtraction.
Definition: ISDOpcodes.h:328
@ STEP_VECTOR
STEP_VECTOR(IMM) - Returns a scalable vector whose lanes are comprised of a linear sequence of unsign...
Definition: ISDOpcodes.h:647
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
Definition: ISDOpcodes.h:501
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
Definition: ISDOpcodes.h:350
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
Definition: ISDOpcodes.h:728
@ VECREDUCE_UMAX
Definition: ISDOpcodes.h:1378
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
Definition: ISDOpcodes.h:628
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
Definition: ISDOpcodes.h:324
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
Definition: ISDOpcodes.h:1371
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
Definition: ISDOpcodes.h:652
@ SHL
Shift and rotation operations.
Definition: ISDOpcodes.h:706
@ FMINNUM_IEEE
FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimumNumber or maximumNumber on two values,...
Definition: ISDOpcodes.h:985
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
Definition: ISDOpcodes.h:536
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
Definition: ISDOpcodes.h:781
@ FP_TO_UINT_SAT
Definition: ISDOpcodes.h:857
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
Definition: ISDOpcodes.h:743
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum or maximum on two values.
Definition: ISDOpcodes.h:972
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
Definition: ISDOpcodes.h:360
@ SMULO
Same for multiplication.
Definition: ISDOpcodes.h:332
@ ANY_EXTEND_VECTOR_INREG
ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register any-extension of the low la...
Definition: ISDOpcodes.h:810
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
Definition: ISDOpcodes.h:799
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
Definition: ISDOpcodes.h:675
@ SDIVFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition: ISDOpcodes.h:387
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition: ISDOpcodes.h:889
@ VSELECT
Select with a vector condition (op #0) and two vector operands (ops #1 and #2), returning a vector re...
Definition: ISDOpcodes.h:737
@ STRICT_SINT_TO_FP
STRICT_[US]INT_TO_FP - Convert a signed or unsigned integer to a floating point value.
Definition: ISDOpcodes.h:450
@ VECREDUCE_UMIN
Definition: ISDOpcodes.h:1379
@ STRICT_FP_TO_UINT
Definition: ISDOpcodes.h:444
@ STRICT_FP_ROUND
X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision ...
Definition: ISDOpcodes.h:466
@ STRICT_FP_TO_SINT
STRICT_FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition: ISDOpcodes.h:443
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
Definition: ISDOpcodes.h:991
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition: ISDOpcodes.h:837
@ STRICT_FP_EXTEND
X = STRICT_FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition: ISDOpcodes.h:471
@ AND
Bitwise operators - logical and, logical or, logical xor.
Definition: ISDOpcodes.h:681
@ VECREDUCE_FMUL
Definition: ISDOpcodes.h:1360
@ STRICT_FADD
Constrained versions of the binary floating point operators.
Definition: ISDOpcodes.h:401
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
Definition: ISDOpcodes.h:52
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
Definition: ISDOpcodes.h:945
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
Definition: ISDOpcodes.h:870
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
Definition: ISDOpcodes.h:832
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
Definition: ISDOpcodes.h:856
@ VECREDUCE_FMINIMUM
Definition: ISDOpcodes.h:1367
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
Definition: ISDOpcodes.h:787
@ VECREDUCE_SEQ_FMUL
Definition: ISDOpcodes.h:1347
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
Definition: ISDOpcodes.h:61
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
Definition: ISDOpcodes.h:494
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
Definition: ISDOpcodes.h:341
@ AssertZext
Definition: ISDOpcodes.h:62
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
Definition: ISDOpcodes.h:516
std::optional< unsigned > getVPMaskIdx(unsigned Opcode)
The operand position of the vector mask.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
Definition: ISDOpcodes.h:1530
LoadExtType
LoadExtType enum - This enum defines the three variants of LOADEXT (load with extension).
Definition: ISDOpcodes.h:1510
bool isVPOpcode(unsigned Opcode)
Whether this is a vector-predicated Opcode.
Libcall
RTLIB::Libcall enum - This enum defines all of the runtime library calls the backend can emit.
Libcall getFPLibCall(EVT VT, Libcall Call_F32, Libcall Call_F64, Libcall Call_F80, Libcall Call_F128, Libcall Call_PPCF128)
GetFPLibCall - Helper to return the right libcall for the given floating point type,...
ManagedStatic< cl::opt< FnT >, OptCreatorT > Action
std::optional< VFInfo > tryDemangleForVFABI(StringRef MangledName, const FunctionType *FTy)
Function to construct a VFInfo out of a mangled names in the following format:
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Offset
Definition: DWP.cpp:456
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1729
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
#define N
Extended Value Type.
Definition: ValueTypes.h:34
EVT changeVectorElementTypeToInteger() const
Return a vector with the same number of elements as this vector, but with the element type converted ...
Definition: ValueTypes.h:93
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
Definition: ValueTypes.h:73
ElementCount getVectorElementCount() const
Definition: ValueTypes.h:340
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
Definition: ValueTypes.h:358
uint64_t getScalarSizeInBits() const
Definition: ValueTypes.h:370
bool isFixedLengthVector() const
Definition: ValueTypes.h:177
bool isVector() const
Return true if this is a vector value type.
Definition: ValueTypes.h:167
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
Definition: ValueTypes.h:313
Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
Definition: ValueTypes.cpp:202
bool isScalableVector() const
Return true if this is a vector type where the runtime length is machine dependent.
Definition: ValueTypes.h:173
EVT getVectorElementType() const
Given a vector type, return the type of each element.
Definition: ValueTypes.h:318
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
Definition: ValueTypes.h:326
bool bitsLE(EVT VT) const
Return true if this has no more bits than VT.
Definition: ValueTypes.h:298
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
This structure contains all information that is necessary for lowering calls.