LLVM 23.0.0git
SPIRVLegalizerInfo.cpp
Go to the documentation of this file.
1//===- SPIRVLegalizerInfo.cpp --- SPIR-V Legalization Rules ------*- C++ -*-==//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the targeting of the Machinelegalizer class for SPIR-V.
10//
11//===----------------------------------------------------------------------===//
12
13#include "SPIRVLegalizerInfo.h"
14#include "SPIRV.h"
15#include "SPIRVGlobalRegistry.h"
16#include "SPIRVSubtarget.h"
17#include "SPIRVUtils.h"
24#include "llvm/IR/IntrinsicsSPIRV.h"
25#include "llvm/Support/Debug.h"
27
28using namespace llvm;
29using namespace llvm::LegalizeActions;
30using namespace llvm::LegalityPredicates;
31
32#define DEBUG_TYPE "spirv-legalizer"
33
34LegalityPredicate typeOfExtendedScalars(unsigned TypeIdx, bool IsExtendedInts) {
35 return [IsExtendedInts, TypeIdx](const LegalityQuery &Query) {
36 const LLT Ty = Query.Types[TypeIdx];
37 return IsExtendedInts && Ty.isValid() && Ty.isScalar();
38 };
39}
40
42 using namespace TargetOpcode;
43
44 this->ST = &ST;
45 GR = ST.getSPIRVGlobalRegistry();
46
47 const LLT s1 = LLT::scalar(1);
48 const LLT s8 = LLT::scalar(8);
49 const LLT s16 = LLT::scalar(16);
50 const LLT s32 = LLT::scalar(32);
51 const LLT s64 = LLT::scalar(64);
52 const LLT s128 = LLT::scalar(128);
53
54 const LLT v16s64 = LLT::fixed_vector(16, 64);
55 const LLT v16s32 = LLT::fixed_vector(16, 32);
56 const LLT v16s16 = LLT::fixed_vector(16, 16);
57 const LLT v16s8 = LLT::fixed_vector(16, 8);
58 const LLT v16s1 = LLT::fixed_vector(16, 1);
59
60 const LLT v8s64 = LLT::fixed_vector(8, 64);
61 const LLT v8s32 = LLT::fixed_vector(8, 32);
62 const LLT v8s16 = LLT::fixed_vector(8, 16);
63 const LLT v8s8 = LLT::fixed_vector(8, 8);
64 const LLT v8s1 = LLT::fixed_vector(8, 1);
65
66 const LLT v4s64 = LLT::fixed_vector(4, 64);
67 const LLT v4s32 = LLT::fixed_vector(4, 32);
68 const LLT v4s16 = LLT::fixed_vector(4, 16);
69 const LLT v4s8 = LLT::fixed_vector(4, 8);
70 const LLT v4s1 = LLT::fixed_vector(4, 1);
71
72 const LLT v3s64 = LLT::fixed_vector(3, 64);
73 const LLT v3s32 = LLT::fixed_vector(3, 32);
74 const LLT v3s16 = LLT::fixed_vector(3, 16);
75 const LLT v3s8 = LLT::fixed_vector(3, 8);
76 const LLT v3s1 = LLT::fixed_vector(3, 1);
77
78 const LLT v2s64 = LLT::fixed_vector(2, 64);
79 const LLT v2s32 = LLT::fixed_vector(2, 32);
80 const LLT v2s16 = LLT::fixed_vector(2, 16);
81 const LLT v2s8 = LLT::fixed_vector(2, 8);
82 const LLT v2s1 = LLT::fixed_vector(2, 1);
83
84 const unsigned PSize = ST.getPointerSize();
85 const LLT p0 = LLT::pointer(0, PSize); // Function
86 const LLT p1 = LLT::pointer(1, PSize); // CrossWorkgroup
87 const LLT p2 = LLT::pointer(2, PSize); // UniformConstant
88 const LLT p3 = LLT::pointer(3, PSize); // Workgroup
89 const LLT p4 = LLT::pointer(4, PSize); // Generic
90 const LLT p5 =
91 LLT::pointer(5, PSize); // Input, SPV_INTEL_usm_storage_classes (Device)
92 const LLT p6 = LLT::pointer(6, PSize); // SPV_INTEL_usm_storage_classes (Host)
93 const LLT p7 = LLT::pointer(7, PSize); // Input
94 const LLT p8 = LLT::pointer(8, PSize); // Output
95 const LLT p9 =
96 LLT::pointer(9, PSize); // CodeSectionINTEL, SPV_INTEL_function_pointers
97 const LLT p10 = LLT::pointer(10, PSize); // Private
98 const LLT p11 = LLT::pointer(11, PSize); // StorageBuffer
99 const LLT p12 = LLT::pointer(12, PSize); // Uniform
100 const LLT p13 = LLT::pointer(13, PSize); // PushConstant
101
102 // TODO: remove copy-pasting here by using concatenation in some way.
103 auto allPtrsScalarsAndVectors = {
104 p0, p1, p2, p3, p4, p5, p6, p7, p8,
105 p9, p10, p11, p12, p13, s1, s8, s16, s32,
106 s64, v2s1, v2s8, v2s16, v2s32, v2s64, v3s1, v3s8, v3s16,
107 v3s32, v3s64, v4s1, v4s8, v4s16, v4s32, v4s64, v8s1, v8s8,
108 v8s16, v8s32, v8s64, v16s1, v16s8, v16s16, v16s32, v16s64};
109
110 auto allVectors = {v2s1, v2s8, v2s16, v2s32, v2s64, v3s1, v3s8,
111 v3s16, v3s32, v3s64, v4s1, v4s8, v4s16, v4s32,
112 v4s64, v8s1, v8s8, v8s16, v8s32, v8s64, v16s1,
113 v16s8, v16s16, v16s32, v16s64};
114
115 auto allShaderVectors = {v2s1, v2s8, v2s16, v2s32, v2s64,
116 v3s1, v3s8, v3s16, v3s32, v3s64,
117 v4s1, v4s8, v4s16, v4s32, v4s64};
118
119 auto allScalars = {s1, s8, s16, s32, s64};
120
121 auto allScalarsAndVectors = {
122 s1, s8, s16, s32, s64, s128, v2s1, v2s8,
123 v2s16, v2s32, v2s64, v3s1, v3s8, v3s16, v3s32, v3s64,
124 v4s1, v4s8, v4s16, v4s32, v4s64, v8s1, v8s8, v8s16,
125 v8s32, v8s64, v16s1, v16s8, v16s16, v16s32, v16s64};
126
127 auto allIntScalarsAndVectors = {
128 s8, s16, s32, s64, s128, v2s8, v2s16, v2s32, v2s64,
129 v3s8, v3s16, v3s32, v3s64, v4s8, v4s16, v4s32, v4s64, v8s8,
130 v8s16, v8s32, v8s64, v16s8, v16s16, v16s32, v16s64};
131
132 auto allBoolScalarsAndVectors = {s1, v2s1, v3s1, v4s1, v8s1, v16s1};
133
134 auto allIntScalars = {s8, s16, s32, s64, s128};
135
136 auto allFloatScalarsAndF16Vector2AndVector4s = {s16, s32, s64, v2s16, v4s16};
137
138 auto allFloatScalarsAndVectors = {
139 s16, s32, s64, v2s16, v2s32, v2s64, v3s16, v3s32, v3s64,
140 v4s16, v4s32, v4s64, v8s16, v8s32, v8s64, v16s16, v16s32, v16s64};
141
142 auto allFloatAndIntScalarsAndPtrs = {s8, s16, s32, s64, p0, p1,
143 p2, p3, p4, p5, p6, p7,
144 p8, p9, p10, p11, p12, p13};
145
146 auto allPtrs = {p0, p1, p2, p3, p4, p5, p6, p7, p8, p9, p10, p11, p12, p13};
147
148 auto &allowedVectorTypes = ST.isShader() ? allShaderVectors : allVectors;
149
150 bool IsExtendedInts =
151 ST.canUseExtension(
152 SPIRV::Extension::SPV_ALTERA_arbitrary_precision_integers) ||
153 ST.canUseExtension(SPIRV::Extension::SPV_KHR_bit_instructions) ||
154 ST.canUseExtension(SPIRV::Extension::SPV_INTEL_int4);
155 auto extendedScalarsAndVectors =
156 [IsExtendedInts](const LegalityQuery &Query) {
157 const LLT Ty = Query.Types[0];
158 return IsExtendedInts && Ty.isValid() && !Ty.isPointerOrPointerVector();
159 };
160 auto extendedScalarsAndVectorsProduct = [IsExtendedInts](
161 const LegalityQuery &Query) {
162 const LLT Ty1 = Query.Types[0], Ty2 = Query.Types[1];
163 return IsExtendedInts && Ty1.isValid() && Ty2.isValid() &&
164 !Ty1.isPointerOrPointerVector() && !Ty2.isPointerOrPointerVector();
165 };
166 auto extendedPtrsScalarsAndVectors =
167 [IsExtendedInts](const LegalityQuery &Query) {
168 const LLT Ty = Query.Types[0];
169 return IsExtendedInts && Ty.isValid();
170 };
171
172 // The universal validation rules in the SPIR-V specification state that
173 // vector sizes are typically limited to 2, 3, or 4. However, larger vector
174 // sizes (8 and 16) are enabled when the Kernel capability is present. For
175 // shader execution models, vector sizes are strictly limited to 4. In
176 // non-shader contexts, vector sizes of 8 and 16 are also permitted, but
177 // arbitrary sizes (e.g., 6 or 11) are not.
178 uint32_t MaxVectorSize = ST.isShader() ? 4 : 16;
179 LLVM_DEBUG(dbgs() << "MaxVectorSize: " << MaxVectorSize << "\n");
180
181 for (auto Opc : getTypeFoldingSupportedOpcodes()) {
182 switch (Opc) {
183 case G_EXTRACT_VECTOR_ELT:
184 case G_UREM:
185 case G_SREM:
186 case G_UDIV:
187 case G_SDIV:
188 case G_FREM:
189 break;
190 default:
192 .customFor(allScalars)
193 .customFor(allowedVectorTypes)
197 0, ElementCount::getFixed(MaxVectorSize)))
198 .custom();
199 break;
200 }
201 }
202
203 getActionDefinitionsBuilder({G_UREM, G_SREM, G_SDIV, G_UDIV, G_FREM})
204 .customFor(allScalars)
205 .customFor(allowedVectorTypes)
209 0, ElementCount::getFixed(MaxVectorSize)))
210 .custom();
211
212 getActionDefinitionsBuilder({G_FMA, G_STRICT_FMA})
213 .legalFor(allScalars)
214 .legalFor(allowedVectorTypes)
218 0, ElementCount::getFixed(MaxVectorSize)))
219 .alwaysLegal();
220
221 getActionDefinitionsBuilder(G_INTRINSIC_W_SIDE_EFFECTS).custom();
222
223 getActionDefinitionsBuilder(G_SHUFFLE_VECTOR)
224 .legalForCartesianProduct(allowedVectorTypes, allowedVectorTypes)
226 .lowerIf(vectorElementCountIsGreaterThan(0, MaxVectorSize))
228 .lowerIf(vectorElementCountIsGreaterThan(1, MaxVectorSize));
229
230 getActionDefinitionsBuilder(G_EXTRACT_VECTOR_ELT)
234 1, ElementCount::getFixed(MaxVectorSize)))
235 .custom();
236
237 getActionDefinitionsBuilder(G_INSERT_VECTOR_ELT)
241 0, ElementCount::getFixed(MaxVectorSize)))
242 .custom();
243
244 // Illegal G_UNMERGE_VALUES instructions should be handled
245 // during the combine phase.
246 getActionDefinitionsBuilder(G_BUILD_VECTOR)
248
249 // When entering the legalizer, there should be no G_BITCAST instructions.
250 // They should all be calls to the `spv_bitcast` intrinsic. The call to
251 // the intrinsic will be converted to a G_BITCAST during legalization if
252 // the vectors are not legal. After using the rules to legalize a G_BITCAST,
253 // we turn it back into a call to the intrinsic with a custom rule to avoid
254 // potential machine verifier failures.
260 0, ElementCount::getFixed(MaxVectorSize)))
261 .lowerIf(vectorElementCountIsGreaterThan(1, MaxVectorSize))
262 .custom();
263
264 // If the result is still illegal, the combiner should be able to remove it.
265 getActionDefinitionsBuilder(G_CONCAT_VECTORS)
266 .legalForCartesianProduct(allowedVectorTypes, allowedVectorTypes);
267
268 getActionDefinitionsBuilder(G_SPLAT_VECTOR)
269 .legalFor(allowedVectorTypes)
273 .alwaysLegal();
274
275 // Vector Reduction Operations
277 {G_VECREDUCE_SMIN, G_VECREDUCE_SMAX, G_VECREDUCE_UMIN, G_VECREDUCE_UMAX,
278 G_VECREDUCE_ADD, G_VECREDUCE_MUL, G_VECREDUCE_FMUL, G_VECREDUCE_FMIN,
279 G_VECREDUCE_FMAX, G_VECREDUCE_FMINIMUM, G_VECREDUCE_FMAXIMUM,
280 G_VECREDUCE_OR, G_VECREDUCE_AND, G_VECREDUCE_XOR})
281 .legalFor(allowedVectorTypes)
282 .scalarize(1)
283 .lower();
284
285 getActionDefinitionsBuilder({G_VECREDUCE_SEQ_FADD, G_VECREDUCE_SEQ_FMUL})
286 .scalarize(2)
287 .lower();
288
289 // Illegal G_UNMERGE_VALUES instructions should be handled
290 // during the combine phase.
291 getActionDefinitionsBuilder(G_UNMERGE_VALUES)
293
294 getActionDefinitionsBuilder({G_MEMCPY, G_MEMMOVE})
295 .unsupportedIf(LegalityPredicates::any(typeIs(0, p9), typeIs(1, p9)))
296 .legalIf(all(typeInSet(0, allPtrs), typeInSet(1, allPtrs)));
297
299 .unsupportedIf(typeIs(0, p9))
300 .legalIf(all(typeInSet(0, allPtrs), typeInSet(1, allIntScalars)));
301
302 getActionDefinitionsBuilder(G_ADDRSPACE_CAST)
305 all(typeIsNot(0, p9), typeIs(1, p9))))
306 .legalForCartesianProduct(allPtrs, allPtrs);
307
308 // Should we be legalizing bad scalar sizes like s5 here instead
309 // of handling them in the instruction selector?
310 getActionDefinitionsBuilder({G_LOAD, G_STORE})
311 .unsupportedIf(typeIs(1, p9))
312 .legalForCartesianProduct(allowedVectorTypes, allPtrs)
313 .legalForCartesianProduct(allPtrs, allPtrs)
314 .legalIf(isScalar(0))
315 .custom();
316
317 getActionDefinitionsBuilder({G_SMIN, G_SMAX, G_UMIN, G_UMAX, G_ABS,
318 G_BITREVERSE, G_SADDSAT, G_UADDSAT, G_SSUBSAT,
319 G_USUBSAT, G_SCMP, G_UCMP})
320 .legalFor(allIntScalarsAndVectors)
321 .legalIf(extendedScalarsAndVectors);
322
323 getActionDefinitionsBuilder(G_STRICT_FLDEXP)
324 .legalForCartesianProduct(allFloatScalarsAndVectors, allIntScalars);
325
326 getActionDefinitionsBuilder({G_FPTOSI, G_FPTOUI})
327 .legalForCartesianProduct(allIntScalarsAndVectors,
328 allFloatScalarsAndVectors);
329
330 getActionDefinitionsBuilder({G_FPTOSI_SAT, G_FPTOUI_SAT})
331 .legalForCartesianProduct(allIntScalarsAndVectors,
332 allFloatScalarsAndVectors);
333
334 getActionDefinitionsBuilder({G_SITOFP, G_UITOFP})
335 .legalForCartesianProduct(allFloatScalarsAndVectors,
336 allScalarsAndVectors);
337
339 .legalForCartesianProduct(allIntScalarsAndVectors)
340 .legalIf(extendedScalarsAndVectorsProduct);
341
342 // Extensions.
343 getActionDefinitionsBuilder({G_TRUNC, G_ZEXT, G_SEXT, G_ANYEXT})
344 .legalForCartesianProduct(allScalarsAndVectors)
345 .legalIf(extendedScalarsAndVectorsProduct);
346
348 .legalFor(allPtrsScalarsAndVectors)
349 .legalIf(extendedPtrsScalarsAndVectors);
350
352 all(typeInSet(0, allPtrsScalarsAndVectors),
353 typeInSet(1, allPtrsScalarsAndVectors)));
354
355 getActionDefinitionsBuilder({G_IMPLICIT_DEF, G_FREEZE})
356 .legalFor({s1, s128})
357 .legalFor(allFloatAndIntScalarsAndPtrs)
358 .legalFor(allowedVectorTypes)
359 .legalIf([](const LegalityQuery &Query) {
360 return Query.Types[0].isPointerVector();
361 })
362 .moreElementsToNextPow2(0)
363 .fewerElementsIf(vectorElementCountIsGreaterThan(0, MaxVectorSize),
365 0, ElementCount::getFixed(MaxVectorSize)));
366
367 getActionDefinitionsBuilder({G_STACKSAVE, G_STACKRESTORE}).alwaysLegal();
368
370 .legalForCartesianProduct(allPtrs, allIntScalars)
371 .legalIf(
372 all(typeInSet(0, allPtrs), typeOfExtendedScalars(1, IsExtendedInts)))
373 .legalIf([](const LegalityQuery &Query) {
374 const LLT DstTy = Query.Types[0];
375 const LLT SrcTy = Query.Types[1];
376 return DstTy.isPointerVector() && SrcTy.isVector() &&
377 !SrcTy.isPointer() &&
378 DstTy.getNumElements() == SrcTy.getNumElements();
379 });
381 .legalForCartesianProduct(allIntScalars, allPtrs)
382 .legalIf(
383 all(typeOfExtendedScalars(0, IsExtendedInts), typeInSet(1, allPtrs)))
384 .legalIf([](const LegalityQuery &Query) {
385 const LLT DstTy = Query.Types[0];
386 const LLT SrcTy = Query.Types[1];
387 return SrcTy.isPointerVector() && DstTy.isVector() &&
388 !DstTy.isPointer() &&
389 DstTy.getNumElements() == SrcTy.getNumElements();
390 });
392 .legalForCartesianProduct(allPtrs, allIntScalars)
393 .legalIf(
394 all(typeInSet(0, allPtrs), typeOfExtendedScalars(1, IsExtendedInts)));
395
396 // ST.canDirectlyComparePointers() for pointer args is supported in
397 // legalizeCustom().
400 all(typeIs(0, p9), typeInSet(1, allPtrs), typeIsNot(1, p9)),
401 all(typeInSet(0, allPtrs), typeIsNot(0, p9), typeIs(1, p9))))
402 .legalIf([IsExtendedInts](const LegalityQuery &Query) {
403 const LLT Ty = Query.Types[1];
404 return IsExtendedInts && Ty.isValid() && !Ty.isPointerOrPointerVector();
405 })
406 .customIf(all(typeInSet(0, allBoolScalarsAndVectors),
407 typeInSet(1, allPtrsScalarsAndVectors)));
408
410 all(typeInSet(0, allBoolScalarsAndVectors),
411 typeInSet(1, allFloatScalarsAndVectors)));
412
413 getActionDefinitionsBuilder({G_ATOMICRMW_OR, G_ATOMICRMW_ADD, G_ATOMICRMW_AND,
414 G_ATOMICRMW_MAX, G_ATOMICRMW_MIN,
415 G_ATOMICRMW_SUB, G_ATOMICRMW_XOR,
416 G_ATOMICRMW_UMAX, G_ATOMICRMW_UMIN})
417 .legalForCartesianProduct(allIntScalars, allPtrs);
418
420 {G_ATOMICRMW_FADD, G_ATOMICRMW_FSUB, G_ATOMICRMW_FMIN, G_ATOMICRMW_FMAX})
421 .legalForCartesianProduct(allFloatScalarsAndF16Vector2AndVector4s,
422 allPtrs);
423
424 getActionDefinitionsBuilder(G_ATOMICRMW_XCHG)
425 .legalForCartesianProduct(allFloatAndIntScalarsAndPtrs, allPtrs);
426
427 getActionDefinitionsBuilder(G_ATOMIC_CMPXCHG_WITH_SUCCESS).lower();
428 // TODO: add proper legalization rules.
429 getActionDefinitionsBuilder(G_ATOMIC_CMPXCHG).alwaysLegal();
430
432 {G_UADDO, G_SADDO, G_USUBO, G_SSUBO, G_UMULO, G_SMULO})
433 .alwaysLegal();
434
435 getActionDefinitionsBuilder({G_LROUND, G_LLROUND})
436 .legalForCartesianProduct(allFloatScalarsAndVectors,
437 allIntScalarsAndVectors);
438
439 // FP conversions.
440 getActionDefinitionsBuilder({G_FPTRUNC, G_FPEXT})
441 .legalForCartesianProduct(allFloatScalarsAndVectors);
442
443 // Pointer-handling.
444 getActionDefinitionsBuilder(G_FRAME_INDEX).legalFor({p0});
445
446 getActionDefinitionsBuilder(G_GLOBAL_VALUE).legalFor(allPtrs);
447
448 // Control-flow. In some cases (e.g. constants) s1 may be promoted to s32.
449 getActionDefinitionsBuilder(G_BRCOND).legalFor({s1, s32});
450
452 allFloatScalarsAndVectors, {s32, v2s32, v3s32, v4s32, v8s32, v16s32});
453
454 // TODO: Review the target OpenCL and GLSL Extended Instruction Set specs to
455 // tighten these requirements. Many of these math functions are only legal on
456 // specific bitwidths, so they are not selectable for
457 // allFloatScalarsAndVectors.
458 // clang-format off
459 getActionDefinitionsBuilder({G_STRICT_FSQRT,
460 G_FPOW,
461 G_FEXP,
462 G_FMODF,
463 G_FSINCOS,
464 G_FEXP2,
465 G_FEXP10,
466 G_FLOG,
467 G_FLOG2,
468 G_FLOG10,
469 G_FABS,
470 G_FMINNUM,
471 G_FMAXNUM,
472 G_FCEIL,
473 G_FCOS,
474 G_FSIN,
475 G_FTAN,
476 G_FACOS,
477 G_FASIN,
478 G_FATAN,
479 G_FATAN2,
480 G_FCOSH,
481 G_FSINH,
482 G_FTANH,
483 G_FSQRT,
484 G_FFLOOR,
485 G_FRINT,
486 G_FNEARBYINT,
487 G_INTRINSIC_ROUND,
488 G_INTRINSIC_TRUNC,
489 G_FMINIMUM,
490 G_FMAXIMUM,
491 G_INTRINSIC_ROUNDEVEN})
492 .legalFor(allFloatScalarsAndVectors);
493 // clang-format on
494
495 getActionDefinitionsBuilder(G_FCOPYSIGN)
496 .legalForCartesianProduct(allFloatScalarsAndVectors,
497 allFloatScalarsAndVectors);
498
500 allFloatScalarsAndVectors, allIntScalarsAndVectors);
501
502 if (ST.canUseExtInstSet(SPIRV::InstructionSet::OpenCL_std)) {
504 {G_CTTZ, G_CTTZ_ZERO_UNDEF, G_CTLZ, G_CTLZ_ZERO_UNDEF})
505 .legalForCartesianProduct(allIntScalarsAndVectors,
506 allIntScalarsAndVectors);
507
508 // Struct return types become a single scalar, so cannot easily legalize.
509 getActionDefinitionsBuilder({G_SMULH, G_UMULH}).alwaysLegal();
510 }
511
512 getActionDefinitionsBuilder(G_IS_FPCLASS).custom();
513
515 verify(*ST.getInstrInfo());
516}
517
520 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
521 Register DstReg = MI.getOperand(0).getReg();
522 Register SrcReg = MI.getOperand(1).getReg();
523 Register IdxReg = MI.getOperand(2).getReg();
524
525 MIRBuilder
526 .buildIntrinsic(Intrinsic::spv_extractelt, ArrayRef<Register>{DstReg})
527 .addUse(SrcReg)
528 .addUse(IdxReg);
529 MI.eraseFromParent();
530 return true;
531}
532
535 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
536 Register DstReg = MI.getOperand(0).getReg();
537 Register SrcReg = MI.getOperand(1).getReg();
538 Register ValReg = MI.getOperand(2).getReg();
539 Register IdxReg = MI.getOperand(3).getReg();
540
541 MIRBuilder
542 .buildIntrinsic(Intrinsic::spv_insertelt, ArrayRef<Register>{DstReg})
543 .addUse(SrcReg)
544 .addUse(ValReg)
545 .addUse(IdxReg);
546 MI.eraseFromParent();
547 return true;
548}
549
551 LegalizerHelper &Helper,
554 Register ConvReg = MRI.createGenericVirtualRegister(ConvTy);
555 MRI.setRegClass(ConvReg, GR->getRegClass(SpvType));
556 GR->assignSPIRVTypeToVReg(SpvType, ConvReg, Helper.MIRBuilder.getMF());
557 Helper.MIRBuilder.buildInstr(TargetOpcode::G_PTRTOINT)
558 .addDef(ConvReg)
559 .addUse(Reg);
560 return ConvReg;
561}
562
563static bool needsVectorLegalization(const LLT &Ty, const SPIRVSubtarget &ST) {
564 if (!Ty.isVector())
565 return false;
566 unsigned NumElements = Ty.getNumElements();
567 unsigned MaxVectorSize = ST.isShader() ? 4 : 16;
568 return (NumElements > 4 && !isPowerOf2_32(NumElements)) ||
569 NumElements > MaxVectorSize;
570}
571
574 MachineRegisterInfo &MRI = MI.getMF()->getRegInfo();
575 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
576 Register DstReg = MI.getOperand(0).getReg();
577 Register PtrReg = MI.getOperand(1).getReg();
578 LLT DstTy = MRI.getType(DstReg);
579
580 if (!DstTy.isVector())
581 return true;
582
583 const SPIRVSubtarget &ST = MI.getMF()->getSubtarget<SPIRVSubtarget>();
584 if (!needsVectorLegalization(DstTy, ST))
585 return true;
586
587 SmallVector<Register, 8> SplitRegs;
588 LLT EltTy = DstTy.getElementType();
589 unsigned NumElts = DstTy.getNumElements();
590
591 LLT PtrTy = MRI.getType(PtrReg);
592 auto Zero = MIRBuilder.buildConstant(LLT::scalar(32), 0);
593
594 for (unsigned i = 0; i < NumElts; ++i) {
595 auto Idx = MIRBuilder.buildConstant(LLT::scalar(32), i);
596 Register EltPtr = MRI.createGenericVirtualRegister(PtrTy);
597
598 MIRBuilder.buildIntrinsic(Intrinsic::spv_gep, ArrayRef<Register>{EltPtr})
599 .addImm(1) // InBounds
600 .addUse(PtrReg)
601 .addUse(Zero.getReg(0))
602 .addUse(Idx.getReg(0));
603
604 MachinePointerInfo EltPtrInfo;
605 Align EltAlign = Align(1);
606 if (!MI.memoperands_empty()) {
607 MachineMemOperand *MMO = *MI.memoperands_begin();
608 EltPtrInfo =
609 MMO->getPointerInfo().getWithOffset(i * EltTy.getSizeInBytes());
610 EltAlign = commonAlignment(MMO->getAlign(), i * EltTy.getSizeInBytes());
611 }
612
613 Register EltReg = MRI.createGenericVirtualRegister(EltTy);
614 MIRBuilder.buildLoad(EltReg, EltPtr, EltPtrInfo, EltAlign);
615 SplitRegs.push_back(EltReg);
616 }
617
618 MIRBuilder.buildBuildVector(DstReg, SplitRegs);
619 MI.eraseFromParent();
620 return true;
621}
622
625 MachineRegisterInfo &MRI = MI.getMF()->getRegInfo();
626 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
627 Register ValReg = MI.getOperand(0).getReg();
628 Register PtrReg = MI.getOperand(1).getReg();
629 LLT ValTy = MRI.getType(ValReg);
630
631 assert(ValTy.isVector() && "Expected vector store");
632
633 SmallVector<Register, 8> SplitRegs;
634 LLT EltTy = ValTy.getElementType();
635 unsigned NumElts = ValTy.getNumElements();
636
637 for (unsigned i = 0; i < NumElts; ++i)
638 SplitRegs.push_back(MRI.createGenericVirtualRegister(EltTy));
639
640 MIRBuilder.buildUnmerge(SplitRegs, ValReg);
641
642 LLT PtrTy = MRI.getType(PtrReg);
643 auto Zero = MIRBuilder.buildConstant(LLT::scalar(32), 0);
644
645 for (unsigned i = 0; i < NumElts; ++i) {
646 auto Idx = MIRBuilder.buildConstant(LLT::scalar(32), i);
647 Register EltPtr = MRI.createGenericVirtualRegister(PtrTy);
648
649 MIRBuilder.buildIntrinsic(Intrinsic::spv_gep, ArrayRef<Register>{EltPtr})
650 .addImm(1) // InBounds
651 .addUse(PtrReg)
652 .addUse(Zero.getReg(0))
653 .addUse(Idx.getReg(0));
654
655 MachinePointerInfo EltPtrInfo;
656 Align EltAlign = Align(1);
657 if (!MI.memoperands_empty()) {
658 MachineMemOperand *MMO = *MI.memoperands_begin();
659 EltPtrInfo =
660 MMO->getPointerInfo().getWithOffset(i * EltTy.getSizeInBytes());
661 EltAlign = commonAlignment(MMO->getAlign(), i * EltTy.getSizeInBytes());
662 }
663
664 MIRBuilder.buildStore(SplitRegs[i], EltPtr, EltPtrInfo, EltAlign);
665 }
666
667 MI.eraseFromParent();
668 return true;
669}
670
673 LostDebugLocObserver &LocObserver) const {
674 MachineRegisterInfo &MRI = MI.getMF()->getRegInfo();
675 switch (MI.getOpcode()) {
676 default:
677 // TODO: implement legalization for other opcodes.
678 return true;
679 case TargetOpcode::G_BITCAST:
680 return legalizeBitcast(Helper, MI);
681 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
682 return legalizeExtractVectorElt(Helper, MI, GR);
683 case TargetOpcode::G_INSERT_VECTOR_ELT:
684 return legalizeInsertVectorElt(Helper, MI, GR);
685 case TargetOpcode::G_INTRINSIC:
686 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS:
687 return legalizeIntrinsic(Helper, MI);
688 case TargetOpcode::G_IS_FPCLASS:
689 return legalizeIsFPClass(Helper, MI, LocObserver);
690 case TargetOpcode::G_ICMP: {
691 assert(GR->getSPIRVTypeForVReg(MI.getOperand(0).getReg()));
692 auto &Op0 = MI.getOperand(2);
693 auto &Op1 = MI.getOperand(3);
694 Register Reg0 = Op0.getReg();
695 Register Reg1 = Op1.getReg();
697 static_cast<CmpInst::Predicate>(MI.getOperand(1).getPredicate());
698 if ((!ST->canDirectlyComparePointers() ||
700 MRI.getType(Reg0).isPointer() && MRI.getType(Reg1).isPointer()) {
701 LLT ConvT = LLT::scalar(ST->getPointerSize());
702 Type *LLVMTy = IntegerType::get(MI.getMF()->getFunction().getContext(),
703 ST->getPointerSize());
704 SPIRVTypeInst SpirvTy = GR->getOrCreateSPIRVType(
705 LLVMTy, Helper.MIRBuilder, SPIRV::AccessQualifier::ReadWrite, true);
706 Op0.setReg(convertPtrToInt(Reg0, ConvT, SpirvTy, Helper, MRI, GR));
707 Op1.setReg(convertPtrToInt(Reg1, ConvT, SpirvTy, Helper, MRI, GR));
708 }
709 return true;
710 }
711 case TargetOpcode::G_LOAD:
712 return legalizeLoad(Helper, MI, GR);
713 case TargetOpcode::G_STORE:
714 return legalizeStore(Helper, MI, GR);
715 }
716}
717
720 Register SrcReg, LLT SrcTy,
721 MachinePointerInfo &PtrInfo, Align &VecAlign) {
722 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
723 MachineRegisterInfo &MRI = *MIRBuilder.getMRI();
724
725 VecAlign = Helper.getStackTemporaryAlignment(SrcTy);
726 auto StackTemp = Helper.createStackTemporary(
727 TypeSize::getFixed(SrcTy.getSizeInBytes()), VecAlign, PtrInfo);
728
729 // Set the type of StackTemp to a pointer to an array of the element type.
730 SPIRVTypeInst SpvSrcTy = GR->getSPIRVTypeForVReg(SrcReg);
731 SPIRVTypeInst EltSpvTy = GR->getScalarOrVectorComponentType(SpvSrcTy);
732 const Type *LLVMEltTy = GR->getTypeForSPIRVType(EltSpvTy);
733 const Type *LLVMArrTy =
734 ArrayType::get(const_cast<Type *>(LLVMEltTy), SrcTy.getNumElements());
735 SPIRVTypeInst ArrSpvTy = GR->getOrCreateSPIRVType(
736 LLVMArrTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite, true);
737 SPIRVTypeInst PtrToArrSpvTy = GR->getOrCreateSPIRVPointerType(
738 ArrSpvTy, MIRBuilder, SPIRV::StorageClass::Function);
739
740 Register StackReg = StackTemp.getReg(0);
741 MRI.setRegClass(StackReg, GR->getRegClass(PtrToArrSpvTy));
742 GR->assignSPIRVTypeToVReg(PtrToArrSpvTy, StackReg, MIRBuilder.getMF());
743
744 return StackTemp;
745}
746
749 LLVM_DEBUG(dbgs() << "Found a bitcast instruction\n");
750 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
751 MachineRegisterInfo &MRI = *MIRBuilder.getMRI();
752 const SPIRVSubtarget &ST = MI.getMF()->getSubtarget<SPIRVSubtarget>();
753
754 Register DstReg = MI.getOperand(0).getReg();
755 Register SrcReg = MI.getOperand(2).getReg();
756 LLT DstTy = MRI.getType(DstReg);
757 LLT SrcTy = MRI.getType(SrcReg);
758
759 // If an spv_bitcast needs to be legalized, we convert it to G_BITCAST to
760 // allow using the generic legalization rules.
761 if (needsVectorLegalization(DstTy, ST) ||
762 needsVectorLegalization(SrcTy, ST)) {
763 LLVM_DEBUG(dbgs() << "Replacing with a G_BITCAST\n");
764 MIRBuilder.buildBitcast(DstReg, SrcReg);
765 MI.eraseFromParent();
766 }
767 return true;
768}
769
772 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
773 MachineRegisterInfo &MRI = *MIRBuilder.getMRI();
774 const SPIRVSubtarget &ST = MI.getMF()->getSubtarget<SPIRVSubtarget>();
775
776 Register DstReg = MI.getOperand(0).getReg();
777 LLT DstTy = MRI.getType(DstReg);
778
779 if (needsVectorLegalization(DstTy, ST)) {
780 Register SrcReg = MI.getOperand(2).getReg();
781 Register ValReg = MI.getOperand(3).getReg();
782 LLT SrcTy = MRI.getType(SrcReg);
783 MachineOperand &IdxOperand = MI.getOperand(4);
784
785 if (getImm(IdxOperand, &MRI)) {
786 uint64_t IdxVal = foldImm(IdxOperand, &MRI);
787 if (IdxVal < SrcTy.getNumElements()) {
789 SPIRVTypeInst ElementType =
791 LLT ElementLLTTy = GR->getRegType(ElementType);
792 for (unsigned I = 0, E = SrcTy.getNumElements(); I < E; ++I) {
793 Register Reg = MRI.createGenericVirtualRegister(ElementLLTTy);
794 MRI.setRegClass(Reg, GR->getRegClass(ElementType));
795 GR->assignSPIRVTypeToVReg(ElementType, Reg, *MI.getMF());
796 Regs.push_back(Reg);
797 }
798 MIRBuilder.buildUnmerge(Regs, SrcReg);
799 Regs[IdxVal] = ValReg;
800 MIRBuilder.buildBuildVector(DstReg, Regs);
801 MI.eraseFromParent();
802 return true;
803 }
804 }
805
806 LLT EltTy = SrcTy.getElementType();
807 Align VecAlign;
808 MachinePointerInfo PtrInfo;
809 auto StackTemp = createStackTemporaryForVector(Helper, GR, SrcReg, SrcTy,
810 PtrInfo, VecAlign);
811
812 MIRBuilder.buildStore(SrcReg, StackTemp, PtrInfo, VecAlign);
813
814 Register IdxReg = IdxOperand.getReg();
815 LLT PtrTy = MRI.getType(StackTemp.getReg(0));
816 Register EltPtr = MRI.createGenericVirtualRegister(PtrTy);
817 auto Zero = MIRBuilder.buildConstant(LLT::scalar(32), 0);
818
819 MIRBuilder.buildIntrinsic(Intrinsic::spv_gep, ArrayRef<Register>{EltPtr})
820 .addImm(1) // InBounds
821 .addUse(StackTemp.getReg(0))
822 .addUse(Zero.getReg(0))
823 .addUse(IdxReg);
824
826 Align EltAlign = Helper.getStackTemporaryAlignment(EltTy);
827 MIRBuilder.buildStore(ValReg, EltPtr, EltPtrInfo, EltAlign);
828
829 MIRBuilder.buildLoad(DstReg, StackTemp, PtrInfo, VecAlign);
830 MI.eraseFromParent();
831 return true;
832 }
833 return true;
834}
835
838 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
839 MachineRegisterInfo &MRI = *MIRBuilder.getMRI();
840 const SPIRVSubtarget &ST = MI.getMF()->getSubtarget<SPIRVSubtarget>();
841
842 Register SrcReg = MI.getOperand(2).getReg();
843 LLT SrcTy = MRI.getType(SrcReg);
844
845 if (needsVectorLegalization(SrcTy, ST)) {
846 Register DstReg = MI.getOperand(0).getReg();
847 MachineOperand &IdxOperand = MI.getOperand(3);
848
849 if (getImm(IdxOperand, &MRI)) {
850 uint64_t IdxVal = foldImm(IdxOperand, &MRI);
851 if (IdxVal < SrcTy.getNumElements()) {
852 LLT DstTy = MRI.getType(DstReg);
854 SPIRVTypeInst DstSpvTy = GR->getSPIRVTypeForVReg(DstReg);
855 for (unsigned I = 0, E = SrcTy.getNumElements(); I < E; ++I) {
856 if (I == IdxVal) {
857 Regs.push_back(DstReg);
858 } else {
860 MRI.setRegClass(Reg, GR->getRegClass(DstSpvTy));
861 GR->assignSPIRVTypeToVReg(DstSpvTy, Reg, *MI.getMF());
862 Regs.push_back(Reg);
863 }
864 }
865 MIRBuilder.buildUnmerge(Regs, SrcReg);
866 MI.eraseFromParent();
867 return true;
868 }
869 }
870
871 LLT EltTy = SrcTy.getElementType();
872 Align VecAlign;
873 MachinePointerInfo PtrInfo;
874 auto StackTemp = createStackTemporaryForVector(Helper, GR, SrcReg, SrcTy,
875 PtrInfo, VecAlign);
876
877 MIRBuilder.buildStore(SrcReg, StackTemp, PtrInfo, VecAlign);
878
879 Register IdxReg = IdxOperand.getReg();
880 LLT PtrTy = MRI.getType(StackTemp.getReg(0));
881 Register EltPtr = MRI.createGenericVirtualRegister(PtrTy);
882 auto Zero = MIRBuilder.buildConstant(LLT::scalar(32), 0);
883
884 MIRBuilder.buildIntrinsic(Intrinsic::spv_gep, ArrayRef<Register>{EltPtr})
885 .addImm(1) // InBounds
886 .addUse(StackTemp.getReg(0))
887 .addUse(Zero.getReg(0))
888 .addUse(IdxReg);
889
891 Align EltAlign = Helper.getStackTemporaryAlignment(EltTy);
892 MIRBuilder.buildLoad(DstReg, EltPtr, EltPtrInfo, EltAlign);
893
894 MI.eraseFromParent();
895 return true;
896 }
897 return true;
898}
899
902 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
903 MachineRegisterInfo &MRI = *MIRBuilder.getMRI();
904 const SPIRVSubtarget &ST = MI.getMF()->getSubtarget<SPIRVSubtarget>();
905
906 Register DstReg = MI.getOperand(0).getReg();
907 LLT DstTy = MRI.getType(DstReg);
908
909 if (!needsVectorLegalization(DstTy, ST))
910 return true;
911
913 if (MI.getNumOperands() == 2) {
914 // The "null" case: no values are attached.
915 LLT EltTy = DstTy.getElementType();
916 auto Zero = MIRBuilder.buildConstant(EltTy, 0);
917 SPIRVTypeInst SpvDstTy = GR->getSPIRVTypeForVReg(DstReg);
918 SPIRVTypeInst SpvEltTy = GR->getScalarOrVectorComponentType(SpvDstTy);
919 GR->assignSPIRVTypeToVReg(SpvEltTy, Zero.getReg(0), MIRBuilder.getMF());
920 for (unsigned i = 0; i < DstTy.getNumElements(); ++i)
921 SrcRegs.push_back(Zero.getReg(0));
922 } else {
923 for (unsigned i = 2; i < MI.getNumOperands(); ++i) {
924 SrcRegs.push_back(MI.getOperand(i).getReg());
925 }
926 }
927 MIRBuilder.buildBuildVector(DstReg, SrcRegs);
928 MI.eraseFromParent();
929 return true;
930}
931
933 MachineInstr &MI) const {
934 LLVM_DEBUG(dbgs() << "legalizeIntrinsic: " << MI);
935 auto IntrinsicID = cast<GIntrinsic>(MI).getIntrinsicID();
936 switch (IntrinsicID) {
937 case Intrinsic::spv_bitcast:
938 return legalizeSpvBitcast(Helper, MI, GR);
939 case Intrinsic::spv_insertelt:
940 return legalizeSpvInsertElt(Helper, MI, GR);
941 case Intrinsic::spv_extractelt:
942 return legalizeSpvExtractElt(Helper, MI, GR);
943 case Intrinsic::spv_const_composite:
944 return legalizeSpvConstComposite(Helper, MI, GR);
945 }
946 return true;
947}
948
949bool SPIRVLegalizerInfo::legalizeBitcast(LegalizerHelper &Helper,
950 MachineInstr &MI) const {
951 // Once the G_BITCAST is using vectors that are allowed, we turn it back into
952 // an spv_bitcast to avoid verifier problems when the register types are the
953 // same for the source and the result. Note that the SPIR-V types associated
954 // with the bitcast can be different even if the register types are the same.
955 MachineIRBuilder &MIRBuilder = Helper.MIRBuilder;
956 Register DstReg = MI.getOperand(0).getReg();
957 Register SrcReg = MI.getOperand(1).getReg();
958 SmallVector<Register, 1> DstRegs = {DstReg};
959 MIRBuilder.buildIntrinsic(Intrinsic::spv_bitcast, DstRegs).addUse(SrcReg);
960 MI.eraseFromParent();
961 return true;
962}
963
964// Note this code was copied from LegalizerHelper::lowerISFPCLASS and adjusted
965// to ensure that all instructions created during the lowering have SPIR-V types
966// assigned to them.
967bool SPIRVLegalizerInfo::legalizeIsFPClass(
969 LostDebugLocObserver &LocObserver) const {
970 auto [DstReg, DstTy, SrcReg, SrcTy] = MI.getFirst2RegLLTs();
971 FPClassTest Mask = static_cast<FPClassTest>(MI.getOperand(2).getImm());
972
973 auto &MIRBuilder = Helper.MIRBuilder;
974 auto &MF = MIRBuilder.getMF();
975 MachineRegisterInfo &MRI = MF.getRegInfo();
976
977 Type *LLVMDstTy =
978 IntegerType::get(MIRBuilder.getContext(), DstTy.getScalarSizeInBits());
979 if (DstTy.isVector())
980 LLVMDstTy = VectorType::get(LLVMDstTy, DstTy.getElementCount());
981 SPIRVTypeInst SPIRVDstTy = GR->getOrCreateSPIRVType(
982 LLVMDstTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
983 /*EmitIR*/ true);
984
985 unsigned BitSize = SrcTy.getScalarSizeInBits();
986 const fltSemantics &Semantics = getFltSemanticForLLT(SrcTy.getScalarType());
987
988 LLT IntTy = LLT::scalar(BitSize);
989 Type *LLVMIntTy = IntegerType::get(MIRBuilder.getContext(), BitSize);
990 if (SrcTy.isVector()) {
991 IntTy = LLT::vector(SrcTy.getElementCount(), IntTy);
992 LLVMIntTy = VectorType::get(LLVMIntTy, SrcTy.getElementCount());
993 }
994 SPIRVTypeInst SPIRVIntTy = GR->getOrCreateSPIRVType(
995 LLVMIntTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
996 /*EmitIR*/ true);
997
998 // Clang doesn't support capture of structured bindings:
999 LLT DstTyCopy = DstTy;
1000 const auto assignSPIRVTy = [&](MachineInstrBuilder &&MI) {
1001 // Assign this MI's (assumed only) destination to one of the two types we
1002 // expect: either the G_IS_FPCLASS's destination type, or the integer type
1003 // bitcast from the source type.
1004 LLT MITy = MRI.getType(MI.getReg(0));
1005 assert((MITy == IntTy || MITy == DstTyCopy) &&
1006 "Unexpected LLT type while lowering G_IS_FPCLASS");
1007 SPIRVTypeInst SPVTy = MITy == IntTy ? SPIRVIntTy : SPIRVDstTy;
1008 GR->assignSPIRVTypeToVReg(SPVTy, MI.getReg(0), MF);
1009 return MI;
1010 };
1011
1012 // Helper to build and assign a constant in one go
1013 const auto buildSPIRVConstant = [&](LLT Ty, auto &&C) -> MachineInstrBuilder {
1014 if (!Ty.isFixedVector())
1015 return assignSPIRVTy(MIRBuilder.buildConstant(Ty, C));
1016 auto ScalarC = MIRBuilder.buildConstant(Ty.getScalarType(), C);
1017 assert((Ty == IntTy || Ty == DstTyCopy) &&
1018 "Unexpected LLT type while lowering constant for G_IS_FPCLASS");
1019 SPIRVTypeInst VecEltTy = GR->getOrCreateSPIRVType(
1020 (Ty == IntTy ? LLVMIntTy : LLVMDstTy)->getScalarType(), MIRBuilder,
1021 SPIRV::AccessQualifier::ReadWrite,
1022 /*EmitIR*/ true);
1023 GR->assignSPIRVTypeToVReg(VecEltTy, ScalarC.getReg(0), MF);
1024 return assignSPIRVTy(MIRBuilder.buildSplatBuildVector(Ty, ScalarC));
1025 };
1026
1027 if (Mask == fcNone) {
1028 MIRBuilder.buildCopy(DstReg, buildSPIRVConstant(DstTy, 0));
1029 MI.eraseFromParent();
1030 return true;
1031 }
1032 if (Mask == fcAllFlags) {
1033 MIRBuilder.buildCopy(DstReg, buildSPIRVConstant(DstTy, 1));
1034 MI.eraseFromParent();
1035 return true;
1036 }
1037
1038 // Note that rather than creating a COPY here (between a floating-point and
1039 // integer type of the same size) we create a SPIR-V bitcast immediately. We
1040 // can't create a G_BITCAST because the LLTs are the same, and we can't seem
1041 // to correctly lower COPYs to SPIR-V bitcasts at this moment.
1042 Register ResVReg = MRI.createGenericVirtualRegister(IntTy);
1043 MRI.setRegClass(ResVReg, GR->getRegClass(SPIRVIntTy));
1044 GR->assignSPIRVTypeToVReg(SPIRVIntTy, ResVReg, Helper.MIRBuilder.getMF());
1045 auto AsInt = MIRBuilder.buildInstr(SPIRV::OpBitcast)
1046 .addDef(ResVReg)
1047 .addUse(GR->getSPIRVTypeID(SPIRVIntTy))
1048 .addUse(SrcReg);
1049 AsInt = assignSPIRVTy(std::move(AsInt));
1050
1051 // Various masks.
1052 APInt SignBit = APInt::getSignMask(BitSize);
1053 APInt ValueMask = APInt::getSignedMaxValue(BitSize); // All bits but sign.
1054 APInt Inf = APFloat::getInf(Semantics).bitcastToAPInt(); // Exp and int bit.
1055 APInt ExpMask = Inf;
1056 APInt AllOneMantissa = APFloat::getLargest(Semantics).bitcastToAPInt() & ~Inf;
1057 APInt QNaNBitMask =
1058 APInt::getOneBitSet(BitSize, AllOneMantissa.getActiveBits() - 1);
1059 APInt InversionMask = APInt::getAllOnes(DstTy.getScalarSizeInBits());
1060
1061 auto SignBitC = buildSPIRVConstant(IntTy, SignBit);
1062 auto ValueMaskC = buildSPIRVConstant(IntTy, ValueMask);
1063 auto InfC = buildSPIRVConstant(IntTy, Inf);
1064 auto ExpMaskC = buildSPIRVConstant(IntTy, ExpMask);
1065 auto ZeroC = buildSPIRVConstant(IntTy, 0);
1066
1067 auto Abs = assignSPIRVTy(MIRBuilder.buildAnd(IntTy, AsInt, ValueMaskC));
1068 auto Sign = assignSPIRVTy(
1069 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_NE, DstTy, AsInt, Abs));
1070
1071 auto Res = buildSPIRVConstant(DstTy, 0);
1072
1073 const auto appendToRes = [&](MachineInstrBuilder &&ToAppend) {
1074 Res = assignSPIRVTy(
1075 MIRBuilder.buildOr(DstTyCopy, Res, assignSPIRVTy(std::move(ToAppend))));
1076 };
1077
1078 // Tests that involve more than one class should be processed first.
1079 if ((Mask & fcFinite) == fcFinite) {
1080 // finite(V) ==> abs(V) u< exp_mask
1081 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_ULT, DstTy, Abs,
1082 ExpMaskC));
1083 Mask &= ~fcFinite;
1084 } else if ((Mask & fcFinite) == fcPosFinite) {
1085 // finite(V) && V > 0 ==> V u< exp_mask
1086 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_ULT, DstTy, AsInt,
1087 ExpMaskC));
1088 Mask &= ~fcPosFinite;
1089 } else if ((Mask & fcFinite) == fcNegFinite) {
1090 // finite(V) && V < 0 ==> abs(V) u< exp_mask && signbit == 1
1091 auto Cmp = assignSPIRVTy(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_ULT,
1092 DstTy, Abs, ExpMaskC));
1093 appendToRes(MIRBuilder.buildAnd(DstTy, Cmp, Sign));
1094 Mask &= ~fcNegFinite;
1095 }
1096
1097 if (FPClassTest PartialCheck = Mask & (fcZero | fcSubnormal)) {
1098 // fcZero | fcSubnormal => test all exponent bits are 0
1099 // TODO: Handle sign bit specific cases
1100 // TODO: Handle inverted case
1101 if (PartialCheck == (fcZero | fcSubnormal)) {
1102 auto ExpBits = assignSPIRVTy(MIRBuilder.buildAnd(IntTy, AsInt, ExpMaskC));
1103 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy,
1104 ExpBits, ZeroC));
1105 Mask &= ~PartialCheck;
1106 }
1107 }
1108
1109 // Check for individual classes.
1110 if (FPClassTest PartialCheck = Mask & fcZero) {
1111 if (PartialCheck == fcPosZero)
1112 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy,
1113 AsInt, ZeroC));
1114 else if (PartialCheck == fcZero)
1115 appendToRes(
1116 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy, Abs, ZeroC));
1117 else // fcNegZero
1118 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy,
1119 AsInt, SignBitC));
1120 }
1121
1122 if (FPClassTest PartialCheck = Mask & fcSubnormal) {
1123 // issubnormal(V) ==> unsigned(abs(V) - 1) u< (all mantissa bits set)
1124 // issubnormal(V) && V>0 ==> unsigned(V - 1) u< (all mantissa bits set)
1125 auto V = (PartialCheck == fcPosSubnormal) ? AsInt : Abs;
1126 auto OneC = buildSPIRVConstant(IntTy, 1);
1127 auto VMinusOne = MIRBuilder.buildSub(IntTy, V, OneC);
1128 auto SubnormalRes = assignSPIRVTy(
1129 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_ULT, DstTy, VMinusOne,
1130 buildSPIRVConstant(IntTy, AllOneMantissa)));
1131 if (PartialCheck == fcNegSubnormal)
1132 SubnormalRes = MIRBuilder.buildAnd(DstTy, SubnormalRes, Sign);
1133 appendToRes(std::move(SubnormalRes));
1134 }
1135
1136 if (FPClassTest PartialCheck = Mask & fcInf) {
1137 if (PartialCheck == fcPosInf)
1138 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy,
1139 AsInt, InfC));
1140 else if (PartialCheck == fcInf)
1141 appendToRes(
1142 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy, Abs, InfC));
1143 else { // fcNegInf
1144 APInt NegInf = APFloat::getInf(Semantics, true).bitcastToAPInt();
1145 auto NegInfC = buildSPIRVConstant(IntTy, NegInf);
1146 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_EQ, DstTy,
1147 AsInt, NegInfC));
1148 }
1149 }
1150
1151 if (FPClassTest PartialCheck = Mask & fcNan) {
1152 auto InfWithQnanBitC =
1153 buildSPIRVConstant(IntTy, std::move(Inf) | QNaNBitMask);
1154 if (PartialCheck == fcNan) {
1155 // isnan(V) ==> abs(V) u> int(inf)
1156 appendToRes(
1157 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_UGT, DstTy, Abs, InfC));
1158 } else if (PartialCheck == fcQNan) {
1159 // isquiet(V) ==> abs(V) u>= (unsigned(Inf) | quiet_bit)
1160 appendToRes(MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_UGE, DstTy, Abs,
1161 InfWithQnanBitC));
1162 } else { // fcSNan
1163 // issignaling(V) ==> abs(V) u> unsigned(Inf) &&
1164 // abs(V) u< (unsigned(Inf) | quiet_bit)
1165 auto IsNan = assignSPIRVTy(
1166 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_UGT, DstTy, Abs, InfC));
1167 auto IsNotQnan = assignSPIRVTy(MIRBuilder.buildICmp(
1168 CmpInst::Predicate::ICMP_ULT, DstTy, Abs, InfWithQnanBitC));
1169 appendToRes(MIRBuilder.buildAnd(DstTy, IsNan, IsNotQnan));
1170 }
1171 }
1172
1173 if (FPClassTest PartialCheck = Mask & fcNormal) {
1174 // isnormal(V) ==> (0 u< exp u< max_exp) ==> (unsigned(exp-1) u<
1175 // (max_exp-1))
1176 APInt ExpLSB = ExpMask & ~(ExpMask.shl(1));
1177 auto ExpMinusOne = assignSPIRVTy(
1178 MIRBuilder.buildSub(IntTy, Abs, buildSPIRVConstant(IntTy, ExpLSB)));
1179 APInt MaxExpMinusOne = std::move(ExpMask) - ExpLSB;
1180 auto NormalRes = assignSPIRVTy(
1181 MIRBuilder.buildICmp(CmpInst::Predicate::ICMP_ULT, DstTy, ExpMinusOne,
1182 buildSPIRVConstant(IntTy, MaxExpMinusOne)));
1183 if (PartialCheck == fcNegNormal)
1184 NormalRes = MIRBuilder.buildAnd(DstTy, NormalRes, Sign);
1185 else if (PartialCheck == fcPosNormal) {
1186 auto PosSign = assignSPIRVTy(MIRBuilder.buildXor(
1187 DstTy, Sign, buildSPIRVConstant(DstTy, InversionMask)));
1188 NormalRes = MIRBuilder.buildAnd(DstTy, NormalRes, PosSign);
1189 }
1190 appendToRes(std::move(NormalRes));
1191 }
1192
1193 MIRBuilder.buildCopy(DstReg, Res);
1194 MI.eraseFromParent();
1195 return true;
1196}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static void scalarize(Instruction *I, SmallVectorImpl< Instruction * > &Worklist)
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
IRTranslator LLVM IR MI
#define I(x, y, z)
Definition MD5.cpp:57
This file declares the MachineIRBuilder class.
Register Reg
Promote Memory to Register
Definition Mem2Reg.cpp:110
ppc ctr loops verify
const SmallVectorImpl< MachineOperand > & Cond
static bool legalizeSpvInsertElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool needsVectorLegalization(const LLT &Ty, const SPIRVSubtarget &ST)
static MachineInstrBuilder createStackTemporaryForVector(LegalizerHelper &Helper, SPIRVGlobalRegistry *GR, Register SrcReg, LLT SrcTy, MachinePointerInfo &PtrInfo, Align &VecAlign)
static Register convertPtrToInt(Register Reg, LLT ConvTy, SPIRVTypeInst SpvType, LegalizerHelper &Helper, MachineRegisterInfo &MRI, SPIRVGlobalRegistry *GR)
LegalityPredicate typeOfExtendedScalars(unsigned TypeIdx, bool IsExtendedInts)
static bool legalizeExtractVectorElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeStore(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvExtractElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvBitcast(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvConstComposite(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeLoad(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeInsertVectorElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
#define LLVM_DEBUG(...)
Definition Debug.h:114
APInt bitcastToAPInt() const
Definition APFloat.h:1408
static APFloat getLargest(const fltSemantics &Sem, bool Negative=false)
Returns the largest finite number in the given semantics.
Definition APFloat.h:1193
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1153
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
static APInt getSignMask(unsigned BitWidth)
Get the SignMask for a specific bit width.
Definition APInt.h:230
unsigned getActiveBits() const
Compute the number of active bits in the value.
Definition APInt.h:1527
static APInt getSignedMaxValue(unsigned numBits)
Gets maximum signed value of APInt for a specific bit width.
Definition APInt.h:210
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
static APInt getOneBitSet(unsigned numBits, unsigned BitNo)
Return an APInt with exactly one bit set in the result.
Definition APInt.h:240
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
static LLVM_ABI ArrayType * get(Type *ElementType, uint64_t NumElements)
This static method is the primary way to construct an ArrayType.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:700
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:701
@ ICMP_NE
not equal
Definition InstrTypes.h:698
static constexpr ElementCount getFixed(ScalarTy MinVal)
Definition TypeSize.h:309
static LLVM_ABI IntegerType * get(LLVMContext &C, unsigned NumBits)
This static method is the primary way of constructing an IntegerType.
Definition Type.cpp:318
static constexpr LLT vector(ElementCount EC, unsigned ScalarSizeInBits)
Get a low-level vector of some number of elements and element width.
constexpr bool isPointerVector() const
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr bool isValid() const
constexpr uint16_t getNumElements() const
Returns the number of elements in a vector LLT.
constexpr bool isVector() const
static constexpr LLT pointer(unsigned AddressSpace, unsigned SizeInBits)
Get a low-level pointer in the given address space.
constexpr bool isPointer() const
constexpr LLT getElementType() const
Returns the vector's element type. Only valid for vector types.
constexpr unsigned getAddressSpace() const
static constexpr LLT fixed_vector(unsigned NumElements, unsigned ScalarSizeInBits)
Get a low-level fixed-width vector of some number of elements and element width.
constexpr bool isPointerOrPointerVector() const
constexpr bool isFixedVector() const
Returns true if the LLT is a fixed vector.
constexpr LLT getScalarType() const
constexpr TypeSize getSizeInBytes() const
Returns the total size of the type in bytes, i.e.
LLVM_ABI void computeTables()
Compute any ancillary tables needed to quickly decide how an operation should be handled.
LegalizeRuleSet & legalFor(std::initializer_list< LLT > Types)
The instruction is legal when type index 0 is any type in the given list.
LegalizeRuleSet & fewerElementsIf(LegalityPredicate Predicate, LegalizeMutation Mutation)
Remove elements to reach the type selected by the mutation if the predicate is true.
LegalizeRuleSet & moreElementsToNextPow2(unsigned TypeIdx)
Add more elements to the vector to reach the next power of two.
LegalizeRuleSet & lower()
The instruction is lowered.
LegalizeRuleSet & scalarizeIf(LegalityPredicate Predicate, unsigned TypeIdx)
LegalizeRuleSet & lowerIf(LegalityPredicate Predicate)
The instruction is lowered if predicate is true.
LegalizeRuleSet & custom()
Unconditionally custom lower.
LegalizeRuleSet & unsupportedIf(LegalityPredicate Predicate)
LegalizeRuleSet & alwaysLegal()
LegalizeRuleSet & scalarize(unsigned TypeIdx)
LegalizeRuleSet & legalForCartesianProduct(std::initializer_list< LLT > Types)
The instruction is legal when type indexes 0 and 1 are both in the given list.
LegalizeRuleSet & legalIf(LegalityPredicate Predicate)
The instruction is legal if predicate is true.
LegalizeRuleSet & customFor(std::initializer_list< LLT > Types)
LLVM_ABI MachineInstrBuilder createStackTemporary(TypeSize Bytes, Align Alignment, MachinePointerInfo &PtrInfo)
Create a stack temporary based on the size in bytes and the alignment.
MachineIRBuilder & MIRBuilder
Expose MIRBuilder so clients can set their own RecordInsertInstruction functions.
LLVM_ABI Align getStackTemporaryAlignment(LLT Type, Align MinAlign=Align()) const
Return the alignment to use for a stack temporary object with the given type.
LegalizeRuleSet & getActionDefinitionsBuilder(unsigned Opcode)
Get the action definition builder for the given opcode.
const LegacyLegalizerInfo & getLegacyLegalizerInfo() const
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Helper class to build MachineInstr.
LLVMContext & getContext() const
MachineInstrBuilder buildUnmerge(ArrayRef< LLT > Res, const SrcOp &Op)
Build and insert Res0, ... = G_UNMERGE_VALUES Op.
MachineInstrBuilder buildAnd(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1)
Build and insert Res = G_AND Op0, Op1.
MachineInstrBuilder buildICmp(CmpInst::Predicate Pred, const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert a Res = G_ICMP Pred, Op0, Op1.
MachineInstrBuilder buildSub(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_SUB Op0, Op1.
MachineInstrBuilder buildIntrinsic(Intrinsic::ID ID, ArrayRef< Register > Res, bool HasSideEffects, bool isConvergent)
Build and insert a G_INTRINSIC instruction.
MachineInstrBuilder buildSplatBuildVector(const DstOp &Res, const SrcOp &Src)
Build and insert Res = G_BUILD_VECTOR with Src replicated to fill the number of elements.
MachineInstrBuilder buildBuildVector(const DstOp &Res, ArrayRef< Register > Ops)
Build and insert Res = G_BUILD_VECTOR Op0, ...
MachineInstrBuilder buildLoad(const DstOp &Res, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert Res = G_LOAD Addr, MMO.
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
MachineInstrBuilder buildInstr(unsigned Opcode)
Build and insert <empty> = Opcode <empty>.
MachineFunction & getMF()
Getter for the function we currently build.
MachineInstrBuilder buildBitcast(const DstOp &Dst, const SrcOp &Src)
Build and insert Dst = G_BITCAST Src.
MachineRegisterInfo * getMRI()
Getter for MRI.
MachineInstrBuilder buildOr(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_OR Op0, Op1.
MachineInstrBuilder buildCopy(const DstOp &Res, const SrcOp &Op)
Build and insert Res = COPY Op.
MachineInstrBuilder buildXor(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1)
Build and insert Res = G_XOR Op0, Op1.
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
const MachineInstrBuilder & addUse(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addDef(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register definition operand.
Representation of each machine instruction.
A description of a memory reference used in the backend.
const MachinePointerInfo & getPointerInfo() const
LLVM_ABI Align getAlign() const
Return the minimum known alignment in bytes of the actual memory reference.
MachineOperand class - Representation of each machine instruction operand.
Register getReg() const
getReg - Returns the register number.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLT getType(Register Reg) const
Get the low-level type of Reg or LLT{} if Reg is not a generic (target independent) virtual register.
LLVM_ABI void setRegClass(Register Reg, const TargetRegisterClass *RC)
setRegClass - Set the register class of the specified virtual register.
LLVM_ABI Register createGenericVirtualRegister(LLT Ty, StringRef Name="")
Create and return a new generic virtual register with low-level type Ty.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
void assignSPIRVTypeToVReg(SPIRVTypeInst Type, Register VReg, const MachineFunction &MF)
const TargetRegisterClass * getRegClass(SPIRVTypeInst SpvType) const
const Type * getTypeForSPIRVType(SPIRVTypeInst Ty) const
LLT getRegType(SPIRVTypeInst SpvType) const
SPIRVTypeInst getOrCreateSPIRVPointerType(const Type *BaseType, MachineIRBuilder &MIRBuilder, SPIRV::StorageClass::StorageClass SC)
SPIRVTypeInst getScalarOrVectorComponentType(SPIRVTypeInst Type) const
SPIRVTypeInst getOrCreateSPIRVType(const Type *Type, MachineInstr &I, SPIRV::AccessQualifier::AccessQualifier AQ, bool EmitIR)
SPIRVTypeInst getSPIRVTypeForVReg(Register VReg, const MachineFunction *MF=nullptr) const
SPIRVLegalizerInfo(const SPIRVSubtarget &ST)
bool legalizeCustom(LegalizerHelper &Helper, MachineInstr &MI, LostDebugLocObserver &LocObserver) const override
Called for instructions with the Custom LegalizationAction.
bool legalizeIntrinsic(LegalizerHelper &Helper, MachineInstr &MI) const override
SPIRVGlobalRegistry * getSPIRVGlobalRegistry() const
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI LegalityPredicate isScalar(unsigned TypeIdx)
True iff the specified type index is a scalar.
LLVM_ABI LegalityPredicate numElementsNotPow2(unsigned TypeIdx)
True iff the specified type index is a vector whose element count is not a power of 2.
LLVM_ABI LegalityPredicate vectorElementCountIsLessThanOrEqualTo(unsigned TypeIdx, unsigned Size)
True iff the specified type index is a vector with a number of elements that's less than or equal to ...
LLVM_ABI LegalityPredicate typeInSet(unsigned TypeIdx, std::initializer_list< LLT > TypesInit)
True iff the given type index is one of the specified types.
LLVM_ABI LegalityPredicate vectorElementCountIsGreaterThan(unsigned TypeIdx, unsigned Size)
True iff the specified type index is a vector with a number of elements that's greater than the given...
Predicate any(Predicate P0, Predicate P1)
True iff P0 or P1 are true.
LegalityPredicate typeIsNot(unsigned TypeIdx, LLT Type)
True iff the given type index is not the specified type.
Predicate all(Predicate P0, Predicate P1)
True iff P0 and P1 are true.
LLVM_ABI LegalityPredicate typeIs(unsigned TypeIdx, LLT TypesInit)
True iff the given type index is the specified type.
LLVM_ABI LegalizeMutation changeElementCountTo(unsigned TypeIdx, unsigned FromTypeIdx)
Keep the same scalar or element type as TypeIdx, but take the number of elements from FromTypeIdx.
LLVM_ABI LegalizeMutation changeElementSizeTo(unsigned TypeIdx, unsigned FromTypeIdx)
Change the scalar size or element size to have the same scalar size as type index FromIndex.
Invariant opcodes: All instruction sets have these as their low opcodes.
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
LLVM_ABI const llvm::fltSemantics & getFltSemanticForLLT(LLT Ty)
Get the appropriate floating point arithmetic semantic based on the bit size of the given scalar LLT.
std::function< bool(const LegalityQuery &)> LegalityPredicate
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
const std::set< unsigned > & getTypeFoldingSupportedOpcodes()
int64_t foldImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition Alignment.h:201
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
The LegalityQuery object bundles together all the information that's needed to decide whether a given...
ArrayRef< LLT > Types
This class contains a discriminated union of information about pointers in memory operands,...
MachinePointerInfo getWithOffset(int64_t O) const