LLVM 17.0.0git
MipsInstructionSelector.cpp
Go to the documentation of this file.
1//===- MipsInstructionSelector.cpp ------------------------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8/// \file
9/// This file implements the targeting of the InstructionSelector class for
10/// Mips.
11/// \todo This should be generated by TableGen.
12//===----------------------------------------------------------------------===//
13
15#include "MipsMachineFunction.h"
17#include "MipsTargetMachine.h"
21#include "llvm/IR/IntrinsicsMips.h"
22
23#define DEBUG_TYPE "mips-isel"
24
25using namespace llvm;
26
27namespace {
28
29#define GET_GLOBALISEL_PREDICATE_BITSET
30#include "MipsGenGlobalISel.inc"
31#undef GET_GLOBALISEL_PREDICATE_BITSET
32
33class MipsInstructionSelector : public InstructionSelector {
34public:
35 MipsInstructionSelector(const MipsTargetMachine &TM, const MipsSubtarget &STI,
36 const MipsRegisterBankInfo &RBI);
37
38 bool select(MachineInstr &I) override;
39 static const char *getName() { return DEBUG_TYPE; }
40
41private:
42 bool selectImpl(MachineInstr &I, CodeGenCoverage &CoverageInfo) const;
43 bool isRegInGprb(Register Reg, MachineRegisterInfo &MRI) const;
44 bool isRegInFprb(Register Reg, MachineRegisterInfo &MRI) const;
45 bool materialize32BitImm(Register DestReg, APInt Imm,
46 MachineIRBuilder &B) const;
49 getRegClassForTypeOnBank(Register Reg, MachineRegisterInfo &MRI) const;
50 unsigned selectLoadStoreOpCode(MachineInstr &I,
52 bool buildUnalignedStore(MachineInstr &I, unsigned Opc,
53 MachineOperand &BaseAddr, unsigned Offset,
54 MachineMemOperand *MMO) const;
55 bool buildUnalignedLoad(MachineInstr &I, unsigned Opc, Register Dest,
56 MachineOperand &BaseAddr, unsigned Offset,
57 Register TiedDest, MachineMemOperand *MMO) const;
58
59 const MipsTargetMachine &TM;
60 const MipsSubtarget &STI;
61 const MipsInstrInfo &TII;
62 const MipsRegisterInfo &TRI;
63 const MipsRegisterBankInfo &RBI;
64
65#define GET_GLOBALISEL_PREDICATES_DECL
66#include "MipsGenGlobalISel.inc"
67#undef GET_GLOBALISEL_PREDICATES_DECL
68
69#define GET_GLOBALISEL_TEMPORARIES_DECL
70#include "MipsGenGlobalISel.inc"
71#undef GET_GLOBALISEL_TEMPORARIES_DECL
72};
73
74} // end anonymous namespace
75
76#define GET_GLOBALISEL_IMPL
77#include "MipsGenGlobalISel.inc"
78#undef GET_GLOBALISEL_IMPL
79
80MipsInstructionSelector::MipsInstructionSelector(
81 const MipsTargetMachine &TM, const MipsSubtarget &STI,
82 const MipsRegisterBankInfo &RBI)
83 : TM(TM), STI(STI), TII(*STI.getInstrInfo()), TRI(*STI.getRegisterInfo()),
84 RBI(RBI),
85
87#include "MipsGenGlobalISel.inc"
90#include "MipsGenGlobalISel.inc"
92{
93}
94
95bool MipsInstructionSelector::isRegInGprb(Register Reg,
96 MachineRegisterInfo &MRI) const {
97 return RBI.getRegBank(Reg, MRI, TRI)->getID() == Mips::GPRBRegBankID;
98}
99
100bool MipsInstructionSelector::isRegInFprb(Register Reg,
101 MachineRegisterInfo &MRI) const {
102 return RBI.getRegBank(Reg, MRI, TRI)->getID() == Mips::FPRBRegBankID;
103}
104
105bool MipsInstructionSelector::selectCopy(MachineInstr &I,
106 MachineRegisterInfo &MRI) const {
107 Register DstReg = I.getOperand(0).getReg();
108 if (DstReg.isPhysical())
109 return true;
110
111 const TargetRegisterClass *RC = getRegClassForTypeOnBank(DstReg, MRI);
112 if (!RBI.constrainGenericRegister(DstReg, *RC, MRI)) {
113 LLVM_DEBUG(dbgs() << "Failed to constrain " << TII.getName(I.getOpcode())
114 << " operand\n");
115 return false;
116 }
117 return true;
118}
119
120const TargetRegisterClass *MipsInstructionSelector::getRegClassForTypeOnBank(
121 Register Reg, MachineRegisterInfo &MRI) const {
122 const LLT Ty = MRI.getType(Reg);
123 const unsigned TySize = Ty.getSizeInBits();
124
125 if (isRegInGprb(Reg, MRI)) {
126 assert((Ty.isScalar() || Ty.isPointer()) && TySize == 32 &&
127 "Register class not available for LLT, register bank combination");
128 return &Mips::GPR32RegClass;
129 }
130
131 if (isRegInFprb(Reg, MRI)) {
132 if (Ty.isScalar()) {
133 assert((TySize == 32 || TySize == 64) &&
134 "Register class not available for LLT, register bank combination");
135 if (TySize == 32)
136 return &Mips::FGR32RegClass;
137 return STI.isFP64bit() ? &Mips::FGR64RegClass : &Mips::AFGR64RegClass;
138 }
139 }
140
141 llvm_unreachable("Unsupported register bank.");
142}
143
144bool MipsInstructionSelector::materialize32BitImm(Register DestReg, APInt Imm,
145 MachineIRBuilder &B) const {
146 assert(Imm.getBitWidth() == 32 && "Unsupported immediate size.");
147 // Ori zero extends immediate. Used for values with zeros in high 16 bits.
148 if (Imm.getHiBits(16).isZero()) {
149 MachineInstr *Inst =
150 B.buildInstr(Mips::ORi, {DestReg}, {Register(Mips::ZERO)})
151 .addImm(Imm.getLoBits(16).getLimitedValue());
152 return constrainSelectedInstRegOperands(*Inst, TII, TRI, RBI);
153 }
154 // Lui places immediate in high 16 bits and sets low 16 bits to zero.
155 if (Imm.getLoBits(16).isZero()) {
156 MachineInstr *Inst = B.buildInstr(Mips::LUi, {DestReg}, {})
157 .addImm(Imm.getHiBits(16).getLimitedValue());
158 return constrainSelectedInstRegOperands(*Inst, TII, TRI, RBI);
159 }
160 // ADDiu sign extends immediate. Used for values with 1s in high 17 bits.
161 if (Imm.isSignedIntN(16)) {
162 MachineInstr *Inst =
163 B.buildInstr(Mips::ADDiu, {DestReg}, {Register(Mips::ZERO)})
164 .addImm(Imm.getLoBits(16).getLimitedValue());
165 return constrainSelectedInstRegOperands(*Inst, TII, TRI, RBI);
166 }
167 // Values that cannot be materialized with single immediate instruction.
168 Register LUiReg = B.getMRI()->createVirtualRegister(&Mips::GPR32RegClass);
169 MachineInstr *LUi = B.buildInstr(Mips::LUi, {LUiReg}, {})
170 .addImm(Imm.getHiBits(16).getLimitedValue());
171 MachineInstr *ORi = B.buildInstr(Mips::ORi, {DestReg}, {LUiReg})
172 .addImm(Imm.getLoBits(16).getLimitedValue());
173 if (!constrainSelectedInstRegOperands(*LUi, TII, TRI, RBI))
174 return false;
175 if (!constrainSelectedInstRegOperands(*ORi, TII, TRI, RBI))
176 return false;
177 return true;
178}
179
180/// When I.getOpcode() is returned, we failed to select MIPS instruction opcode.
181unsigned
182MipsInstructionSelector::selectLoadStoreOpCode(MachineInstr &I,
183 MachineRegisterInfo &MRI) const {
184 const Register ValueReg = I.getOperand(0).getReg();
185 const LLT Ty = MRI.getType(ValueReg);
186 const unsigned TySize = Ty.getSizeInBits();
187 const unsigned MemSizeInBytes = (*I.memoperands_begin())->getSize();
188 unsigned Opc = I.getOpcode();
189 const bool isStore = Opc == TargetOpcode::G_STORE;
190
191 if (isRegInGprb(ValueReg, MRI)) {
192 assert(((Ty.isScalar() && TySize == 32) ||
193 (Ty.isPointer() && TySize == 32 && MemSizeInBytes == 4)) &&
194 "Unsupported register bank, LLT, MemSizeInBytes combination");
195 (void)TySize;
196 if (isStore)
197 switch (MemSizeInBytes) {
198 case 4:
199 return Mips::SW;
200 case 2:
201 return Mips::SH;
202 case 1:
203 return Mips::SB;
204 default:
205 return Opc;
206 }
207 else
208 // Unspecified extending load is selected into zeroExtending load.
209 switch (MemSizeInBytes) {
210 case 4:
211 return Mips::LW;
212 case 2:
213 return Opc == TargetOpcode::G_SEXTLOAD ? Mips::LH : Mips::LHu;
214 case 1:
215 return Opc == TargetOpcode::G_SEXTLOAD ? Mips::LB : Mips::LBu;
216 default:
217 return Opc;
218 }
219 }
220
221 if (isRegInFprb(ValueReg, MRI)) {
222 if (Ty.isScalar()) {
223 assert(((TySize == 32 && MemSizeInBytes == 4) ||
224 (TySize == 64 && MemSizeInBytes == 8)) &&
225 "Unsupported register bank, LLT, MemSizeInBytes combination");
226
227 if (MemSizeInBytes == 4)
228 return isStore ? Mips::SWC1 : Mips::LWC1;
229
230 if (STI.isFP64bit())
231 return isStore ? Mips::SDC164 : Mips::LDC164;
232 return isStore ? Mips::SDC1 : Mips::LDC1;
233 }
234
235 if (Ty.isVector()) {
236 assert(STI.hasMSA() && "Vector instructions require target with MSA.");
237 assert((TySize == 128 && MemSizeInBytes == 16) &&
238 "Unsupported register bank, LLT, MemSizeInBytes combination");
239 switch (Ty.getElementType().getSizeInBits()) {
240 case 8:
241 return isStore ? Mips::ST_B : Mips::LD_B;
242 case 16:
243 return isStore ? Mips::ST_H : Mips::LD_H;
244 case 32:
245 return isStore ? Mips::ST_W : Mips::LD_W;
246 case 64:
247 return isStore ? Mips::ST_D : Mips::LD_D;
248 default:
249 return Opc;
250 }
251 }
252 }
253
254 return Opc;
255}
256
257bool MipsInstructionSelector::buildUnalignedStore(
258 MachineInstr &I, unsigned Opc, MachineOperand &BaseAddr, unsigned Offset,
259 MachineMemOperand *MMO) const {
260 MachineInstr *NewInst =
261 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(Opc))
262 .add(I.getOperand(0))
263 .add(BaseAddr)
264 .addImm(Offset)
265 .addMemOperand(MMO);
266 if (!constrainSelectedInstRegOperands(*NewInst, TII, TRI, RBI))
267 return false;
268 return true;
269}
270
271bool MipsInstructionSelector::buildUnalignedLoad(
272 MachineInstr &I, unsigned Opc, Register Dest, MachineOperand &BaseAddr,
273 unsigned Offset, Register TiedDest, MachineMemOperand *MMO) const {
274 MachineInstr *NewInst =
275 BuildMI(*I.getParent(), I, I.getDebugLoc(), TII.get(Opc))
276 .addDef(Dest)
277 .add(BaseAddr)
278 .addImm(Offset)
279 .addUse(TiedDest)
280 .addMemOperand(*I.memoperands_begin());
281 if (!constrainSelectedInstRegOperands(*NewInst, TII, TRI, RBI))
282 return false;
283 return true;
284}
285
286bool MipsInstructionSelector::select(MachineInstr &I) {
287
288 MachineBasicBlock &MBB = *I.getParent();
291
292 if (!isPreISelGenericOpcode(I.getOpcode())) {
293 if (I.isCopy())
294 return selectCopy(I, MRI);
295
296 return true;
297 }
298
299 if (I.getOpcode() == Mips::G_MUL &&
300 isRegInGprb(I.getOperand(0).getReg(), MRI)) {
301 MachineInstr *Mul = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::MUL))
302 .add(I.getOperand(0))
303 .add(I.getOperand(1))
304 .add(I.getOperand(2));
306 return false;
307 Mul->getOperand(3).setIsDead(true);
308 Mul->getOperand(4).setIsDead(true);
309
310 I.eraseFromParent();
311 return true;
312 }
313
314 if (selectImpl(I, *CoverageInfo))
315 return true;
316
317 MachineInstr *MI = nullptr;
318 using namespace TargetOpcode;
319
320 switch (I.getOpcode()) {
321 case G_UMULH: {
322 Register PseudoMULTuReg = MRI.createVirtualRegister(&Mips::ACC64RegClass);
323 MachineInstr *PseudoMULTu, *PseudoMove;
324
325 PseudoMULTu = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::PseudoMULTu))
326 .addDef(PseudoMULTuReg)
327 .add(I.getOperand(1))
328 .add(I.getOperand(2));
329 if (!constrainSelectedInstRegOperands(*PseudoMULTu, TII, TRI, RBI))
330 return false;
331
332 PseudoMove = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::PseudoMFHI))
333 .addDef(I.getOperand(0).getReg())
334 .addUse(PseudoMULTuReg);
335 if (!constrainSelectedInstRegOperands(*PseudoMove, TII, TRI, RBI))
336 return false;
337
338 I.eraseFromParent();
339 return true;
340 }
341 case G_PTR_ADD: {
342 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDu))
343 .add(I.getOperand(0))
344 .add(I.getOperand(1))
345 .add(I.getOperand(2));
346 break;
347 }
348 case G_INTTOPTR:
349 case G_PTRTOINT: {
350 I.setDesc(TII.get(COPY));
351 return selectCopy(I, MRI);
352 }
353 case G_FRAME_INDEX: {
354 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDiu))
355 .add(I.getOperand(0))
356 .add(I.getOperand(1))
357 .addImm(0);
358 break;
359 }
360 case G_BRCOND: {
361 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::BNE))
362 .add(I.getOperand(0))
363 .addUse(Mips::ZERO)
364 .add(I.getOperand(1));
365 break;
366 }
367 case G_BRJT: {
368 unsigned EntrySize =
370 assert(isPowerOf2_32(EntrySize) &&
371 "Non-power-of-two jump-table entry size not supported.");
372
373 Register JTIndex = MRI.createVirtualRegister(&Mips::GPR32RegClass);
374 MachineInstr *SLL = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::SLL))
375 .addDef(JTIndex)
376 .addUse(I.getOperand(2).getReg())
377 .addImm(Log2_32(EntrySize));
378 if (!constrainSelectedInstRegOperands(*SLL, TII, TRI, RBI))
379 return false;
380
381 Register DestAddress = MRI.createVirtualRegister(&Mips::GPR32RegClass);
382 MachineInstr *ADDu = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDu))
383 .addDef(DestAddress)
384 .addUse(I.getOperand(0).getReg())
385 .addUse(JTIndex);
386 if (!constrainSelectedInstRegOperands(*ADDu, TII, TRI, RBI))
387 return false;
388
389 Register Dest = MRI.createVirtualRegister(&Mips::GPR32RegClass);
390 MachineInstr *LW =
391 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LW))
392 .addDef(Dest)
393 .addUse(DestAddress)
394 .addJumpTableIndex(I.getOperand(1).getIndex(), MipsII::MO_ABS_LO)
398 return false;
399
400 if (MF.getTarget().isPositionIndependent()) {
401 Register DestTmp = MRI.createVirtualRegister(&Mips::GPR32RegClass);
402 LW->getOperand(0).setReg(DestTmp);
403 MachineInstr *ADDu = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDu))
404 .addDef(Dest)
405 .addUse(DestTmp)
408 if (!constrainSelectedInstRegOperands(*ADDu, TII, TRI, RBI))
409 return false;
410 }
411
413 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::PseudoIndirectBranch))
414 .addUse(Dest);
415 if (!constrainSelectedInstRegOperands(*Branch, TII, TRI, RBI))
416 return false;
417
418 I.eraseFromParent();
419 return true;
420 }
421 case G_BRINDIRECT: {
422 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::PseudoIndirectBranch))
423 .add(I.getOperand(0));
424 break;
425 }
426 case G_PHI: {
427 const Register DestReg = I.getOperand(0).getReg();
428
429 const TargetRegisterClass *DefRC = nullptr;
430 if (DestReg.isPhysical())
431 DefRC = TRI.getRegClass(DestReg);
432 else
433 DefRC = getRegClassForTypeOnBank(DestReg, MRI);
434
435 I.setDesc(TII.get(TargetOpcode::PHI));
436 return RBI.constrainGenericRegister(DestReg, *DefRC, MRI);
437 }
438 case G_STORE:
439 case G_LOAD:
440 case G_ZEXTLOAD:
441 case G_SEXTLOAD: {
442 auto MMO = *I.memoperands_begin();
443 MachineOperand BaseAddr = I.getOperand(1);
444 int64_t SignedOffset = 0;
445 // Try to fold load/store + G_PTR_ADD + G_CONSTANT
446 // %SignedOffset:(s32) = G_CONSTANT i32 16_bit_signed_immediate
447 // %Addr:(p0) = G_PTR_ADD %BaseAddr, %SignedOffset
448 // %LoadResult/%StoreSrc = load/store %Addr(p0)
449 // into:
450 // %LoadResult/%StoreSrc = NewOpc %BaseAddr(p0), 16_bit_signed_immediate
451
452 MachineInstr *Addr = MRI.getVRegDef(I.getOperand(1).getReg());
453 if (Addr->getOpcode() == G_PTR_ADD) {
454 MachineInstr *Offset = MRI.getVRegDef(Addr->getOperand(2).getReg());
455 if (Offset->getOpcode() == G_CONSTANT) {
456 APInt OffsetValue = Offset->getOperand(1).getCImm()->getValue();
457 if (OffsetValue.isSignedIntN(16)) {
458 BaseAddr = Addr->getOperand(1);
459 SignedOffset = OffsetValue.getSExtValue();
460 }
461 }
462 }
463
464 // Unaligned memory access
465 if (MMO->getAlign() < MMO->getSize() &&
466 !STI.systemSupportsUnalignedAccess()) {
467 if (MMO->getSize() != 4 || !isRegInGprb(I.getOperand(0).getReg(), MRI))
468 return false;
469
470 if (I.getOpcode() == G_STORE) {
471 if (!buildUnalignedStore(I, Mips::SWL, BaseAddr, SignedOffset + 3, MMO))
472 return false;
473 if (!buildUnalignedStore(I, Mips::SWR, BaseAddr, SignedOffset, MMO))
474 return false;
475 I.eraseFromParent();
476 return true;
477 }
478
479 if (I.getOpcode() == G_LOAD) {
480 Register ImplDef = MRI.createVirtualRegister(&Mips::GPR32RegClass);
481 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::IMPLICIT_DEF))
482 .addDef(ImplDef);
483 Register Tmp = MRI.createVirtualRegister(&Mips::GPR32RegClass);
484 if (!buildUnalignedLoad(I, Mips::LWL, Tmp, BaseAddr, SignedOffset + 3,
485 ImplDef, MMO))
486 return false;
487 if (!buildUnalignedLoad(I, Mips::LWR, I.getOperand(0).getReg(),
488 BaseAddr, SignedOffset, Tmp, MMO))
489 return false;
490 I.eraseFromParent();
491 return true;
492 }
493
494 return false;
495 }
496
497 const unsigned NewOpc = selectLoadStoreOpCode(I, MRI);
498 if (NewOpc == I.getOpcode())
499 return false;
500
501 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(NewOpc))
502 .add(I.getOperand(0))
503 .add(BaseAddr)
504 .addImm(SignedOffset)
505 .addMemOperand(MMO);
506 break;
507 }
508 case G_UDIV:
509 case G_UREM:
510 case G_SDIV:
511 case G_SREM: {
512 Register HILOReg = MRI.createVirtualRegister(&Mips::ACC64RegClass);
513 bool IsSigned = I.getOpcode() == G_SREM || I.getOpcode() == G_SDIV;
514 bool IsDiv = I.getOpcode() == G_UDIV || I.getOpcode() == G_SDIV;
515
516 MachineInstr *PseudoDIV, *PseudoMove;
517 PseudoDIV = BuildMI(MBB, I, I.getDebugLoc(),
518 TII.get(IsSigned ? Mips::PseudoSDIV : Mips::PseudoUDIV))
519 .addDef(HILOReg)
520 .add(I.getOperand(1))
521 .add(I.getOperand(2));
522 if (!constrainSelectedInstRegOperands(*PseudoDIV, TII, TRI, RBI))
523 return false;
524
525 PseudoMove = BuildMI(MBB, I, I.getDebugLoc(),
526 TII.get(IsDiv ? Mips::PseudoMFLO : Mips::PseudoMFHI))
527 .addDef(I.getOperand(0).getReg())
528 .addUse(HILOReg);
529 if (!constrainSelectedInstRegOperands(*PseudoMove, TII, TRI, RBI))
530 return false;
531
532 I.eraseFromParent();
533 return true;
534 }
535 case G_SELECT: {
536 // Handle operands with pointer type.
537 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::MOVN_I_I))
538 .add(I.getOperand(0))
539 .add(I.getOperand(2))
540 .add(I.getOperand(1))
541 .add(I.getOperand(3));
542 break;
543 }
544 case G_UNMERGE_VALUES: {
545 if (I.getNumOperands() != 3)
546 return false;
547 Register Src = I.getOperand(2).getReg();
548 Register Lo = I.getOperand(0).getReg();
549 Register Hi = I.getOperand(1).getReg();
550 if (!isRegInFprb(Src, MRI) ||
551 !(isRegInGprb(Lo, MRI) && isRegInGprb(Hi, MRI)))
552 return false;
553
554 unsigned Opcode =
555 STI.isFP64bit() ? Mips::ExtractElementF64_64 : Mips::ExtractElementF64;
556
557 MachineInstr *ExtractLo = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Opcode))
558 .addDef(Lo)
559 .addUse(Src)
560 .addImm(0);
561 if (!constrainSelectedInstRegOperands(*ExtractLo, TII, TRI, RBI))
562 return false;
563
564 MachineInstr *ExtractHi = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Opcode))
565 .addDef(Hi)
566 .addUse(Src)
567 .addImm(1);
568 if (!constrainSelectedInstRegOperands(*ExtractHi, TII, TRI, RBI))
569 return false;
570
571 I.eraseFromParent();
572 return true;
573 }
574 case G_IMPLICIT_DEF: {
575 Register Dst = I.getOperand(0).getReg();
576 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::IMPLICIT_DEF))
577 .addDef(Dst);
578
579 // Set class based on register bank, there can be fpr and gpr implicit def.
580 MRI.setRegClass(Dst, getRegClassForTypeOnBank(Dst, MRI));
581 break;
582 }
583 case G_CONSTANT: {
585 if (!materialize32BitImm(I.getOperand(0).getReg(),
586 I.getOperand(1).getCImm()->getValue(), B))
587 return false;
588
589 I.eraseFromParent();
590 return true;
591 }
592 case G_FCONSTANT: {
593 const APFloat &FPimm = I.getOperand(1).getFPImm()->getValueAPF();
594 APInt APImm = FPimm.bitcastToAPInt();
595 unsigned Size = MRI.getType(I.getOperand(0).getReg()).getSizeInBits();
596
597 if (Size == 32) {
598 Register GPRReg = MRI.createVirtualRegister(&Mips::GPR32RegClass);
600 if (!materialize32BitImm(GPRReg, APImm, B))
601 return false;
602
604 B.buildInstr(Mips::MTC1, {I.getOperand(0).getReg()}, {GPRReg});
605 if (!MTC1.constrainAllUses(TII, TRI, RBI))
606 return false;
607 }
608 if (Size == 64) {
609 Register GPRRegHigh = MRI.createVirtualRegister(&Mips::GPR32RegClass);
610 Register GPRRegLow = MRI.createVirtualRegister(&Mips::GPR32RegClass);
612 if (!materialize32BitImm(GPRRegHigh, APImm.getHiBits(32).trunc(32), B))
613 return false;
614 if (!materialize32BitImm(GPRRegLow, APImm.getLoBits(32).trunc(32), B))
615 return false;
616
617 MachineInstrBuilder PairF64 = B.buildInstr(
618 STI.isFP64bit() ? Mips::BuildPairF64_64 : Mips::BuildPairF64,
619 {I.getOperand(0).getReg()}, {GPRRegLow, GPRRegHigh});
620 if (!PairF64.constrainAllUses(TII, TRI, RBI))
621 return false;
622 }
623
624 I.eraseFromParent();
625 return true;
626 }
627 case G_FABS: {
628 unsigned Size = MRI.getType(I.getOperand(0).getReg()).getSizeInBits();
629 unsigned FABSOpcode =
630 Size == 32 ? Mips::FABS_S
631 : STI.isFP64bit() ? Mips::FABS_D64 : Mips::FABS_D32;
632 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(FABSOpcode))
633 .add(I.getOperand(0))
634 .add(I.getOperand(1));
635 break;
636 }
637 case G_FPTOSI: {
638 unsigned FromSize = MRI.getType(I.getOperand(1).getReg()).getSizeInBits();
639 unsigned ToSize = MRI.getType(I.getOperand(0).getReg()).getSizeInBits();
640 (void)ToSize;
641 assert((ToSize == 32) && "Unsupported integer size for G_FPTOSI");
642 assert((FromSize == 32 || FromSize == 64) &&
643 "Unsupported floating point size for G_FPTOSI");
644
645 unsigned Opcode;
646 if (FromSize == 32)
647 Opcode = Mips::TRUNC_W_S;
648 else
649 Opcode = STI.isFP64bit() ? Mips::TRUNC_W_D64 : Mips::TRUNC_W_D32;
650 Register ResultInFPR = MRI.createVirtualRegister(&Mips::FGR32RegClass);
651 MachineInstr *Trunc = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Opcode))
652 .addDef(ResultInFPR)
653 .addUse(I.getOperand(1).getReg());
654 if (!constrainSelectedInstRegOperands(*Trunc, TII, TRI, RBI))
655 return false;
656
657 MachineInstr *Move = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::MFC1))
658 .addDef(I.getOperand(0).getReg())
659 .addUse(ResultInFPR);
660 if (!constrainSelectedInstRegOperands(*Move, TII, TRI, RBI))
661 return false;
662
663 I.eraseFromParent();
664 return true;
665 }
666 case G_GLOBAL_VALUE: {
667 const llvm::GlobalValue *GVal = I.getOperand(1).getGlobal();
668 if (MF.getTarget().isPositionIndependent()) {
669 MachineInstr *LWGOT = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LW))
670 .addDef(I.getOperand(0).getReg())
673 .addGlobalAddress(GVal);
674 // Global Values that don't have local linkage are handled differently
675 // when they are part of call sequence. MipsCallLowering::lowerCall
676 // creates G_GLOBAL_VALUE instruction as part of call sequence and adds
677 // MO_GOT_CALL flag when Callee doesn't have local linkage.
678 if (I.getOperand(1).getTargetFlags() == MipsII::MO_GOT_CALL)
680 else
682 LWGOT->addMemOperand(
685 if (!constrainSelectedInstRegOperands(*LWGOT, TII, TRI, RBI))
686 return false;
687
688 if (GVal->hasLocalLinkage()) {
689 Register LWGOTDef = MRI.createVirtualRegister(&Mips::GPR32RegClass);
690 LWGOT->getOperand(0).setReg(LWGOTDef);
691
692 MachineInstr *ADDiu =
693 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDiu))
694 .addDef(I.getOperand(0).getReg())
695 .addReg(LWGOTDef)
696 .addGlobalAddress(GVal);
698 if (!constrainSelectedInstRegOperands(*ADDiu, TII, TRI, RBI))
699 return false;
700 }
701 } else {
702 Register LUiReg = MRI.createVirtualRegister(&Mips::GPR32RegClass);
703
704 MachineInstr *LUi = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LUi))
705 .addDef(LUiReg)
706 .addGlobalAddress(GVal);
708 if (!constrainSelectedInstRegOperands(*LUi, TII, TRI, RBI))
709 return false;
710
711 MachineInstr *ADDiu =
712 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDiu))
713 .addDef(I.getOperand(0).getReg())
714 .addUse(LUiReg)
715 .addGlobalAddress(GVal);
717 if (!constrainSelectedInstRegOperands(*ADDiu, TII, TRI, RBI))
718 return false;
719 }
720 I.eraseFromParent();
721 return true;
722 }
723 case G_JUMP_TABLE: {
724 if (MF.getTarget().isPositionIndependent()) {
725 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LW))
726 .addDef(I.getOperand(0).getReg())
729 .addJumpTableIndex(I.getOperand(1).getIndex(), MipsII::MO_GOT)
732 Align(4)));
733 } else {
734 MI =
735 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LUi))
736 .addDef(I.getOperand(0).getReg())
737 .addJumpTableIndex(I.getOperand(1).getIndex(), MipsII::MO_ABS_HI);
738 }
739 break;
740 }
741 case G_ICMP: {
742 struct Instr {
743 unsigned Opcode;
745 Instr(unsigned Opcode, Register Def, Register LHS, Register RHS)
746 : Opcode(Opcode), Def(Def), LHS(LHS), RHS(RHS){};
747
748 bool hasImm() const {
749 if (Opcode == Mips::SLTiu || Opcode == Mips::XORi)
750 return true;
751 return false;
752 }
753 };
754
756 Register ICMPReg = I.getOperand(0).getReg();
757 Register Temp = MRI.createVirtualRegister(&Mips::GPR32RegClass);
758 Register LHS = I.getOperand(2).getReg();
759 Register RHS = I.getOperand(3).getReg();
761 static_cast<CmpInst::Predicate>(I.getOperand(1).getPredicate());
762
763 switch (Cond) {
764 case CmpInst::ICMP_EQ: // LHS == RHS -> (LHS ^ RHS) < 1
765 Instructions.emplace_back(Mips::XOR, Temp, LHS, RHS);
766 Instructions.emplace_back(Mips::SLTiu, ICMPReg, Temp, 1);
767 break;
768 case CmpInst::ICMP_NE: // LHS != RHS -> 0 < (LHS ^ RHS)
769 Instructions.emplace_back(Mips::XOR, Temp, LHS, RHS);
770 Instructions.emplace_back(Mips::SLTu, ICMPReg, Mips::ZERO, Temp);
771 break;
772 case CmpInst::ICMP_UGT: // LHS > RHS -> RHS < LHS
773 Instructions.emplace_back(Mips::SLTu, ICMPReg, RHS, LHS);
774 break;
775 case CmpInst::ICMP_UGE: // LHS >= RHS -> !(LHS < RHS)
776 Instructions.emplace_back(Mips::SLTu, Temp, LHS, RHS);
777 Instructions.emplace_back(Mips::XORi, ICMPReg, Temp, 1);
778 break;
779 case CmpInst::ICMP_ULT: // LHS < RHS -> LHS < RHS
780 Instructions.emplace_back(Mips::SLTu, ICMPReg, LHS, RHS);
781 break;
782 case CmpInst::ICMP_ULE: // LHS <= RHS -> !(RHS < LHS)
783 Instructions.emplace_back(Mips::SLTu, Temp, RHS, LHS);
784 Instructions.emplace_back(Mips::XORi, ICMPReg, Temp, 1);
785 break;
786 case CmpInst::ICMP_SGT: // LHS > RHS -> RHS < LHS
787 Instructions.emplace_back(Mips::SLT, ICMPReg, RHS, LHS);
788 break;
789 case CmpInst::ICMP_SGE: // LHS >= RHS -> !(LHS < RHS)
790 Instructions.emplace_back(Mips::SLT, Temp, LHS, RHS);
791 Instructions.emplace_back(Mips::XORi, ICMPReg, Temp, 1);
792 break;
793 case CmpInst::ICMP_SLT: // LHS < RHS -> LHS < RHS
794 Instructions.emplace_back(Mips::SLT, ICMPReg, LHS, RHS);
795 break;
796 case CmpInst::ICMP_SLE: // LHS <= RHS -> !(RHS < LHS)
797 Instructions.emplace_back(Mips::SLT, Temp, RHS, LHS);
798 Instructions.emplace_back(Mips::XORi, ICMPReg, Temp, 1);
799 break;
800 default:
801 return false;
802 }
803
805 for (const struct Instr &Instruction : Instructions) {
806 MachineInstrBuilder MIB = B.buildInstr(
807 Instruction.Opcode, {Instruction.Def}, {Instruction.LHS});
808
809 if (Instruction.hasImm())
810 MIB.addImm(Instruction.RHS);
811 else
812 MIB.addUse(Instruction.RHS);
813
814 if (!MIB.constrainAllUses(TII, TRI, RBI))
815 return false;
816 }
817
818 I.eraseFromParent();
819 return true;
820 }
821 case G_FCMP: {
822 unsigned MipsFCMPCondCode;
823 bool isLogicallyNegated;
824 switch (CmpInst::Predicate Cond = static_cast<CmpInst::Predicate>(
825 I.getOperand(1).getPredicate())) {
826 case CmpInst::FCMP_UNO: // Unordered
827 case CmpInst::FCMP_ORD: // Ordered (OR)
828 MipsFCMPCondCode = Mips::FCOND_UN;
829 isLogicallyNegated = Cond != CmpInst::FCMP_UNO;
830 break;
831 case CmpInst::FCMP_OEQ: // Equal
832 case CmpInst::FCMP_UNE: // Not Equal (NEQ)
833 MipsFCMPCondCode = Mips::FCOND_OEQ;
834 isLogicallyNegated = Cond != CmpInst::FCMP_OEQ;
835 break;
836 case CmpInst::FCMP_UEQ: // Unordered or Equal
837 case CmpInst::FCMP_ONE: // Ordered or Greater Than or Less Than (OGL)
838 MipsFCMPCondCode = Mips::FCOND_UEQ;
839 isLogicallyNegated = Cond != CmpInst::FCMP_UEQ;
840 break;
841 case CmpInst::FCMP_OLT: // Ordered or Less Than
842 case CmpInst::FCMP_UGE: // Unordered or Greater Than or Equal (UGE)
843 MipsFCMPCondCode = Mips::FCOND_OLT;
844 isLogicallyNegated = Cond != CmpInst::FCMP_OLT;
845 break;
846 case CmpInst::FCMP_ULT: // Unordered or Less Than
847 case CmpInst::FCMP_OGE: // Ordered or Greater Than or Equal (OGE)
848 MipsFCMPCondCode = Mips::FCOND_ULT;
849 isLogicallyNegated = Cond != CmpInst::FCMP_ULT;
850 break;
851 case CmpInst::FCMP_OLE: // Ordered or Less Than or Equal
852 case CmpInst::FCMP_UGT: // Unordered or Greater Than (UGT)
853 MipsFCMPCondCode = Mips::FCOND_OLE;
854 isLogicallyNegated = Cond != CmpInst::FCMP_OLE;
855 break;
856 case CmpInst::FCMP_ULE: // Unordered or Less Than or Equal
857 case CmpInst::FCMP_OGT: // Ordered or Greater Than (OGT)
858 MipsFCMPCondCode = Mips::FCOND_ULE;
859 isLogicallyNegated = Cond != CmpInst::FCMP_ULE;
860 break;
861 default:
862 return false;
863 }
864
865 // Default compare result in gpr register will be `true`.
866 // We will move `false` (MIPS::Zero) to gpr result when fcmp gives false
867 // using MOVF_I. When orignal predicate (Cond) is logically negated
868 // MipsFCMPCondCode, result is inverted i.e. MOVT_I is used.
869 unsigned MoveOpcode = isLogicallyNegated ? Mips::MOVT_I : Mips::MOVF_I;
870
871 Register TrueInReg = MRI.createVirtualRegister(&Mips::GPR32RegClass);
872 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::ADDiu))
873 .addDef(TrueInReg)
874 .addUse(Mips::ZERO)
875 .addImm(1);
876
877 unsigned Size = MRI.getType(I.getOperand(2).getReg()).getSizeInBits();
878 unsigned FCMPOpcode =
879 Size == 32 ? Mips::FCMP_S32
880 : STI.isFP64bit() ? Mips::FCMP_D64 : Mips::FCMP_D32;
881 MachineInstr *FCMP = BuildMI(MBB, I, I.getDebugLoc(), TII.get(FCMPOpcode))
882 .addUse(I.getOperand(2).getReg())
883 .addUse(I.getOperand(3).getReg())
884 .addImm(MipsFCMPCondCode);
885 if (!constrainSelectedInstRegOperands(*FCMP, TII, TRI, RBI))
886 return false;
887
888 MachineInstr *Move = BuildMI(MBB, I, I.getDebugLoc(), TII.get(MoveOpcode))
889 .addDef(I.getOperand(0).getReg())
890 .addUse(Mips::ZERO)
891 .addUse(Mips::FCC0)
892 .addUse(TrueInReg);
893 if (!constrainSelectedInstRegOperands(*Move, TII, TRI, RBI))
894 return false;
895
896 I.eraseFromParent();
897 return true;
898 }
899 case G_FENCE: {
900 MI = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::SYNC)).addImm(0);
901 break;
902 }
903 case G_VASTART: {
905 int FI = FuncInfo->getVarArgsFrameIndex();
906
907 Register LeaReg = MRI.createVirtualRegister(&Mips::GPR32RegClass);
908 MachineInstr *LEA_ADDiu =
909 BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::LEA_ADDiu))
910 .addDef(LeaReg)
911 .addFrameIndex(FI)
912 .addImm(0);
913 if (!constrainSelectedInstRegOperands(*LEA_ADDiu, TII, TRI, RBI))
914 return false;
915
916 MachineInstr *Store = BuildMI(MBB, I, I.getDebugLoc(), TII.get(Mips::SW))
917 .addUse(LeaReg)
918 .addUse(I.getOperand(0).getReg())
919 .addImm(0);
920 if (!constrainSelectedInstRegOperands(*Store, TII, TRI, RBI))
921 return false;
922
923 I.eraseFromParent();
924 return true;
925 }
926 default:
927 return false;
928 }
929
930 I.eraseFromParent();
932}
933
934namespace llvm {
936 MipsSubtarget &Subtarget,
938 return new MipsInstructionSelector(TM, Subtarget, RBI);
939}
940} // end namespace llvm
unsigned const MachineRegisterInfo * MRI
static bool selectCopy(MachineInstr &I, const TargetInstrInfo &TII, MachineRegisterInfo &MRI, const TargetRegisterInfo &TRI, const RegisterBankInfo &RBI)
MachineBasicBlock & MBB
static bool isStore(int Opcode)
SmallVector< MachineOperand, 4 > Cond
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
#define LLVM_DEBUG(X)
Definition: Debug.h:101
uint64_t Addr
uint64_t Size
#define DEBUG_TYPE
const HexagonInstrInfo * TII
IRTranslator LLVM IR MI
#define I(x, y, z)
Definition: MD5.cpp:58
This file declares the MachineIRBuilder class.
unsigned const TargetRegisterInfo * TRI
typename CallsiteContextGraph< DerivedCCG, FuncTy, CallTy >::FuncInfo FuncInfo
static unsigned getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
#define GET_GLOBALISEL_PREDICATES_INIT
#define GET_GLOBALISEL_TEMPORARIES_INIT
This file declares the targeting of the RegisterBankInfo class for Mips.
const char LLVMTargetMachineRef TM
static StringRef getName(Value *V)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
Value * RHS
Value * LHS
BinaryOperator * Mul
static unsigned getSize(unsigned Kind)
APInt bitcastToAPInt() const
Definition: APFloat.h:1174
Class for arbitrary precision integers.
Definition: APInt.h:75
APInt getLoBits(unsigned numBits) const
Compute an APInt containing numBits lowbits from this APInt.
Definition: APInt.cpp:605
APInt getHiBits(unsigned numBits) const
Compute an APInt containing numBits highbits from this APInt.
Definition: APInt.cpp:600
APInt trunc(unsigned width) const
Truncate to new width.
Definition: APInt.cpp:898
bool isSignedIntN(unsigned N) const
Check if this APInt has an N-bits signed integer value.
Definition: APInt.h:427
int64_t getSExtValue() const
Get sign extended value.
Definition: APInt.h:1516
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition: InstrTypes.h:718
@ FCMP_OEQ
0 0 0 1 True if ordered and equal
Definition: InstrTypes.h:721
@ ICMP_SLT
signed less than
Definition: InstrTypes.h:747
@ ICMP_SLE
signed less or equal
Definition: InstrTypes.h:748
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition: InstrTypes.h:724
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition: InstrTypes.h:733
@ FCMP_OGT
0 0 1 0 True if ordered and greater than
Definition: InstrTypes.h:722
@ FCMP_OGE
0 0 1 1 True if ordered and greater than or equal
Definition: InstrTypes.h:723
@ ICMP_UGE
unsigned greater or equal
Definition: InstrTypes.h:742
@ ICMP_UGT
unsigned greater than
Definition: InstrTypes.h:741
@ ICMP_SGT
signed greater than
Definition: InstrTypes.h:745
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition: InstrTypes.h:732
@ FCMP_ONE
0 1 1 0 True if ordered and operands are unequal
Definition: InstrTypes.h:726
@ FCMP_UEQ
1 0 0 1 True if unordered or equal
Definition: InstrTypes.h:729
@ ICMP_ULT
unsigned less than
Definition: InstrTypes.h:743
@ FCMP_UGT
1 0 1 0 True if unordered or greater than
Definition: InstrTypes.h:730
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition: InstrTypes.h:725
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition: InstrTypes.h:727
@ ICMP_EQ
equal
Definition: InstrTypes.h:739
@ ICMP_NE
not equal
Definition: InstrTypes.h:740
@ ICMP_SGE
signed greater or equal
Definition: InstrTypes.h:746
@ FCMP_UNE
1 1 1 0 True if unordered or not equal
Definition: InstrTypes.h:734
@ ICMP_ULE
unsigned less or equal
Definition: InstrTypes.h:744
@ FCMP_UGE
1 0 1 1 True if unordered, greater than, or equal
Definition: InstrTypes.h:731
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition: InstrTypes.h:728
bool hasLocalLinkage() const
Definition: GlobalValue.h:523
Provides the logic to select generic machine instructions.
virtual bool select(MachineInstr &I)=0
Select the (possibly generic) instruction I to only use target-specific opcodes.
constexpr bool isScalar() const
constexpr bool isVector() const
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
constexpr bool isPointer() const
constexpr LLT getElementType() const
Returns the vector's element type. Only valid for vector types.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, uint64_t s, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
const LLVMTargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
Helper class to build MachineInstr.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
bool constrainAllUses(const TargetInstrInfo &TII, const TargetRegisterInfo &TRI, const RegisterBankInfo &RBI) const
const MachineInstrBuilder & addJumpTableIndex(unsigned Idx, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
Representation of each machine instruction.
Definition: MachineInstr.h:68
const MachineOperand & getOperand(unsigned i) const
Definition: MachineInstr.h:526
void addMemOperand(MachineFunction &MF, MachineMemOperand *MO)
Add a MachineMemOperand to the machine instruction.
unsigned getEntrySize(const DataLayout &TD) const
getEntrySize - Return the size of each entry in the jump table.
A description of a memory reference used in the backend.
uint64_t getSize() const
Return the size in bytes of the memory reference.
@ MOLoad
The memory access reads data.
Align getAlign() const
Return the minimum known alignment in bytes of the actual memory reference.
MachineOperand class - Representation of each machine instruction operand.
void setReg(Register Reg)
Change the register this operand corresponds to.
void setTargetFlags(unsigned F)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
MipsFunctionInfo - This class is derived from MachineFunction private Mips target-specific informatio...
Register getGlobalBaseRegForGlobalISel(MachineFunction &MF)
This class provides the information for the target register banks.
Wrapper class representing virtual and physical registers.
Definition: Register.h:19
bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Definition: Register.h:97
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1200
bool isPositionIndependent() const
Value * getOperand(unsigned i) const
Definition: User.h:169
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ MO_GOT_CALL
MO_GOT_CALL - Represents the offset into the global offset table at which the address of a call site ...
Definition: MipsBaseInfo.h:44
@ MO_GOT
MO_GOT - Represents the offset into the global offset table at which the address the relocation entry...
Definition: MipsBaseInfo.h:38
@ MO_ABS_HI
MO_ABS_HI/LO - Represents the hi or low part of an absolute symbol address.
Definition: MipsBaseInfo.h:52
bool hasImm(uint64_t TSFlags)
Definition: X86BaseInfo.h:994
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Offset
Definition: DWP.cpp:406
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool constrainSelectedInstRegOperands(MachineInstr &I, const TargetInstrInfo &TII, const TargetRegisterInfo &TRI, const RegisterBankInfo &RBI)
Mutate the newly-selected instruction I to constrain its (possibly generic) virtual register operands...
Definition: Utils.cpp:152
bool isPreISelGenericOpcode(unsigned Opcode)
Check whether the given Opcode is a generic opcode that is not supposed to appear after ISel.
Definition: TargetOpcodes.h:30
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition: MathExtras.h:382
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition: MathExtras.h:292
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
InstructionSelector * createMipsInstructionSelector(const MipsTargetMachine &, MipsSubtarget &, MipsRegisterBankInfo &)
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition: Alignment.h:39
This class contains a discriminated union of information about pointers in memory operands,...
static MachinePointerInfo getGOT(MachineFunction &MF)
Return a MachinePointerInfo record that refers to a GOT entry.