Searched refs:OffsetImm (Results 1 – 6 of 6) sorted by relevance
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/AsmParser/ |
H A D | ARMAsmParser.cpp | 889 const MCExpr *OffsetImm; // Offset immediate value member 1147 if(!Memory.OffsetImm || Memory.OffsetRegNum) return false; in isThumbMemPC() 1149 if (const auto *CE = dyn_cast<MCConstantExpr>(Memory.OffsetImm)) in isThumbMemPC() 1486 return Memory.OffsetRegNum == 0 && Memory.OffsetImm == nullptr && in isMemNoOffset() 1498 return Memory.OffsetRegNum == 0 && Memory.OffsetImm == nullptr && in isMemNoOffsetT2() 1510 return Memory.OffsetRegNum == 0 && Memory.OffsetImm == nullptr && in isMemNoOffsetT2NoSp() 1522 return Memory.OffsetRegNum == 0 && Memory.OffsetImm == nullptr && in isMemNoOffsetT() 1532 if (!Memory.OffsetImm) return true; in isMemPCRelImm12() 1533 if (const auto *CE = dyn_cast<MCConstantExpr>(Memory.OffsetImm)) { in isMemPCRelImm12() 1620 if (!Memory.OffsetImm) return true; in isAddrMode2() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64LoadStoreOptimizer.cpp | 777 int64_t OffsetImm; in mergeNarrowZeroStores() local 779 OffsetImm = MIOffsetInBytes; in mergeNarrowZeroStores() 781 OffsetImm = IOffsetInBytes; in mergeNarrowZeroStores() 789 assert(((OffsetImm % NewOffsetStride) == 0) && in mergeNarrowZeroStores() 791 OffsetImm = OffsetImm / NewOffsetStride; in mergeNarrowZeroStores() 801 .addImm(OffsetImm) in mergeNarrowZeroStores() 1024 int OffsetImm = AArch64InstrInfo::getLdStOffsetOp(*RtMI).getImm(); in mergePairedInsns() local 1027 assert(!(OffsetImm % TII->getMemScale(*RtMI)) && in mergePairedInsns() 1029 OffsetImm /= TII->getMemScale(*RtMI); in mergePairedInsns() 1076 .addImm(OffsetImm) in mergePairedInsns()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | Thumb2SizeReduction.cpp | 578 unsigned OffsetImm = 0; in ReduceLoadStore() local 580 OffsetImm = MI->getOperand(2).getImm(); in ReduceLoadStore() 583 if ((OffsetImm & (Scale - 1)) || OffsetImm > MaxOffset) in ReduceLoadStore() 602 MIB.addImm(OffsetImm / Scale); in ReduceLoadStore()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCInstrInfo.cpp | 3536 int64_t OffsetImm = 0; in foldFrameOffset() local 3541 if (!isImmInstrEligibleForFolding(MI, ToBeDeletedReg, XFormOpcode, OffsetImm, in foldFrameOffset() 3564 if (isValidToBeChangedReg(ADDMI, 1, ADDIMI, OffsetAddi, OffsetImm)) in foldFrameOffset() 3566 else if (isValidToBeChangedReg(ADDMI, 2, ADDIMI, OffsetAddi, OffsetImm)) in foldFrameOffset() 3603 ADDIMI->getOperand(2).setImm(OffsetAddi + OffsetImm); in foldFrameOffset() 3650 int64_t &OffsetImm, in isImmInstrEligibleForFolding() argument 3685 OffsetImm = ImmOperand.getImm(); in isImmInstrEligibleForFolding() 3693 int64_t OffsetImm) const { in isValidToBeChangedReg() 3727 if (isInt<16>(OffsetAddi + OffsetImm)) in isValidToBeChangedReg()
|
H A D | PPCInstrInfo.h | 641 int64_t OffsetImm) const;
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | SIInstrInfo.cpp | 438 const MachineOperand *OffsetImm = in getMemOperandsWithOffsetWidth() local 440 Offset = OffsetImm->getImm(); in getMemOperandsWithOffsetWidth()
|