/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
H A D | TargetInstrInfo.h | 1446 int64_t &Offset2) const { in areLoadsFromSameBasePtr() 1459 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear() 1564 int64_t Offset2, bool OffsetIsScalable2, in shouldClusterMemOps()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/Mips/ |
H A D | MicroMipsSizeReduction.cpp | 400 int64_t Offset1, Offset2; in ConsecutiveInstr() local
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | ScheduleDAGSDNodes.cpp | 247 int64_t Offset1, Offset2; in ClusterNeighboringLoads() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | SeparateConstOffsetFromGEP.cpp | 1369 Value *Offset2 = Second->getOperand(1); in swapGEPOperand() local
|
H A D | ConstraintElimination.cpp | 667 int64_t Offset2 = BDec.Offset; in getConstraint() local
|
/freebsd/contrib/llvm-project/llvm/lib/IR/ |
H A D | Value.cpp | 1033 APInt Offset2(DL.getIndexTypeSizeInBits(Ptr2->getType()), 0); in getPointerOffsetFrom() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | AggressiveInstCombine.cpp | 682 APInt Offset2(DL.getIndexTypeSizeInBits(Load2Ptr->getType()), 0); in foldLoadsRecursive() local
|
/freebsd/contrib/llvm-project/clang/lib/StaticAnalyzer/Checkers/ |
H A D | ContainerModeling.cpp | 970 SymbolRef Offset2, in invalidateIteratorPositions()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
H A D | SystemZInstrInfo.cpp | 1898 int64_t Offset2 = (MCID.TSFlags & SystemZII::Is128Bit ? Offset + 8 : Offset); in getOpcodeForOffset() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64InstrInfo.cpp | 4284 int64_t Offset2, unsigned Opcode2) { in shouldClusterFI() 4355 int64_t Offset2 = SecondLdSt.getOperand(2).getImm(); in shouldClusterMemOps() local
|
H A D | AArch64ISelLowering.cpp | 22610 TypeSize Offset2 = TypeSize::getFixed(2); in combineV3I8LoadExt() local 22894 TypeSize Offset2 = TypeSize::getFixed(2); in combineI8TruncStore() local
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | MachinePipeliner.cpp | 873 int64_t Offset1, Offset2; in addLoopCarriedDependences() local
|
H A D | CodeGenPrepare.cpp | 2409 uint64_t Offset2 = Offset.getLimitedValue(); in optimizeCallInst() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/ |
H A D | RISCVInstrInfo.cpp | 2684 shouldClusterMemOps(ArrayRef<const MachineOperand * > BaseOps1,int64_t Offset1,bool OffsetIsScalable1,ArrayRef<const MachineOperand * > BaseOps2,int64_t Offset2,bool OffsetIsScalable2,unsigned ClusterSize,unsigned NumBytes) const shouldClusterMemOps() argument
|
H A D | RISCVISelLowering.cpp | 19544 unsigned Offset2 = State.AllocateStack(2, Align(2)); CC_RISCV_FastCC() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCInstrInfo.cpp | 2930 int64_t Offset1 = 0, Offset2 = 0; in shouldClusterMemOps() local
|
H A D | PPCISelLowering.cpp | 13835 int64_t Offset1 = 0, Offset2 = 0; in isConsecutiveLSLoc() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineAndOrXor.cpp | 1281 const APInt *Offset1 = nullptr, *Offset2 = nullptr; in foldAndOrOfICmpsUsingRanges() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | ARMBaseInstrInfo.cpp | 2016 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86InstrInfo.cpp | 8778 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | SIInstrInfo.cpp | 552 int64_t Offset2, bool OffsetIsScalable2, in shouldClusterMemOps()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
H A D | CombinerHelper.cpp | 7039 std::optional<APInt> Offset2; in tryFoldAndOrOrICmpsUsingRanges() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
H A D | SLPVectorizer.cpp | 5979 int Offset2 = Pair2.second; in canFormVector() local
|