| /src/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUImageIntrinsicOptimizer.cpp | 210 unsigned NumLoads = IIList.size(); in optimizeSection() local 212 unsigned NumVAddrLoads = 3 * NumLoads; in optimizeSection() 213 unsigned NumVDataLoads = divideCeil(NumElts, isD16 ? 2 : 1) * NumLoads; in optimizeSection() 217 if (NumLoads < NumMsaas || in optimizeSection()
|
| H A D | AMDGPURegisterBankInfo.cpp | 1350 int NumLoads = 1; in applyMappingSBufferLoad() local 1352 NumLoads = LoadSize / 128; in applyMappingSBufferLoad() 1353 Ty = Ty.divide(NumLoads); in applyMappingSBufferLoad() 1358 const Align Alignment = NumLoads > 1 ? Align(16 * NumLoads) : Align(1); in applyMappingSBufferLoad() 1388 SmallVector<Register, 4> LoadParts(NumLoads); in applyMappingSBufferLoad() 1393 for (int i = 0; i < NumLoads; ++i) { in applyMappingSBufferLoad() 1394 if (NumLoads == 1) { in applyMappingSBufferLoad() 1433 if (NumLoads != 1) { in applyMappingSBufferLoad()
|
| H A D | SIInstrInfo.h | 254 int64_t Offset1, unsigned NumLoads) const override;
|
| H A D | SIISelLowering.cpp | 8329 unsigned NumLoads = 1; in lowerSBuffer() local 8336 NumLoads = NumElts / 4; in lowerSBuffer() 8345 NumLoads > 1 ? Align(16 * NumLoads) : Align(4)); in lowerSBuffer() 8348 for (unsigned i = 0; i < NumLoads; ++i) { in lowerSBuffer()
|
| H A D | SIInstrInfo.cpp | 596 unsigned NumLoads) const { in shouldScheduleLoadsNear() 603 return (NumLoads <= 16 && (Offset1 - Offset0) < 64); in shouldScheduleLoadsNear()
|
| /src/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
| H A D | HexagonTargetTransformInfo.cpp | 196 unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth; in getMemoryOpCost() local 197 return 3 * NumLoads; in getMemoryOpCost() 208 unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth; in getMemoryOpCost() local 210 return Cost * NumLoads; in getMemoryOpCost() 214 return (3 - LogA) * Cost * NumLoads; in getMemoryOpCost()
|
| /src/contrib/llvm-project/llvm/lib/Target/NVPTX/ |
| H A D | NVPTXLowerAggrCopies.cpp | 108 unsigned NumLoads = DL.getTypeStoreSize(LI->getType()); in runOnFunction() local 110 ConstantInt::get(Type::getInt32Ty(Context), NumLoads); in runOnFunction()
|
| /src/contrib/llvm-project/llvm/lib/Target/X86/ |
| H A D | X86InterleavedAccess.cpp | 197 unsigned int NumLoads = NumSubVectors; in decompose() local 205 NumLoads = NumSubVectors * (VecLength / 384); in decompose() 216 for (unsigned i = 0; i < NumLoads; i++) { in decompose()
|
| H A D | X86FastPreTileConfig.cpp | 40 STATISTIC(NumLoads, "Number of loads added"); 266 ++NumLoads; in reload()
|
| H A D | X86InstrInfo.h | 496 unsigned NumLoads) const override;
|
| H A D | X86InstrInfo.cpp | 8779 unsigned NumLoads) const { in shouldScheduleLoadsNear() 8806 if (NumLoads >= 3) in shouldScheduleLoadsNear() 8808 } else if (NumLoads) { in shouldScheduleLoadsNear() 8818 if (NumLoads) in shouldScheduleLoadsNear()
|
| /src/contrib/llvm-project/llvm/lib/CodeGen/ |
| H A D | ExpandMemCmp.cpp | 418 const unsigned NumLoads = in getCompareLoadPairs() local 432 NumLoads == 1 ? nullptr in getCompareLoadPairs() 435 for (unsigned i = 0; i < NumLoads; ++i, ++LoadIndex) { in getCompareLoadPairs() 441 if (NumLoads != 1) { in getCompareLoadPairs()
|
| H A D | RegAllocFast.cpp | 54 STATISTIC(NumLoads, "Number of loads added"); 632 ++NumLoads; in reload()
|
| /src/contrib/llvm-project/llvm/include/llvm/Analysis/ |
| H A D | LoopAccessAnalysis.h | 662 unsigned getNumLoads() const { return NumLoads;} in getNumLoads() 753 unsigned NumLoads = 0; variable
|
| /src/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | ScheduleDAGSDNodes.cpp | 274 unsigned NumLoads = 0; in ClusterNeighboringLoads() local 281 if (!TII->shouldScheduleLoadsNear(BaseLoad, Load, BaseOff, Offset,NumLoads)) in ClusterNeighboringLoads() 284 ++NumLoads; in ClusterNeighboringLoads() 287 if (NumLoads == 0) in ClusterNeighboringLoads()
|
| /src/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | LowerMatrixIntrinsics.cpp | 336 unsigned NumLoads = 0; member 346 NumLoads += RHS.NumLoads; in operator +=() 440 OpInfo.NumLoads += N; in addNumLoads() 444 void setNumLoads(unsigned N) { OpInfo.NumLoads = N; } in setNumLoads() 462 unsigned getNumLoads() const { return OpInfo.NumLoads; } in getNumLoads() 2586 << ore::NV("NumLoads", Counts.NumLoads) << " loads, " in emitRemarks() 2592 if (SharedCounts.NumStores > 0 || SharedCounts.NumLoads > 0 || in emitRemarks() 2596 << ore::NV("NumLoads", SharedCounts.NumLoads) << " loads, " in emitRemarks()
|
| /src/contrib/llvm-project/llvm/lib/Target/RISCV/ |
| H A D | RISCVTargetTransformInfo.cpp | 734 unsigned NumLoads = getEstimatedVLFor(&VTy); in getGatherScatterOpCost() local 735 return NumLoads * MemOpCost; in getGatherScatterOpCost() 757 unsigned NumLoads = getEstimatedVLFor(&VTy); in getStridedMemoryOpCost() local 758 return NumLoads * MemOpCost; in getStridedMemoryOpCost()
|
| /src/contrib/llvm-project/llvm/lib/Target/ARM/ |
| H A D | ARMBaseInstrInfo.h | 265 unsigned NumLoads) const override;
|
| H A D | ARMBaseInstrInfo.cpp | 2017 unsigned NumLoads) const { in shouldScheduleLoadsNear() 2039 if (NumLoads >= 3) in shouldScheduleLoadsNear()
|
| H A D | ARMISelLowering.cpp | 21697 unsigned NumLoads = getNumInterleavedAccesses(VecTy, DL); in lowerInterleavedLoad() local 21709 if (NumLoads > 1) { in lowerInterleavedLoad() 21713 VecTy->getNumElements() / NumLoads); in lowerInterleavedLoad() 21754 for (unsigned LoadCount = 0; LoadCount < NumLoads; ++LoadCount) { in lowerInterleavedLoad()
|
| /src/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
| H A D | TargetInstrInfo.h | 1460 unsigned NumLoads) const { in shouldScheduleLoadsNear() argument
|
| /src/contrib/llvm-project/llvm/utils/TableGen/Common/ |
| H A D | CodeGenDAGPatterns.cpp | 4158 unsigned NumLoads = 0; in VerifyInstructionFlags() local 4163 NumLoads += InstInfo.mayLoad; in VerifyInstructionFlags() 4184 if (!PatInfo.mayStore && PatInfo.mayLoad && !NumLoads) in VerifyInstructionFlags()
|
| /src/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | LoopAccessAnalysis.cpp | 2475 NumLoads++; in analyzeLoop()
|
| /src/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
| H A D | LoopVectorize.cpp | 5033 unsigned NumLoads = Legal->getNumLoads(); in selectInterleaveCount() local 5035 unsigned LoadsIC = IC / (NumLoads ? NumLoads : 1); in selectInterleaveCount()
|
| /src/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64ISelLowering.cpp | 16753 unsigned NumLoads = getNumInterleavedAccesses(VTy, DL, UseScalable); in lowerInterleavedLoad() local 16767 FVTy->getNumElements() / NumLoads); in lowerInterleavedLoad() 16804 for (unsigned LoadCount = 0; LoadCount < NumLoads; ++LoadCount) { in lowerInterleavedLoad() 17077 unsigned NumLoads = getNumInterleavedAccesses(VTy, DL, UseScalable); in lowerDeinterleaveIntrinsicToLoad() local 17081 VTy->getElementCount().divideCoefficientBy(NumLoads)); in lowerDeinterleaveIntrinsicToLoad() 17096 if (NumLoads > 1) { in lowerDeinterleaveIntrinsicToLoad() 17100 for (unsigned I = 0; I < NumLoads; ++I) { in lowerDeinterleaveIntrinsicToLoad()
|