| /src/contrib/llvm-project/llvm/lib/CodeGen/ |
| H A D | TypePromotion.cpp | 264 if (auto *ZExt = dyn_cast<ZExtInst>(V)) in isSink() 578 if (auto ZExt = dyn_cast<ZExtInst>(I)) in TruncateSinks() 598 if (!isa<ZExtInst>(V)) in Cleanup() 601 auto ZExt = cast<ZExtInst>(V); in Cleanup() 972 if (isa<ZExtInst>(&I) && isa<PHINode>(I.getOperand(0)) && in run()
|
| H A D | HardwareLoops.cpp | 429 isa<ZExtInst>(Count) ? cast<ZExtInst>(Count)->getOperand(0) : nullptr; in CanGenerateTest()
|
| /src/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineShifts.cpp | 962 return new ZExtInst(Overflow, Ty); in foldLShrOverflowBit() 1050 return new ZExtInst(Builder.CreateShl(X, ShAmtC), Ty); in visitShl() 1276 return new ZExtInst(Builder.CreateIsNotNeg(X, "isnotneg"), Ty); in visitLShr() 1356 return new ZExtInst(Cmp, Ty); in visitLShr() 1420 return new ZExtInst(NewLShr, Ty); in visitLShr() 1438 return new ZExtInst(NewLShr, Ty); in visitLShr() 1446 return new ZExtInst(AShr, Ty); in visitLShr() 1454 return new ZExtInst(Builder.CreateIsNotNull(X), Ty); in visitLShr() 1458 return new ZExtInst(Builder.CreateICmpSLT(X, Y), Ty); in visitLShr() 1552 return new ZExtInst(NewShift, Ty); in visitLShr() [all …]
|
| H A D | InstCombineAndOrXor.cpp | 1727 return new ZExtInst(NewOp, DestTy); in foldLogicCastConstant() 1769 auto *ICmpR = cast<ZExtInst>(Op1)->getOperand(0); in foldCastedBitwiseLogic() 1772 return new ZExtInst(BitwiseOp, Op0->getType()); in foldCastedBitwiseLogic() 1946 return new ZExtInst(Builder.CreateAnd(NewBO, X), Ty); in narrowMaskedBinOp() 2354 return new ZExtInst(IsZero, Ty); in visitAnd() 2458 Value *BinOp = isa<ZExtInst>(BO->getOperand(0)) in visitAnd() 2463 return new ZExtInst(And, Ty); in visitAnd() 2476 return new ZExtInst(NewBO, Ty); in visitAnd() 2486 return new ZExtInst(NewBO, Ty); in visitAnd() 2960 } else if (isa<ZExtInst>(Or0) || isa<ZExtInst>(Or1)) { in convertOrOfShiftsToFunnelShift() [all …]
|
| H A D | InstCombineAddSub.cpp | 827 return new ZExtInst(X, Ty); in foldNoWrapAdd() 830 return new ZExtInst( in foldNoWrapAdd() 906 return new ZExtInst(Builder.CreateIsNotNeg(X, "isnotneg"), Ty); in foldAddWithConstant() 985 return new ZExtInst(Builder.CreateNot(X), Ty); in foldAddWithConstant() 1003 return new ZExtInst(X, Ty); in foldAddWithConstant() 1618 return new ZExtInst(B, LHS->getType()); in visitAdd() 1720 Value *Add = isa<ZExtInst>(Ext) ? InstCombiner::AddOne(C) in visitAdd()
|
| H A D | InstCombineInternal.h | 137 Instruction *visitZExt(ZExtInst &Zext); 308 Instruction *transformZExtICmp(ICmpInst *Cmp, ZExtInst &Zext);
|
| H A D | InstCombineMulDivRem.cpp | 1292 return new ZExtInst(Builder.CreateICmpEQ(Op1, Op0), Ty); in commonIDivTransforms() 1484 return new ZExtInst(NarrowOp, Ty); in narrowUDivURem() 1497 return new ZExtInst(IC.Builder.CreateBinOp(Opcode, X, TruncC), Ty); in narrowUDivURem() 1508 return new ZExtInst(IC.Builder.CreateBinOp(Opcode, TruncC, X), Ty); in narrowUDivURem() 1606 return new ZExtInst(Builder.CreateICmpEQ(Op0, Op1), Ty); in visitSDiv()
|
| H A D | InstCombineCasts.cpp | 925 ZExtInst &Zext) { in transformZExtICmp() 1138 Instruction *InstCombinerImpl::visitZExt(ZExtInst &Zext) { in visitZExt() 1213 return new ZExtInst(And, DestTy); in visitZExt() 1945 return new ZExtInst(X, DestType); in foldItoFPtoI() 2693 if (isa<TruncInst>(Src) || isa<ZExtInst>(Src)) { in visitBitCast()
|
| /src/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | LowerExpectIntrinsic.cpp | 147 if (ZExtInst *ZExt = dyn_cast<ZExtInst>(V)) { in handlePhiDef()
|
| H A D | LoopFlatten.cpp | 241 (isa<SExtInst>(MatchedItCount) || isa<ZExtInst>(MatchedItCount))) { in matchLinearIVUser() 246 : dyn_cast<ZExtInst>(MatchedItCount)->getOperand(0); in matchLinearIVUser() 266 (isa<SExtInst>(InnerTripCount) || isa<ZExtInst>(InnerTripCount))) in checkInnerInductionPhiUsers() 378 if ((!isa<ZExtInst>(TripCountInst) && !isa<SExtInst>(TripCountInst)) || in verifyTripCount()
|
| H A D | NaryReassociate.cpp | 366 } else if (ZExtInst *ZExt = dyn_cast<ZExtInst>(IndexToSplit)) { in tryReassociateGEPAtIndex()
|
| H A D | SeparateConstOffsetFromGEP.cpp | 641 } else if (isa<ZExtInst>(V)) { in find() 704 (isa<SExtInst>(Cast) || isa<ZExtInst>(Cast) || isa<TruncInst>(Cast)) && in distributeExtsAndCloneChain()
|
| H A D | CorrelatedValuePropagation.cpp | 1124 static bool processZExt(ZExtInst *ZExt, LazyValueInfo *LVI) { in processZExt() 1248 BBChanged |= processZExt(cast<ZExtInst>(&II), LVI); in runImpl()
|
| /src/contrib/llvm-project/llvm/lib/Target/BPF/ |
| H A D | BPFCheckAndAdjustIR.cpp | 180 ZExtInst *ZExt; 196 if (auto *ZExt = dyn_cast<ZExtInst>(V)) { in sinkMinMaxInBB()
|
| /src/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
| H A D | TruncInstCombine.cpp | 279 bool IsExtInst = (isa<ZExtInst>(I) || isa<SExtInst>(I)); in getBestTruncatedType() 520 assert((isa<SExtInst>(I.first) || isa<ZExtInst>(I.first)) && in ReduceExpressionGraph()
|
| /src/contrib/llvm-project/llvm/lib/ExecutionEngine/Interpreter/ |
| H A D | Interpreter.h | 138 void visitZExtInst(ZExtInst &I);
|
| /src/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPULowerKernelAttributes.cpp | 278 auto *ZextGroupSize = dyn_cast<ZExtInst>(U); in processUse()
|
| /src/contrib/llvm-project/llvm/include/llvm/IR/ |
| H A D | InstVisitor.h | 177 RetTy visitZExtInst(ZExtInst &I) { DELEGATE(CastInst);} in visitZExtInst()
|
| H A D | Instruction.def | 185 HANDLE_CAST_INST(39, ZExt , ZExtInst ) // Zero extend integers
|
| H A D | PatternMatch.h | 1937 if (auto *I = dyn_cast<ZExtInst>(V)) in match() 2045 inline CastInst_match<OpTy, ZExtInst> m_ZExt(const OpTy &Op) { in m_ZExt() 2046 return CastInst_match<OpTy, ZExtInst>(Op); in m_ZExt() 2055 inline match_combine_or<CastInst_match<OpTy, ZExtInst>, OpTy> 2074 inline match_combine_or<CastInst_match<OpTy, ZExtInst>, 2081 inline match_combine_or<match_combine_or<CastInst_match<OpTy, ZExtInst>,
|
| /src/contrib/llvm-project/llvm/lib/Target/ARM/ |
| H A D | MVEGatherScatterLowering.cpp | 282 ZExtInst *ZextOffs = dyn_cast<ZExtInst>(Offsets); in decomposeGEP() 520 } else if (isa<ZExtInst>(User) && in tryCreateMaskedGatherOffset()
|
| /src/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64FastISel.cpp | 301 assert((isa<ZExtInst>(I) || isa<SExtInst>(I)) && in isIntExtFree() 305 bool IsZExt = isa<ZExtInst>(I); in isIntExtFree() 754 if (const auto *ZE = dyn_cast<ZExtInst>(I)) { in computeAddress() 836 if (const auto *ZE = dyn_cast<ZExtInst>(I)) { in computeAddress() 894 if (const auto *ZE = dyn_cast<ZExtInst>(U)) { in computeAddress() 1978 if (const auto *ZE = dyn_cast<ZExtInst>(I->use_begin()->getUser())) { in selectLoad() 4530 bool IsZExt = isa<ZExtInst>(I); in optimizeIntExtLoad() 4568 assert((isa<ZExtInst>(I) || isa<SExtInst>(I)) && in selectIntExt() 4587 bool IsZExt = isa<ZExtInst>(I); in selectIntExt() 4674 if (const auto *ZExt = dyn_cast<ZExtInst>(Src0)) { in selectMul() [all …]
|
| /src/contrib/llvm-project/llvm/lib/Target/X86/ |
| H A D | X86PartialReduction.cpp | 248 if (auto *ZExt = dyn_cast<ZExtInst>(Op)) in trySADReplacement()
|
| /src/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
| H A D | SystemZTargetTransformInfo.cpp | 1086 else if (isa<ZExtInst>(UserI)) in isFoldableLoad() 1187 isa<ZExtInst>(OtherOp))) in getMemoryOpCost()
|
| /src/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | VectorUtils.cpp | 619 if (TTI && (isa<ZExtInst>(&I) || isa<SExtInst>(&I)) && in computeMinimumValueSizes() 664 if (isa<SExtInst>(I) || isa<ZExtInst>(I) || isa<LoadInst>(I) || in computeMinimumValueSizes()
|