| /external/llvm/lib/Analysis/ |
| ConstantFolding.cpp | 232 unsigned BitWidth = TD.getPointerTypeSizeInBits(GV->getType()); 233 Offset = APInt(BitWidth, 0); 251 unsigned BitWidth = TD.getPointerTypeSizeInBits(GEP->getType()); 252 APInt TmpOffset(BitWidth, 0); 626 unsigned BitWidth = DL->getTypeSizeInBits(Op0->getType()->getScalarType()); 627 APInt KnownZero0(BitWidth, 0), KnownOne0(BitWidth, 0); 628 APInt KnownZero1(BitWidth, 0), KnownOne1(BitWidth, 0); 659 // PtrToInt may change the bitwidth so we have convert to the right siz [all...] |
| InstructionSimplify.cpp | [all...] |
| ValueTracking.cpp | 39 /// getBitWidth - Returns the bitwidth of the given scalar or pointer type (if 40 /// unknown returns 0). For vector types, returns the element type's bitwidth. 42 if (unsigned BitWidth = Ty->getScalarSizeInBits()) 43 return BitWidth; 58 unsigned BitWidth = KnownZero.getBitWidth(); 60 // NLZ can't be BitWidth with no sign bit 61 APInt MaskV = APInt::getHighBitsSet(BitWidth, NLZ+1); 70 KnownZero = APInt::getHighBitsSet(BitWidth, NLZ2); 76 unsigned BitWidth = KnownZero.getBitWidth(); 82 APInt LHSKnownZero(BitWidth, 0), LHSKnownOne(BitWidth, 0) [all...] |
| ScalarEvolution.cpp | 693 // Suppose, W is the bitwidth of the return value. We must be prepared for [all...] |
| DependenceAnalysis.cpp | [all...] |
| /external/llvm/lib/ExecutionEngine/ |
| ExecutionEngine.cpp | 639 uint32_t BitWidth = cast<IntegerType>(CE->getType())->getBitWidth(); 640 GV.IntVal = GV.IntVal.trunc(BitWidth); 645 uint32_t BitWidth = cast<IntegerType>(CE->getType())->getBitWidth(); 646 GV.IntVal = GV.IntVal.zext(BitWidth); 651 uint32_t BitWidth = cast<IntegerType>(CE->getType())->getBitWidth(); 652 GV.IntVal = GV.IntVal.sext(BitWidth); 700 uint32_t BitWidth = cast<IntegerType>(CE->getType())->getBitWidth(); 702 GV.IntVal = APIntOps::RoundFloatToAPInt(GV.FloatVal, BitWidth); 704 GV.IntVal = APIntOps::RoundDoubleToAPInt(GV.DoubleVal, BitWidth); 709 (void)apf.convertToInteger(&v, BitWidth, [all...] |
| /external/llvm/lib/IR/ |
| ConstantFold.cpp | 664 uint32_t BitWidth = cast<IntegerType>(DestTy)->getBitWidth(); 666 CI->getValue().zext(BitWidth)); 671 uint32_t BitWidth = cast<IntegerType>(DestTy)->getBitWidth(); 673 CI->getValue().sext(BitWidth)); [all...] |
| Instructions.cpp | [all...] |
| /external/llvm/lib/Target/CppBackend/ |
| CPPBackend.cpp | 378 unsigned BitWidth = cast<IntegerType>(Ty)->getBitWidth(); 379 return "IntegerType::get(mod->getContext(), " + utostr(BitWidth) + ")"; [all...] |
| /external/llvm/lib/Transforms/IPO/ |
| MergeFunctions.cpp | 158 /// If both types are vectors, then vector with greater bitwidth is 160 /// If both types are vectors with the same bitwidth, then types 185 /// Ints with greater bitwidth are greater. Ints with same bitwidths 232 /// 2.1. All vectors with equal bitwidth to vector A, has equal bitwidth to [all...] |
| /external/clang/lib/CodeGen/ |
| CGExprConstant.cpp | 275 unsigned BitWidth = FieldValue.getBitWidth(); 277 FieldValue = FieldValue.zext(CharWidth) << (CharWidth - BitWidth); [all...] |
| /external/llvm/lib/CodeGen/AsmPrinter/ |
| AsmPrinter.cpp | [all...] |
| /external/llvm/lib/CodeGen/SelectionDAG/ |
| TargetLowering.cpp | 322 unsigned BitWidth, 342 unsigned DemandedSize = BitWidth - Demanded.countLeadingZeros(); 346 for (; SmallVTBits < BitWidth; SmallVTBits = NextPowerOf2(SmallVTBits)) { 378 unsigned BitWidth = DemandedMask.getBitWidth(); 379 assert(Op.getValueType().getScalarType().getSizeInBits() == BitWidth && 385 KnownZero = KnownOne = APInt(BitWidth, 0); 397 NewMask = APInt::getAllOnesValue(BitWidth); 454 if (TLO.ShrinkDemandedOp(Op, BitWidth, NewMask, dl)) 488 if (TLO.ShrinkDemandedOp(Op, BitWidth, NewMask, dl)) 513 if (TLO.ShrinkDemandedOp(Op, BitWidth, NewMask, dl) [all...] |
| /external/llvm/lib/Target/AArch64/ |
| AArch64ISelDAGToDAG.cpp | [all...] |
| /external/llvm/lib/Target/R600/ |
| SIInstrInfo.cpp | [all...] |
| /external/llvm/lib/Target/XCore/ |
| XCoreISelLowering.cpp | [all...] |
| /external/llvm/lib/Transforms/InstCombine/ |
| InstCombineAddSub.cpp | 880 int BitWidth = Op0KnownZero.getBitWidth(); 882 Op0KnownZeroTemp.clearBit(BitWidth - 1); 883 int Op0ZeroPosition = BitWidth - Op0KnownZeroTemp.countLeadingZeros() - 1; 885 int Op1OnePosition = BitWidth - Op1MaybeOne.countLeadingZeros() - 1; 920 int BitWidth = IT->getBitWidth(); 921 APInt LHSKnownZero(BitWidth, 0); 922 APInt LHSKnownOne(BitWidth, 0); 925 APInt RHSKnownZero(BitWidth, 0); 926 APInt RHSKnownOne(BitWidth, 0); 931 if ((LHSKnownOne[BitWidth - 1] && RHSKnownZero[BitWidth - 1]) | [all...] |
| InstCombineAndOrXor.cpp | 201 uint32_t BitWidth = AndRHS->getType()->getBitWidth(); 202 uint32_t OpRHSVal = OpRHS->getLimitedValue(BitWidth); 203 APInt ShlMask(APInt::getHighBitsSet(BitWidth, BitWidth-OpRHSVal)); 221 uint32_t BitWidth = AndRHS->getType()->getBitWidth(); 222 uint32_t OpRHSVal = OpRHS->getLimitedValue(BitWidth); 223 APInt ShrMask(APInt::getLowBitsSet(BitWidth, BitWidth - OpRHSVal)); 241 uint32_t BitWidth = AndRHS->getType()->getBitWidth(); 242 uint32_t OpRHSVal = OpRHS->getLimitedValue(BitWidth); [all...] |
| InstCombineCalls.cpp | 322 uint32_t BitWidth = IT->getBitWidth(); 323 APInt KnownZero(BitWidth, 0); 324 APInt KnownOne(BitWidth, 0); 327 APInt Mask(APInt::getLowBitsSet(BitWidth, TrailingZeros)); 330 APInt(BitWidth, TrailingZeros))); 340 uint32_t BitWidth = IT->getBitWidth(); 341 APInt KnownZero(BitWidth, 0); 342 APInt KnownOne(BitWidth, 0); 345 APInt Mask(APInt::getHighBitsSet(BitWidth, LeadingZeros)); 348 APInt(BitWidth, LeadingZeros))) [all...] |
| InstCombineCasts.cpp | 374 uint32_t BitWidth = Ty->getScalarSizeInBits(); 375 if (BitWidth < OrigBitWidth) { 376 APInt Mask = APInt::getHighBitsSet(OrigBitWidth, OrigBitWidth-BitWidth); 389 uint32_t BitWidth = Ty->getScalarSizeInBits(); 390 if (CI->getLimitedValue(BitWidth) < BitWidth) 400 uint32_t BitWidth = Ty->getScalarSizeInBits(); 402 APInt::getHighBitsSet(OrigBitWidth, OrigBitWidth-BitWidth)) && 403 CI->getLimitedValue(BitWidth) < BitWidth) { [all...] |
| InstructionCombining.cpp | [all...] |
| /external/llvm/lib/Transforms/Scalar/ |
| ScalarReplAggregates.cpp | 360 unsigned BitWidth = AllocaSize * 8; 364 if (BitWidth > ScalarLoadThreshold) 368 !HadNonMemTransferAccess && !DL.fitsInLegalInteger(BitWidth)) 378 NewTy = IntegerType::get(AI->getContext(), BitWidth); [all...] |
| /external/clang/lib/AST/ |
| ASTImporter.cpp | [all...] |
| Decl.cpp | [all...] |
| /external/clang/lib/Sema/ |
| SemaTemplateInstantiateDecl.cpp | 464 Expr *BitWidth = D->getBitWidth(); 466 BitWidth = nullptr; 467 else if (BitWidth) { 473 = SemaRef.SubstExpr(BitWidth, TemplateArgs); 476 BitWidth = nullptr; 478 BitWidth = InstantiatedBitWidth.getAs<Expr>(); 486 BitWidth, [all...] |