/external/llvm/lib/Target/ARM/ |
ARMConstantIslandPass.cpp | 86 /// BasicBlockInfo - Information about the offset and size of a single 100 /// Size - Size of the basic block in bytes. If the block contains 103 /// The size does not include any alignment padding whether from the 105 unsigned Size; 112 /// of unknown size. The real size may be smaller than Size bytes by a 121 BasicBlockInfo() : Offset(0), Size(0), KnownBits(0), Unalign(0), 129 // If the block size isn't a multiple of the known bits, assume th [all...] |
ARMISelDAGToDAG.cpp | [all...] |
ARMLoadStoreOptimizer.cpp | 125 int Opcode, unsigned Size, 421 unsigned NumRegs = Regs.size(); 546 for (unsigned i = 0, e = ImpDefs.size(); i != e; ++i) 572 for (unsigned i = 0; i < MemOps.size(); ++i) { 625 for (unsigned i = 0, e = memOps.size(); i != e; ++i) { 709 unsigned Base, int Opcode, unsigned Size, 743 for (unsigned i = SIndex+1, e = MemOps.size(); i != e; ++i) { 752 NewOffset == Offset + (int)Size && 758 Offset += Size; 765 MergeLDR_STR(MBB, i, Base, Opcode, Size, Pred, PredReg, Scratch [all...] |
ARMBaseInstrInfo.cpp | 409 assert((Cond.size() == 2 || Cond.size() == 0) && 482 if (Pred1.size() > 2 || Pred2.size() > 2) 564 assert(JTI < JT.size()); 565 return JT[JTI].MBBs.size(); 568 /// GetInstSize - Return the size of the specified MachineInstr. 598 // If this machine instr is a constant pool entry, its size is recorded as 620 // jumptable. The size is 4 + 4 * number of entries. For TBB, each 631 assert(JTI < JT.size()); [all...] |
ARMFastISel.cpp | 274 // Do we optionally set a predicate? Preds is size > 0 iff the predicate [all...] |
/external/llvm/lib/Target/R600/ |
SIInstrInfo.cpp | [all...] |
AMDGPUISelLowering.cpp | 93 assert(StoreSize % 32 == 0 && "Store size not a multiple of 32"); 689 uint64_t Size = TD->getTypeAllocSize(GV->getType()->getElementType()); 693 MFI->LDSSize += Size; 703 unsigned Size = TD->getTypeAllocSize(EltType); 709 int FI = FrameInfo->CreateStackObject(Size, Alignment, false); [all...] |
/external/llvm/lib/Target/X86/AsmParser/ |
X86AsmParser.cpp | 114 unsigned Idx = InfixOperatorStack.size() - 1; 129 Idx = InfixOperatorStack.size() - 1; 165 for (unsigned i = 0, e = PostfixStack.size(); i != e; ++i) { 170 assert (OperandStack.size() > 1 && "Too few operands."); 226 assert (OperandStack.size() == 1 && "Expected a single result."); 665 ParseIntelSegmentOverride(unsigned SegReg, SMLoc Start, unsigned Size); 667 ParseIntelMemOperand(int64_t ImmDisp, SMLoc StartLoc, unsigned Size); 672 unsigned Size); 682 SMLoc End, unsigned Size, StringRef Identifier, 685 bool ParseDirectiveWord(unsigned Size, SMLoc L) [all...] |
/external/llvm/lib/Target/XCore/ |
XCoreISelLowering.cpp | 293 // We can only fold positive offsets that are a multiple of the word size. 362 unsigned NumEntries = MJTI->getJumpTables()[JTI].MBBs.size(); [all...] |
/external/llvm/lib/Transforms/IPO/ |
GlobalOpt.cpp | 246 for (int i = 0, e = Dead.size(); i != e; ++i) { 564 if (Val >= NewGlobals.size()) Val = 0; // Out of bound array access. 602 for (unsigned i = 0, e = NewGlobals.size(); i != e; ++i) 608 return FirstGlobal != NewGlobals.size() ? NewGlobals[FirstGlobal] : nullptr; 722 if (Idxs.size() == GEPI->getNumOperands()-1) [all...] |
/external/llvm/lib/Transforms/InstCombine/ |
InstCombineCalls.cpp | 78 // if the size is something we can handle with a single primitive load/store. 81 uint64_t Size = MemOpLength->getLimitedValue(); 82 assert(Size && "0-sized memory transferring should be removed already."); 84 if (Size > 8 || (Size&(Size-1))) 93 IntegerType* IntType = IntegerType::get(MI->getContext(), Size<<3); 108 if (DL && SrcETy->isSized() && DL->getTypeStoreSize(SrcETy) == Size) { 126 cast<ConstantInt>(M->getOperand(1))->getValue() == Size && 151 // Set the size of the copy to 0, it will be deleted on the next iteration [all...] |
/external/llvm/lib/Transforms/Instrumentation/ |
DataFlowSanitizer.cpp | 290 Value *loadShadow(Value *ShadowAddr, uint64_t Size, uint64_t Align, 292 void storeShadow(Value *Addr, uint64_t Size, uint64_t Align, Value *Shadow, 478 Asm.replace(Pos, SearchStr.size(), [all...] |
/external/llvm/lib/Transforms/Scalar/ |
ScalarReplAggregates.cpp | 78 // Do not limit the scalar integer load size if no threshold is given. 132 /// SRThreshold - The maximum alloca size to considered for SROA. 143 /// ScalarLoadThreshold - The maximum size in bits of scalars to load when 161 bool TypeHasComponent(Type *T, uint64_t Offset, uint64_t Size); 260 /// AllocaSize - The size of the alloca being considered in bytes. 305 explicit ConvertToScalarInfo(unsigned Size, const DataLayout &DL, 307 : AllocaSize(Size), DL(DL), ScalarLoadThreshold(SLT), IsNotTrivial(false), 362 // Do not convert to scalar integer if the alloca size exceeds the 391 /// 1) A union of vector types of the same size and potentially its elements. 408 // If the In type is a vector that is the same size as the alloca, see if i [all...] |
/external/llvm/utils/TableGen/ |
CodeGenRegisters.cpp | 38 Size = R->getValueAsInt("Size"); 44 : TheDef(nullptr), Name(N), Namespace(Nspace), Size(-1), Offset(-1), 62 if (Comps.size() != 2) 75 if (Parts.size() < 2) 79 for (unsigned i = 0, e = Parts.size(); i != e; ++i) 121 if (SRIs.size() != SRs.size()) 123 "SubRegs and SubRegIndices must have the same size"); 125 for (unsigned i = 0, e = SRIs.size(); i != e; ++i) [all...] |
/external/lzma/CPP/7zip/UI/Common/ |
Bench.cpp | 157 size_t Size;
160 void Init(const Byte *data, size_t size)
163 Size = size;
166 STDMETHOD(Read)(void *data, UInt32 size, UInt32 *processedSize);
169 STDMETHODIMP CBenchmarkInStream::Read(void *data, UInt32 size, UInt32 *processedSize)
171 size_t remain = Size - Pos;
173 if (size > kMaxBlockSize)
174 size = kMaxBlockSize;
175 if (size > remain) 1025 UInt64 size = (UInt64)numCycles * totalSize; local [all...] |
/external/mesa3d/src/mesa/tnl/ |
t_context.h | 196 GLuint Size; /**< Max vertices per vertex buffer, constant */ 247 /* Called on any statechange or input array size change or 304 GLuint vertattrsize; /* size of the attribute in bytes */ 336 GLuint size; member in struct:tnl_attr_type
|
/external/chromium_org/sandbox/win/src/ |
nt_internals.h | 246 LARGE_INTEGER Size; 416 IN SIZE_T Size);
|
/external/clang/include/clang/Analysis/Analyses/ |
ThreadSafetyTIL.h | 123 // require special handling; they use BT_ValueRef, and size ST_0. 151 : Base(B), Size(Sz), Signed(S), VectSize(VS) 155 SizeType Size; 667 switch (ValType.Size) { 693 switch (ValType.Size) { [all...] |
/external/clang/lib/AST/ |
Decl.cpp | [all...] |
ExprConstant.cpp | 113 /// path, and find the size of the containing array, if any. 120 for (unsigned I = 0, N = Path.size(); I != N; ++I) { 164 /// The size of the array of which the most-derived object is an element, or 231 MostDerivedPathLength = Entries.size(); 245 MostDerivedPathLength = Entries.size(); 258 MostDerivedPathLength = Entries.size(); 264 if (MostDerivedPathLength == Entries.size() && MostDerivedArraySize) { 364 *Diag << StringRef(Buffer.data(), Buffer.size()); 382 *Diag << StringRef(Buffer.data(), Buffer.size()); 766 : Info(Info), OldStackSize(Info.CleanupStack.size()) {} [all...] |
/external/clang/lib/CodeGen/ |
CGCall.cpp | 95 RequiredArgs required = RequiredArgs::forPrototypePlus(FTP, prefix.size()); 240 (D->isVariadic() ? RequiredArgs(argTypes.size()) : RequiredArgs::All); 351 (MD->isVariadic() ? RequiredArgs(argTys.size()) : RequiredArgs::All); 379 assert(args.size() >= numExtraRequiredArgs); 397 required = RequiredArgs(args.size()); 462 (isVariadic ? RequiredArgs(args.size()) : RequiredArgs::All); 534 sizeof(ArgInfo) * (argTypes.size() + 1)); 546 FI->NumArgs = argTypes.size(); 548 for (unsigned i = 0, e = argTypes.size(); i != e; ++i) 659 /// with an in-memory size smaller than DstSize [all...] |
CodeGenFunction.h | 255 /// The size of the following cleanup object. 256 size_t Size : 29; 260 size_t getSize() const { return Size; } 406 size_t OldSize = LifetimeExtendedCleanupStack.size(); 408 LifetimeExtendedCleanupStack.size() + sizeof(Header) + Header.Size); 479 CGF.LifetimeExtendedCleanupStack.size(); 557 /// \brief Takes the old cleanup stack size and emits the cleanup blocks 561 /// \brief Takes the old cleanup stack size and emits the cleanup blocks 862 // VLASizeMap - This keeps track of the associated size for each VLA type [all...] |
/external/libnfc-nxp/inc/ |
phNfcTypes.h | 162 #define PHHAL_MAX_DATASIZE 0xFBU /* 256 * Maximum Data size sent 216 * size. 247 uint32_t Size;/**< Size of the datablock*/ 410 ? sRecvData : Shall contain buffer of size 16 422 ? sSendData : Shall contain buffer of size 16 containing 433 ? sSendData : Shall contain buffer of size 4 containing [all...] |
/external/libnfc-nxp/src/ |
phFriNfc_NdefMap.h | 160 #define PH_FRINFC_NDEFMAP_TOPAZ_MAX_SIZE 256 /**< \internal Total Memory size = 96 bytes (newer version have mode) */ 161 #define PH_FRINFC_NDEFMAP_TOPAZ_UID_SIZE 0x04 /**< \internal UID size returned by READID command = 4 bytes */ 237 * \name Buffer Size Definitions 241 /** \ingroup grp_fri_nfc_ndef_map Minimum size of the TRX buffer required */ 243 /** \internal The size of s MIFARE block */ 264 /**< \internal NDEF TLV size */ 266 /**< \internal NDEF TLV size */ 277 /**< \internal Remaining size that can be read */ 461 /** To store the remaining size of the Mifare 1k or 4k card */ 463 /** To update the remaining size when writing to the Mifare 1k or 4k card * [all...] |
/external/llvm/lib/Target/Sparc/ |
SparcISelLowering.cpp | 89 unsigned size = (LocVT == MVT::f128) ? 16 : 8; local 91 unsigned Offset = State.AllocateStack(size, alignment); 205 for (unsigned i = 0; i != RVLocs.size(); ++i) { 267 for (unsigned i = 0; i != RVLocs.size(); ++i) { 296 if (i+1 < RVLocs.size() && RVLocs[i+1].getLocReg() == VA.getLocReg()) { 359 for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i, ++InIdx) { 559 for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i) { 705 // Get the size of the outgoing arguments stack space requirement. 715 for (unsigned i = 0, e = Outs.size(); i != e; ++i) { 721 unsigned Size = Flags.getByValSize() [all...] |