/external/skia/src/sfnt/ |
SkOTTable_loca.h | 23 union Offsets { 26 } offsets; member in struct:SkOTTableIndexToLocation
|
/external/llvm/tools/lli/ |
lli.cpp | 216 SmallVector<std::pair<const void*, uint64_t>, 16> Offsets; 228 Offsets.push_back(std::pair<const void*,uint64_t>(I->first.base(), CurOffset)); 234 unsigned FirstDataIndex = Offsets.size(); 245 Offsets.push_back(std::pair<const void*,uint64_t>(I->first.base(), CurOffset)); 256 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 257 uint64_t Addr = RemoteAddr + Offsets[i].second; 258 EE->mapSectionAddress(const_cast<void*>(Offsets[i].first), Addr); 260 DEBUG(dbgs() << " Mapping local: " << Offsets[i].first 269 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 270 uint64_t Addr = RemoteAddr + Offsets[i].second [all...] |
/external/llvm/include/llvm/CodeGen/ |
Analysis.h | 52 /// If Offsets is non-null, it points to a vector to be filled in 53 /// with the in-memory offsets of each of the individual values. 57 SmallVectorImpl<uint64_t> *Offsets = 0,
|
/external/llvm/lib/Support/ |
IntervalMap.cpp | 19 void Path::replaceRoot(void *Root, unsigned Size, IdxPair Offsets) { 21 path.front() = Entry(Root, Size, Offsets.first); 22 path.insert(path.begin() + 1, Entry(subtree(0), Offsets.second));
|
/external/llvm/lib/Target/PowerPC/ |
PPCFrameLowering.h | 131 // With the SVR4 ABI, callee-saved registers have fixed offsets on the stack. 151 // Note that the offsets here overlap, but this is fixed up in 154 static const SpillSlot Offsets[] = { 155 // Floating-point register save area offsets. 175 // General register save area offsets. 220 // Floating-point register save area offsets. 240 // General register save area offsets. 283 NumEntries = array_lengthof(Offsets); 285 return Offsets;
|
/external/llvm/lib/CodeGen/ |
Analysis.cpp | 71 /// If Offsets is non-null, it points to a vector to be filled in 72 /// with the in-memory offsets of each of the individual values. 76 SmallVectorImpl<uint64_t> *Offsets, 85 ComputeValueVTs(TLI, *EI, ValueVTs, Offsets, 94 ComputeValueVTs(TLI, EltTy, ValueVTs, Offsets, 103 if (Offsets) 104 Offsets->push_back(StartingOffset);
|
/external/llvm/lib/Object/ |
Archive.cpp | 193 const char *Offsets = Buf + 4; 196 Offset = *(reinterpret_cast<const support::ubig32_t*>(Offsets) 203 // Skip offsets. 226 Offset = *(reinterpret_cast<const support::ulittle32_t*>(Offsets) 259 buf += 4 + (member_count * 4); // Skip offsets. 278 buf += 4 + (member_count * 4); // Skip offsets.
|
/external/llvm/lib/CodeGen/SelectionDAG/ |
ScheduleDAGSDNodes.cpp | 203 /// This function finds loads of the same base and different offsets. If the 204 /// offsets are not far apart (target specific), it add MVT::Glue inputs and 216 // the same base pointer and different offsets. 218 SmallVector<int64_t, 4> Offsets; 234 Offsets.push_back(Offset1); 236 Offsets.push_back(Offset2); 246 std::sort(Offsets.begin(), Offsets.end()); 251 int64_t BaseOff = Offsets[0]; 254 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) [all...] |
SelectionDAGBuilder.cpp | [all...] |
/external/llvm/lib/Transforms/Scalar/ |
SROA.cpp | 86 /// This provides an ordering over ranges such that start offsets are 87 /// always increasing, and within equal start offsets, the end offsets are 289 /// The destination begin and end offsets when the destination is within 294 /// The source begin and end offsets when the source is within this alloca. 584 MemTransferOffsets &Offsets = P.MemTransferInstData[&II]; 587 Offsets.IsSplittable = Length; 590 Offsets.DestBegin = RawOffset; 591 Offsets.DestEnd = RawOffset + Size; 594 Offsets.SourceBegin = RawOffset [all...] |
LoopStrengthReduce.cpp | 817 const SmallVectorImpl<int64_t> &Offsets, 907 const SmallVectorImpl<int64_t> &Offsets, [all...] |
/external/clang/lib/AST/ |
VTableBuilder.cpp | 103 /// as a record decl and a subobject number) and its offsets in the most 110 /// ComputeBaseOffsets - Compute the offsets for all base subobjects of the 159 // Compute base offsets. 249 // even change the base offsets in ASTRecordLayout to be specified in 439 /// VCallOffsetMap - Keeps track of vcall offsets when building a vtable. 444 /// Offsets - Keeps track of methods and their offsets. 446 SmallVector<MethodAndOffsetPairTy, 16> Offsets; 464 bool empty() const { return Offsets.empty(); } 513 for (unsigned I = 0, E = Offsets.size(); I != E; ++I) [all...] |
/external/chromium/chrome/browser/resources/file_manager/js/ |
exif_reader.js | 75 // Offsets inside the EXIF block are based after this bit of
|
/external/clang/lib/Sema/ |
SemaStmtAsm.cpp | 436 ArrayRef<unsigned> Offsets) 437 : SemaRef(Ref), AsmLoc(Loc), AsmToks(Toks), TokOffsets(Offsets) { }
|
/external/webkit/Source/WebKit/chromium/tests/ |
KURLTest.cpp | 561 TEST(KURLTest, Offsets)
|
/external/clang/lib/Serialization/ |
ASTReaderDecl.cpp | [all...] |
ASTReader.cpp | 694 const std::pair<uint64_t, uint64_t> &Offsets, 698 if (Offsets.first != 0) { 699 Cursor.JumpToBit(Offsets.first); 715 if (Offsets.second != 0) { 716 Cursor.JumpToBit(Offsets.second); [all...] |
/external/clang/include/clang/Serialization/ |
ASTReader.h | 300 /// \brief A map of global bit offsets to the module that stores entities 301 /// at those bit offsets. 310 /// SourceLocation offsets to the modules containing them. 405 const std::pair<uint64_t, uint64_t> &Offsets, [all...] |
/external/zlib/src/contrib/masmx64/ |
gvmat64.asm | 114 ; all the +4 offsets are due to the addition of pending_buf_size (in zlib
125 ;;; Offsets for fields in the deflate_state structure. These numbers
132 ; all the +zlib1222add offsets are due to the addition of fields
|
/external/zlib/src/contrib/masmx86/ |
match686.asm | 74 ; all the +zlib1222add offsets are due to the addition of fields
118 ;;; stack frame offsets
140 ;;; Offsets for fields in the deflate_state structure. These numbers
|
/external/llvm/include/llvm/ADT/ |
IntervalMap.h | 863 /// @param Offsets Offsets into the root and first branch nodes. 864 void replaceRoot(void *Root, unsigned Size, IdxPair Offsets); [all...] |
/external/qemu-pc-bios/bochs/bios/ |
rombios.c | 55 // $fef3 ; Initial Interrupt Vector Offsets Loaded by POST 100 // - needs to be reworked. Uses direct [bp] offsets. (?) [all...] |