/external/skia/src/sfnt/ |
SkOTTable_loca.h | 23 union Offsets { 26 } offsets; member in struct:SkOTTableIndexToLocation
|
/external/llvm/unittests/Transforms/IPO/ |
LowerBitSets.cpp | 17 std::vector<uint64_t> Offsets; 55 for (auto Offset : T.Offsets) 67 for (auto Offset : T.Offsets) 70 auto I = T.Offsets.begin(); 72 if (I != T.Offsets.end() && *I == NonOffset) {
|
/external/clang/tools/clang-format/ |
ClangFormat.cpp | 38 Offsets("offset", 128 if (!Offsets.empty() || !Lengths.empty()) { 152 if (Offsets.empty()) 153 Offsets.push_back(0); 154 if (Offsets.size() != Lengths.size() && 155 !(Offsets.size() == 1 && Lengths.empty())) { 160 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 161 if (Offsets[i] >= Code->getBufferSize()) { 162 llvm::errs() << "error: offset " << Offsets[i] 167 Sources.getLocForStartOfFile(ID).getLocWithOffset(Offsets[i]) [all...] |
/external/llvm/include/llvm/Transforms/IPO/ |
LowerBitSets.h | 63 SmallVector<uint64_t, 16> Offsets; 74 Offsets.push_back(Offset); 144 /// loading from indexed offsets into the byte array and applying a mask, a
|
/external/llvm/tools/lli/ |
RemoteMemoryManager.cpp | 94 SmallVector<std::pair<Allocation, uint64_t>, 16> Offsets; 109 Offsets.push_back(std::pair<Allocation,uint64_t>(Section, CurOffset)); 126 Offsets.push_back(std::pair<Allocation,uint64_t>(Section, CurOffset)); 138 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 139 uint64_t Addr = RemoteAddr + Offsets[i].second; 140 EE->mapSectionAddress(const_cast<void*>(Offsets[i].first.MB.base()), Addr); 142 DEBUG(dbgs() << " Mapping local: " << Offsets[i].first.MB.base() 145 MappedSections[Addr] = Offsets[i].first;
|
/art/runtime/ |
handle_scope_test.cc | 38 // Test the offsets computed for members of HandleScope. Because of cross-compiling 41 TEST(HandleScopeTest, Offsets) NO_THREAD_SAFETY_ANALYSIS { 43 // here. So do the inverse: set some data, and access it through pointers created from the offsets.
|
/external/llvm/lib/Target/R600/MCTargetDesc/ |
R600MCCodeEmitter.cpp | 118 int64_t Offsets[3] = { 127 SrcSelect[ELEMENT_W] << 29 | Offsets[0] << 0 | Offsets[1] << 5 | 128 Offsets[2] << 10;
|
/external/llvm/lib/Support/ |
IntervalMap.cpp | 19 void Path::replaceRoot(void *Root, unsigned Size, IdxPair Offsets) { 21 path.front() = Entry(Root, Size, Offsets.first); 22 path.insert(path.begin() + 1, Entry(subtree(0), Offsets.second));
|
/external/llvm/include/llvm/CodeGen/ |
Analysis.h | 64 /// If Offsets is non-null, it points to a vector to be filled in 65 /// with the in-memory offsets of each of the individual values. 69 SmallVectorImpl<uint64_t> *Offsets = nullptr,
|
/external/llvm/lib/Object/ |
Archive.cpp | 360 const char *Offsets = Buf; 362 Offsets += sizeof(uint64_t); 364 Offsets += sizeof(uint32_t); 367 Offset = read32be(Offsets + SymbolIndex * 4); 369 Offset = read64be(Offsets + SymbolIndex * 8); 372 // Offsets (the first uint32_t is the number of bytes of the ranlib 377 Offset = read32le(Offsets + SymbolIndex * 8 + 4); 379 // Skip offsets. 399 Offset = read32le(Offsets + OffsetIndex * 4); 479 buf += 4 + (member_count * 4); // Skip offsets [all...] |
/external/llvm/lib/CodeGen/SelectionDAG/ |
ScheduleDAGSDNodes.cpp | 198 /// This function finds loads of the same base and different offsets. If the 199 /// offsets are not far apart (target specific), it add MVT::Glue inputs and 211 // the same base pointer and different offsets. 213 SmallVector<int64_t, 4> Offsets; 232 Offsets.push_back(Offset1); 234 Offsets.push_back(Offset2); 246 std::sort(Offsets.begin(), Offsets.end()); 251 int64_t BaseOff = Offsets[0]; 254 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) [all...] |
SelectionDAGBuilder.cpp | [all...] |
/external/llvm/lib/CodeGen/ |
Analysis.cpp | 81 /// If Offsets is non-null, it points to a vector to be filled in 82 /// with the in-memory offsets of each of the individual values. 86 SmallVectorImpl<uint64_t> *Offsets, 95 ComputeValueVTs(TLI, *EI, ValueVTs, Offsets, 104 ComputeValueVTs(TLI, EltTy, ValueVTs, Offsets, 113 if (Offsets) 114 Offsets->push_back(StartingOffset);
|
/external/llvm/lib/Transforms/Scalar/ |
SROA.cpp | 136 /// stores both the begin and end offsets of this use, a pointer to the use 169 /// This provides an ordering over ranges such that start offsets are 170 /// always increasing, and within equal start offsets, the end offsets are 273 /// \brief The begining and ending offsets of the alloca for this partition. 524 /// offsets of the slices and the ability to split them. It will visit "empty" [all...] |
LoopStrengthReduce.cpp | 846 /// additional base registers and/or unfolded offsets to derive an 897 const SmallVectorImpl<int64_t> &Offsets, [all...] |
/external/pdfium/core/src/fxcodec/lcms2/lcms2-2.6/src/ |
cmstypes.c | 164 // A table of n elements is readed, where first comes n records containing offsets and sizes and 177 // Let's take the offsets to each element 233 // Keep starting position of curve offsets [all...] |
/external/llvm/lib/Target/NVPTX/ |
NVPTXISelLowering.cpp | 85 SmallVectorImpl<uint64_t> *Offsets = nullptr, 97 if (Offsets) 98 Offsets->push_back(Off+j*VT.getVectorElementType().getStoreSize()); 102 if (Offsets) 103 Offsets->push_back(Off); 2122 SmallVector<uint64_t, 16> offsets; local [all...] |
/external/llvm/lib/Target/R600/ |
SIInstrInfo.h | 364 /// Offsets in bytes from the start of the input buffer 365 enum Offsets {
|
/external/clang/lib/AST/ |
VTableBuilder.cpp | 110 /// as a record decl and a subobject number) and its offsets in the most 117 /// ComputeBaseOffsets - Compute the offsets for all base subobjects of the 164 // Compute base offsets. 252 // even change the base offsets in ASTRecordLayout to be specified in 442 /// VCallOffsetMap - Keeps track of vcall offsets when building a vtable. 447 /// Offsets - Keeps track of methods and their offsets. 449 SmallVector<MethodAndOffsetPairTy, 16> Offsets; 467 bool empty() const { return Offsets.empty(); } 516 for (unsigned I = 0, E = Offsets.size(); I != E; ++I) [all...] |
/external/clang/lib/Parse/ |
ParseStmtAsm.cpp | 50 ArrayRef<Token> Toks, ArrayRef<unsigned> Offsets) 52 AsmTokOffsets(Offsets) {
|
/external/llvm/lib/Transforms/IPO/ |
LowerBitSets.cpp | 98 // the bitwise OR of each of the offsets. The number of trailing zeros 99 // in the mask gives us the log2 of the alignment of all offsets, which 103 for (uint64_t &Offset : Offsets) { 115 // Build the compressed bitset while normalizing the offsets against the 118 for (uint64_t Offset : Offsets) { 522 // Compute the offsets of the original globals within the new global. 544 // Build aliases pointing to offsets into the combined global for each
|
/external/llvm/lib/Target/PowerPC/ |
PPCFrameLowering.cpp | 93 // With the SVR4 ABI, callee-saved registers have fixed offsets on the stack. 113 // Note that the offsets here overlap, but this is fixed up in 116 static const SpillSlot Offsets[] = { 117 // Floating-point register save area offsets. 137 // General register save area offsets. 181 // Floating-point register save area offsets. 201 // General register save area offsets. 243 NumEntries = array_lengthof(Offsets); 245 return Offsets; 876 // Describe where callee saved registers were saved, at fixed offsets fro [all...] |
/external/clang/lib/CodeGen/ |
MicrosoftCXXABI.cpp | [all...] |
/external/clang/lib/Serialization/ |
ASTReaderDecl.cpp | [all...] |
/external/zlib/src/contrib/masmx64/ |
gvmat64.asm | 114 ; all the +4 offsets are due to the addition of pending_buf_size (in zlib
125 ;;; Offsets for fields in the deflate_state structure. These numbers
132 ; all the +zlib1222add offsets are due to the addition of fields
|