/external/skia/src/sfnt/ |
SkOTTable_loca.h | 23 union Offsets { 26 } offsets; member in struct:SkOTTableIndexToLocation
|
/external/llvm/unittests/Transforms/IPO/ |
LowerBitSets.cpp | 17 std::vector<uint64_t> Offsets; 55 for (auto Offset : T.Offsets) 67 for (auto Offset : T.Offsets) 70 auto I = T.Offsets.begin(); 72 if (I != T.Offsets.end() && *I == NonOffset) {
|
/external/clang/tools/clang-format/ |
ClangFormat.cpp | 39 Offsets("offset", 142 if (!Offsets.empty() || !Lengths.empty()) { 168 if (Offsets.empty()) 169 Offsets.push_back(0); 170 if (Offsets.size() != Lengths.size() && 171 !(Offsets.size() == 1 && Lengths.empty())) { 175 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 176 if (Offsets[i] >= Code->getBufferSize()) { 177 errs() << "error: offset " << Offsets[i] << " is outside the file\n"; 181 Sources.getLocForStartOfFile(ID).getLocWithOffset(Offsets[i]) [all...] |
/external/llvm/include/llvm/Transforms/IPO/ |
LowerBitSets.h | 66 SmallVector<uint64_t, 16> Offsets; 77 Offsets.push_back(Offset); 147 /// loading from indexed offsets into the byte array and applying a mask, a
|
/external/llvm/tools/lli/ |
RemoteMemoryManager.cpp | 94 SmallVector<std::pair<Allocation, uint64_t>, 16> Offsets; 109 Offsets.push_back(std::pair<Allocation,uint64_t>(Section, CurOffset)); 126 Offsets.push_back(std::pair<Allocation,uint64_t>(Section, CurOffset)); 138 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { 139 uint64_t Addr = RemoteAddr + Offsets[i].second; 140 EE->mapSectionAddress(const_cast<void*>(Offsets[i].first.MB.base()), Addr); 142 DEBUG(dbgs() << " Mapping local: " << Offsets[i].first.MB.base() 145 MappedSections[Addr] = Offsets[i].first;
|
/external/llvm/lib/Support/ |
Signals.cpp | 53 const char **Modules, intptr_t *Offsets, 90 std::vector<intptr_t> Offsets(Depth, 0); 91 if (!findModulesAndOffsets(StackTrace, Depth, Modules.data(), Offsets.data(), 105 Input << Modules[i] << " " << (void*)Offsets[i] << "\n"; 159 OS << "(" << Modules[i] << '+' << format_hex(Offsets[i], 0) << ")";
|
IntervalMap.cpp | 19 void Path::replaceRoot(void *Root, unsigned Size, IdxPair Offsets) { 21 path.front() = Entry(Root, Size, Offsets.first); 22 path.insert(path.begin() + 1, Entry(subtree(0), Offsets.second));
|
/external/llvm/lib/Target/Hexagon/ |
HexagonFrameLowering.h | 60 static const SpillSlot Offsets[] = { 68 NumEntries = array_lengthof(Offsets); 69 return Offsets;
|
/art/runtime/ |
handle_scope_test.cc | 38 // Test the offsets computed for members of HandleScope. Because of cross-compiling 41 TEST(HandleScopeTest, Offsets) NO_THREAD_SAFETY_ANALYSIS { 43 // here. So do the inverse: set some data, and access it through pointers created from the offsets.
|
/external/llvm/lib/Target/AMDGPU/MCTargetDesc/ |
R600MCCodeEmitter.cpp | 117 int64_t Offsets[3] = { 126 SrcSelect[ELEMENT_W] << 29 | Offsets[0] << 0 | Offsets[1] << 5 | 127 Offsets[2] << 10;
|
/external/llvm/include/llvm/CodeGen/ |
Analysis.h | 67 /// If Offsets is non-null, it points to a vector to be filled in 68 /// with the in-memory offsets of each of the individual values. 72 SmallVectorImpl<uint64_t> *Offsets = nullptr,
|
/external/llvm/lib/Object/ |
Archive.cpp | 445 const char *Offsets = Buf; 447 Offsets += sizeof(uint64_t); 449 Offsets += sizeof(uint32_t); 452 Offset = read32be(Offsets + SymbolIndex * 4); 454 Offset = read64be(Offsets + SymbolIndex * 8); 457 // Offsets (the first uint32_t is the number of bytes of the ranlib 462 Offset = read32le(Offsets + SymbolIndex * 8 + 4); 464 // Skip offsets. 484 Offset = read32le(Offsets + OffsetIndex * 4); 566 buf += 4 + (member_count * 4); // Skip offsets [all...] |
/external/llvm/lib/CodeGen/SelectionDAG/ |
ScheduleDAGSDNodes.cpp | 198 /// This function finds loads of the same base and different offsets. If the 199 /// offsets are not far apart (target specific), it add MVT::Glue inputs and 211 // the same base pointer and different offsets. 213 SmallVector<int64_t, 4> Offsets; 232 Offsets.push_back(Offset1); 234 Offsets.push_back(Offset2); 246 std::sort(Offsets.begin(), Offsets.end()); 251 int64_t BaseOff = Offsets[0]; 254 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) [all...] |
/external/llvm/lib/CodeGen/ |
Analysis.cpp | 83 /// If Offsets is non-null, it points to a vector to be filled in 84 /// with the in-memory offsets of each of the individual values. 88 SmallVectorImpl<uint64_t> *Offsets, 97 ComputeValueVTs(TLI, DL, *EI, ValueVTs, Offsets, 106 ComputeValueVTs(TLI, DL, EltTy, ValueVTs, Offsets, 115 if (Offsets) 116 Offsets->push_back(StartingOffset);
|
/external/llvm/lib/Transforms/Scalar/ |
SROA.cpp | 130 /// stores both the begin and end offsets of this use, a pointer to the use 163 /// This provides an ordering over ranges such that start offsets are 164 /// always increasing, and within equal start offsets, the end offsets are 337 /// \brief The beginning and ending offsets of the alloca for this 588 /// offsets of the slices and the ability to split them. It will visit "empty" [all...] |
LoopStrengthReduce.cpp | 851 /// additional base registers and/or unfolded offsets to derive an 902 const SmallVectorImpl<int64_t> &Offsets, [all...] |
/external/pdfium/third_party/lcms2-2.6/src/ |
cmstypes.c | 164 // A table of n elements is readed, where first comes n records containing offsets and sizes and 177 // Let's take the offsets to each element 233 // Keep starting position of curve offsets [all...] |
/toolchain/binutils/binutils-2.25/ld/testsuite/ld-cris/ |
tls-e-20a.d | 14 # offset for all TLS data we care about. Offsets different for TLS
|
/external/llvm/lib/Target/NVPTX/ |
NVPTXISelLowering.cpp | 85 SmallVectorImpl<uint64_t> *Offsets = nullptr, 97 if (Offsets) 98 Offsets->push_back(Off+j*VT.getVectorElementType().getStoreSize()); 102 if (Offsets) 103 Offsets->push_back(Off); [all...] |
/external/llvm/lib/MC/ |
ELFObjectWriter.cpp | [all...] |
/art/compiler/linker/ |
multi_oat_relative_patcher_test.cc | 125 TEST_F(MultiOatRelativePatcherTest, Offsets) {
|
/external/llvm/lib/Target/AMDGPU/ |
SIInstrInfo.h | 497 /// Offsets in bytes from the start of the input buffer 498 enum Offsets {
|
/external/clang/lib/Parse/ |
ParseStmtAsm.cpp | 50 ArrayRef<Token> Toks, ArrayRef<unsigned> Offsets) 52 AsmTokOffsets(Offsets) {
|
/external/llvm/lib/Transforms/IPO/ |
LowerBitSets.cpp | 117 // the bitwise OR of each of the offsets. The number of trailing zeros 118 // in the mask gives us the log2 of the alignment of all offsets, which 122 for (uint64_t &Offset : Offsets) { 134 // Build the compressed bitset while normalizing the offsets against the 137 for (uint64_t Offset : Offsets) { 562 // Compute the offsets of the original globals within the new global. 570 // Build aliases pointing to offsets into the combined global for each 808 // Build aliases pointing to offsets into the jump table, and replace [all...] |
/external/llvm/lib/Target/PowerPC/ |
PPCFrameLowering.cpp | 93 // With the SVR4 ABI, callee-saved registers have fixed offsets on the stack. 113 // Note that the offsets here overlap, but this is fixed up in 116 static const SpillSlot Offsets[] = { 117 // Floating-point register save area offsets. 137 // General register save area offsets. 181 // Floating-point register save area offsets. 201 // General register save area offsets. 243 NumEntries = array_lengthof(Offsets); 245 return Offsets; [all...] |