/prebuilts/clang/host/linux-x86/clang-3977809/prebuilt_include/llvm/include/llvm/CodeGen/ |
Analysis.h | 69 /// If Offsets is non-null, it points to a vector to be filled in 70 /// with the in-memory offsets of each of the individual values. 74 SmallVectorImpl<uint64_t> *Offsets = nullptr,
|
/prebuilts/clang/host/linux-x86/clang-4053586/prebuilt_include/llvm/include/llvm/CodeGen/ |
Analysis.h | 69 /// If Offsets is non-null, it points to a vector to be filled in 70 /// with the in-memory offsets of each of the individual values. 74 SmallVectorImpl<uint64_t> *Offsets = nullptr,
|
/external/swiftshader/third_party/LLVM/lib/CodeGen/ |
Analysis.cpp | 72 /// If Offsets is non-null, it points to a vector to be filled in 73 /// with the in-memory offsets of each of the individual values. 77 SmallVectorImpl<uint64_t> *Offsets, 86 ComputeValueVTs(TLI, *EI, ValueVTs, Offsets, 95 ComputeValueVTs(TLI, EltTy, ValueVTs, Offsets, 104 if (Offsets) 105 Offsets->push_back(StartingOffset);
|
/external/llvm/lib/Object/ |
Archive.cpp | 495 const char *Offsets = Buf; 497 Offsets += sizeof(uint64_t); 499 Offsets += sizeof(uint32_t); 502 Offset = read32be(Offsets + SymbolIndex * 4); 504 Offset = read64be(Offsets + SymbolIndex * 8); 507 // Offsets (the first uint32_t is the number of bytes of the ranlib 512 Offset = read32le(Offsets + SymbolIndex * 8 + 4); 515 // Offsets (the first uint64_t is the number of bytes of the ranlib_64 520 Offset = read64le(Offsets + SymbolIndex * 16 + 8); 522 // Skip offsets [all...] |
/external/llvm/lib/Target/Hexagon/ |
HexagonFrameLowering.h | 62 static const SpillSlot Offsets[] = { 70 NumEntries = array_lengthof(Offsets); 71 return Offsets;
|
/external/swiftshader/third_party/LLVM/lib/Target/PowerPC/ |
PPCFrameLowering.h | 112 // With the SVR4 ABI, callee-saved registers have fixed offsets on the stack. 132 static const SpillSlot Offsets[] = { 133 // Floating-point register save area offsets. 153 // General register save area offsets. 210 // Floating-point register save area offsets. 230 // General register save area offsets. 312 NumEntries = array_lengthof(Offsets); 314 return Offsets;
|
/external/abi-compliance-checker/modules/Internals/ |
CallConv.pm | 173 my %Offsets = (); 198 $Offsets{$Pos} = $Pos; 207 $Offsets{$Pos} = $Offset; 235 $GroupClasses{$Offsets{$Pos}+$Offset} = $SubClasses{$Offset}; 281 my @Offsets = sort {int($a)<=>int($b)} keys(%{$PreClasses}); 282 if($#Offsets==0) { 288 while($Num<=$#Offsets-1) 290 my $Offset1 = $Offsets[$Num]; 291 my $Offset2 = $Offsets[$Num+1]; 337 if($Num==$#Offsets) { [all...] |
XmlDump.pm | 154 if(my @Offsets = keys(%{$TInfo{"VTable"}})) 157 foreach my $Offset (sort { $a<=>$b } @Offsets)
|
/external/llvm/lib/CodeGen/SelectionDAG/ |
ScheduleDAGSDNodes.cpp | 198 /// This function finds loads of the same base and different offsets. If the 199 /// offsets are not far apart (target specific), it add MVT::Glue inputs and 211 // the same base pointer and different offsets. 213 SmallVector<int64_t, 4> Offsets; 232 Offsets.push_back(Offset1); 234 Offsets.push_back(Offset2); 246 std::sort(Offsets.begin(), Offsets.end()); 251 int64_t BaseOff = Offsets[0]; 254 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) [all...] |
/external/swiftshader/third_party/LLVM/lib/CodeGen/SelectionDAG/ |
ScheduleDAGSDNodes.cpp | 167 /// This function finds loads of the same base and different offsets. If the 168 /// offsets are not far apart (target specific), it add MVT::Glue inputs and 180 // the same base pointer and different offsets. 182 SmallVector<int64_t, 4> Offsets; 198 Offsets.push_back(Offset1); 200 Offsets.push_back(Offset2); 210 std::sort(Offsets.begin(), Offsets.end()); 215 int64_t BaseOff = Offsets[0]; 218 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) [all...] |
/external/llvm/lib/CodeGen/ |
Analysis.cpp | 82 /// If Offsets is non-null, it points to a vector to be filled in 83 /// with the in-memory offsets of each of the individual values. 87 SmallVectorImpl<uint64_t> *Offsets, 96 ComputeValueVTs(TLI, DL, *EI, ValueVTs, Offsets, 105 ComputeValueVTs(TLI, DL, EltTy, ValueVTs, Offsets, 114 if (Offsets) 115 Offsets->push_back(StartingOffset);
|
/external/llvm/lib/Transforms/Scalar/ |
SROA.cpp | 117 /// stores both the begin and end offsets of this use, a pointer to the use 150 /// This provides an ordering over ranges such that start offsets are 151 /// always increasing, and within equal start offsets, the end offsets are 324 /// \brief The beginning and ending offsets of the alloca for this 575 /// offsets of the slices and the ability to split them. It will visit "empty" [all...] |
LoopStrengthReduce.cpp | 835 /// additional base registers and/or unfolded offsets to derive an 886 const SmallVectorImpl<int64_t> &Offsets, [all...] |
/art/runtime/ |
handle_scope_test.cc | 42 // Test the offsets computed for members of HandleScope. Because of cross-compiling 45 TEST_F(HandleScopeTest, Offsets) { 49 // here. So do the inverse: set some data, and access it through pointers created from the offsets.
|
/device/linaro/bootloader/edk2/UefiCpuPkg/ResetVector/Vtf0/ |
SerialDebug.asm | 17 ;// UART Register Offsets
|
/external/pdfium/third_party/lcms2-2.6/src/ |
cmstypes.c | 164 // A table of n elements is readed, where first comes n records containing offsets and sizes and 177 // Let's take the offsets to each element 233 // Keep starting position of curve offsets [all...] |
/toolchain/binutils/binutils-2.25/ld/testsuite/ld-cris/ |
tls-e-20a.d | 14 # offset for all TLS data we care about. Offsets different for TLS
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Scalar/ |
LoopStrengthReduce.cpp | 707 const SmallVectorImpl<int64_t> &Offsets, 807 const SmallVectorImpl<int64_t> &Offsets, 837 for (SmallVectorImpl<int64_t>::const_iterator I = Offsets.begin(), 838 E = Offsets.end(); I != E; ++I) { 919 /// offsets, for example in an unrolled loop. [all...] |
/external/swiftshader/third_party/subzero/src/ |
IceCfg.cpp | [all...] |
/external/llvm/lib/Target/NVPTX/ |
NVPTXISelLowering.cpp | 85 SmallVectorImpl<uint64_t> *Offsets = nullptr, 97 if (Offsets) 98 Offsets->push_back(Off+j*VT.getVectorElementType().getStoreSize()); 102 if (Offsets) 103 Offsets->push_back(Off); [all...] |
/external/llvm/lib/MC/ |
ELFObjectWriter.cpp | [all...] |
/art/compiler/linker/ |
multi_oat_relative_patcher_test.cc | 131 TEST_F(MultiOatRelativePatcherTest, Offsets) {
|
/external/clang/lib/Parse/ |
ParseStmtAsm.cpp | 51 ArrayRef<Token> Toks, ArrayRef<unsigned> Offsets) 53 AsmTokOffsets(Offsets) {
|
/external/llvm/lib/Transforms/IPO/ |
LowerTypeTests.cpp | 118 // the bitwise OR of each of the offsets. The number of trailing zeros 119 // in the mask gives us the log2 of the alignment of all offsets, which 123 for (uint64_t &Offset : Offsets) { 135 // Build the compressed bitset while normalizing the offsets against the 138 for (uint64_t Offset : Offsets) { 526 // Compute the offsets of the original globals within the new global. 534 // Build aliases pointing to offsets into the combined global for each 765 // Build aliases pointing to offsets into the jump table, and replace [all...] |
/external/llvm/lib/Target/AMDGPU/ |
SIInstrInfo.h | 574 /// Offsets in bytes from the start of the input buffer 575 enum Offsets {
|