/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
sequential_thunk.cc | 23 SequentialThunk::SequentialThunk(std::vector<std::unique_ptr<Thunk>>&& thunks, 25 : Thunk(Kind::kSequential, hlo), thunks_(std::move(thunks)) {} 29 for (auto& thunk : thunks_) { 30 TF_RETURN_IF_ERROR(thunk->Initialize(executable)); 38 for (const auto& thunk : thunks_) { 39 TF_RETURN_IF_ERROR(thunk->ExecuteOnStream(buffer_allocations, stream));
|
thunk_schedule.cc | 25 const Thunk& thunk, const HloInstruction& operand, 26 const std::unordered_map<const HloInstruction*, Thunk*>& hlo_to_thunk) { 28 // If `operand` is mapped to a thunk, adds `operand` to `thunk`'s dependency 32 stream_assignment_->StreamNumberForHlo(*thunk.hlo_instruction())) { 33 depends_on_[&thunk].push_back(FindOrDie(hlo_to_thunk, &operand)); 36 // If `operand` doesn't need a thunk (e.g. bitcast), continue with its 39 AddDependenciesOnTransitiveOperands(thunk, *operand_of_operand, 51 std::unordered_map<const HloInstruction*, Thunk*> hlo_to_thunk [all...] |
thunk_schedule.h | 25 #include "tensorflow/compiler/xla/service/gpu/thunk.h" 35 // 1. A stream assignment indicating which stream each thunk is executed on. 52 const std::vector<Thunk*>& TotalOrder() const { return thunk_total_order_; } 54 // Thunks that `thunk` depends on. 55 const std::list<const Thunk*>& DependsOn(const Thunk* thunk) const; 56 // Whether `thunk` is depended by another thunk. 57 bool Depended(const Thunk* thunk) const { return depended_by_.count(thunk); [all...] |
gpu_executable.cc | 164 // The next event enqueued on stream N must not run until the thunk at 166 std::map<int32, const Thunk*> last_blocking_thunk_for_stream; 167 std::map<const Thunk*, std::unique_ptr<se::Event>> thunk_to_finish_event; 168 for (Thunk* thunk : thunk_schedule_->TotalOrder()) { 169 TF_RETURN_IF_ERROR(thunk->Initialize(*this)); 171 thunk_schedule_->StreamNumberForHlo(*thunk->hlo_instruction()); 175 for (const Thunk* dependency : thunk_schedule_->DependsOn(thunk)) { 186 // If this thunk requests it, wait for all currently-executing thunks t [all...] |
ir_emitter_unnested.h | 20 #include "tensorflow/compiler/xla/service/gpu/thunk.h" 84 // Same as `EmitTargetElementLoop`, but in given `thunk` rather than 88 KernelThunk* thunk); 91 // Builds the appropriate thunk for the instruction hlo and returns the owning 93 // of the returned Thunk object. 94 std::unique_ptr<Thunk> BuildThunk(const HloInstruction* hlo); 151 // Emits code to initialize buffer of `inst` in given `thunk`. 152 Status EmitInitializer(const HloInstruction* inst, KernelThunk* thunk); 156 // Thunk object. 157 std::unique_ptr<Thunk> BuildKernelThunk(const HloInstruction* inst) [all...] |
ir_emitter_unnested.cc | 50 #include "tensorflow/compiler/xla/service/gpu/thunk.h" 117 // Updates the launch dimensions in "thunk" and annotate the launch dimensions 119 // Precondition: "thunk" must be a KernelThunk. 120 void UpdateLaunchDimensions(const LaunchDimensions& launch_dims, Thunk* thunk, 122 CHECK(Thunk::Kind::kKernel == thunk->kind()); 123 KernelThunk* kernel_thunk = static_cast<KernelThunk*>(thunk); 432 std::unique_ptr<ConvolutionThunk> thunk; local 434 thunk = MakeUnique<ConvolutionThunk> 2294 thunk); member in namespace:xla [all...] |
/libcore/ojluni/src/main/java/sun/misc/ |
Cleaner.java | 44 * <p> A cleaner tracks a referent object and encapsulates a thunk of arbitrary 112 private final Runnable thunk; field in class:Cleaner 114 private Cleaner(Object referent, Runnable thunk) { 116 this.thunk = thunk; 123 * @param thunk 130 public static Cleaner create(Object ob, Runnable thunk) { 131 if (thunk == null) 133 return add(new Cleaner(ob, thunk)); 143 thunk.run() [all...] |
/libcore/luni/src/main/java/libcore/util/ |
NativeAllocationRegistry.java | 126 CleanerThunk thunk; local 129 thunk = new CleanerThunk(); 130 Cleaner cleaner = Cleaner.create(referent, thunk); 138 thunk.setNativePtr(nativePtr); 179 CleanerThunk thunk = new CleanerThunk(); local 180 Cleaner cleaner = Cleaner.create(referent, thunk); 188 thunk.setNativePtr(nativePtr);
|
/bionic/libc/upstream-freebsd/lib/libc/stdlib/ |
qsort.c | 99 med3(char *a, char *b, char *c, cmp_t *cmp, void *thunk 105 return CMP(thunk, a, b) < 0 ? 106 (CMP(thunk, b, c) < 0 ? b : (CMP(thunk, a, c) < 0 ? c : a )) 107 :(CMP(thunk, b, c) > 0 ? b : (CMP(thunk, a, c) < 0 ? a : c )); 112 qsort_r(void *a, size_t n, size_t es, void *thunk, cmp_t *cmp) 114 #define thunk NULL 130 pl > (char *)a && CMP(thunk, pl - es, pl) > 0; 142 pl = med3(pl, pl + d, pl + 2 * d, cmp, thunk); [all...] |
/art/compiler/linker/ |
relative_patcher.cc | 142 bool RelativePatcher::WriteThunk(OutputStream* out, const ArrayRef<const uint8_t>& thunk) { 143 if (UNLIKELY(!out->WriteFully(thunk.data(), thunk.size()))) { 146 size_relative_call_thunks_ += thunk.size(); 150 bool RelativePatcher::WriteMiscThunk(OutputStream* out, const ArrayRef<const uint8_t>& thunk) { 151 if (UNLIKELY(!out->WriteFully(thunk.data(), thunk.size()))) { 154 size_misc_thunks_ += thunk.size();
|
relative_patcher.h | 54 * ReserveSpace() may need to allocate space for a special dispatch thunk. 96 // The caller may use this method to preemptively force thunk space reservation and 117 // Patch a branch to a Baker read barrier thunk. 133 bool WriteThunk(OutputStream* out, const ArrayRef<const uint8_t>& thunk); 134 bool WriteMiscThunk(OutputStream* out, const ArrayRef<const uint8_t>& thunk);
|
/external/tensorflow/tensorflow/contrib/kfac/python/ops/ |
estimator.py | 167 self.cov_update_ops = [thunk() for thunk in self.cov_update_thunks] 175 self.inv_update_ops = [thunk() for thunk in self.inv_update_thunks] 265 """Constructs a covariance update thunk for a single FisherFactor.""" 267 def thunk(): function in function:FisherEstimator._create_cov_update_thunk 272 return thunk 275 """Constructs an inverse update thunk for a single FisherFactor.""" 277 def thunk(): function in function:FisherEstimator._create_inv_update_thunk 282 return thunk [all...] |
utils.py | 386 thunks: List of thunks. Each thunk encapsulates one op. Return values are 396 def true_fn(thunk): 397 """Ensures thunk is executed and returns an Op (not a Tensor).""" 400 with ops.control_dependencies([thunk()]): 414 true_fns = [true_fn(thunk) for thunk in thunks] 415 false_fns = [false_fn(thunk) for thunk in thunks]
|
/external/turbine/java/com/google/turbine/binder/lookup/ |
ImportIndex.java | 140 Supplier<ImportScope> thunk = thunks.get(lookup.first()); 141 if (thunk == null) { 144 ImportScope scope = thunk.get();
|
/art/tools/ahat/src/main/com/android/ahat/heapdump/ |
AhatClassInstance.java | 265 Value vthunk = getField("thunk"); 270 AhatClassInstance thunk = vthunk.asAhatInstance().asClassInstance(); local 271 if (thunk == null 272 || !thunk.isInstanceOfClass("libcore.util.NativeAllocationRegistry$CleanerThunk")) { 276 Value vregistry = thunk.getField("this$0");
|
/device/linaro/bootloader/edk2/MdeModulePkg/Universal/EbcDxe/X64/ |
EbcLowLevel.S | 29 # This instruction requires that we thunk out to external native
|
/external/tensorflow/tensorflow/contrib/kfac/python/kernel_tests/ |
estimator_test.py | 145 [(math_ops.equal(global_step, i), thunk) 146 for i, thunk in enumerate(cov_update_op_thunks)]) 193 [(math_ops.equal(global_step, i), thunk) 194 for i, thunk in enumerate(inv_update_op_thunks)])
|
/art/dex2oat/linker/ |
multi_oat_relative_patcher_test.cc | 64 std::vector<uint8_t> thunk(next_write_call_thunk_, 'c'); 65 bool success = WriteThunk(out, ArrayRef<const uint8_t>(thunk)); 71 std::vector<uint8_t> thunk(next_write_misc_thunk_, 'm'); 72 bool success = WriteMiscThunk(out, ArrayRef<const uint8_t>(thunk));
|
/external/libmojo/mojo/android/javatests/src/org/chromium/mojo/bindings/ |
RouterTest.java | 163 RouterImpl.ResponderThunk thunk = 166 thunk.finalize();
|
/libcore/ojluni/src/main/java/sun/nio/fs/ |
PollingWatchService.java | 288 Runnable thunk = new Runnable() { public void run() { poll(); }}; 290 .scheduleAtFixedRate(thunk, period, period, TimeUnit.SECONDS);
|
/art/compiler/linker/arm/ |
relative_patcher_arm_base.cc | 38 DCHECK(NeedsNextThunk()); // The data is constructed only when we expect to need the thunk. 70 // Adjust the MaxNextOffset() down if needed to fit the code before the next thunk. 145 std::vector<uint8_t> code_; // The code of the thunk. 146 std::vector<uint32_t> offsets_; // Offsets at which the thunk needs to be written. 147 uint32_t max_next_offset_; // The maximum offset at which the next thunk can be placed. 157 // the thunk with the smallest pending offset, so use the reverse ordering, i.e. ">". 171 // to place thunk will be soon enough, we need to reserve all needed thunks now. Code for 214 // Mark the thunk as written at the pending offset and update the `pending_thunks_` heap. 233 // called before reserving thunk positions for subsequent oat files. 248 // Get the base name to use for the first occurrence of the thunk 299 ThunkData* thunk = unreserved_thunks_.front(); local [all...] |
/art/runtime/gc/accounting/ |
space_bitmap.h | 162 uintptr_t max, SweepCallback* thunk, void* arg);
|
/prebuilts/go/darwin-x86/src/cmd/vendor/github.com/ianlancetaylor/demangle/ |
demangle_test.go | 160 "non-virtual thunk to C::D::E() const", 161 "non-virtual thunk to C::D::E() const", 162 "non-virtual thunk to C::D::E() const", 163 "non-virtual thunk to C::D::E() const", 167 "virtual thunk to C::D::E::F() const", 168 "virtual thunk to C::D::E::F() const", 169 "virtual thunk to C::D::E::F() const", 170 "virtual thunk to C::D::E::F() const",
|
/prebuilts/go/linux-x86/src/cmd/vendor/github.com/ianlancetaylor/demangle/ |
demangle_test.go | 160 "non-virtual thunk to C::D::E() const", 161 "non-virtual thunk to C::D::E() const", 162 "non-virtual thunk to C::D::E() const", 163 "non-virtual thunk to C::D::E() const", 167 "virtual thunk to C::D::E::F() const", 168 "virtual thunk to C::D::E::F() const", 169 "virtual thunk to C::D::E::F() const", 170 "virtual thunk to C::D::E::F() const",
|
/device/linaro/bootloader/edk2/EdkCompatibilityPkg/Compatibility/BootScriptSaveOnS3SaveStateThunk/X64/ |
AsmDispatchExecute.S | 39 # Description: A thunk function to execute 32-bit code in long mode.
|