/external/v8/src/compiler/ |
basic-block-instrumentor.h | 19 class Schedule; 24 Graph* graph, Schedule* schedule);
|
machine-graph-verifier.h | 17 class Schedule; 23 static void Run(Graph* graph, Schedule const* const schedule,
|
verifier.h | 18 class Schedule; 57 // Verifies properties of a schedule, such as dominance, phi placement, etc. 60 static void Run(Schedule* schedule);
|
pipeline.h | 31 class Schedule; 48 // Run the pipeline on a machine graph and generate code. The {schedule} must 52 Graph* graph, Schedule* schedule, 60 // Run the pipeline on a machine graph and generate code. If {schedule} is 61 // {nullptr}, then compute a new schedule for code generation. 64 Schedule* schedule = nullptr); 71 // Run the pipeline on a machine graph and generate code. If {schedule} is 72 // {nullptr}, then compute a new schedule for code generation [all...] |
scheduler.h | 11 #include "src/compiler/schedule.h" 27 // Computes a schedule from a graph, placing nodes into basic blocks and 35 // The complete scheduling algorithm. Creates a new schedule and places all 37 static Schedule* ComputeSchedule(Zone* zone, Graph* graph, Flags flags); 39 // Compute the RPO of blocks in an existing schedule. 40 static BasicBlockVector* ComputeSpecialRPO(Zone* zone, Schedule* schedule); 66 Schedule* schedule_; 76 Scheduler(Zone* zone, Graph* graph, Schedule* schedule, Flags flags) [all...] |
schedule.cc | 5 #include "src/compiler/schedule.h" 128 Schedule::Schedule(Zone* zone, size_t node_count_hint) 139 BasicBlock* Schedule::block(Node* node) const { 147 bool Schedule::IsScheduled(Node* node) { 153 BasicBlock* Schedule::GetBlockById(BasicBlock::Id block_id) { 159 bool Schedule::SameBasicBlock(Node* a, Node* b) const { 165 BasicBlock* Schedule::NewBasicBlock() { 173 void Schedule::PlanNode(BasicBlock* block, Node* node) { 184 void Schedule::AddNode(BasicBlock* block, Node* node) [all...] |
graph-visualizer.h | 24 class Schedule; 53 AsC1V(const char* phase, const Schedule* schedule, 56 : schedule_(schedule), 60 const Schedule* schedule_;
|
/external/eigen/unsupported/Eigen/CXX11/src/ThreadPool/ |
ThreadPoolInterface.h | 19 virtual void Schedule(std::function<void()> fn) = 0;
|
/external/tensorflow/tensorflow/core/common_runtime/ |
eigen_thread_pool.h | 29 void Schedule(std::function<void()> fn) override { 30 pool_->Schedule(std::move(fn));
|
/external/eigen/unsupported/test/ |
cxx11_tensor_notification.cpp | 42 thread_pool.Schedule(func); 66 thread_pool.Schedule(func); 67 thread_pool.Schedule(func); 68 thread_pool.Schedule(func); 69 thread_pool.Schedule(func);
|
cxx11_non_blocking_thread_pool.cpp | 36 // Schedule kThreads tasks and ensure that they all are running. 38 tp.Schedule([&]() { 52 // Now, while the previous tasks exit, schedule another kThreads tasks and 55 tp.Schedule([&, i]() { 60 // continue running and quarter of tasks schedule another 2 tasks each. 72 tp.Schedule([&]() { 88 tp.Schedule([&]() {
|
/external/tensorflow/tensorflow/stream_executor/host/ |
host_stream.cc | 36 host_executor_->Schedule([this, task]() {
|
/external/tensorflow/tensorflow/core/lib/core/ |
blocking_counter_test.cc | 43 thread_pool->Schedule([&bc] { bc.DecrementCount(); }); 62 thread_pool->Schedule([&bc, shards_per_thread] {
|
notification_test.cc | 33 thread_pool->Schedule([&start, &proceed, &counter] { 64 thread_pool->Schedule([&n, &lock, &counter] {
|
threadpool.h | 35 // wait. Conversely, if the threadpool is used to schedule high-latency 60 void Schedule(std::function<void()> fn);
|
threadpool_test.cc | 48 pool.Schedule([&work, i]() { 132 pool.Schedule(work); 155 pool.Schedule([&count, &done_lock, &done, &done_flag]() {
|
/external/tensorflow/tensorflow/core/distributed_runtime/rpc/ |
grpc_worker_service.cc | 187 void Schedule(std::function<void()> f) { 188 worker_->env()->compute_pool->Schedule(std::move(f)); 205 Schedule([this, call]() { 215 Schedule([this, call]() { 226 Schedule([this, call]() { 236 Schedule([this, call]() { 245 Schedule([this, call]() { 254 Schedule([this, call]() { 262 Schedule([this, call]() { 284 Schedule([this, call]() [all...] |
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
basic_batch_scheduler.h | 35 // could use two BasicBatchScheduler instances to schedule the two model/GPU 56 // When Schedule() is called, if the queue is full the call will fail with an 172 // The maximum allowable number of enqueued (accepted by Schedule() but 174 // If this limit is reached, Schedule() will return an UNAVAILABLE error. 191 Status Schedule(std::unique_ptr<TaskType>* task) override; 245 Status BasicBatchScheduler<TaskType>::Schedule( 247 return shared_scheduler_queue_->Schedule(task);
|
/external/llvm/lib/CodeGen/ |
MachinePipeliner.cpp | 38 // 3) Attempt to schedule the nodes in the specified order using the MII. 53 // where an instruction may be inserted in the pipelined schedule. 195 /// and attempts to schedule the instructions using the SMS algorithm. 198 /// The minimum initiation interval between iterations for this schedule. 200 /// Set to true if a valid pipelined schedule is found for the loop. 227 /// Instructions to change when emitting the final schedule. 267 void schedule(); 332 /// Set the Minimum Initiation Interval for this schedule attempt. 335 MachineInstr *applyInstrChange(MachineInstr *MI, SMSchedule &Schedule, 365 bool schedulePipeline(SMSchedule &Schedule); 819 void SwingSchedulerDAG::schedule() { function in class:SwingSchedulerDAG [all...] |
/external/tensorflow/tensorflow/contrib/boosted_trees/lib/utils/ |
parallel_for.cc | 38 thread_pool->Schedule([&do_work, &counter, start, end]() {
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
TensorDeviceThreadPool.h | 154 pool_->Schedule(std::bind(&FunctionWrapperWithNotification<Function, Args...>::run, n, f, args...)); 162 pool_->Schedule(std::bind( 168 pool_->Schedule(std::bind(f, args...)); 258 pool_->Schedule([=, &handleRange]() { handleRange(mid, last); }); 259 pool_->Schedule([=, &handleRange]() { handleRange(first, mid); });
|
/external/guava/guava-tests/test/com/google/common/util/concurrent/ |
AbstractScheduledServiceTest.java | 301 Scheduler schedule = Scheduler.newFixedRateSchedule(initialDelay, delay, unit); local 302 schedule.schedule(null, new ScheduledThreadPoolExecutor(1) { 314 Scheduler schedule = Scheduler.newFixedDelaySchedule(initialDelay, delay, unit); local 315 schedule.schedule(null, new ScheduledThreadPoolExecutor(10) { 329 protected Schedule getNextSchedule() throws Exception { 331 return new Schedule(0, TimeUnit.SECONDS); 352 Future<?> future = scheduler.schedule(null, Executors.newScheduledThreadPool(10), task); 381 protected Schedule getNextSchedule() throws Exception [all...] |
/external/tensorflow/tensorflow/cc/client/ |
client_session_test.cc | 79 thread_pool.Schedule([&session, a]() { 85 thread_pool.Schedule([&session, b]() {
|
/external/guava/guava/src/com/google/common/util/concurrent/ |
AbstractScheduledService.java | 46 * the {@link #startUp} and {@link #shutDown} methods and also uses that service to schedule the 54 * #runOneIteration} takes longer than its schedule defines, then subsequent executions may start 105 * service with a fixed schedule. If more flexibility is needed then consider subclassing 125 public Future<?> schedule(AbstractService service, ScheduledExecutorService executor, 144 public Future<?> schedule(AbstractService service, ScheduledExecutorService executor, 152 abstract Future<?> schedule(AbstractService service, ScheduledExecutorService executor, 201 runningTask = scheduler().schedule(delegate, executorService, task); 391 * use a dynamically changing schedule. After every execution of the task, assuming it hasn't 453 final Schedule schedule = CustomScheduler.this.getNextSchedule() [all...] |
/external/tensorflow/tensorflow/contrib/ffmpeg/default/ |
ffmpeg_lib_utility_test.cc | 50 pool.Schedule([&mu, &temp_filenames, environment]() {
|