/system/extras/memory_replay/ |
Threads.h | 37 size_t num_threads() { return num_threads_; } function in class:Threads
|
/external/tensorflow/tensorflow/contrib/slim/python/slim/data/ |
prefetch_queue.py | 35 num_threads=1, 49 images, labels = tf.train.batch([image, label], batch_size=32, num_threads=4) 58 num_threads: An integer. Number of threads running the enqueue op. 88 queue_runner.QueueRunner(queue, [enqueue_op] * num_threads))
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
backend.cc | 54 int num_threads) { 55 intra_op_parallelism_threads_ = num_threads; 66 explicit EigenThreadPoolWrapper(const int num_threads) 68 "XLAEigen", num_threads)), 148 const int num_threads = intra_op_parallelism_threads > 0 local 152 new EigenThreadPoolWrapper(num_threads));
|
/external/tensorflow/tensorflow/core/kernels/ |
concat_lib_cpu.h | 46 int num_threads = std::min(4, worker_threads->num_threads); local 50 num_threads = 51 static_cast<int>(std::min<int64>(num_threads, output->size() / 4096)); 55 if (num_threads == 0) { 125 Shard(worker_threads->num_threads, worker_threads->workers, output->size(),
|
/external/fio/t/ |
dedupe.c | 68 static unsigned int num_threads; variable 349 for (i = 0; i < num_threads; i++) { 388 threads = malloc(num_threads * sizeof(struct worker_thread)); 389 for (i = 0; i < num_threads; i++) { 405 for (i = 0; i < num_threads; i++) { 412 printf("Threads(%u): %lu items processed\n", num_threads, nitems); 458 printf("Will check <%s>, size <%llu>, using %u threads\n", filename, (unsigned long long) dev_size, num_threads); 549 num_threads = atoi(optarg); 575 if (!num_threads) 576 num_threads = cpus_online() [all...] |
/external/mesa3d/src/gallium/drivers/llvmpipe/ |
lp_screen.c | 654 screen->num_threads = util_cpu_caps.nr_cpus > 1 ? util_cpu_caps.nr_cpus : 0; 656 screen->num_threads = 0; 658 screen->num_threads = debug_get_num_option("LP_NUM_THREADS", screen->num_threads); 659 screen->num_threads = MIN2(screen->num_threads, LP_MAX_THREADS); 661 screen->rast = lp_rast_create(screen->num_threads);
|
lp_query.c | 99 unsigned num_threads = MAX2(1, screen->num_threads); local 123 for (i = 0; i < num_threads; i++) { 128 for (i = 0; i < num_threads; i++) { 134 for (i = 0; i < num_threads; i++) { 171 for (i = 0; i < num_threads; i++) {
|
lp_rast.c | 725 if (rast->num_threads == 0) { 751 for (i = 0; i < rast->num_threads; i++) { 763 if (rast->num_threads == 0) { 770 for (i = 0; i < rast->num_threads; i++) { 863 /* NOTE: if num_threads is zero, we won't use any threads */ 864 for (i = 0; i < rast->num_threads; i++) { 875 * Create new lp_rasterizer. If num_threads is zero, don't create any 877 * \param num_threads number of rasterizer threads to create 880 lp_rast_create( unsigned num_threads ) 895 for (i = 0; i < MAX2(1, num_threads); i++) [all...] |
/external/tensorflow/tensorflow/contrib/learn/python/learn/estimators/ |
state_saving_rnn_estimator.py | 243 num_threads=3, 267 num_threads: The Python integer number of threads enqueuing input examples 296 num_threads=num_threads, 387 num_threads, 412 num_threads: The Python integer number of threads enqueuing input examples 479 num_threads=num_threads, 550 num_threads=3, 595 num_threads: The Python integer number of threads enqueuing input example [all...] |
/system/extras/tests/memtest/ |
bandwidth.cpp | 56 { "num_threads", true }, 258 bool runThreadedTest(thread_arg_t args[], int num_threads, int run_time) { 259 pthread_t threads[num_threads]; 263 for (int i = 0; i < num_threads; i++) { 283 for (int i = 0; i < num_threads; i++) { 341 if (values.count("num_threads") == 0) { 342 printf("Must specify the num_threads value.\n"); 345 int num_threads = values["num_threads"].int_value; local 347 thread_arg_t args[num_threads]; [all...] |
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/test/ |
test_thread.py | 135 def __init__(self, num_threads):
136 self.num_threads = num_threads
145 if self.waiting == self.num_threads:
146 self.waiting = self.num_threads - 1
|
/external/ltp/testcases/kernel/io/ltp-aiodio/ |
aio-stress.c | 101 int num_threads = 1; variable 1011 int setup_shared_mem(int num_threads, int num_files, int depth, 1019 total_ram = num_files * depth * padded_reclen + num_threads; 1090 for (i = 0; i < num_threads; i++) { 1126 if (num_threads > 1) { 1129 if (threads_starting == num_threads) { 1134 while (threads_starting != num_threads) 1201 if (num_threads > 1) { 1204 if (threads_ending == num_threads) { 1209 while (threads_ending != num_threads) [all...] |
/external/autotest/client/tests/aiostress/ |
aio-stress.c | 100 int num_threads = 1; variable 994 int setup_shared_mem(int num_threads, int num_files, int depth, 1002 total_ram = num_files * depth * padded_reclen + num_threads; 1069 for (i = 0 ; i < num_threads ; i++) { 1106 if (num_threads > 1) { 1109 if (threads_starting == num_threads) { 1114 while (threads_starting != num_threads) 1178 if (num_threads > 1) { 1181 if (threads_ending == num_threads) { 1186 while(threads_ending != num_threads) [all...] |
/external/python/cpython3/Lib/concurrent/futures/ |
thread.py | 126 num_threads = len(self._threads) 127 if num_threads < self._max_workers: 129 num_threads)
|
/external/python/futures/concurrent/futures/ |
thread.py | 141 num_threads = len(self._threads) 142 if num_threads < self._max_workers: 144 num_threads)
|
/external/tensorflow/tensorflow/python/estimator/inputs/ |
numpy_io.py | 97 num_threads=1): 127 num_threads: Integer, number of threads used for reading and enqueueing. In 129 such as in prediction and evaluation mode, `num_threads` should be 1. 194 num_threads=num_threads,
|
/system/extras/ioshark/ |
ioshark_bench.c | 83 fprintf(stderr, "%s [-b blockdev_name] [-d preserve_delays] [-n num_iterations] [-t num_threads] -q -v | -s <list of parsed input files>\n", 607 wait_for_threads(int num_threads) 611 for (i = 0; i < num_threads; i++) { 658 int num_threads = 0; local 681 num_threads = atoi(optarg); 702 if (num_threads > MAX_THREADS) 767 if (num_threads == 0 || num_threads > num_files) 768 num_threads = num_files; 773 for (i = 0; i < num_threads; i++) [all...] |
/external/clang/test/OpenMP/ |
target_parallel_for_ast_print.cpp | 78 #pragma omp target parallel for private(argc, b), firstprivate(c, d), lastprivate(d, f) collapse(N) schedule(static, N) ordered(N) if (parallel :argc) num_threads(N) default(shared) shared(e) reduction(+ : h) 90 // CHECK-NEXT: #pragma omp target parallel for private(argc,b) firstprivate(c,d) lastprivate(d,f) collapse(N) schedule(static, N) ordered(N) if(parallel: argc) num_threads(N) default(shared) shared(e) reduction(+: h) 102 #pragma omp target parallel for default(none), private(argc,b) firstprivate(argv) shared (d) if (parallel:argc > 0) num_threads(N) proc_bind(master) reduction(+:c, arr1[argc]) reduction(max:e, arr[:N][0:10]) 104 // CHECK-NEXT: #pragma omp target parallel for default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(N) proc_bind(master) reduction(+: c,arr1[argc]) reduction(max: e,arr[:N][0:10]) 107 #pragma omp target parallel for if (N) num_threads(s) proc_bind(close) reduction(^:e, f, arr[0:N][:argc]) reduction(&& : h) 108 // CHECK-NEXT: #pragma omp target parallel for if(N) num_threads(s) proc_bind(close) reduction(^: e,f,arr[0:N][:argc]) reduction(&&: h) 181 #pragma omp target parallel for private(argc, b), firstprivate(argv, c), lastprivate(d, f) collapse(2) schedule(auto) ordered if (target: argc) num_threads(a) default(shared) shared(e) reduction(+ : h) linear(a:-5) 185 // CHECK-NEXT: #pragma omp target parallel for private(argc,b) firstprivate(argv,c) lastprivate(d,f) collapse(2) schedule(auto) ordered if(target: argc) num_threads(a) default(shared) shared(e) reduction(+: h) linear(a: -5) 189 #pragma omp target parallel for default(none), private(argc,b) firstprivate(argv) shared (d) if (parallel:argc > 0) num_threads(5) proc_bind(master) reduction(+:c, arr1[argc]) reduction(max:e, arr[:5][0:10]) 191 // CHECK-NEXT: #pragma omp target parallel for default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(5) proc_bind(master) reduction(+: c,arr1[argc]) reduction(max: e,arr[:5][0:10] [all...] |
target_parallel_for_simd_ast_print.cpp | 80 #pragma omp target parallel for simd private(argc, b), firstprivate(c, d), lastprivate(d, f) collapse(N) schedule(static, N) ordered(N) if (parallel :argc) num_threads(N) default(shared) shared(e) reduction(+ : h) 87 // CHECK-NEXT: #pragma omp target parallel for simd private(argc,b) firstprivate(c,d) lastprivate(d,f) collapse(N) schedule(static, N) ordered(N) if(parallel: argc) num_threads(N) default(shared) shared(e) reduction(+: h) 95 #pragma omp target parallel for simd default(none), private(argc,b) firstprivate(argv) shared (d) if(parallel:argc > 0) num_threads(N) proc_bind(master) reduction(+:c, arr1[argc]) reduction(max:e, arr[:N][0:10]) 97 // CHECK: #pragma omp target parallel for simd default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(N) proc_bind(master) reduction(+: c,arr1[argc]) reduction(max: e,arr[:N][0:10]) 101 #pragma omp target parallel for simd if(N) num_threads(s) proc_bind(close) reduction(^:e, f, arr[0:N][:argc]) reduction(&& : h) 103 // CHECK: #pragma omp target parallel for simd if(N) num_threads(s) proc_bind(close) reduction(^: e,f,arr[0:N][:argc]) reduction(&&: h) 206 #pragma omp target parallel for simd private(argc, b), firstprivate(argv, c), lastprivate(d, f) collapse(2) schedule(auto) ordered if (target: argc) num_threads(a) default(shared) shared(e) reduction(+ : h) linear(a:-5) 210 // CHECK: #pragma omp target parallel for simd private(argc,b) firstprivate(argv,c) lastprivate(d,f) collapse(2) schedule(auto) ordered if(target: argc) num_threads(a) default(shared) shared(e) reduction(+: h) linear(a: -5) 215 #pragma omp target parallel for simd default(none), private(argc,b) firstprivate(argv) shared (d) if (parallel:argc > 0) num_threads(5) proc_bind(master) reduction(+:c, arr1[argc]) reduction(max:e, arr[:5][0:10]) 217 // CHECK: #pragma omp target parallel for simd default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(5) proc_bind(master) reduction(+: c,arr1[argc]) reduction(max: e,arr[:5][0:10] [all...] |
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
TensorExecutor.h | 160 size_t num_threads = device.numThreads(); local 161 if (num_threads > 1) { 162 num_threads = TensorCostModel<ThreadPoolDevice>::numThreads( 163 size, evaluator.costPerCoeff(Vectorizable), num_threads); 165 if (num_threads == 1) { 169 Index blocksz = std::ceil<Index>(static_cast<float>(size)/num_threads) + PacketSize - 1;
|
/external/google-benchmark/src/ |
mutex.h | 101 Barrier(int num_threads) : running_threads_(num_threads) {}
|
/external/libcxx/utils/google-benchmark/src/ |
mutex.h | 101 Barrier(int num_threads) : running_threads_(num_threads) {}
|
/external/ltp/utils/ffsb-6.0-rc2/ |
ffsb_tg.h | 63 unsigned num_threads; member in struct:ffsb_tg 108 void init_ffsb_tg(ffsb_tg_t *tg, unsigned num_threads, unsigned tg_num);
|
/prebuilts/ndk/r16/sources/cxx-stl/llvm-libc++/utils/google-benchmark/src/ |
mutex.h | 101 Barrier(int num_threads) : running_threads_(num_threads) {}
|
/art/runtime/ |
thread_pool.cc | 123 ThreadPool::ThreadPool(const char* name, size_t num_threads, bool create_peers) 134 creation_barier_(num_threads + 1), 135 max_active_workers_(num_threads), 138 while (GetThreadCount() < num_threads) {
|