HomeSort by relevance Sort by last modified time
    Searched refs:num_threads (Results 51 - 75 of 323) sorted by null

1 23 4 5 6 7 8 91011>>

  /external/google-breakpad/src/common/linux/tests/
crash_generator.cc 172 unsigned num_threads, unsigned crash_thread, int crash_signal,
174 if (num_threads == 0 || crash_thread >= num_threads) {
175 fprintf(stderr, "CrashGenerator: Invalid thread counts; num_threads=%u"
176 " crash_thread=%u\n", num_threads, crash_thread);
180 if (!MapSharedMemory(num_threads * sizeof(pid_t))) {
192 CreateThreadsInChildProcess(num_threads);
275 void CrashGenerator::CreateThreadsInChildProcess(unsigned num_threads) {
278 if (num_threads <= 1)
283 ThreadData* thread_data = new ThreadData[num_threads];
    [all...]
  /external/ltp/testcases/misc/math/float/
main.c 59 int num_threads = DEFAULT_NUM_THREADS; variable
151 num_threads = atoi(optarg);
182 if (num_threads <= 0) {
184 "num_threads undefined or incorrect, using \"1\"");
185 num_threads = 1;
188 if (nb_func * num_threads > PTHREAD_THREADS_MAX - 2)
189 while (nb_func * num_threads > PTHREAD_THREADS_MAX - 2)
190 num_threads--;
194 argv[0], nb_func, num_threads);
207 threads = malloc(nb_func * num_threads * sizeof(pthread_t))
    [all...]
tfloat.h 60 extern int true, num_threads;
67 * th_num thread # (range: 0 ... num_threads-1)
  /external/tensorflow/tensorflow/contrib/learn/python/learn/learn_io/
graph_io.py 48 num_threads=1,
78 num_threads: The number of threads enqueuing examples. In order to have
80 prediction and evaluation mode, `num_threads` should be 1.
101 num_threads=num_threads,
115 num_threads=1,
145 num_threads: The number of threads enqueuing examples. In order to have
147 prediction and evaluation mode, `num_threads` should be 1.
170 num_threads=num_threads,
    [all...]
pandas_io.py 56 num_threads=1,
65 num_threads=num_threads,
generator_io.py 34 num_threads=1,
72 num_threads: Integer, number of threads used for reading and enqueueing.
121 num_threads=num_threads,
  /external/eigen/unsupported/Eigen/CXX11/src/ThreadPool/
NonBlockingThreadPool.h 22 NonBlockingThreadPoolTempl(int num_threads, Environment env = Environment())
24 threads_(num_threads),
25 queues_(num_threads),
26 coprimes_(num_threads),
27 waiters_(num_threads),
32 waiters_.resize(num_threads);
34 // Calculate coprimes of num_threads.
37 // a walk starting thread index t and calculate num_threads - 1 subsequent
38 // indices as (t + coprime) % num_threads, we will cover all threads without
41 for (int i = 1; i <= num_threads; i++)
    [all...]
SimpleThreadPool.h 23 // Construct a pool that contains "num_threads" threads.
24 explicit SimpleThreadPoolTempl(int num_threads, Environment env = Environment())
25 : env_(env), threads_(num_threads), waiters_(num_threads) {
26 for (int i = 0; i < num_threads; i++) {
  /external/clang/test/OpenMP/
parallel_sections_ast_print.cpp 40 #pragma omp parallel sections default(none), private(argc, b) firstprivate(argv) shared(d) if (parallel: argc > 0) num_threads(C) copyin(S < T > ::TS) proc_bind(master) reduction(+ : c) reduction(max : e)
44 #pragma omp parallel sections if (C) num_threads(s) proc_bind(close) reduction(^ : e, f) reduction(&& : g) lastprivate(b, c)
61 // CHECK-NEXT: #pragma omp parallel sections default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(5) copyin(S<int>::TS) proc_bind(master) reduction(+: c) reduction(max: e)
65 // CHECK-NEXT: #pragma omp parallel sections if(5) num_threads(s) proc_bind(close) reduction(^: e,f) reduction(&&: g) lastprivate(b,c)
79 // CHECK-NEXT: #pragma omp parallel sections default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(1) copyin(S<long>::TS) proc_bind(master) reduction(+: c) reduction(max: e)
83 // CHECK-NEXT: #pragma omp parallel sections if(1) num_threads(s) proc_bind(close) reduction(^: e,f) reduction(&&: g) lastprivate(b,c)
97 // CHECK-NEXT: #pragma omp parallel sections default(none) private(argc,b) firstprivate(argv) shared(d) if(parallel: argc > 0) num_threads(C) copyin(S<T>::TS) proc_bind(master) reduction(+: c) reduction(max: e)
101 // CHECK-NEXT: #pragma omp parallel sections if(C) num_threads(s) proc_bind(close) reduction(^: e,f) reduction(&&: g) lastprivate(b,c)
125 #pragma omp parallel sections default(none), private(argc, b) firstprivate(argv) if (argc > 0) num_threads(ee) copyin(a) proc_bind(spread) reduction(| : c, d) reduction(* : e) lastprivate(argv)
126 // CHECK-NEXT: #pragma omp parallel sections default(none) private(argc,b) firstprivate(argv) if(argc > 0) num_threads(ee) copyin(a) proc_bind(spread) reduction(|: c,d) reduction(*: e) lastprivate(argv
    [all...]
  /external/ltp/utils/ffsb-6.0-rc2/
ffsb_tg.c 26 void init_ffsb_tg(ffsb_tg_t * tg, unsigned num_threads, unsigned tg_num)
31 tg->threads = ffsb_malloc(sizeof(ffsb_thread_t) * num_threads);
33 tg->num_threads = num_threads;
38 for (i = 0; i < num_threads; i++)
45 for (i = 0; i < tg->num_threads; i++)
74 for (i = 0; i < tg->num_threads; i++) {
91 for (i = 0; i < tg->num_threads; i++)
150 return tg->num_threads;
159 for (i = 0; i < tg->num_threads; i++
    [all...]
  /external/tensorflow/tensorflow/core/kernels/batching_util/
fake_clock_env.h 51 // Blocks until there are at least num_threads sleeping.
52 void BlockUntilThreadsAsleep(int num_threads);
fake_clock_env.cc 56 void FakeClockEnv::BlockUntilThreadsAsleep(int num_threads) {
60 if (num_threads <= sleeping_threads_.size()) {
  /external/tensorflow/tensorflow/core/kernels/
diag_op_gpu.cu.cc 31 __global__ void DiagCudaKernel(const int num_threads, const int64 size,
33 CUDA_1D_KERNEL_LOOP(index, num_threads) {
85 __global__ void DiagPartCudaKernel(const int num_threads, const int64 size,
87 CUDA_1D_KERNEL_LOOP(index, num_threads) {
matrix_set_diag_op_gpu.cu.cc 30 __global__ void MatrixSetDiagKernel(const int num_threads, const int m,
34 CUDA_1D_KERNEL_LOOP(index, num_threads) {
44 const int num_threads, const int m, const int n, const int minsize,
46 CUDA_1D_KERNEL_LOOP(index, num_threads) {
eye_functor_gpu.cu.cc 33 __global__ void EyeKernel(int num_threads, int batch_size, int m, int n,
37 CUDA_1D_KERNEL_LOOP(index, num_threads) {
matrix_band_part_op_gpu.cu.cc 32 __global__ void MatrixBandPartKernel(const int num_threads,
38 CUDA_1D_KERNEL_LOOP(index, num_threads) {
xsmm_conv2d.cc 223 int num_threads = worker_threads->num_threads; local
228 if (blocksofm > num_threads) {
230 BlockingCounter count(num_threads);
231 for (int i = 0; i < num_threads; ++i) {
233 int start = work / num_threads * i;
234 int end = (start + work / num_threads) > work
236 : start + work / num_threads;
246 int num_threads = work; local
248 BlockingCounter count(num_threads);
    [all...]
  /external/eigen/unsupported/Eigen/CXX11/src/Tensor/
TensorContractionThreadPool.h 43 const Index num_threads; member in struct:Eigen::internal::packRhsAndKernelArg
205 int num_threads = TensorCostModel<ThreadPoolDevice>::numThreads( local
210 if (n == 1) num_threads = 1;
212 if (num_threads == 1) {
226 shard_by_col = shardByCol(m, n, num_threads);
230 blocking(k, m, n, num_threads);
237 blocking(k, m, n, num_threads);
257 gm = coarsenM(m, n, bm, bn, bk, gn, num_threads, shard_by_col);
258 gn = coarsenN(m, n, bm, bn, bk, gm, num_threads, shard_by_col);
260 gn = coarsenN(m, n, bm, bn, bk, gm, num_threads, shard_by_col)
796 const Index num_threads = this->m_device.numThreads(); local
    [all...]
  /external/mesa3d/src/gallium/auxiliary/util/
u_queue.c 178 unsigned num_threads)
184 queue->num_threads = num_threads;
198 queue->threads = (pipe_thread*)CALLOC(num_threads, sizeof(pipe_thread));
203 for (i = 0; i < num_threads; i++) {
218 queue->num_threads = i;
252 for (i = 0; i < queue->num_threads; i++)
254 queue->num_threads = 0;
  /external/tensorflow/tensorflow/core/common_runtime/
local_device.cc 42 eigen_worker_threads_.num_threads = intra_op_parallelism_threads;
48 eigen_threadpool_wrapper_.get(), eigen_worker_threads_.num_threads));
  /external/gemmlowp/internal/
block_params.h 46 void Init(int rows, int cols, int depth, int num_threads, int l1_bytes_to_use,
48 FindL2BlockSizes<KernelFormat>(rows, cols, depth, num_threads,
56 static void FindL2BlockSizes(int rows, int cols, int depth, int num_threads,
65 std::max(1, RoundUp<KernelFormat::kRows>(rows) / num_threads);
91 (num_threads * (l2_depth + 4 * l2_cols)));
  /external/autotest/client/tests/sysbench/
sysbench.py 32 num_threads = utils.count_cpus(), max_time = 60, \
53 self.execute_pgsql(build, num_threads, max_time, read_only, args)
55 self.execute_mysql(build, num_threads, max_time, read_only, args)
58 def execute_pgsql(self, build, num_threads, max_time, read_only, args):
86 ' --num-threads=' + str(num_threads) + \
103 def execute_mysql(self, build, num_threads, max_time, read_only, args):
131 ' --num-threads=' + str(num_threads) + \
  /external/autotest/client/tests/signaltest/src/
signaltest.c 223 static int num_threads = 2; variable
255 case 't': num_threads = atoi(optarg); break;
264 if (num_threads < 2)
337 par = calloc(num_threads, sizeof(struct thread_param));
340 stat = calloc(num_threads, sizeof(struct thread_stat));
344 for (i = 0; i < num_threads; i++) {
371 for (i = 0; i < num_threads; i++) {
378 for (i = 0; i < num_threads - 1; i++)
413 for (i = 0; i < num_threads; i++) {
  /external/tensorflow/tensorflow/python/estimator/inputs/
pandas_io.py 45 num_threads=1,
60 num_threads: Integer, number of threads used for reading and enqueueing. In
62 such as in prediction and evaluation mode, `num_threads` should be 1.
108 num_threads=num_threads,
  /external/tensorflow/tensorflow/tools/benchmark/
benchmark_model.h 34 Status InitializeSession(int num_threads, const string& graph,

Completed in 877 milliseconds

1 23 4 5 6 7 8 91011>>