/external/eigen/bench/tensors/ |
tensor_benchmarks_sycl.cc | 27 static void BM_##FUNC(int iters, int N) { \ 32 suite.FUNC(iters); \
|
/prebuilts/go/darwin-x86/src/cmd/compile/internal/gc/ |
reproduciblebuilds_test.go | 19 iters := 10 21 iters = 4 31 for i := 0; i < iters; i++ { 44 t.Fatalf("builds produced different output after %d iters (%d bytes vs %d bytes)", i, len(want), len(obj))
|
/prebuilts/go/linux-x86/src/cmd/compile/internal/gc/ |
reproduciblebuilds_test.go | 19 iters := 10 21 iters = 4 31 for i := 0; i < iters; i++ { 44 t.Fatalf("builds produced different output after %d iters (%d bytes vs %d bytes)", i, len(want), len(obj))
|
/art/test/055-enum-performance/src/ |
Main.java | 96 static public int basis(int iters) { 97 for (int i = iters; i > 0; i--) { 120 return iters * 20; 132 static public int test1(int iters) { 134 for (int i = iters; i > 0; i--) { 157 return iters * 20; 160 static public int test2(int iters) { 162 for (int i = iters; i > 0; i--) { 185 return iters * 20; 188 static public int test3(int iters) { [all...] |
/external/autotest/client/common_lib/ |
decorators_unittest.py | 30 iters = 100 38 args=(self.inc_count, iters)) 44 self.assertEquals(iters * num_threads, self.count)
|
/external/tensorflow/tensorflow/core/kernels/ |
adjust_contrast_op_benchmark_test.cc | 40 static void BM_AdjustContrast_##DEVICE##_##B##_##W##_##H(int iters) { \ 41 testing::ItemsProcessed(iters* B* W* H * 3); \ 42 test::Benchmark(#DEVICE, BM_AdjustContrast(B, W, H)).Run(iters); \
|
basic_ops_benchmark_test.cc | 67 static void BM_MulChain(int iters, int chain_length) { 68 const int64 tot = static_cast<int64>(iters) * chain_length; 73 test::Benchmark("cpu", run, GetOptions(), init).Run(iters);
|
dequantize_op_test.cc | 154 static void BM_DequantizeMinCombinedCpu(int iters) { 167 test::Benchmark("cpu", g).Run(iters); 168 testing::BytesProcessed(iters * num_values * (sizeof(float) + sizeof(T))); 169 testing::ItemsProcessed(iters); 172 static void BM_DequantizeMinCombinedCpuQuint16(int iters) { 173 BM_DequantizeMinCombinedCpu<quint16>(iters); 176 static void BM_DequantizeMinCombinedCpuQint16(int iters) { 177 BM_DequantizeMinCombinedCpu<qint16>(iters); 180 static void BM_DequantizeMinCombinedCpuQuint8(int iters) { 181 BM_DequantizeMinCombinedCpu<quint8>(iters); [all...] |
diag_op_test.cc | 34 static void BM_Diag##_##N##_##TFTYPE##_##DEVICE(int iters) { \ 36 testing::ItemsProcessed(static_cast<int64>(iters) * N * N); \ 37 test::Benchmark(#DEVICE, Diag<T>(N, TFTYPE)).Run(iters); \
|
multinomial_op_test.cc | 44 static void BM_Multinomial_##DEVICE##_##B##_##C##_##S(int iters) { \ 45 test::Benchmark(#DEVICE, Multinomial(B, C, S)).Run(iters); \ 46 testing::ItemsProcessed(static_cast<int64>(B) * C * S * iters); \
|
resize_op_benchmark_test.cc | 45 static void BM_Resize_##ALGORITHM##_##DEVICE##_##B##_##W##_##H(int iters) { \ 46 testing::ItemsProcessed(iters* B* W* H * 3); \ 47 test::Benchmark(#DEVICE, BM_Resize(#ALGORITHM, B, W, H)).Run(iters); \
|
xent_op_test.cc | 37 static void BM_Xent##_##BATCH##_##CLASS##_##DEVICE(int iters) { \ 38 testing::ItemsProcessed(static_cast<int64>(iters) * BATCH * CLASS); \ 39 test::Benchmark(#DEVICE, Xent(BATCH, CLASS)).Run(iters); \
|
training_ops_test.cc | 77 static void BM_SGD(int iters, int params) { 78 const int64 tot = static_cast<int64>(iters) * params; 84 test::Benchmark("cpu", train, GetOptions(), init).Run(iters); 109 static void BM_Adagrad(int iters, int params) { 110 const int64 tot = static_cast<int64>(iters) * params; 116 test::Benchmark("cpu", train, GetOptions(), init).Run(iters); 143 static void BM_Momentum(int iters, int params) { 144 const int64 tot = static_cast<int64>(iters) * params; 150 test::Benchmark("cpu", train, GetOptions(), init).Run(iters); 186 static void BM_Adam(int iters, int params) [all...] |
scatter_nd_op_test.cc | 257 static void BM_ScatterNdHelper(int iters, int embedding_size, const char* op) { 284 iters); 286 while (iters-- > 0) { 292 static void BM_ScatterNdUpdateInt32(int iters, int embedding_size) { 293 BM_ScatterNdHelper<int32>(iters, embedding_size, "ScatterNdUpdate"); 295 static void BM_ScatterNdUpdateInt64(int iters, int embedding_size) { 296 BM_ScatterNdHelper<int64>(iters, embedding_size, "ScatterNdUpdate"); 299 static void BM_ScatterNdAddInt32(int iters, int embedding_size) { 300 BM_ScatterNdHelper<int32>(iters, embedding_size, "ScatterNdAdd"); 302 static void BM_ScatterNdAddInt64(int iters, int embedding_size) [all...] |
quantized_concat_op_test.cc | 251 static void ConcatHelper(int iters, int concat_dimension, bool same_limits, 281 testing::BytesProcessed(static_cast<int64>(iters) * 284 test::Benchmark("cpu", g).Run(iters); 288 static void BM_QConcatDim0SameLimitQInt32(int iters, int dim2) { 289 ConcatHelper<qint32>(iters, 0 /* concat_dimension */, true /* same_limits */, 293 static void BM_QConcatDim1SameLimitQInt32(int iters, int dim2) { 294 ConcatHelper<qint32>(iters, 1 /* concat_dimension */, true /* same_limits */, 298 static void BM_QConcatDim0DifferLimitQInt32(int iters, int dim2) { 299 ConcatHelper<qint32>(iters, 0 /* concat_dimension */, false /* same_limits */, 303 static void BM_QConcatDim1DifferLimitQInt32(int iters, int dim2) [all...] |
cwise_ops_test.cc | 46 void BM_##DEVICE##_##FUNC##_##TYPE(int iters, int num) { \ 47 const int64 tot = static_cast<int64>(iters) * num; \ 50 test::Benchmark(#DEVICE, Unary<T>(#FUNC, num, TYPE)).Run(iters); \ 101 void BM_##DEVICE##_##FUNC##_scalar(int iters, int num) { \ 102 const int64 tot = static_cast<int64>(iters) * num; \ 105 test::Benchmark(#DEVICE, BinaryScalar(num, #FUNC)).Run(iters); \ 145 void BM_##DEVICE##_##C_TYPE##_BiasAdd_R##R##_C##C(int iters, int arg) { \ 148 const int64 tot = static_cast<int64>(iters) * rows * cols; \ 151 test::Benchmark(#DEVICE, BiasAdd<C_TYPE>(rows, cols, TF_TYPE)).Run(iters); \ 196 int iters, int arg, int channels) { [all...] |
nn_ops_test.cc | 105 static void BM_ConvFloat(int iters, int batch, int rows, int cols, int in_depth, 223 test::Benchmark(device, g, &options).Run(iters); 224 testing::ItemsProcessed(num_ops * iters); 235 static void BM_ConvFloatFwdCPU1_##LABEL(int iters) { \ 236 BM_ConvFloat(iters, BS, R, C, ID, OD, KR, KC, CONV_OP_FORWARD, 1, STR, \ 241 static void BM_ConvFloatFwdCPU4_##LABEL(int iters) { \ 242 BM_ConvFloat(iters, BS, R, C, ID, OD, KR, KC, CONV_OP_FORWARD, 4, STR, \ 247 static void BM_ConvFloatFusedCPU1_##LABEL(int iters) { \ 248 BM_ConvFloat(iters, BS, R, C, ID, OD, KR, KC, CONV_OP_FUSED, 1, STR, PAD, \ 253 static void BM_ConvFloatFusedCPU4_##LABEL(int iters) { \ [all...] |
split_v_op_test.cc | 76 static void BM_SplitV_1d_##num_split##_##total_size(int iters) { \ 78 testing::ItemsProcessed(static_cast<int64>(iters) * total_size); \ 87 test::Benchmark("cpu", g).Run(iters); \ 94 int iters) { \ 97 testing::ItemsProcessed(static_cast<int64>(iters) * total_size0 * \ 109 test::Benchmark("cpu", g).Run(iters); \
|
/external/tensorflow/tensorflow/contrib/factorization/kernels/ |
clustering_ops_test.cc | 75 void BM_KmeansPlusPlusInitialization(int iters) { 77 testing::ItemsProcessed(static_cast<int64>(iters) * num_points * num_dims * 83 test::Benchmark("cpu", g).Run(iters); 87 void BM_KmeansPlusPlusInitialization_##p##_##c##_##d##_##r(int iters) { \ 88 BM_KmeansPlusPlusInitialization<p, c, d, r>(iters); \ 135 void BM_KMC2Initialization(int iters) { 137 testing::ItemsProcessed(static_cast<int64>(iters) * num_points * num_dims * 142 test::Benchmark("cpu", g).Run(iters); 145 void BM_KMC2Initialization_##p##_##c##_##d(int iters) { \ 146 BM_KMC2Initialization<p, c, d>(iters); \ [all...] |
/external/tensorflow/tensorflow/core/lib/random/ |
distribution_sampler_test.cc | 35 int iters = weights.size() * trials_per_bin; local 41 for (int i = 0; i < iters; i++) { 49 counts[i] /= iters; 85 static void BM_DistributionSampler(int iters, int n) { 96 for (int i = 0; i < iters; i++) {
|
/external/tensorflow/tensorflow/c/eager/ |
runtime_test.cc | 109 void BM_CreateGraph(int iters) { 110 for (int i = 0; i < iters; ++i) { 119 void BM_RunGraph(int iters) { 130 for (int i = 0; i < iters; ++i) { 137 void BM_CreateAndDestroySession(int iters) { 143 for (int i = 0; i < iters; ++i) { 149 void BM_KernelAndDeviceInit(int iters) { 160 for (int i = 0; i < iters; ++i) { 167 void BM_KernelAndDeviceRun(int iters) { 185 for (int i = 0; i < iters; ++i) [all...] |
/prebuilts/go/darwin-x86/src/testing/ |
benchmark_test.go | 69 iters := uint64(0) 74 atomic.AddUint64(&iters, 1) 80 if iters != uint64(b.N) { 81 t.Errorf("got %v iters, want %v", iters, b.N)
|
/prebuilts/go/linux-x86/src/testing/ |
benchmark_test.go | 69 iters := uint64(0) 74 atomic.AddUint64(&iters, 1) 80 if iters != uint64(b.N) { 81 t.Errorf("got %v iters, want %v", iters, b.N)
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
basic_batch_scheduler_benchmark_test.cc | 118 void RunBenchmark(int iters); 139 void ThroughputBenchmark::RunBenchmark(int iters) { 140 CHECK_GE(iters, 1); 149 testing::ItemsProcessed(iters * kNumTasksPerIteration); 154 for (int i = 0; i < iters; ++i) { 341 static void RunThroughputBenchmark(int iters, int64 batch_timeout_micros, 350 benchmark.RunBenchmark(iters); 353 static void ThroughputBM_ZeroTimeout(int iters, int num_batch_threads) { 354 RunThroughputBenchmark(iters, 0 /* 0 ms timeout */, num_batch_threads); 365 static void ThroughputBM_SmallTimeout(int iters, int num_batch_threads) [all...] |
/external/tensorflow/tensorflow/core/framework/ |
bfloat16_test.cc | 119 static void BM_FloatToBFloat16(int iters) { 122 const int64 tot = static_cast<int64>(iters) * N; 130 while (iters--) { 138 static void BM_BFloat16ToFloat(int iters) { 141 const int64 tot = static_cast<int64>(iters) * N; 149 while (iters--) {
|