HomeSort by relevance Sort by last modified time
    Searched full:benchmark (Results 476 - 500 of 2250) sorted by null

<<11121314151617181920>>

  /system/libhwbinder/vts/performance/
Benchmark_binder.cpp 17 #include <benchmark/benchmark.h>
73 static void BM_sendVec_binder(benchmark::State& state) {
85 state.SkipWithError("Failed to retrieve benchmark service.");
93 BENCHMARK(BM_sendVec_binder)->RangeMultiplier(2)->Range(4, 65536);
96 ::benchmark::Initialize(&argc, argv);
101 ::benchmark::RunSpecifiedBenchmarks();
  /external/drrickorang/LoopbackApp/app/src/main/java/org/drrickorang/loopback/
PerformanceMeasurement.java 30 // this is used to enlarge the benchmark, so that it can be displayed with better accuracy on
64 * then a benchmark and a count on outliers can be produced as a measurement of performance.
65 * The benchmark should be as small as possible, so is the number of outliers.
94 // compute benchmark and count the number of outliers
95 double benchmark = computeWeightedBenchmark(); local
100 log("benchmark: " + benchmark);
127 * Compute a benchmark using the following formula:
130 * Also, the benchmark is additionally multiplied by mMultiplicationFactor. This is not in the
131 * original formula, and it is used only because the original benchmark will be too small t
137 double benchmark; local
    [all...]
  /external/google-benchmark/tools/
compare.py 4 compare.py - versatile benchmark output compare tool
37 description='versatile benchmark output compare tool')
46 'baseline', 'The benchmark baseline')
52 help='A benchmark executable or JSON output file')
54 'contender', 'The benchmark that will be compared against the baseline')
60 help='A benchmark executable or JSON output file')
65 help='Arguments to pass when running benchmark executables')
68 'filters', help='Compare filter one with the filter two of benchmark')
70 'baseline', 'The benchmark baseline')
76 help='A benchmark executable or JSON output file'
    [all...]
  /external/libcxx/utils/google-benchmark/tools/
compare.py 4 compare.py - versatile benchmark output compare tool
37 description='versatile benchmark output compare tool')
46 'baseline', 'The benchmark baseline')
52 help='A benchmark executable or JSON output file')
54 'contender', 'The benchmark that will be compared against the baseline')
60 help='A benchmark executable or JSON output file')
65 help='Arguments to pass when running benchmark executables')
68 'filters', help='Compare filter one with the filter two of benchmark')
70 'baseline', 'The benchmark baseline')
76 help='A benchmark executable or JSON output file'
    [all...]
  /external/tensorflow/tensorflow/tools/test/
upload_test_benchmarks.py 16 """Command to upload benchmark test results to a cloud datastore.
19 in a specified data directory, files that contain benchmark test results. The
32 The "TestResults" object contains test metadata and multiple benchmark entries.
35 each related benchmark entry in a separate "Entry" Entity. Datastore create a
46 - entry: unique name of this benchmark entry within this test (string)
52 retrieval of benchmark data and reduced I/O to the client without adding a lot
58 fetch the recent (start, timing) data for a given test/entry benchmark.
66 * Get the recent timings for a given benchmark:
120 # Note: The file locking code uses flock() instead of lockf() because benchmark
141 """Parse benchmark data and use the client to upload it to the datastore
    [all...]
  /external/perfetto/buildtools/
BUILD.gn 480 include_dirs = [ "benchmark/include" ]
484 source_set("benchmark") {
487 "benchmark/include/benchmark/benchmark.h",
488 "benchmark/include/benchmark/benchmark_api.h",
489 "benchmark/include/benchmark/reporter.h",
490 "benchmark/src/arraysize.h"
    [all...]
  /bionic/benchmarks/
semaphore_benchmark.cpp 23 #include <benchmark/benchmark.h>
26 static void BM_semaphore_sem_getvalue(benchmark::State& state) {
37 static void BM_semaphore_sem_wait_sem_post(benchmark::State& state) {
62 // with the benchmark clock running. This ensures nothing else in the system
81 class SemaphoreFixture : public benchmark::Fixture {
83 void SetUp(const benchmark::State&) {
123 // This is commented out because dynamic benchmark registering doesn't currently support fixtures.
124 // Uncomment it and recompile to run this benchmark on every run.
125 /* BENCHMARK_F(SemaphoreFixture, semaphore_sem_post)(benchmark::State& state)
    [all...]
  /external/tensorflow/tensorflow/core/util/
test_log.proto 20 // Each unit test or benchmark in a test or benchmark run provides
25 // This BenchmarkEntry should be emitted by each unit test or benchmark
28 // The name of the specific benchmark or test
32 // If a benchmark, how many iterations it was run for
119 // Host name of machine that ran the benchmark.
140 // Run-specific items such as arguments to the test / benchmark.
145 // The output of one benchmark / test run. Each run contains a list of
178 // Benchmark target identifier.
181 // The type of benchmark
    [all...]
  /external/trappy/trappy/wa/
results.py 27 """A DataFrame-like class for storing benchmark results"""
62 def plot_results_benchmark(self, benchmark, title=None):
63 """Plot the results of the execution of a given benchmark
69 title = benchmark.replace('_', ' ')
72 self[benchmark].plot(ax=self.ax, kind="bar", title=title)
100 benchmark name and the second the id within it. For benchmarks
154 has two column indexes. The first one is the benchmark and the
160 for benchmark in data[0].columns.levels[0]:
161 concat_objs = [d[benchmark] for d in data]
162 res_dict[benchmark] = pd.concat(concat_objs, axis=1
    [all...]
  /external/v8/tools/
perf-to-html.py 113 class Benchmark:
171 def RenderOneBenchmark(self, benchmark):
173 self.Print("<a name=\"" + benchmark.name() + "\">")
174 self.Print(benchmark.name() + "</a> <a href=\"#top\">(top)</a>")
176 self.Print("<table class=\"benchmark\">")
184 tests = benchmark.tests()
185 for test in benchmark.SortedTestKeys():
207 for benchmark in data:
208 if benchmark != "errors":
209 self.Print("<li><a href=\"#" + benchmark + "\">" + benchmark + "</a></li>"
    [all...]
try_perf.py 79 for benchmark in options.benchmarks:
80 if benchmark not in PUBLIC_BENCHMARKS:
81 print ('%s not found in our benchmark list. The respective trybot might '
83 'Available public benchmarks: %s' % (benchmark, PUBLIC_BENCHMARKS))
99 benchmarks = ['"%s"' % benchmark for benchmark in options.benchmarks]
  /cts/suite/cts/utils/
grapher.py 24 A simple script to render the data from the benchmark as a graph.
44 for benchmark in tests:
45 if benchmark.startswith('com.android.cts.opengl.primitive'):
46 results = tests[benchmark]
51 plt.title(benchmark[benchmark.index('#') + 1:])
  /cts/tests/openglperf2/src/android/opengl2/cts/primitive/
GLPrimitiveBenchmark.java 113 * @param benchmark An enum representing the benchmark to run.
117 * @param timeout The milliseconds to wait for an iteration of the benchmark before timing out.
119 * @throws Exception If the benchmark could not be run.
121 private void runBenchmark(BenchmarkName benchmark, boolean offscreen, int numFrames,
123 String benchmarkName = benchmark.toString();
  /external/caliper/caliper/src/test/java/com/google/caliper/runner/
ExperimentingRunnerModuleTest.java 24 import com.google.caliper.Benchmark;
108 fail("should have thrown for invalid benchmark method name");
115 @Benchmark void a() {}
116 @Benchmark void b() {}
117 @Benchmark void c() {}
136 public void dryRun(Object benchmark) throws InvalidBenchmarkException {}
  /external/guava/guava-tests/benchmark/com/google/common/math/
IntMathBenchmark.java 27 import com.google.caliper.Benchmark;
55 @Benchmark int pow(int reps) {
64 @Benchmark int mod(int reps) {
73 @Benchmark int gCD(int reps) {
82 @Benchmark int factorial(int reps) {
91 @Benchmark int binomial(int reps) {
LongMathBenchmark.java 27 import com.google.caliper.Benchmark;
58 @Benchmark int pow(int reps) {
67 @Benchmark int mod(int reps) {
76 @Benchmark int gCD(int reps) {
85 @Benchmark int factorial(int reps) {
94 @Benchmark int binomial(int reps) {
  /external/lz4/tests/
README.md 23 Then it compiles all branches of lz4 and performs a speed benchmark for a given list of files (the `testFileNames` parameter).
25 If a new commit is found it is compiled and a speed benchmark for this commit is performed.
26 The results of the speed benchmark are compared to the previous results.
27 If compression or decompression speed for one of lz4 levels is lower than `lowerLimit` (an optional parameter, default 0.98) the speed benchmark is restarted.
33 - The speed benchmark is not performed until computers' load average is lower than `maxLoadAvg` (an optional parameter, default 0.75)
50 testFileNames file names list for speed benchmark
  /external/skia/bench/
SortBench.cpp 8 #include "Benchmark.h"
101 class SortBench : public Benchmark {
141 typedef Benchmark INHERITED;
146 static Benchmark* NewSkQSort(Type t) {
149 static Benchmark* NewSkHeap(Type t) {
152 static Benchmark* NewQSort(Type t) {
  /external/skqp/bench/
SortBench.cpp 8 #include "Benchmark.h"
101 class SortBench : public Benchmark {
141 typedef Benchmark INHERITED;
146 static Benchmark* NewSkQSort(Type t) {
149 static Benchmark* NewSkHeap(Type t) {
152 static Benchmark* NewQSort(Type t) {
  /external/tensorflow/tensorflow/core/kernels/
sendrecv_ops_test.cc 26 // clouding the benchmark results with the time spent in the various
28 // errors that would arise from running either benchmark in a loop.
60 test::Benchmark("cpu", Send(), nullptr, nullptr, new DummyRendezvous)
63 BENCHMARK(BM_Send);
68 test::Benchmark("cpu", Recv(), nullptr, nullptr, new DummyRendezvous)
71 BENCHMARK(BM_Recv);
  /frameworks/base/tests/JankBench/app/src/main/java/com/android/benchmark/ui/
EditTextInputActivity.java 17 package com.android.benchmark.ui;
26 import com.android.benchmark.R;
27 import com.android.benchmark.ui.automation.Automator;
28 import com.android.benchmark.ui.automation.Interaction;
37 final int runId = getIntent().getIntExtra("com.android.benchmark.RUN_ID", 0);
38 final int iteration = getIntent().getIntExtra("com.android.benchmark.ITERATION", -1);
ListViewScrollActivity.java 17 package com.android.benchmark.ui;
28 import com.android.benchmark.R;
29 import com.android.benchmark.ui.automation.Automator;
30 import com.android.benchmark.ui.automation.Interaction;
45 final int runId = getIntent().getIntExtra("com.android.benchmark.RUN_ID", 0);
46 final int iteration = getIntent().getIntExtra("com.android.benchmark.ITERATION", -1);
ShadowGridActivity.java 16 package com.android.benchmark.ui;
27 import com.android.benchmark.R;
28 import com.android.benchmark.ui.automation.Automator;
29 import com.android.benchmark.ui.automation.Interaction;
44 final int runId = getIntent().getIntExtra("com.android.benchmark.RUN_ID", 0);
45 final int iteration = getIntent().getIntExtra("com.android.benchmark.ITERATION", -1);
TextScrollActivity.java 17 package com.android.benchmark.ui;
25 import com.android.benchmark.registry.BenchmarkRegistry;
26 import com.android.benchmark.ui.automation.Automator;
27 import com.android.benchmark.ui.automation.Interaction;
46 final int runId = getIntent().getIntExtra("com.android.benchmark.RUN_ID", 0);
47 final int iteration = getIntent().getIntExtra("com.android.benchmark.ITERATION", -1);
  /cts/tests/openglperf2/src/android/opengl2/cts/reference/
GLReferenceActivity.java 44 // Start benchmark
68 // Benchmark passed
73 setException(new Exception("Benchmark failed to run"));

Completed in 867 milliseconds

<<11121314151617181920>>