/external/autotest/client/site_tests/platform_LibCBench/ |
platform_LibCBench.py | 46 results = {} 60 results.setdefault(current_benchmark, []).append(time) 63 for benchmark in results: 64 average = sum(results[benchmark]) / len(results[benchmark]) 65 minimum = min(results[benchmark]) 66 maximum = max(results[benchmark])
|
/external/icu/icu4c/source/i18n/ |
csrmbcs.h | 76 UBool match(InputText* input, CharsetMatch *results) const = 0; 105 UBool match(InputText* input, CharsetMatch *results) const; 127 UBool match(InputText* input, CharsetMatch *results) const = 0; 149 UBool match(InputText* input, CharsetMatch *results) const; 164 UBool match(InputText* input, CharsetMatch *results) const; 182 UBool match(InputText* input, CharsetMatch *results) const; 201 UBool match(InputText* input, CharsetMatch *results) const;
|
/prebuilts/gdb/darwin-x86/lib/python2.7/lib2to3/fixes/ |
fix_has_key.py | 72 def transform(self, node, results): 73 assert results 80 negation = results.get("negation") 81 anchor = results["anchor"] 83 before = [n.clone() for n in results["before"]] 84 arg = results["arg"].clone() 85 after = results.get("after")
|
/prebuilts/gdb/linux-x86/lib/python2.7/lib2to3/fixes/ |
fix_has_key.py | 72 def transform(self, node, results): 73 assert results 80 negation = results.get("negation") 81 anchor = results["anchor"] 83 before = [n.clone() for n in results["before"]] 84 arg = results["arg"].clone() 85 after = results.get("after")
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/lib2to3/fixes/ |
fix_has_key.py | 72 def transform(self, node, results): 73 assert results 80 negation = results.get("negation") 81 anchor = results["anchor"] 83 before = [n.clone() for n in results["before"]] 84 arg = results["arg"].clone() 85 after = results.get("after")
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/lib2to3/fixes/ |
fix_has_key.py | 72 def transform(self, node, results): 73 assert results 80 negation = results.get("negation") 81 anchor = results["anchor"] 83 before = [n.clone() for n in results["before"]] 84 arg = results["arg"].clone() 85 after = results.get("after")
|
/external/ImageMagick/PerlMagick/demo/ |
compose-specials.pl | 48 my $results=Image::Magick->new(); 63 push(@$results, $clone); 72 push(@$results, $clone); 81 push(@$results, $clone); 90 push(@$results, $clone); 103 push(@$results, $clone); 113 push(@$results, $clone); 123 push(@$results, $clone); 133 push(@$results, $clone); 147 push(@$results, $clone) [all...] |
/external/eclipse-basebuilder/basebuilder-3.6.2/org.eclipse.releng.basebuilder/plugins/org.eclipse.test.performance.ui/src/org/eclipse/test/internal/performance/results/model/ |
ComponentResultsElement.java | 11 package org.eclipse.test.internal.performance.results.model; 21 import org.eclipse.test.internal.performance.results.db.AbstractResults; 22 import org.eclipse.test.internal.performance.results.db.ComponentResults; 23 import org.eclipse.test.internal.performance.results.db.PerformanceResults; 24 import org.eclipse.test.internal.performance.results.db.ScenarioResults; 25 import org.eclipse.test.internal.performance.results.utils.IPerformancesConstants; 26 import org.eclipse.test.internal.performance.results.utils.Util; 54 // Results category 56 NAME_DESCRIPTOR.setCategory("Results"); 58 CURRENT_BUILD_DESCRIPTOR.setCategory("Results"); [all...] |
/frameworks/base/core/tests/bandwidthtests/src/com/android/bandwidthtest/ |
BandwidthTest.java | 128 Bundle results = new Bundle(); local 129 results.putString("device_id", mDeviceId); 130 results.putString("timestamp", ts); 131 results.putInt("size", FILE_SIZE); 132 addStatsToResults(PROF_LABEL, prof_stats, results, mUid); 133 addStatsToResults(PROC_LABEL, proc_stats, results, mUid); 134 getInstrumentation().sendStatus(INSTRUMENTATION_IN_PROGRESS, results); 182 Bundle results = new Bundle(); local 183 results.putString("device_id", mDeviceId); 184 results.putString("timestamp", ts) 239 Bundle results = new Bundle(); local [all...] |
/external/chromium-trace/catapult/telemetry/telemetry/web_perf/metrics/ |
blob_timeline.py | 17 The following metrics are added to the results: 44 def AddResults(self, model, renderer_thread, interactions, results): 57 self._AddWriteResultsInternal(write_events, interactions, results) 58 self._AddReadResultsInternal(read_events, interactions, results) 60 def _AddWriteResultsInternal(self, events, interactions, results): 68 results.AddValue(list_of_scalar_values.ListOfScalarValues( 69 page=results.current_page, 77 results.AddValue(list_of_scalar_values.ListOfScalarValues( 78 page=results.current_page, 86 def _AddReadResultsInternal(self, events, interactions, results) [all...] |
trace_event_stats_unittest.py | 62 results = test_page_test_results.TestPageTestResults(self) 64 interactions, results) 65 return results 81 results = self.RunAggregator(aggregator, interactions) 82 results.AssertHasPageSpecificScalarValue('metric-name-count', 'count', 2) 83 results.AssertHasPageSpecificListOfScalarValues( 109 results = self.RunAggregator(aggregator, interactions) 110 results.AssertHasPageSpecificScalarValue('metric-name-count', 'count', 3) 111 results.AssertHasPageSpecificListOfScalarValues( 130 results = self.RunAggregator(aggregator, interactions [all...] |
jitter_timeline.py | 18 results. 28 def AddResults(self, model, renderer_thread, interactions, results): 36 self._AddJitterResultsInternal(jitter_events, interactions, results) 38 def _AddJitterResultsInternal(self, events, interactions, results): 44 results.AddValue(list_of_scalar_values.ListOfScalarValues( 45 page=results.current_page,
|
single_event.py | 21 def AddResults(self, model, renderer_thread, interactions, results): 25 interactions, results) 27 def _AddResultsInternal(self, events, interactions, results): 39 results.AddValue(list_of_scalar_values.ListOfScalarValues( 40 page=results.current_page,
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/test/ |
test_profile.py | 22 results = []
26 results.append(timer() - start_timer)
32 results.append(s.getvalue())
33 return results
36 results = self.do_profiling()
37 self.assertEqual(results[0], 1000)
39 self.assertEqual(results[i+1], self.expected_output[method],
66 results = cls.do_profiling()
79 cls.__name__, method, results[i+1]))
|
/external/chromium-trace/catapult/telemetry/telemetry/internal/platform/power_monitor/ |
cros_power_monitor_unittest.py | 149 def _assertPowerEqual(self, results, expected): 150 battery = results['component_utilization']['battery'] 156 self.assertAlmostEqual(results['energy_consumption_mwh'], 158 self.assertAlmostEqual(results['power_samples_mw'][0], 160 self.assertAlmostEqual(results['power_samples_mw'][1], 164 results = cros_power_monitor.CrosPowerMonitor.ParsePower( 166 self._assertPowerEqual(results, self.expected_power) 175 results = cros_power_monitor.CrosPowerMonitor.ParsePower( 177 self._assertPowerEqual(results, self.expected_incomplete_power)
|
/external/chromium-trace/catapult/telemetry/telemetry/page/ |
page_run_end_to_end_unittest.py | 21 from telemetry.internal.results import results_options 83 def GetSuccessfulPageRuns(results): 84 return [run for run in results.all_page_runs if run.ok or run.skipped] 172 def ValidateAndMeasurePage(self, page, tab, results): 180 results = results_options.CreateResults(EmptyMetadataForTest(), options) 181 story_runner.Run(test, story_set, options, results) 183 self.assertEquals(2, len(GetSuccessfulPageRuns(results))) 184 self.assertEquals(1, len(results.failures)) 207 def ValidateAndMeasurePage(self, page, tab, results): 215 results = results_options.CreateResults(EmptyMetadataForTest(), options 563 results = results_options.CreateResults(EmptyMetadataForTest(), options) variable in class:ActualPageRunEndToEndTests.testRunPageWithProfilingFlag.Measurement [all...] |
/prebuilts/gdb/darwin-x86/lib/python2.7/test/ |
test_profile.py | 22 results = [] 26 results.append(timer() - start_timer) 32 results.append(s.getvalue()) 33 return results 36 results = self.do_profiling() 37 self.assertEqual(results[0], 1000) 39 self.assertEqual(results[i+1], self.expected_output[method], 66 results = cls.do_profiling() 79 cls.__name__, method, results[i+1]))
|
/prebuilts/gdb/linux-x86/lib/python2.7/test/ |
test_profile.py | 22 results = [] 26 results.append(timer() - start_timer) 32 results.append(s.getvalue()) 33 return results 36 results = self.do_profiling() 37 self.assertEqual(results[0], 1000) 39 self.assertEqual(results[i+1], self.expected_output[method], 66 results = cls.do_profiling() 79 cls.__name__, method, results[i+1]))
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/test/ |
test_profile.py | 22 results = [] 26 results.append(timer() - start_timer) 32 results.append(s.getvalue()) 33 return results 36 results = self.do_profiling() 37 self.assertEqual(results[0], 1000) 39 self.assertEqual(results[i+1], self.expected_output[method], 66 results = cls.do_profiling() 79 cls.__name__, method, results[i+1]))
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/test/ |
test_profile.py | 22 results = [] 26 results.append(timer() - start_timer) 32 results.append(s.getvalue()) 33 return results 36 results = self.do_profiling() 37 self.assertEqual(results[0], 1000) 39 self.assertEqual(results[i+1], self.expected_output[method], 66 results = cls.do_profiling() 79 cls.__name__, method, results[i+1]))
|
/development/samples/SampleSyncAdapter/src/com/example/android/samplesync/platform/ |
BatchOperation.java | 64 ContentProviderResult[] results = mResolver.applyBatch(ContactsContract.AUTHORITY, local 66 if ((results != null) && (results.length > 0)){ 67 for (int i = 0; i < results.length; i++){ 68 resultUris.add(results[i].uri);
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/lib2to3/fixes/ |
fix_apply.py | 31 def transform(self, node, results):
33 assert results
34 func = results["func"]
35 args = results["args"]
36 kwds = results.get("kwds")
|
fix_execfile.py | 24 def transform(self, node, results):
25 assert results
26 filename = results["filename"]
27 globals = results.get("globals")
28 locals = results.get("locals")
|
fix_intern.py | 28 def transform(self, node, results):
30 obj = results["obj"].clone()
35 after = results["after"]
41 [results["lpar"].clone(),
43 results["rpar"].clone()])] + after)
|
/external/autotest/client/site_tests/power_Resume/ |
power_Resume.py | 10 # In cases like crosbug.com/p/26289, we want results, but also want 12 # of extra suspend time to make sure we can get some results, even 38 (results, device_times) = \ 48 results[dev] = device_times[dev] 51 value=results['seconds_system_suspend'], 54 value=results['seconds_system_resume'], 56 self.write_perf_keyval(results)
|