/external/chromium_org/tools/perf/metrics/ |
v8_object_stats.py | 167 self._results = None 207 self._results = V8ObjectStatsMetric.GetV8StatsTable(tab, self._counters) 208 if not self._results: 213 assert self._results != None, 'Must call Stop() first' 214 for counter_name in self._results: 216 results.Add(display_name, 'kb', self._results[counter_name] / 1024.0)
|
power.py | 16 self._results = None 32 self._results = self._browser.platform.StopMonitoringPower() 33 if self._results: # StopMonitoringPower() can return None. 34 self._results['cpu_stats'] = ( 41 self._results = None 65 if not self._results: 68 energy_consumption_mwh = self._results.get('energy_consumption_mwh') 74 component_utilization = self._results.get('component_utilization', {}) 84 for (process_type, stats) in self._results.get('cpu_stats', {}).items(): 98 self._results = Non [all...] |
cpu.py | 14 self._results = None 27 self._results = _SubtractCpuStats(self._browser.cpu_stats, self._start_cpu) 32 assert self._results, 'Must call Stop() first' 34 for process_type in self._results: 36 cpu_percent = 100 * self._results[process_type]
|
media.py | 25 self._results = None 35 self._results = tab.EvaluateJavaScript('window.__getAllMetrics()') 40 for media_metric in self._results:
|
/external/chromium_org/third_party/webrtc/modules/video_coding/codecs/test_framework/ |
benchmark.cc | 88 _results.open(_resultsFileName.c_str(), std::fstream::out); 89 _results << GetMagicStr() << std::endl; 90 _results << _codecName << std::endl; 119 _results << (*it)->GetName() << "," << VideoSource::GetSizeString(size[i]) 154 _results << "," << actualBitRate; 168 _results << std::endl << "Y-PSNR [dB]"; 172 _results << "," << psnr[k].average; 178 _results << std::endl << "SSIM "; 182 _results << "," << ssim[k].average; 188 _results << std::endl << "Encode Time[ms]" [all...] |
benchmark.h | 35 std::ofstream _results; member in class:Benchmark
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/unittest2/ |
signals.py | 24 for result in _results.keys(): 27 _results = weakref.WeakKeyDictionary() variable 29 _results[result] = 1 32 return bool(_results.pop(result, None))
|
/external/lldb/test/unittest2/ |
signals.py | 24 for result in _results.keys(): 27 _results = weakref.WeakKeyDictionary() variable 29 _results[result] = 1 32 return bool(_results.pop(result, None))
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/unittest/ |
signals.py | 38 for result in _results.keys(): 41 _results = weakref.WeakKeyDictionary() variable 43 _results[result] = 1 46 return bool(_results.pop(result, None))
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/unittest/ |
signals.py | 38 for result in _results.keys(): 41 _results = weakref.WeakKeyDictionary() variable 43 _results[result] = 1 46 return bool(_results.pop(result, None))
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/common/net/ |
layouttestresults.py | 101 self._results = parsed_json 104 return self._results["interrupted"] 107 return self._results["builder_name"] 110 return int(self._results["blink_revision"]) 113 result = result_for_test(self._results["tests"], test)
|
/external/opencv/ml/src/ |
mlknearest.cpp | 214 CvMat* _results, CvMat* _neighbor_responses, 220 int rstep = _results && !CV_IS_MAT_CONT(_results->type) ? _results->step/sizeof(result) : 1; 227 if( _results || start+i == 0 ) 276 if( _results ) 277 _results->data.fl[(start + i)*rstep] = r; 305 float CvKNearest::find_nearest( const CvMat* _samples, int k, CvMat* _results, 327 if( _results && (!CV_IS_MAT(_results) || [all...] |
/external/chromium_org/build/android/pylib/base/ |
base_test_result.py | 77 self._results = set() 118 s.append('ALL: %d' % len(self._results)) 133 self._results.add(result) 152 self._results.update(results._results) 156 return self._results.copy() 160 return set(t for t in self._results if t.GetType() == test_type)
|
/external/chromium_org/build/android/pylib/perf/ |
surface_stats_collector.py | 44 self._results = [] 73 self._results = [] 77 return self._results or self._GetEmptyResults() 149 self._results.append(SurfaceStatsCollector.Result( 160 self._results.append(SurfaceStatsCollector.Result( 162 self._results += self._CalculateResults(refresh_period, timestamps, '') 163 self._results += self._CalculateBuckets(refresh_period, timestamps)
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/test/ |
runner_unittest.py | 62 self._results = {} 66 self._results['%s.%s' % (m.group(2), m.group(1))] = tuple([test_name, result, msg]) 72 return FakeModuleSuite(*self._results[name])
|
/external/chromium_org/tools/telemetry/telemetry/web_perf/ |
timeline_based_measurement.py | 35 self._results = results 40 self._results.Add(trace_name, units, value, chart_name, data_type) 45 self._results.AddSummary(trace_name, units, value, chart_name, data_type)
|
/external/lldb/test/unittest2/test/ |
test_break.py | 22 unittest2.signals._results = weakref.WeakKeyDictionary() 43 for ref in unittest2.signals._results: 148 self.assertIn(result, unittest2.signals._results)
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/unittest/test/ |
test_break.py | 27 unittest.signals._results = weakref.WeakKeyDictionary() 48 for ref in unittest.signals._results: 161 self.assertIn(result, unittest.signals._results)
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/unittest/test/ |
test_break.py | 27 unittest.signals._results = weakref.WeakKeyDictionary() 48 for ref in unittest.signals._results: 161 self.assertIn(result, unittest.signals._results)
|
/external/chromium_org/third_party/skia/gm/rebaseline_server/ |
results.py | 78 return self._results[results_type] 93 response_dict = self._results[results_type]
|
server.py | 226 # 1. self._results 229 # self._results will be filled in by calls to update_results() 230 self._results = None 236 return self._results 256 """ Create or update self._results, based on the latest expectations and 260 thread attempts to update either self._results or the underlying files at 264 invalidate: if True, invalidate self._results immediately upon entry; 270 self._results = None 337 self._results = compare_to_expectations.ExpectationComparisons(
|
compare_configs.py | 68 and stores the summary in self._results. 166 self._results = {
|
/external/skia/gm/rebaseline_server/ |
results.py | 78 return self._results[results_type] 93 response_dict = self._results[results_type]
|
server.py | 226 # 1. self._results 229 # self._results will be filled in by calls to update_results() 230 self._results = None 236 return self._results 256 """ Create or update self._results, based on the latest expectations and 260 thread attempts to update either self._results or the underlying files at 264 invalidate: if True, invalidate self._results immediately upon entry; 270 self._results = None 337 self._results = compare_to_expectations.ExpectationComparisons(
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/ |
perftestsrunner.py | 283 for test, metrics in self._results: 363 self._results = [] 370 self._results.append((test, metrics))
|