HomeSort by relevance Sort by last modified time
    Searched refs:_results (Results 1 - 25 of 33) sorted by null

1 2

  /external/chromium_org/tools/perf/metrics/
v8_object_stats.py 167 self._results = None
207 self._results = V8ObjectStatsMetric.GetV8StatsTable(tab, self._counters)
208 if not self._results:
213 assert self._results != None, 'Must call Stop() first'
214 for counter_name in self._results:
216 results.Add(display_name, 'kb', self._results[counter_name] / 1024.0)
power.py 16 self._results = None
32 self._results = self._browser.platform.StopMonitoringPower()
33 if self._results: # StopMonitoringPower() can return None.
34 self._results['cpu_stats'] = (
41 self._results = None
65 if not self._results:
68 energy_consumption_mwh = self._results.get('energy_consumption_mwh')
74 component_utilization = self._results.get('component_utilization', {})
84 for (process_type, stats) in self._results.get('cpu_stats', {}).items():
98 self._results = Non
    [all...]
cpu.py 14 self._results = None
27 self._results = _SubtractCpuStats(self._browser.cpu_stats, self._start_cpu)
32 assert self._results, 'Must call Stop() first'
34 for process_type in self._results:
36 cpu_percent = 100 * self._results[process_type]
media.py 25 self._results = None
35 self._results = tab.EvaluateJavaScript('window.__getAllMetrics()')
40 for media_metric in self._results:
  /external/chromium_org/third_party/webrtc/modules/video_coding/codecs/test_framework/
benchmark.cc 88 _results.open(_resultsFileName.c_str(), std::fstream::out);
89 _results << GetMagicStr() << std::endl;
90 _results << _codecName << std::endl;
119 _results << (*it)->GetName() << "," << VideoSource::GetSizeString(size[i])
154 _results << "," << actualBitRate;
168 _results << std::endl << "Y-PSNR [dB]";
172 _results << "," << psnr[k].average;
178 _results << std::endl << "SSIM ";
182 _results << "," << ssim[k].average;
188 _results << std::endl << "Encode Time[ms]"
    [all...]
benchmark.h 35 std::ofstream _results; member in class:Benchmark
  /external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/unittest2/
signals.py 24 for result in _results.keys():
27 _results = weakref.WeakKeyDictionary() variable
29 _results[result] = 1
32 return bool(_results.pop(result, None))
  /external/lldb/test/unittest2/
signals.py 24 for result in _results.keys():
27 _results = weakref.WeakKeyDictionary() variable
29 _results[result] = 1
32 return bool(_results.pop(result, None))
  /prebuilts/python/darwin-x86/2.7.5/lib/python2.7/unittest/
signals.py 38 for result in _results.keys():
41 _results = weakref.WeakKeyDictionary() variable
43 _results[result] = 1
46 return bool(_results.pop(result, None))
  /prebuilts/python/linux-x86/2.7.5/lib/python2.7/unittest/
signals.py 38 for result in _results.keys():
41 _results = weakref.WeakKeyDictionary() variable
43 _results[result] = 1
46 return bool(_results.pop(result, None))
  /external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/common/net/
layouttestresults.py 101 self._results = parsed_json
104 return self._results["interrupted"]
107 return self._results["builder_name"]
110 return int(self._results["blink_revision"])
113 result = result_for_test(self._results["tests"], test)
  /external/opencv/ml/src/
mlknearest.cpp 214 CvMat* _results, CvMat* _neighbor_responses,
220 int rstep = _results && !CV_IS_MAT_CONT(_results->type) ? _results->step/sizeof(result) : 1;
227 if( _results || start+i == 0 )
276 if( _results )
277 _results->data.fl[(start + i)*rstep] = r;
305 float CvKNearest::find_nearest( const CvMat* _samples, int k, CvMat* _results,
327 if( _results && (!CV_IS_MAT(_results) ||
    [all...]
  /external/chromium_org/build/android/pylib/base/
base_test_result.py 77 self._results = set()
118 s.append('ALL: %d' % len(self._results))
133 self._results.add(result)
152 self._results.update(results._results)
156 return self._results.copy()
160 return set(t for t in self._results if t.GetType() == test_type)
  /external/chromium_org/build/android/pylib/perf/
surface_stats_collector.py 44 self._results = []
73 self._results = []
77 return self._results or self._GetEmptyResults()
149 self._results.append(SurfaceStatsCollector.Result(
160 self._results.append(SurfaceStatsCollector.Result(
162 self._results += self._CalculateResults(refresh_period, timestamps, '')
163 self._results += self._CalculateBuckets(refresh_period, timestamps)
  /external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/test/
runner_unittest.py 62 self._results = {}
66 self._results['%s.%s' % (m.group(2), m.group(1))] = tuple([test_name, result, msg])
72 return FakeModuleSuite(*self._results[name])
  /external/chromium_org/tools/telemetry/telemetry/web_perf/
timeline_based_measurement.py 35 self._results = results
40 self._results.Add(trace_name, units, value, chart_name, data_type)
45 self._results.AddSummary(trace_name, units, value, chart_name, data_type)
  /external/lldb/test/unittest2/test/
test_break.py 22 unittest2.signals._results = weakref.WeakKeyDictionary()
43 for ref in unittest2.signals._results:
148 self.assertIn(result, unittest2.signals._results)
  /prebuilts/python/darwin-x86/2.7.5/lib/python2.7/unittest/test/
test_break.py 27 unittest.signals._results = weakref.WeakKeyDictionary()
48 for ref in unittest.signals._results:
161 self.assertIn(result, unittest.signals._results)
  /prebuilts/python/linux-x86/2.7.5/lib/python2.7/unittest/test/
test_break.py 27 unittest.signals._results = weakref.WeakKeyDictionary()
48 for ref in unittest.signals._results:
161 self.assertIn(result, unittest.signals._results)
  /external/chromium_org/third_party/skia/gm/rebaseline_server/
results.py 78 return self._results[results_type]
93 response_dict = self._results[results_type]
server.py 226 # 1. self._results
229 # self._results will be filled in by calls to update_results()
230 self._results = None
236 return self._results
256 """ Create or update self._results, based on the latest expectations and
260 thread attempts to update either self._results or the underlying files at
264 invalidate: if True, invalidate self._results immediately upon entry;
270 self._results = None
337 self._results = compare_to_expectations.ExpectationComparisons(
compare_configs.py 68 and stores the summary in self._results.
166 self._results = {
  /external/skia/gm/rebaseline_server/
results.py 78 return self._results[results_type]
93 response_dict = self._results[results_type]
server.py 226 # 1. self._results
229 # self._results will be filled in by calls to update_results()
230 self._results = None
236 return self._results
256 """ Create or update self._results, based on the latest expectations and
260 thread attempts to update either self._results or the underlying files at
264 invalidate: if True, invalidate self._results immediately upon entry;
270 self._results = None
337 self._results = compare_to_expectations.ExpectationComparisons(
  /external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/
perftestsrunner.py 283 for test, metrics in self._results:
363 self._results = []
370 self._results.append((test, metrics))

Completed in 508 milliseconds

1 2