HomeSort by relevance Sort by last modified time
    Searched refs:math_utils (Results 1 - 5 of 5) sorted by null

  /external/chromium_org/tools/auto_bisect/
math_utils_test.py 8 import math_utils namespace
17 math_utils.TruncatedMean([], 0)
21 self.assertEqual(3.0, math_utils.Mean([3]))
25 self.assertEqual(0.5, math_utils.Mean([-3, 0, 1, 4]))
37 math_utils.Mean(values))
42 self.assertEqual(0.5, math_utils.RelativeChange(1.0, 1.5))
43 self.assertEqual(0.5, math_utils.RelativeChange(2.0, 1.0))
48 self.assertEqual(0, math_utils.RelativeChange(0, 0))
50 math.isnan(math_utils.RelativeChange(0, 1)))
52 math.isnan(math_utils.RelativeChange(0, -1))
    [all...]
ttest.py 8 to use definitions of mean and variance from math_utils instead of numpy.
28 import math_utils namespace
45 mean1 = math_utils.Mean(sample1)
46 mean2 = math_utils.Mean(sample2)
47 v1 = math_utils.Variance(sample1)
48 v2 = math_utils.Variance(sample2)
bisect_results.py 9 import math_utils namespace
79 mean_of_prev_runs = math_utils.Mean(sum(previous_values, []))
80 mean_of_current_runs = math_utils.Mean(current_values)
182 mean_of_bad_runs = math_utils.Mean(broken_mean)
183 mean_of_good_runs = math_utils.Mean(working_mean)
185 regression_size = 100 * math_utils.RelativeChange(mean_of_good_runs,
190 regression_std_err = math.fabs(math_utils.PooledStandardError(
bisect_perf_regression.py 55 import math_utils namespace
    [all...]
  /external/chromium_org/tools/
run-bisect-perf-regression.py 23 from auto_bisect import math_utils namespace
316 std_err = math_utils.PooledStandardError(

Completed in 191 milliseconds