HomeSort by relevance Sort by last modified time
    Searched refs:anomalies (Results 1 - 20 of 20) sorted by null

  /external/chromium-trace/catapult/dashboard/dashboard/models/
alert_group_test.py 62 anomalies = ndb.get_multi(anomaly_keys)
64 # Add these anomalies to groups and put them again. When anomalies are
66 # the groups of the anomalies will be updated.
67 anomalies[0].group = group_keys[0]
68 anomalies[0].put()
69 anomalies[1].group = group_keys[0]
70 anomalies[1].put()
71 anomalies[2].group = group_keys[1]
72 anomalies[2].put(
    [all...]
alert_test.py 24 anomalies = anomaly.Anomaly.query().fetch()
25 bot_names = alert.GetBotNamesFromAlerts(anomalies)
26 self.assertEqual(2, len(anomalies))
35 anomalies = anomaly.Anomaly.query().fetch()
36 bot_names = alert.GetBotNamesFromAlerts(anomalies)
  /external/chromium-trace/catapult/dashboard/dashboard/
email_summary.py 19 <br><b><u>%d Total Anomalies</b></u><br>
23 _EMAIL_SUBJECT = '%s: %d anomalies found at %d:%d.'
30 anomalies related to that sheriff that were triggered in the past 24 hours.
34 """Emails sheriffs with anomalies identified in most-recent 24 hours."""
40 # Start time after which to get anomalies.
52 start_time: A starting datetime for anomalies to fetch.
55 anomalies = _RecentUntriagedAnomalies(sheriff_entity, start_time)
56 if not anomalies:
58 subject = _EmailSubject(sheriff_entity, anomalies)
59 html, text = _EmailBody(anomalies)
    [all...]
main.py 5 """URL endpoint for the main page which lists recent anomalies."""
30 days: Number of days to show anomalies for (optional).
31 sheriff: Sheriff to show anomalies for (optional)
42 anomalies = _GetRecentAnomalies(days, sheriff)
44 top_improvements = _TopImprovements(anomalies, num_changes)
45 top_regressions = _TopRegressions(anomalies, num_changes)
59 """Fetches recent Anomalies from the datastore.
62 days: Number of days old of the oldest Anomalies to fetch.
63 sheriff: The ndb.Key of the Sheriff to fetch Anomalies for.
72 anomalies = anomalies_query.fetch(limit=_ANOMALY_FETCH_LIMIT
    [all...]
group_report.py 63 raise request_handler.InvalidInputError('No anomalies specified.')
79 anomalies = anomaly_query.fetch(limit=_DISPLAY_LIMIT)
83 self._ShowAlerts(anomalies + stoppage_alerts, bug_id)
100 anomalies = anomaly_query.fetch(limit=_QUERY_LIMIT)
101 anomalies = [a for a in anomalies if a.start_revision <= rev]
105 self._ShowAlerts(anomalies + stoppage_alerts)
110 Query for anomalies with overlapping revision. The |keys|
134 raise request_handler.InvalidInputError('No anomalies found.')
142 anomalies = query.fetch(limit=_QUERY_LIMIT
    [all...]
auto_bisect.py 149 anomalies = anomaly.Anomaly.query(anomaly.Anomaly.bug_id == bug_id).fetch()
150 if not anomalies:
153 good_revision, bad_revision = _ChooseRevisionRange(anomalies)
159 test = _ChooseTest(anomalies, run_count)
208 def _ChooseTest(anomalies, index=0):
212 chosen. The test to choose could depend on which of the anomalies has the
225 anomalies: A non-empty list of Anomaly entities.
227 than the number of Anomalies, it will wrap around. This is used to
234 if not anomalies:
236 index %= len(anomalies)
    [all...]
auto_triage.py 31 # This is the max queried untriaged anomalies per sheriff.
32 # Takes about 30 seconds to fetch 2000 anomalies per sheriff.
46 """URL endpoint for a cron job to automatically triage anomalies and bugs."""
74 """Class for triaging anomalies."""
78 """Processes anomalies."""
79 # Check for recovered anomalies that are untriaged.
80 anomalies = cls._FetchUntriagedAnomalies()
81 recovered_anomalies = _FindAndUpdateRecoveredAnomalies(anomalies)
86 """Fetches recent untriaged anomalies asynchronously from all sheriffs."""
87 anomalies = [
    [all...]
find_anomalies_test.py 147 anomalies = anomaly.Anomaly.query().fetch()
148 self.assertEqual(len(anomalies), 3)
151 anomalies, test, percent_changed, direction,
153 for a in anomalies:
166 anomalies, test.key, percent_changed=100, direction=anomaly.UP,
172 anomalies, test.key, percent_changed=-50, direction=anomaly.DOWN,
178 anomalies, test.key, percent_changed=sys.float_info.max,
185 anomalies, test.key, percent_changed=100, direction=anomaly.DOWN,
203 anomalies = anomaly.Anomaly.query().fetch()
204 self.assertEqual(len(anomalies), 1
    [all...]
alerts.py 25 """Shows an overview of recent anomalies for perf sheriffing."""
37 improvements: Whether to include improvement anomalies.
55 anomalies = ndb.get_multi(anomaly_keys[:_MAX_ANOMALIES_TO_SHOW])
59 'anomaly_list': AnomalyDicts(anomalies),
83 include_improvements: Whether to include improvement Anomalies.
84 include_triaged: Whether to include Anomalies with a bug ID already set.
135 def AnomalyDicts(anomalies):
137 bisect_statuses = _GetBisectStatusDict(anomalies)
138 return [GetAnomalyDict(a, bisect_statuses.get(a.bug_id)) for a in anomalies]
203 def _GetBisectStatusDict(anomalies)
    [all...]
find_anomalies.py 30 """Processes a test to find new anomalies.
59 # Get anomalies and check if they happen in ref build also.
63 anomalies = [_MakeAnomalyEntity(c, test, rows) for c in change_points]
65 # If no new anomalies were found, then we're done.
66 if not anomalies:
72 test.last_alerted_revision = anomalies[-1].end_revision
76 anomalies, utils.TestSuiteName(test.key), 'Anomaly')
79 for anomaly_entity in anomalies:
85 ndb.put_multi(anomalies)
123 """Filters out the anomalies that match the anomalies in ref build
    [all...]
change_internal_only.py 163 anomalies = query.fetch()
164 for anomaly_entity in anomalies:
167 ndb.put_multi(anomalies)
dump_graph_json_test.py 112 anomalies = _EntitiesOfKind(entities, 'Anomaly')
114 self.assertEqual(1, len(anomalies))
152 anomalies = _EntitiesOfKind(entities, 'Anomaly')
155 self.assertEqual(1, len(anomalies))
change_internal_only_test.py 98 anomalies = anomaly.Anomaly.query().fetch()
99 for a in anomalies:
migrate_test_names_test.py 86 """Checks whether the anomalies match the ones added in _AddMockData.
89 test_path: The test path for the Test which the Anomalies are on.
94 anomalies = anomaly.Anomaly.query(
96 self.assertEqual(2, len(anomalies))
97 self.assertEqual(r1, anomalies[0].end_revision)
98 self.assertEqual(r2, anomalies[1].end_revision)
dump_graph_json.py 104 anomalies = self._FetchAnomalies(sheriff, num_anomalies)
105 test_keys = [a.test for a in anomalies]
116 entities += anomalies
153 """Fetches recent anomalies for 'sheriff'."""
update_bug_with_results.py 216 """Maps anomalies from source bug to destination bug.
224 anomalies = query.fetch()
225 for anomaly_entity in anomalies:
227 ndb.put_multi(anomalies)
alerts_test.py 49 # Here we set the improvement direction so that some anomalies are
155 # anomalies to be this new sheriff.
160 anomalies = anomaly.Anomaly.query(
162 for anomaly_entity in anomalies:
update_bug_with_results_test.py 332 # Add anomalies.
344 # Map anomalies to base(dest_bug_id) bug.
347 anomalies = anomaly.Anomaly.query(
349 self.assertEqual(0, len(anomalies))
bench_find_anomalies.py 5 """A tool to compare functions for finding anomalies to the current default.
9 results to the base results and to the existing anomalies in the datastore.
59 _TASK_QUEUE_NAME = 'find-anomalies-bench-queue'
61 _FIND_ANOMALIES_BENCH_CACHE_KEY = 'find-anomalies-bench'
63 # Bench name to path of allowable find anomalies function to benchmark.
248 by the experimental find-anomalies function -- they were not "confirmed"
394 raise ValueError('%s is not a valid find anomalies bench function.' %
472 """Updates TestBench entity with invalid and confirmed anomalies revs."""
474 # Start rev for getting Anomalies should be at min_segment_size.
483 anomalies = query.fetch(
    [all...]
graph_json.py 168 anomalies = anomaly.Anomaly.query().filter(
170 return dict((a.end_revision, a) for a in anomalies)

Completed in 82 milliseconds