/external/chromium-trace/catapult/dashboard/dashboard/models/ |
alert_group_test.py | 62 anomalies = ndb.get_multi(anomaly_keys) 64 # Add these anomalies to groups and put them again. When anomalies are 66 # the groups of the anomalies will be updated. 67 anomalies[0].group = group_keys[0] 68 anomalies[0].put() 69 anomalies[1].group = group_keys[0] 70 anomalies[1].put() 71 anomalies[2].group = group_keys[1] 72 anomalies[2].put( [all...] |
alert_test.py | 24 anomalies = anomaly.Anomaly.query().fetch() 25 bot_names = alert.GetBotNamesFromAlerts(anomalies) 26 self.assertEqual(2, len(anomalies)) 35 anomalies = anomaly.Anomaly.query().fetch() 36 bot_names = alert.GetBotNamesFromAlerts(anomalies)
|
/external/chromium-trace/catapult/dashboard/dashboard/ |
email_summary.py | 19 <br><b><u>%d Total Anomalies</b></u><br> 23 _EMAIL_SUBJECT = '%s: %d anomalies found at %d:%d.' 30 anomalies related to that sheriff that were triggered in the past 24 hours. 34 """Emails sheriffs with anomalies identified in most-recent 24 hours.""" 40 # Start time after which to get anomalies. 52 start_time: A starting datetime for anomalies to fetch. 55 anomalies = _RecentUntriagedAnomalies(sheriff_entity, start_time) 56 if not anomalies: 58 subject = _EmailSubject(sheriff_entity, anomalies) 59 html, text = _EmailBody(anomalies) [all...] |
main.py | 5 """URL endpoint for the main page which lists recent anomalies.""" 30 days: Number of days to show anomalies for (optional). 31 sheriff: Sheriff to show anomalies for (optional) 42 anomalies = _GetRecentAnomalies(days, sheriff) 44 top_improvements = _TopImprovements(anomalies, num_changes) 45 top_regressions = _TopRegressions(anomalies, num_changes) 59 """Fetches recent Anomalies from the datastore. 62 days: Number of days old of the oldest Anomalies to fetch. 63 sheriff: The ndb.Key of the Sheriff to fetch Anomalies for. 72 anomalies = anomalies_query.fetch(limit=_ANOMALY_FETCH_LIMIT [all...] |
group_report.py | 63 raise request_handler.InvalidInputError('No anomalies specified.') 79 anomalies = anomaly_query.fetch(limit=_DISPLAY_LIMIT) 83 self._ShowAlerts(anomalies + stoppage_alerts, bug_id) 100 anomalies = anomaly_query.fetch(limit=_QUERY_LIMIT) 101 anomalies = [a for a in anomalies if a.start_revision <= rev] 105 self._ShowAlerts(anomalies + stoppage_alerts) 110 Query for anomalies with overlapping revision. The |keys| 134 raise request_handler.InvalidInputError('No anomalies found.') 142 anomalies = query.fetch(limit=_QUERY_LIMIT [all...] |
auto_bisect.py | 149 anomalies = anomaly.Anomaly.query(anomaly.Anomaly.bug_id == bug_id).fetch() 150 if not anomalies: 153 good_revision, bad_revision = _ChooseRevisionRange(anomalies) 159 test = _ChooseTest(anomalies, run_count) 208 def _ChooseTest(anomalies, index=0): 212 chosen. The test to choose could depend on which of the anomalies has the 225 anomalies: A non-empty list of Anomaly entities. 227 than the number of Anomalies, it will wrap around. This is used to 234 if not anomalies: 236 index %= len(anomalies) [all...] |
auto_triage.py | 31 # This is the max queried untriaged anomalies per sheriff. 32 # Takes about 30 seconds to fetch 2000 anomalies per sheriff. 46 """URL endpoint for a cron job to automatically triage anomalies and bugs.""" 74 """Class for triaging anomalies.""" 78 """Processes anomalies.""" 79 # Check for recovered anomalies that are untriaged. 80 anomalies = cls._FetchUntriagedAnomalies() 81 recovered_anomalies = _FindAndUpdateRecoveredAnomalies(anomalies) 86 """Fetches recent untriaged anomalies asynchronously from all sheriffs.""" 87 anomalies = [ [all...] |
find_anomalies_test.py | 147 anomalies = anomaly.Anomaly.query().fetch() 148 self.assertEqual(len(anomalies), 3) 151 anomalies, test, percent_changed, direction, 153 for a in anomalies: 166 anomalies, test.key, percent_changed=100, direction=anomaly.UP, 172 anomalies, test.key, percent_changed=-50, direction=anomaly.DOWN, 178 anomalies, test.key, percent_changed=sys.float_info.max, 185 anomalies, test.key, percent_changed=100, direction=anomaly.DOWN, 203 anomalies = anomaly.Anomaly.query().fetch() 204 self.assertEqual(len(anomalies), 1 [all...] |
alerts.py | 25 """Shows an overview of recent anomalies for perf sheriffing.""" 37 improvements: Whether to include improvement anomalies. 55 anomalies = ndb.get_multi(anomaly_keys[:_MAX_ANOMALIES_TO_SHOW]) 59 'anomaly_list': AnomalyDicts(anomalies), 83 include_improvements: Whether to include improvement Anomalies. 84 include_triaged: Whether to include Anomalies with a bug ID already set. 135 def AnomalyDicts(anomalies): 137 bisect_statuses = _GetBisectStatusDict(anomalies) 138 return [GetAnomalyDict(a, bisect_statuses.get(a.bug_id)) for a in anomalies] 203 def _GetBisectStatusDict(anomalies) [all...] |
find_anomalies.py | 30 """Processes a test to find new anomalies. 59 # Get anomalies and check if they happen in ref build also. 63 anomalies = [_MakeAnomalyEntity(c, test, rows) for c in change_points] 65 # If no new anomalies were found, then we're done. 66 if not anomalies: 72 test.last_alerted_revision = anomalies[-1].end_revision 76 anomalies, utils.TestSuiteName(test.key), 'Anomaly') 79 for anomaly_entity in anomalies: 85 ndb.put_multi(anomalies) 123 """Filters out the anomalies that match the anomalies in ref build [all...] |
change_internal_only.py | 163 anomalies = query.fetch() 164 for anomaly_entity in anomalies: 167 ndb.put_multi(anomalies)
|
dump_graph_json_test.py | 112 anomalies = _EntitiesOfKind(entities, 'Anomaly') 114 self.assertEqual(1, len(anomalies)) 152 anomalies = _EntitiesOfKind(entities, 'Anomaly') 155 self.assertEqual(1, len(anomalies))
|
change_internal_only_test.py | 98 anomalies = anomaly.Anomaly.query().fetch() 99 for a in anomalies:
|
migrate_test_names_test.py | 86 """Checks whether the anomalies match the ones added in _AddMockData. 89 test_path: The test path for the Test which the Anomalies are on. 94 anomalies = anomaly.Anomaly.query( 96 self.assertEqual(2, len(anomalies)) 97 self.assertEqual(r1, anomalies[0].end_revision) 98 self.assertEqual(r2, anomalies[1].end_revision)
|
dump_graph_json.py | 104 anomalies = self._FetchAnomalies(sheriff, num_anomalies) 105 test_keys = [a.test for a in anomalies] 116 entities += anomalies 153 """Fetches recent anomalies for 'sheriff'."""
|
update_bug_with_results.py | 216 """Maps anomalies from source bug to destination bug. 224 anomalies = query.fetch() 225 for anomaly_entity in anomalies: 227 ndb.put_multi(anomalies)
|
alerts_test.py | 49 # Here we set the improvement direction so that some anomalies are 155 # anomalies to be this new sheriff. 160 anomalies = anomaly.Anomaly.query( 162 for anomaly_entity in anomalies:
|
update_bug_with_results_test.py | 332 # Add anomalies. 344 # Map anomalies to base(dest_bug_id) bug. 347 anomalies = anomaly.Anomaly.query( 349 self.assertEqual(0, len(anomalies))
|
bench_find_anomalies.py | 5 """A tool to compare functions for finding anomalies to the current default. 9 results to the base results and to the existing anomalies in the datastore. 59 _TASK_QUEUE_NAME = 'find-anomalies-bench-queue' 61 _FIND_ANOMALIES_BENCH_CACHE_KEY = 'find-anomalies-bench' 63 # Bench name to path of allowable find anomalies function to benchmark. 248 by the experimental find-anomalies function -- they were not "confirmed" 394 raise ValueError('%s is not a valid find anomalies bench function.' % 472 """Updates TestBench entity with invalid and confirmed anomalies revs.""" 474 # Start rev for getting Anomalies should be at min_segment_size. 483 anomalies = query.fetch( [all...] |
graph_json.py | 168 anomalies = anomaly.Anomaly.query().filter( 170 return dict((a.end_revision, a) for a in anomalies)
|