/external/chromium-trace/catapult/dashboard/dashboard/models/ |
alert_test.py | 12 from dashboard.models import anomaly namespace 16 """Test case for some functions in anomaly.""" 24 anomaly.Anomaly(test=utils.TestKey('SuperGPU/Bot1/foo/bar')).put() 25 anomaly.Anomaly(test=utils.TestKey('SuperGPU/Bot1/foo/bar')).put() 26 anomalies = anomaly.Anomaly.query().fetch() 34 anomaly.Anomaly(test=utils.TestKey('SuperGPU/Bot1/foo/bar')).put( [all...] |
alert_group_test.py | 12 from dashboard.models import anomaly namespace 18 """Test case for the behavior of updating anomaly groups.""" 28 first_paint_test.improvement_direction = anomaly.DOWN 35 alert_kind='Anomaly', 40 alert_kind='Anomaly', 45 anomaly.Anomaly( 50 anomaly.Anomaly( 55 anomaly.Anomaly [all...] |
anomaly.py | 5 """The database model for an "Anomaly", which represents a step up or down.""" 16 # Possible improvement directions for a change. An Anomaly will always have a 21 class Anomaly(alert.Alert): 24 An Anomaly can be an upward or downward change, and can represent an 27 # The number of points before and after this anomaly that were looked at 28 # when finding this anomaly. 32 # The medians of the segments before and after the anomaly. 36 # The standard deviation of the segments before the anomaly. 43 # In order to estimate how likely it is that this anomaly is due to noise, 50 # Whether this anomaly represents an improvement; if false, this anomaly i [all...] |
anomaly_config.py | 5 """The database models for anomaly alerting threshold configs.""" 11 # The parameters to use from anomaly threshold config dict. 24 """Represents a set of parameters for the anomaly detection function. 26 The anomaly detection module uses set of parameters to determine the 27 thresholds for what is considered an anomaly. 52 """Gets the anomaly threshold config for the given test.
|
/external/chromium-trace/catapult/dashboard/dashboard/ |
units_to_direction_test.py | 9 from dashboard.models import anomaly namespace 21 anomaly.DOWN, units_to_direction.GetImprovementDirection('ms')) 23 anomaly.UP, units_to_direction.GetImprovementDirection('score')) 25 anomaly.UNKNOWN, 33 anomaly.UNKNOWN, units_to_direction.GetImprovementDirection('runs/s')) 39 anomaly.UP, units_to_direction.GetImprovementDirection('runs/s')) 47 anomaly.UP, units_to_direction.GetImprovementDirection('score')) 52 anomaly.UNKNOWN, units_to_direction.GetImprovementDirection('score')) 59 anomaly.DOWN, units_to_direction.GetImprovementDirection('ms')) 64 anomaly.UP, units_to_direction.GetImprovementDirection('ms') [all...] |
email_sheriff.py | 14 def EmailSheriff(sheriff, test, anomaly): 15 """Sends an email to the sheriff on duty about the given anomaly. 19 test: The graph_data.TestMetadata entity associated with the anomaly. 20 anomaly: The anomaly.Anomaly entity. 26 anomaly_info = email_template.GetAlertInfo(anomaly, test)
|
auto_triage.py | 24 from dashboard.models import anomaly namespace 55 This will include updating Anomaly entities, and checking whether they 95 query = anomaly.Anomaly.query( 96 anomaly.Anomaly.sheriff == key, 97 anomaly.Anomaly.bug_id == None, 98 anomaly.Anomaly.is_improvement == False [all...] |
group_report.py | 17 from dashboard.models import anomaly namespace 41 keys: A comma-separated list of urlsafe Anomaly keys (optional). 76 anomaly_query = anomaly.Anomaly.query( 77 anomaly.Anomaly.bug_id == bug_id) 96 # Anomaly entities based on one of these, then filter the resulting list. 97 anomaly_query = anomaly.Anomaly.query(anomaly.Anomaly.end_revision >= rev [all...] |
associate_alerts_test.py | 19 from dashboard.models import anomaly namespace 53 """Adds sample Anomaly data and returns a dict of revision to key.""" 61 anomaly_key = anomaly.Anomaly( 67 # Add an anomaly that overlaps. 68 anomaly_key = anomaly.Anomaly( 74 # Add an anomaly that overlaps and has bug ID. 75 anomaly_key = anomaly.Anomaly( [all...] |
units_to_direction.py | 9 from dashboard.models import anomaly namespace 20 return anomaly.UNKNOWN 23 return anomaly.UNKNOWN 25 return anomaly.UP 26 return anomaly.DOWN
|
alerts.py | 15 from dashboard.models import anomaly namespace 79 """Fetches the list of Anomaly keys that may be shown. 87 A list of Anomaly keys, in reverse-chronological order. 89 query = anomaly.Anomaly.query( 90 anomaly.Anomaly.sheriff == sheriff_key) 94 anomaly.Anomaly.is_improvement == False) 98 anomaly.Anomaly.bug_id == None [all...] |
main.py | 15 from dashboard.models import anomaly namespace 66 A list of Anomaly entities sorted from large to small relative change. 69 anomalies_query = anomaly.Anomaly.query( 70 anomaly.Anomaly.timestamp > oldest_time, 71 anomaly.Anomaly.sheriff == sheriff) 87 """Returns a CSS class name for the anomaly, based on percent changed.""" 143 recent_anomalies: A list of Anomaly entities sorted from large to small [all...] |
find_anomalies.py | 5 """Processes tests and creates new Anomaly entities. 9 new Anomaly entities. 20 from dashboard.models import anomaly namespace 25 # may be used if a test has a "max_window_size" anomaly config parameter. 69 logging.info('Found at least one anomaly in: %s', test.test_path) 76 anomalies, utils.TestSuiteName(test.key), 'Anomaly') 130 in the same environment. If an anomaly happens in the ref build results at 131 the same time as an anomaly happened in the test build, that suggests that 144 an anomaly found in the corresponding ref test will be filtered out. 160 # Log information about what anomaly got filtered and what did not [all...] |
dump_graph_json.py | 19 from dashboard.models import anomaly namespace 24 # entities for each anomaly. 51 TestMetadata, Row, Anomaly and Sheriff entities). 76 # Get the Anomaly and Sheriff entities. 77 alerts = anomaly.Anomaly.GetAlertsForTest(test_key) 88 """Dumps Anomaly data for all sheriffs. 93 num_alerts: Max number of Anomaly entities (optional). 98 TestMetadata, Row, Anomaly and Sheriff entities). 119 # Add the Anomaly and Sheriff entities [all...] |
main_test.py | 13 from dashboard.models import anomaly namespace 38 foo_anomaly = anomaly.Anomaly( 68 foo_anomaly = anomaly.Anomaly(
|
email_summary.py | 14 from dashboard.models import anomaly namespace 67 recent_anomalies = anomaly.Anomaly.query( 68 anomaly.Anomaly.sheriff == sheriff_entity.key, 69 anomaly.Anomaly.timestamp > start_time).fetch()
|
email_sheriff_test.py | 16 from dashboard.models import anomaly namespace 33 test.improvement_direction = anomaly.DOWN 42 """Adds an Anomaly and returns arguments for email_sheriff.EmailSheriff.""" 45 anomaly_entity = anomaly.Anomaly( 55 'anomaly': anomaly_entity 136 args['anomaly'].median_before_anomaly = 0.0 140 self.assertIn(anomaly.FREAKIN_HUGE, str(messages[0].subject)) 142 self.assertIn(anomaly.FREAKIN_HUGE, str(messages[0].body)) 144 self.assertIn(anomaly.FREAKIN_HUGE, str(messages[0].html) [all...] |
auto_bisect_test.py | 16 from dashboard.models import anomaly namespace 36 anomaly.Anomaly( 98 anomaly.Anomaly( 110 anomaly.Anomaly( 124 anomaly.Anomaly( 149 anomaly.Anomaly [all...] |
change_internal_only_test.py | 13 from dashboard.models import anomaly namespace 47 anomaly.Anomaly( 97 # Verify that Anomaly entities were changed. 98 anomalies = anomaly.Anomaly.query().fetch()
|
find_anomalies_test.py | 14 from dashboard.models import anomaly namespace 63 """Custom matcher to test if an anomaly matches a given end rev.""" 150 anomalies = anomaly.Anomaly.query().fetch() 169 anomalies, test.key, percent_changed=100, direction=anomaly.UP, 175 anomalies, test.key, percent_changed=-50, direction=anomaly.DOWN, 182 direction=anomaly.UP, start_revision=10057, end_revision=10061, 188 anomalies, test.key, percent_changed=100, direction=anomaly.DOWN, 203 test.improvement_direction = anomaly.DOWN 206 anomalies = anomaly.Anomaly.query().fetch( [all...] |
alerts_test.py | 14 from dashboard.models import anomaly namespace 53 test.improvement_direction = anomaly.DOWN 59 anomaly_entity = anomaly.Anomaly( 71 anomaly_entity = anomaly.Anomaly( 84 anomaly_entity = anomaly.Anomaly( 160 anomalies = anomaly.Anomaly.query [all...] |
/external/chromium-trace/catapult/dashboard/dashboard/templates/ |
main.html | 8 {% macro link(anomaly) -%} 9 <a href="{{anomaly.dashboard_link}}" 28 {% for anomaly in anomalies %} 29 <tr class="{% if anomaly.improvement %}improvement{% else %}regression{% endif %}-{{anomaly.color_class}}"> 30 <td>{{link(anomaly)}}{{anomaly.percent_changed}}</a></td> 31 <td>{{link(anomaly)}}{{anomaly.start_revision}}-{{anomaly.end_revision}}</a></td [all...] |
debug_alert.html | 11 <title>Anomaly Detection Function Debugger</title> 38 <h1>Anomaly Detection Function Debugger</h1> 40 <p>This page shows the results of running the anomaly detection function 41 on some test data for some anomaly threshold configuration.</p> 49 Above each detected anomaly is a label showing the revision number 50 and percent change from the median before the anomaly to the median 51 after the anomaly.</p> 133 <h3>Recent Anomaly entities in the datastore for this test:</h3> 139 <th>Median before anomaly</th> 140 <th>Median after anomaly</th [all...] |
edit_anomaly_configs.html | 11 <title>Edit Anomaly Configs</title> 14 // This should be an Object mapping anomaly config names to Objects 33 * Updates the interface when a new anomaly config is selected to be edited. 63 <h1>Edit Anomaly Configs</h1> 65 <p><b>Note</b>: If the test path patterns of two different anomaly 66 threshold configurations both match some test, then the anomaly config 68 order will apply. You can use this to create "higher priority" anomaly
|
/external/autotest/client/cros/audio/ |
audio_analysis_unittest.py | 108 """Inserts an anomaly to the test signal. 110 The anomaly self.anomaly_samples should be created before calling this 130 """Creates an anomaly of constant samples. 141 """Runs the anomaly detection.""" 148 """Verifies that there is no anomaly in detection result.""" 154 """Verifies that there is anomaly in detection result. 156 The detection result should contain anomaly time stamps that are 157 close to where anomaly was inserted. There can be multiple anomalies 163 # Anomaly can be detected as long as the detection window of block size 164 # overlaps with anomaly [all...] |