Home | History | Annotate | Download | only in perf_upload
      1 #!/usr/bin/python
      2 
      3 """Unit tests for the perf_uploader.py module.
      4 
      5 """
      6 
      7 import json, unittest
      8 
      9 import common
     10 from autotest_lib.tko import models as tko_models
     11 from autotest_lib.tko.perf_upload import perf_uploader
     12 
     13 
     14 class test_aggregate_iterations(unittest.TestCase):
     15     """Tests for the aggregate_iterations function."""
     16 
     17     _PERF_ITERATION_DATA = {
     18         '1': [
     19             {
     20                 'description': 'metric1',
     21                 'value': 1,
     22                 'stddev': 0.0,
     23                 'units': 'units1',
     24                 'higher_is_better': True,
     25                 'graph': None
     26             },
     27             {
     28                 'description': 'metric2',
     29                 'value': 10,
     30                 'stddev': 0.0,
     31                 'units': 'units2',
     32                 'higher_is_better': True,
     33                 'graph': 'graph1',
     34             },
     35             {
     36                 'description': 'metric2',
     37                 'value': 100,
     38                 'stddev': 1.7,
     39                 'units': 'units3',
     40                 'higher_is_better': False,
     41                 'graph': 'graph2',
     42             }
     43         ],
     44         '2': [
     45             {
     46                 'description': 'metric1',
     47                 'value': 2,
     48                 'stddev': 0.0,
     49                 'units': 'units1',
     50                 'higher_is_better': True,
     51                 'graph': None,
     52             },
     53             {
     54                 'description': 'metric2',
     55                 'value': 20,
     56                 'stddev': 0.0,
     57                 'units': 'units2',
     58                 'higher_is_better': True,
     59                 'graph': 'graph1',
     60             },
     61             {
     62                 'description': 'metric2',
     63                 'value': 200,
     64                 'stddev': 21.2,
     65                 'units': 'units3',
     66                 'higher_is_better': False,
     67                 'graph': 'graph2',
     68             }
     69         ],
     70     }
     71 
     72 
     73     def setUp(self):
     74         """Sets up for each test case."""
     75         self._perf_values = []
     76         for iter_num, iter_data in self._PERF_ITERATION_DATA.iteritems():
     77             self._perf_values.append(
     78                     tko_models.perf_value_iteration(iter_num, iter_data))
     79 
     80 
     81     def test_one_iteration(self):
     82         """Tests that data for 1 iteration is aggregated properly."""
     83         result = perf_uploader._aggregate_iterations([self._perf_values[0]])
     84         self.assertEqual(len(result), 3, msg='Expected results for 3 metrics.')
     85         key = [('metric1', None), ('metric2', 'graph1'), ('metric2', 'graph2')]
     86         self.assertTrue(
     87             all([x in result for x in key]),
     88             msg='Parsed metrics not as expected.')
     89         msg = 'Perf values for metric not aggregated properly.'
     90         self.assertEqual(result[('metric1', None)]['value'], [1], msg=msg)
     91         self.assertEqual(result[('metric2', 'graph1')]['value'], [10], msg=msg)
     92         self.assertEqual(result[('metric2', 'graph2')]['value'], [100], msg=msg)
     93         msg = 'Standard deviation values not retained properly.'
     94         self.assertEqual(result[('metric1', None)]['stddev'], 0.0, msg=msg)
     95         self.assertEqual(result[('metric2', 'graph1')]['stddev'], 0.0, msg=msg)
     96         self.assertEqual(result[('metric2', 'graph2')]['stddev'], 1.7, msg=msg)
     97 
     98 
     99     def test_two_iterations(self):
    100         """Tests that data for 2 iterations is aggregated properly."""
    101         result = perf_uploader._aggregate_iterations(self._perf_values)
    102         self.assertEqual(len(result), 3, msg='Expected results for 3 metrics.')
    103         key = [('metric1', None), ('metric2', 'graph1'), ('metric2', 'graph2')]
    104         self.assertTrue(
    105             all([x in result for x in key]),
    106             msg='Parsed metrics not as expected.')
    107         msg = 'Perf values for metric not aggregated properly.'
    108         self.assertEqual(result[('metric1', None)]['value'], [1, 2], msg=msg)
    109         self.assertEqual(result[('metric2', 'graph1')]['value'], [10, 20],
    110                          msg=msg)
    111         self.assertEqual(result[('metric2', 'graph2')]['value'], [100, 200],
    112                          msg=msg)
    113 
    114 
    115 class test_compute_avg_stddev(unittest.TestCase):
    116     """Tests for the compute_avg_stddev function."""
    117 
    118     def setUp(self):
    119         """Sets up for each test case."""
    120         self._perf_values = {
    121             'metric1': {'value': [10, 20, 30], 'stddev': 0.0},
    122             'metric2': {'value': [2.0, 3.0, 4.0], 'stddev': 0.0},
    123             'metric3': {'value': [1], 'stddev': 1.7},
    124         }
    125 
    126 
    127     def test_avg_stddev(self):
    128         """Tests that averages and standard deviations are computed properly."""
    129         perf_uploader._compute_avg_stddev(self._perf_values)
    130         result = self._perf_values  # The input dictionary itself is modified.
    131         self.assertEqual(len(result), 3, msg='Expected results for 3 metrics.')
    132         self.assertTrue(
    133             all([x in result for x in ['metric1', 'metric2', 'metric3']]),
    134             msg='Parsed metrics not as expected.')
    135         msg = 'Average value not computed properly.'
    136         self.assertEqual(result['metric1']['value'], 20, msg=msg)
    137         self.assertEqual(result['metric2']['value'], 3.0, msg=msg)
    138         self.assertEqual(result['metric3']['value'], 1, msg=msg)
    139         msg = 'Standard deviation value not computed properly.'
    140         self.assertEqual(result['metric1']['stddev'], 10.0, msg=msg)
    141         self.assertEqual(result['metric2']['stddev'], 1.0, msg=msg)
    142         self.assertEqual(result['metric3']['stddev'], 1.7, msg=msg)
    143 
    144 
    145 class test_json_config_file_sanity(unittest.TestCase):
    146     """Sanity tests for the JSON-formatted presentation config file."""
    147 
    148     def test_parse_json(self):
    149         """Verifies _parse_config_file function."""
    150         perf_uploader._parse_config_file(
    151                 perf_uploader._PRESENTATION_CONFIG_FILE)
    152 
    153 
    154     def test_proper_json(self):
    155         """Verifies the file can be parsed as proper JSON."""
    156         try:
    157             with open(perf_uploader._PRESENTATION_CONFIG_FILE, 'r') as fp:
    158                 json.load(fp)
    159         except:
    160             self.fail('Presentation config file could not be parsed as JSON.')
    161 
    162 
    163     def test_unique_test_names(self):
    164         """Verifies that each test name appears only once in the JSON file."""
    165         json_obj = []
    166         try:
    167             with open(perf_uploader._PRESENTATION_CONFIG_FILE, 'r') as fp:
    168                 json_obj = json.load(fp)
    169         except:
    170             self.fail('Presentation config file could not be parsed as JSON.')
    171 
    172         name_set = set([x['autotest_name'] for x in json_obj])
    173         self.assertEqual(len(name_set), len(json_obj),
    174                          msg='Autotest names not unique in the JSON file.')
    175 
    176 
    177     def test_required_master_name(self):
    178         """Verifies that master name must be specified."""
    179         json_obj = []
    180         try:
    181             with open(perf_uploader._PRESENTATION_CONFIG_FILE, 'r') as fp:
    182                 json_obj = json.load(fp)
    183         except:
    184             self.fail('Presentation config file could not be parsed as JSON.')
    185 
    186         for entry in json_obj:
    187             if not 'master_name' in entry:
    188                 self.fail('Missing master field for test %s.' %
    189                           entry['autotest_name'])
    190 
    191 
    192 class test_gather_presentation_info(unittest.TestCase):
    193     """Tests for the gather_presentation_info function."""
    194 
    195     _PRESENT_INFO = {
    196         'test_name': {
    197             'master_name': 'new_master_name',
    198             'dashboard_test_name': 'new_test_name',
    199         }
    200     }
    201 
    202     _PRESENT_INFO_MISSING_MASTER = {
    203         'test_name': {
    204             'dashboard_test_name': 'new_test_name',
    205         }
    206     }
    207 
    208 
    209     def test_test_name_specified(self):
    210         """Verifies gathers presentation info correctly."""
    211         result = perf_uploader._gather_presentation_info(
    212                 self._PRESENT_INFO, 'test_name')
    213         self.assertTrue(
    214                 all([key in result for key in
    215                      ['test_name', 'master_name']]),
    216                 msg='Unexpected keys in resulting dictionary: %s' % result)
    217         self.assertEqual(result['master_name'], 'new_master_name',
    218                          msg='Unexpected "master_name" value: %s' %
    219                              result['master_name'])
    220         self.assertEqual(result['test_name'], 'new_test_name',
    221                          msg='Unexpected "test_name" value: %s' %
    222                              result['test_name'])
    223 
    224 
    225     def test_test_name_not_specified(self):
    226         """Verifies exception raised if test is not there."""
    227         self.assertRaises(
    228                 perf_uploader.PerfUploadingError,
    229                 perf_uploader._gather_presentation_info,
    230                         self._PRESENT_INFO, 'other_test_name')
    231 
    232 
    233     def test_master_not_specified(self):
    234         """Verifies exception raised if master is not there."""
    235         self.assertRaises(
    236                 perf_uploader.PerfUploadingError,
    237                 perf_uploader._gather_presentation_info,
    238                     self._PRESENT_INFO_MISSING_MASTER, 'test_name')
    239 
    240 
    241 class test_get_id_from_version(unittest.TestCase):
    242     """Tests for the _get_id_from_version function."""
    243 
    244     def test_correctly_formatted_versions(self):
    245         """Verifies that the expected ID is returned when input is OK."""
    246         chrome_version = '27.0.1452.2'
    247         cros_version = '27.3906.0.0'
    248         # 1452.2 + 3906.0.0
    249         # --> 01452 + 002 + 03906 + 000 + 00
    250         # --> 14520020390600000
    251         self.assertEqual(
    252                 14520020390600000,
    253                 perf_uploader._get_id_from_version(
    254                         chrome_version, cros_version))
    255 
    256         chrome_version = '25.10.1000.0'
    257         cros_version = '25.1200.0.0'
    258         # 1000.0 + 1200.0.0
    259         # --> 01000 + 000 + 01200 + 000 + 00
    260         # --> 10000000120000000
    261         self.assertEqual(
    262                 10000000120000000,
    263                 perf_uploader._get_id_from_version(
    264                         chrome_version, cros_version))
    265 
    266     def test_returns_none_when_given_invalid_input(self):
    267         """Checks the return value when invalid input is given."""
    268         chrome_version = '27.0'
    269         cros_version = '27.3906.0.0'
    270         self.assertIsNone(perf_uploader._get_id_from_version(
    271                 chrome_version, cros_version))
    272 
    273 
    274 class test_get_version_numbers(unittest.TestCase):
    275     """Tests for the _get_version_numbers function."""
    276 
    277     def test_with_valid_versions(self):
    278       """Checks the version numbers used when data is formatted as expected."""
    279       self.assertEqual(
    280               ('34.5678.9.0', '34.5.678.9'),
    281               perf_uploader._get_version_numbers(
    282                   {
    283                       'CHROME_VERSION': '34.5.678.9',
    284                       'CHROMEOS_RELEASE_VERSION': '5678.9.0',
    285                   }))
    286 
    287     def test_with_missing_version_raises_error(self):
    288       """Checks that an error is raised when a version is missing."""
    289       with self.assertRaises(perf_uploader.PerfUploadingError):
    290           perf_uploader._get_version_numbers(
    291               {
    292                   'CHROMEOS_RELEASE_VERSION': '5678.9.0',
    293               })
    294 
    295     def test_with_unexpected_version_format_raises_error(self):
    296       """Checks that an error is raised when there's a rN suffix."""
    297       with self.assertRaises(perf_uploader.PerfUploadingError):
    298           perf_uploader._get_version_numbers(
    299               {
    300                   'CHROME_VERSION': '34.5.678.9',
    301                   'CHROMEOS_RELEASE_VERSION': '5678.9.0r1',
    302               })
    303 
    304 
    305 class test_format_for_upload(unittest.TestCase):
    306     """Tests for the format_for_upload function."""
    307 
    308     _PERF_DATA = {
    309         ('metric1', 'graph_name'): {
    310             'value': 2.7,
    311             'stddev': 0.2,
    312             'units': 'msec',
    313             'graph': 'graph_name',
    314             'higher_is_better': False,
    315         },
    316         ('metric2', None): {
    317             'value': 101.35,
    318             'stddev': 5.78,
    319             'units': 'frames_per_sec',
    320             'graph': None,
    321             'higher_is_better': True,
    322         },
    323     }
    324 
    325     _PRESENT_INFO = {
    326         'master_name': 'new_master_name',
    327         'test_name': 'new_test_name',
    328     }
    329 
    330     def setUp(self):
    331         self._perf_data = self._PERF_DATA
    332 
    333     def _verify_result_string(self, actual_result, expected_result):
    334         """Verifies a JSON string matches the expected result.
    335 
    336         This function compares JSON objects rather than strings, because of
    337         possible floating-point values that need to be compared using
    338         assertAlmostEqual().
    339 
    340         @param actual_result: The candidate JSON string.
    341         @param expected_result: The reference JSON string that the candidate
    342             must match.
    343 
    344         """
    345         actual = json.loads(actual_result)
    346         expected = json.loads(expected_result)
    347 
    348         fail_msg = 'Unexpected result string: %s' % actual_result
    349         self.assertEqual(len(actual), len(expected), msg=fail_msg)
    350         # Make sure the dictionaries in 'expected' are in the same order
    351         # as the dictionaries in 'actual' before comparing their values.
    352         actual = sorted(actual, key=lambda x: x['test'])
    353         expected = sorted(expected, key=lambda x: x['test'])
    354         # Now compare the results.
    355         for idx in xrange(len(actual)):
    356             keys_actual = set(actual[idx].keys())
    357             keys_expected = set(expected[idx].keys())
    358             self.assertEqual(len(keys_actual), len(keys_expected),
    359                              msg=fail_msg)
    360             self.assertTrue(all([key in keys_actual for key in keys_expected]),
    361                             msg=fail_msg)
    362 
    363             self.assertEqual(
    364                     actual[idx]['supplemental_columns']['r_cros_version'],
    365                     expected[idx]['supplemental_columns']['r_cros_version'],
    366                     msg=fail_msg)
    367             self.assertEqual(
    368                     actual[idx]['supplemental_columns']['r_chrome_version'],
    369                     expected[idx]['supplemental_columns']['r_chrome_version'],
    370                     msg=fail_msg)
    371             self.assertEqual(
    372                     actual[idx]['supplemental_columns']['a_default_rev'],
    373                     expected[idx]['supplemental_columns']['a_default_rev'],
    374                     msg=fail_msg)
    375             self.assertEqual(
    376                     actual[idx]['supplemental_columns']['a_hardware_identifier'],
    377                     expected[idx]['supplemental_columns']['a_hardware_identifier'],
    378                     msg=fail_msg)
    379             self.assertEqual(
    380                     actual[idx]['supplemental_columns']['a_hardware_hostname'],
    381                     expected[idx]['supplemental_columns']['a_hardware_hostname'],
    382                     msg=fail_msg)
    383             self.assertEqual(
    384                     actual[idx]['bot'], expected[idx]['bot'], msg=fail_msg)
    385             self.assertEqual(
    386                     actual[idx]['revision'], expected[idx]['revision'], msg=fail_msg)
    387             self.assertAlmostEqual(
    388                     actual[idx]['value'], expected[idx]['value'], 4,
    389                     msg=fail_msg)
    390             self.assertEqual(
    391                     actual[idx]['units'], expected[idx]['units'], msg=fail_msg)
    392             self.assertEqual(
    393                     actual[idx]['master'], expected[idx]['master'],
    394                     msg=fail_msg)
    395             self.assertAlmostEqual(
    396                     actual[idx]['error'], expected[idx]['error'], 4,
    397                     msg=fail_msg)
    398             self.assertEqual(
    399                     actual[idx]['test'], expected[idx]['test'], msg=fail_msg)
    400             self.assertEqual(
    401                     actual[idx]['higher_is_better'],
    402                     expected[idx]['higher_is_better'], msg=fail_msg)
    403 
    404 
    405     def test_format_for_upload(self):
    406         """Verifies format_for_upload generates correct json data."""
    407         result = perf_uploader._format_for_upload(
    408                 'platform', '25.1200.0.0', '25.10.1000.0', 'WINKY E2A-F2K-Q35',
    409                 'i7', 'test_machine', self._perf_data, self._PRESENT_INFO)
    410         expected_result_string = (
    411                 '[{"supplemental_columns": {"r_cros_version": "25.1200.0.0", '
    412                 '"a_default_rev" : "r_chrome_version",'
    413                 '"a_hardware_identifier" : "WINKY E2A-F2K-Q35",'
    414                 '"a_hardware_hostname" : "test_machine",'
    415                 '"r_chrome_version": "25.10.1000.0"}, "bot": "cros-platform-i7", '
    416                 '"higher_is_better": false, "value": 2.7, '
    417                 '"revision": 10000000120000000, '
    418                 '"units": "msec", "master": "new_master_name", '
    419                 '"error": 0.2, "test": "new_test_name/graph_name/metric1"}, '
    420                 '{"supplemental_columns": {"r_cros_version": "25.1200.0.0", '
    421                 '"a_default_rev" : "r_chrome_version",'
    422                 '"a_hardware_identifier" : "WINKY E2A-F2K-Q35",'
    423                 '"a_hardware_hostname" : "test_machine",'
    424                 '"r_chrome_version": "25.10.1000.0"}, "bot": "cros-platform-i7", '
    425                 '"higher_is_better": true, "value": 101.35, '
    426                 '"revision": 10000000120000000, '
    427                 '"units": "frames_per_sec", "master": "new_master_name", '
    428                 '"error": 5.78, "test": "new_test_name/metric2"}]')
    429 
    430         self._verify_result_string(result['data'], expected_result_string)
    431 
    432 
    433 if __name__ == '__main__':
    434     unittest.main()
    435