Home | History | Annotate | Download | only in buildbot
      1 # Copyright (c) 2009, Google Inc. All rights reserved.
      2 #
      3 # Redistribution and use in source and binary forms, with or without
      4 # modification, are permitted provided that the following conditions are
      5 # met:
      6 #
      7 #     * Redistributions of source code must retain the above copyright
      8 # notice, this list of conditions and the following disclaimer.
      9 #     * Redistributions in binary form must reproduce the above
     10 # copyright notice, this list of conditions and the following disclaimer
     11 # in the documentation and/or other materials provided with the
     12 # distribution.
     13 #     * Neither the name of Google Inc. nor the names of its
     14 # contributors may be used to endorse or promote products derived from
     15 # this software without specific prior written permission.
     16 #
     17 # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     18 # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     19 # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
     20 # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
     21 # OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
     22 # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
     23 # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     24 # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     25 # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     26 # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
     27 # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     28 #
     29 # WebKit's Python module for interacting with WebKit's buildbot
     30 
     31 try:
     32     import json
     33 except ImportError:
     34     # python 2.5 compatibility
     35     import webkitpy.thirdparty.simplejson as json
     36 
     37 import operator
     38 import re
     39 import urllib
     40 import urllib2
     41 
     42 from webkitpy.common.net.failuremap import FailureMap
     43 from webkitpy.common.net.layouttestresults import LayoutTestResults
     44 from webkitpy.common.net.regressionwindow import RegressionWindow
     45 from webkitpy.common.net.testoutputset import TestOutputSet
     46 from webkitpy.common.system.logutils import get_logger
     47 from webkitpy.common.system.zipfileset import ZipFileSet
     48 from webkitpy.thirdparty.BeautifulSoup import BeautifulSoup
     49 from webkitpy.thirdparty.autoinstalled.mechanize import Browser
     50 
     51 _log = get_logger(__file__)
     52 
     53 
     54 class Builder(object):
     55     def __init__(self, name, buildbot):
     56         self._name = name
     57         self._buildbot = buildbot
     58         self._builds_cache = {}
     59         self._revision_to_build_number = None
     60         self._browser = Browser()
     61         self._browser.set_handle_robots(False) # The builder pages are excluded by robots.txt
     62 
     63     def name(self):
     64         return self._name
     65 
     66     def results_url(self):
     67         return "http://%s/results/%s" % (self._buildbot.buildbot_host, self.url_encoded_name())
     68 
     69     def url_encoded_name(self):
     70         return urllib.quote(self._name)
     71 
     72     def url(self):
     73         return "http://%s/builders/%s" % (self._buildbot.buildbot_host, self.url_encoded_name())
     74 
     75     # This provides a single place to mock
     76     def _fetch_build(self, build_number):
     77         build_dictionary = self._buildbot._fetch_build_dictionary(self, build_number)
     78         if not build_dictionary:
     79             return None
     80         return Build(self,
     81             build_number=int(build_dictionary['number']),
     82             revision=int(build_dictionary['sourceStamp']['revision']),
     83             is_green=(build_dictionary['results'] == 0) # Undocumented, 0 seems to mean "pass"
     84         )
     85 
     86     def build(self, build_number):
     87         if not build_number:
     88             return None
     89         cached_build = self._builds_cache.get(build_number)
     90         if cached_build:
     91             return cached_build
     92 
     93         build = self._fetch_build(build_number)
     94         self._builds_cache[build_number] = build
     95         return build
     96 
     97     def latest_cached_build(self):
     98         revision_build_pairs = self.revision_build_pairs_with_results()
     99         revision_build_pairs.sort(key=lambda i: i[1])
    100         latest_build_number = revision_build_pairs[-1][1]
    101         return self.build(latest_build_number)
    102 
    103     def force_build(self, username="webkit-patch", comments=None):
    104         def predicate(form):
    105             try:
    106                 return form.find_control("username")
    107             except Exception, e:
    108                 return False
    109         self._browser.open(self.url())
    110         self._browser.select_form(predicate=predicate)
    111         self._browser["username"] = username
    112         if comments:
    113             self._browser["comments"] = comments
    114         return self._browser.submit()
    115 
    116     file_name_regexp = re.compile(r"r(?P<revision>\d+) \((?P<build_number>\d+)\)")
    117     def _revision_and_build_for_filename(self, filename):
    118         # Example: "r47483 (1)/" or "r47483 (1).zip"
    119         match = self.file_name_regexp.match(filename)
    120         return (int(match.group("revision")), int(match.group("build_number")))
    121 
    122     def _fetch_revision_to_build_map(self):
    123         # All _fetch requests go through _buildbot for easier mocking
    124         # FIXME: This should use NetworkTransaction's 404 handling instead.
    125         try:
    126             # FIXME: This method is horribly slow due to the huge network load.
    127             # FIXME: This is a poor way to do revision -> build mapping.
    128             # Better would be to ask buildbot through some sort of API.
    129             print "Loading revision/build list from %s." % self.results_url()
    130             print "This may take a while..."
    131             result_files = self._buildbot._fetch_twisted_directory_listing(self.results_url())
    132         except urllib2.HTTPError, error:
    133             if error.code != 404:
    134                 raise
    135             result_files = []
    136 
    137         # This assumes there was only one build per revision, which is false but we don't care for now.
    138         return dict([self._revision_and_build_for_filename(file_info["filename"]) for file_info in result_files])
    139 
    140     def _revision_to_build_map(self):
    141         if not self._revision_to_build_number:
    142             self._revision_to_build_number = self._fetch_revision_to_build_map()
    143         return self._revision_to_build_number
    144 
    145     def revision_build_pairs_with_results(self):
    146         return self._revision_to_build_map().items()
    147 
    148     # This assumes there can be only one build per revision, which is false, but we don't care for now.
    149     def build_for_revision(self, revision, allow_failed_lookups=False):
    150         # NOTE: This lookup will fail if that exact revision was never built.
    151         build_number = self._revision_to_build_map().get(int(revision))
    152         if not build_number:
    153             return None
    154         build = self.build(build_number)
    155         if not build and allow_failed_lookups:
    156             # Builds for old revisions with fail to lookup via buildbot's json api.
    157             build = Build(self,
    158                 build_number=build_number,
    159                 revision=revision,
    160                 is_green=False,
    161             )
    162         return build
    163 
    164     def find_regression_window(self, red_build, look_back_limit=30):
    165         if not red_build or red_build.is_green():
    166             return RegressionWindow(None, None)
    167         common_failures = None
    168         current_build = red_build
    169         build_after_current_build = None
    170         look_back_count = 0
    171         while current_build:
    172             if current_build.is_green():
    173                 # current_build can't possibly have any failures in common
    174                 # with red_build because it's green.
    175                 break
    176             results = current_build.layout_test_results()
    177             # We treat a lack of results as if all the test failed.
    178             # This occurs, for example, when we can't compile at all.
    179             if results:
    180                 failures = set(results.failing_tests())
    181                 if common_failures == None:
    182                     common_failures = failures
    183                 else:
    184                     common_failures = common_failures.intersection(failures)
    185                     if not common_failures:
    186                         # current_build doesn't have any failures in common with
    187                         # the red build we're worried about.  We assume that any
    188                         # failures in current_build were due to flakiness.
    189                         break
    190             look_back_count += 1
    191             if look_back_count > look_back_limit:
    192                 return RegressionWindow(None, current_build, failing_tests=common_failures)
    193             build_after_current_build = current_build
    194             current_build = current_build.previous_build()
    195         # We must iterate at least once because red_build is red.
    196         assert(build_after_current_build)
    197         # Current build must either be green or have no failures in common
    198         # with red build, so we've found our failure transition.
    199         return RegressionWindow(current_build, build_after_current_build, failing_tests=common_failures)
    200 
    201     def find_blameworthy_regression_window(self, red_build_number, look_back_limit=30, avoid_flakey_tests=True):
    202         red_build = self.build(red_build_number)
    203         regression_window = self.find_regression_window(red_build, look_back_limit)
    204         if not regression_window.build_before_failure():
    205             return None  # We ran off the limit of our search
    206         # If avoid_flakey_tests, require at least 2 bad builds before we
    207         # suspect a real failure transition.
    208         if avoid_flakey_tests and regression_window.failing_build() == red_build:
    209             return None
    210         return regression_window
    211 
    212 
    213 class Build(object):
    214     def __init__(self, builder, build_number, revision, is_green):
    215         self._builder = builder
    216         self._number = build_number
    217         self._revision = revision
    218         self._is_green = is_green
    219         self._layout_test_results = None
    220 
    221     @staticmethod
    222     def build_url(builder, build_number):
    223         return "%s/builds/%s" % (builder.url(), build_number)
    224 
    225     def url(self):
    226         return self.build_url(self.builder(), self._number)
    227 
    228     def results_url(self):
    229         results_directory = "r%s (%s)" % (self.revision(), self._number)
    230         return "%s/%s" % (self._builder.results_url(), urllib.quote(results_directory))
    231 
    232     def results_zip_url(self):
    233         return "%s.zip" % self.results_url()
    234 
    235     def results(self):
    236         return TestOutputSet(self._builder.name(), None, ZipFileSet(self.results_zip_url()), include_expected=False)
    237 
    238     def _fetch_results_html(self):
    239         results_html = "%s/results.html" % (self.results_url())
    240         # FIXME: This should use NetworkTransaction's 404 handling instead.
    241         try:
    242             # It seems this can return None if the url redirects and then returns 404.
    243             return urllib2.urlopen(results_html)
    244         except urllib2.HTTPError, error:
    245             if error.code != 404:
    246                 raise
    247 
    248     def layout_test_results(self):
    249         if not self._layout_test_results:
    250             # FIXME: This should cache that the result was a 404 and stop hitting the network.
    251             self._layout_test_results = LayoutTestResults.results_from_string(self._fetch_results_html())
    252         return self._layout_test_results
    253 
    254     def builder(self):
    255         return self._builder
    256 
    257     def revision(self):
    258         return self._revision
    259 
    260     def is_green(self):
    261         return self._is_green
    262 
    263     def previous_build(self):
    264         # previous_build() allows callers to avoid assuming build numbers are sequential.
    265         # They may not be sequential across all master changes, or when non-trunk builds are made.
    266         return self._builder.build(self._number - 1)
    267 
    268 
    269 class BuildBot(object):
    270     # FIXME: This should move into common.config.urls.
    271     default_host = "build.webkit.org"
    272 
    273     def __init__(self, host=default_host):
    274         self.buildbot_host = host
    275         self._builder_by_name = {}
    276 
    277         # If any core builder is red we should not be landing patches.  Other
    278         # builders should be added to this list once they are known to be
    279         # reliable.
    280         # See https://bugs.webkit.org/show_bug.cgi?id=33296 and related bugs.
    281         self.core_builder_names_regexps = [
    282             "SnowLeopard.*Build",
    283             "SnowLeopard.*\(Test",
    284             "SnowLeopard.*\(WebKit2 Test",
    285             "Leopard.*",
    286             "Windows.*Build",
    287             "Windows.*\(Test",
    288             "WinCairo",
    289             "WinCE",
    290             "EFL",
    291             "GTK.*32",
    292             "GTK.*64.*Debug",  # Disallow the 64-bit Release bot which is broken.
    293             "Qt",
    294             "Chromium.*Release$",
    295         ]
    296 
    297     def _parse_last_build_cell(self, builder, cell):
    298         status_link = cell.find('a')
    299         if status_link:
    300             # Will be either a revision number or a build number
    301             revision_string = status_link.string
    302             # If revision_string has non-digits assume it's not a revision number.
    303             builder['built_revision'] = int(revision_string) \
    304                                         if not re.match('\D', revision_string) \
    305                                         else None
    306 
    307             # FIXME: We treat slave lost as green even though it is not to
    308             # work around the Qts bot being on a broken internet connection.
    309             # The real fix is https://bugs.webkit.org/show_bug.cgi?id=37099
    310             builder['is_green'] = not re.search('fail', cell.renderContents()) or \
    311                                   not not re.search('lost', cell.renderContents())
    312 
    313             status_link_regexp = r"builders/(?P<builder_name>.*)/builds/(?P<build_number>\d+)"
    314             link_match = re.match(status_link_regexp, status_link['href'])
    315             builder['build_number'] = int(link_match.group("build_number"))
    316         else:
    317             # We failed to find a link in the first cell, just give up.  This
    318             # can happen if a builder is just-added, the first cell will just
    319             # be "no build"
    320             # Other parts of the code depend on is_green being present.
    321             builder['is_green'] = False
    322             builder['built_revision'] = None
    323             builder['build_number'] = None
    324 
    325     def _parse_current_build_cell(self, builder, cell):
    326         activity_lines = cell.renderContents().split("<br />")
    327         builder["activity"] = activity_lines[0] # normally "building" or "idle"
    328         # The middle lines document how long left for any current builds.
    329         match = re.match("(?P<pending_builds>\d) pending", activity_lines[-1])
    330         builder["pending_builds"] = int(match.group("pending_builds")) if match else 0
    331 
    332     def _parse_builder_status_from_row(self, status_row):
    333         status_cells = status_row.findAll('td')
    334         builder = {}
    335 
    336         # First cell is the name
    337         name_link = status_cells[0].find('a')
    338         builder["name"] = unicode(name_link.string)
    339 
    340         self._parse_last_build_cell(builder, status_cells[1])
    341         self._parse_current_build_cell(builder, status_cells[2])
    342         return builder
    343 
    344     def _matches_regexps(self, builder_name, name_regexps):
    345         for name_regexp in name_regexps:
    346             if re.match(name_regexp, builder_name):
    347                 return True
    348         return False
    349 
    350     # FIXME: Should move onto Builder
    351     def _is_core_builder(self, builder_name):
    352         return self._matches_regexps(builder_name, self.core_builder_names_regexps)
    353 
    354     # FIXME: This method needs to die, but is used by a unit test at the moment.
    355     def _builder_statuses_with_names_matching_regexps(self, builder_statuses, name_regexps):
    356         return [builder for builder in builder_statuses if self._matches_regexps(builder["name"], name_regexps)]
    357 
    358     def red_core_builders(self):
    359         return [builder for builder in self.core_builder_statuses() if not builder["is_green"]]
    360 
    361     def red_core_builders_names(self):
    362         return [builder["name"] for builder in self.red_core_builders()]
    363 
    364     def idle_red_core_builders(self):
    365         return [builder for builder in self.red_core_builders() if builder["activity"] == "idle"]
    366 
    367     def core_builders_are_green(self):
    368         return not self.red_core_builders()
    369 
    370     # FIXME: These _fetch methods should move to a networking class.
    371     def _fetch_build_dictionary(self, builder, build_number):
    372         try:
    373             base = "http://%s" % self.buildbot_host
    374             path = urllib.quote("json/builders/%s/builds/%s" % (builder.name(),
    375                                                                 build_number))
    376             url = "%s/%s" % (base, path)
    377             jsondata = urllib2.urlopen(url)
    378             return json.load(jsondata)
    379         except urllib2.URLError, err:
    380             build_url = Build.build_url(builder, build_number)
    381             _log.error("Error fetching data for %s build %s (%s): %s" % (builder.name(), build_number, build_url, err))
    382             return None
    383         except ValueError, err:
    384             build_url = Build.build_url(builder, build_number)
    385             _log.error("Error decoding json data from %s: %s" % (build_url, err))
    386             return None
    387 
    388     def _fetch_one_box_per_builder(self):
    389         build_status_url = "http://%s/one_box_per_builder" % self.buildbot_host
    390         return urllib2.urlopen(build_status_url)
    391 
    392     def _file_cell_text(self, file_cell):
    393         """Traverses down through firstChild elements until one containing a string is found, then returns that string"""
    394         element = file_cell
    395         while element.string is None and element.contents:
    396             element = element.contents[0]
    397         return element.string
    398 
    399     def _parse_twisted_file_row(self, file_row):
    400         string_or_empty = lambda string: unicode(string) if string else u""
    401         file_cells = file_row.findAll('td')
    402         return {
    403             "filename": string_or_empty(self._file_cell_text(file_cells[0])),
    404             "size": string_or_empty(self._file_cell_text(file_cells[1])),
    405             "type": string_or_empty(self._file_cell_text(file_cells[2])),
    406             "encoding": string_or_empty(self._file_cell_text(file_cells[3])),
    407         }
    408 
    409     def _parse_twisted_directory_listing(self, page):
    410         soup = BeautifulSoup(page)
    411         # HACK: Match only table rows with a class to ignore twisted header/footer rows.
    412         file_rows = soup.find('table').findAll('tr', {'class': re.compile(r'\b(?:directory|file)\b')})
    413         return [self._parse_twisted_file_row(file_row) for file_row in file_rows]
    414 
    415     # FIXME: There should be a better way to get this information directly from twisted.
    416     def _fetch_twisted_directory_listing(self, url):
    417         return self._parse_twisted_directory_listing(urllib2.urlopen(url))
    418 
    419     def builders(self):
    420         return [self.builder_with_name(status["name"]) for status in self.builder_statuses()]
    421 
    422     # This method pulls from /one_box_per_builder as an efficient way to get information about
    423     def builder_statuses(self):
    424         soup = BeautifulSoup(self._fetch_one_box_per_builder())
    425         return [self._parse_builder_status_from_row(status_row) for status_row in soup.find('table').findAll('tr')]
    426 
    427     def core_builder_statuses(self):
    428         return [builder for builder in self.builder_statuses() if self._is_core_builder(builder["name"])]
    429 
    430     def builder_with_name(self, name):
    431         builder = self._builder_by_name.get(name)
    432         if not builder:
    433             builder = Builder(name, self)
    434             self._builder_by_name[name] = builder
    435         return builder
    436 
    437     def failure_map(self, only_core_builders=True):
    438         builder_statuses = self.core_builder_statuses() if only_core_builders else self.builder_statuses()
    439         failure_map = FailureMap()
    440         revision_to_failing_bots = {}
    441         for builder_status in builder_statuses:
    442             if builder_status["is_green"]:
    443                 continue
    444             builder = self.builder_with_name(builder_status["name"])
    445             regression_window = builder.find_blameworthy_regression_window(builder_status["build_number"])
    446             if regression_window:
    447                 failure_map.add_regression_window(builder, regression_window)
    448         return failure_map
    449 
    450     # This makes fewer requests than calling Builder.latest_build would.  It grabs all builder
    451     # statuses in one request using self.builder_statuses (fetching /one_box_per_builder instead of builder pages).
    452     def _latest_builds_from_builders(self, only_core_builders=True):
    453         builder_statuses = self.core_builder_statuses() if only_core_builders else self.builder_statuses()
    454         return [self.builder_with_name(status["name"]).build(status["build_number"]) for status in builder_statuses]
    455 
    456     def _build_at_or_before_revision(self, build, revision):
    457         while build:
    458             if build.revision() <= revision:
    459                 return build
    460             build = build.previous_build()
    461 
    462     def last_green_revision(self, only_core_builders=True):
    463         builds = self._latest_builds_from_builders(only_core_builders)
    464         target_revision = builds[0].revision()
    465         # An alternate way to do this would be to start at one revision and walk backwards
    466         # checking builder.build_for_revision, however build_for_revision is very slow on first load.
    467         while True:
    468             # Make builds agree on revision
    469             builds = [self._build_at_or_before_revision(build, target_revision) for build in builds]
    470             if None in builds: # One of the builds failed to load from the server.
    471                 return None
    472             min_revision = min(map(lambda build: build.revision(), builds))
    473             if min_revision != target_revision:
    474                 target_revision = min_revision
    475                 continue # Builds don't all agree on revision, keep searching
    476             # Check to make sure they're all green
    477             all_are_green = reduce(operator.and_, map(lambda build: build.is_green(), builds))
    478             if not all_are_green:
    479                 target_revision -= 1
    480                 continue
    481             return min_revision
    482