Example #1
0
 def init_core(self, branch):
     name = 'core'
     return GitRepo(dirname=os.path.join(self.repos_root, name),
                    repo_name=name,
                    upstream_name='cfengine',
                    my_name=self.username,
                    checkout_branch=branch)
Example #2
0
 def init_repos(self, checkout_branch, checkout_tag):
     repos = (GitRepo(dirname=os.path.join(self.repos_root, name),
                      repo_name=name,
                      upstream_name='cfengine',
                      my_name=self.username,
                      checkout_branch=checkout_branch,
                      checkout_tag=checkout_tag)
              for name in self.repo_names)
     return repos
Example #3
0
 def run(self, branch):
     """Run the dependency update for a branch, creating PR in the end"""
     self.slack.reply("Running dependency updates for " + branch)
     # prepare repo
     repo_name = 'buildscripts'
     upstream_name = 'cfengine'
     local_path = "../" + repo_name
     self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                 self.username, branch)
     timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
     new_branchname = '{}-deps-{}'.format(branch, timestamp)
     self.buildscripts.checkout(new_branchname, new=True)
     self.readme_file_path = 'deps-packaging/README.md'
     readme_file = self.buildscripts.get_file(self.readme_file_path)
     self.readme_lines = readme_file.split('\n')
     self.monitoring_file_path = 'deps-packaging/release-monitoring.json'
     self.monitoring_ids = json.loads(
         self.buildscripts.get_file(self.monitoring_file_path))
     updates_summary = []
     only_deps = self.get_deps_list(branch)
     for dep in only_deps:
         single_result = self.update_single_dep(dep)
         if single_result:
             updates_summary.append(single_result)
             self.slack.reply(single_result)
     if len(updates_summary) == 0:
         self.slack.reply("Dependency checked, nothing to update")
         return
     self.buildscripts.push(new_branchname)
     updates_summary = '\n'.join(updates_summary)
     pr_text = self.github.create_pr(
         target_repo='{}/{}'.format(upstream_name, repo_name),
         target_branch=branch,
         source_user=self.username,
         source_branch=new_branchname,
         title='Dependency updates for ' + branch,
         text=updates_summary)
     self.slack.reply(
         "Dependency updates:\n```\n{}\n```\n{}".format(
             updates_summary, pr_text), True)
Example #4
0
 def run(self, branch):
     """Generate changelogs on a branch, creating PR in the end"""
     self.slack.reply("Generating changelogs on " + branch)
     repo_names = [
         'core', 'masterfiles', 'enterprise', 'nova', 'mission-portal',
         'buildscripts'
     ]
     # checkout all repos to the required branch
     repos = (GitRepo(dirname=os.path.join(self.repos_root, name),
                      repo_name=name,
                      upstream_name='cfengine',
                      my_name=self.username,
                      checkout_branch=branch) for name in repo_names)
     # generate changelogs
     prs = (self.generate_changelog_in_repo(branch, repo) for repo in repos)
     # currently prs has False elements corresponding to repos which don't
     # have changelogs. Let's filter them out
     prs = (x for x in prs if x)
     self.slack.reply("Changelog PRs:\n{}".format('\n'.join(prs)), True)
Example #5
0
    def update_deps_version(self, branches):
        # prepare repo
        repo_name = 'buildscripts'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                    self.username, 'master')

        branches = branches.split(',')

        # fetch versions from all branches
        deps_table = {}
        # deps_table is a 2d dict: deps_table[dep][branch]=version
        branch_column_widths = {}
        for branch in branches:
            branch_column_widths[branch] = len(branch)
            self.buildscripts.checkout(branch)
            self.buildscripts.run_command('pull')
            deps_versions = self.collect_deps(branch)
            # deps_versions is a dict: deps_versions[dep]=version
            for dep in deps_versions:
                if not dep in deps_table:
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                deps_table[dep][branch] = deps_versions[dep]
                branch_column_widths[branch] = max(
                    branch_column_widths[branch], len(deps_versions[dep]))

        # patch the readme
        self.buildscripts.checkout('master')
        self.readme_file_path = 'README.md'
        readme_file = self.buildscripts.get_file(self.readme_file_path)
        readme_lines = readme_file.split('\n')
        has_notes = False  # flag to say that we're in a table that has "Notes" column
        in_hub = False  # flag that we're in Hub section
        for i, line in enumerate(readme_lines):
            if ' Hub ' in line:
                in_hub = True
            if not line.startswith('| '):
                continue
            if line.startswith('| CFEngine version '):
                has_notes = 'Notes' in line
                # Desired output row: ['CFEngine version', branches..., 'Notes']
                # Also note that list addition is concatenation: [1] + [2] == [1, 2]
                row = (["CFEngine version"] + [branch for branch in branches] +
                       (['Notes'] if has_notes else []))
                # Width of source columns
                column_widths = [len(x) for x in line.split('|')]
                # Note that first and last column widths are zero, since line
                # begins and ends with '|'. We're actually interested in widths
                # of first column (with words "CFEngine version" in it) and,
                # possibly, last ("Notes", which is now second-to-last).
                # Between them are branch column widths, calculated earlier.
                # Also we substract 2 to remove column "padding".
                column_widths = (
                    [column_widths[1] - 2] +  # "CFEngine version"
                    [branch_column_widths[branch] for branch in branches] +
                    ([column_widths[-2] - 2] if has_notes else []
                     )  # "Notes", if exists
                )
                line = '| ' + (" | ".join(
                    (val.ljust(width)
                     for val, width in zip(row, column_widths)))) + ' |'
            elif line.startswith('| --'):
                line = '| ' + (' | '.join(
                    ('-' * width for width in column_widths))) + ' |'
            else:
                # Sample line:
                # | [PHP](http://php.net/) ...
                # For it, in regexp below,
                # \[([a-z0-9-]*)\] will match [PHP]
                # \((.*?)\) will match (http://php.net/)
                match = re.match('\| \[([a-z0-9-]*)\]\((.*?)\) ',
                                 line,
                                 flags=re.IGNORECASE)
                if match:
                    dep_title = match.group(1)
                    dep = dep_title.lower()
                    url = match.group(2)
                else:
                    log.warn("didn't find dep in line [%s]", line)
                    continue
                if dep not in deps_table:
                    log.warn(
                        "unknown dependency in README: [%s] line [%s], will be EMPTY",
                        dep, line)
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                if has_notes:
                    note = re.search('\| ([^|]*) \|$', line)
                    if not note:
                        log.warn("didn't find note in line [%s]", line)
                        note = ''
                    else:
                        note = note.group(1)
                if in_hub:
                    dep = re.sub('-hub$', '', dep)
                row = (["[%s](%s)" % (dep_title, url)] +
                       [deps_table[dep][branch] for branch in branches] +
                       ([note] if has_notes else []))
                line = '| ' + (" | ".join(
                    (val.ljust(width)
                     for val, width in zip(row, column_widths)))) + ' |'
            readme_lines[i] = line

        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = 'deptables-{}'.format(timestamp)
        self.buildscripts.checkout(new_branchname, new=True)
        readme_file = '\n'.join(readme_lines)
        self.buildscripts.put_file(self.readme_file_path, readme_file)
        self.buildscripts.commit('Update dependency tables')
        self.buildscripts.push(new_branchname)
        pr_text = self.github.create_pr(target_repo='{}/{}'.format(
            upstream_name, repo_name),
                                        target_branch='master',
                                        source_user=self.username,
                                        source_branch=new_branchname,
                                        title='Update dependency tables',
                                        text='')
        self.slack.reply("Dependency tables:\n{}".format(pr_text), True)
Example #6
0
class UpdateChecker():
    """Class responsible for doing dependency updates
    Currently it's working only with cfengine/buildscripts repo, as described at
    https://github.com/mendersoftware/infra/blob/master/files/buildcache/release-scripts/RELEASE_PROCESS.org#minor-dependencies-update
    """
    def __init__(self, github, slack, dispatcher, username):
        self.github = github
        self.slack = slack
        self.username = username
        dispatcher.register_command(
            keyword='deps',
            callback=lambda branch: self.run(branch),
            parameter_name='branch',
            short_help='Run dependency updates',
            long_help=
            'Try to find new versions of dependencies on given branch and create PR with them'
        )
        dispatcher.register_command(
            keyword='depstable',
            callback=lambda branches: self.update_deps_version(branches),
            parameter_name='branches',
            short_help='Rebuild dependencies table',
            long_help=
            'Enumerate used dependency versions and update dependency table. Argument is comma-separated list of branches, NO SPACES'
        )

    def get_deps_list(self, branch='master'):
        """Get list of dependencies for given branch.
           Assumes proper branch checked out by `self.buildscripts` repo.
           Returns a list, like this: ["lcov", "pthreads-w32", "libgnurx"]
        """
        # TODO: get value of $EMBEDDED_DB from file
        embedded_db = 'lmdb'
        if branch == '3.7.x':
            options_file = self.buildscripts.get_file(
                'build-scripts/install-dependencies')
        else:
            options_file = self.buildscripts.get_file(
                'build-scripts/compile-options')
        options_lines = options_file.splitlines()
        if branch == '3.7.x':
            filtered_lines = (x for x in options_lines
                              if re.match('\s*DEPS=".*\\$DEPS', x))
            only_deps = (re.sub('\\$?DEPS', '', x) for x in filtered_lines)
            only_deps = (re.sub('[=";]', '', x) for x in only_deps)
            only_deps = (x.strip() for x in only_deps)
        else:
            filtered_lines = (x for x in options_lines
                              if 'var_append DEPS' in x)
            only_deps = (re.sub('.*DEPS "(.*)".*', "\\1", x)
                         for x in filtered_lines)
        # currently only_deps is generator of space-separated deps,
        # i.e. each item can contain several items, like this:
        # list(only_deps) = ["lcov", "pthreads-w32 libgnurx"]
        # to "flattern" it we first join using spaces and then split on spaces
        # in the middle we also do some clean-ups
        only_deps = ' '.join(only_deps)\
                .replace('$EMBEDDED_DB', embedded_db)\
                .replace('libgcc ','')\
                .split(' ')
        # now only_deps looks like this: ["lcov", "pthreads-w32", "libgnurx"]
        log.debug(pretty(only_deps))
        return only_deps

    def increase_version(self, version, increment, separator='.'):
        """increase last part of version - so 1.2.9 becomes 1.2.10
           Args:
               version - old version represented as string
               increment - by how much to increase
               separator - separator character between version parts. Typical
                   values are '.' and '-'. Special case: if separator is 'char'
                   string, then increase last character by 1 - so version
                   '1.2b' becomes '1.2c'
                   (we assume that we never meet version ending with 'z')
           Returns:
             new version as a string
        """
        if separator == 'char':
            return version[:-1] + chr(ord(version[-1]) + increment)
        version_components = version.split(separator)
        version_components[-1] = str(int(version_components[-1]) + increment)
        return separator.join(version_components)

    def checkfile(self, url, sha256=False):
        """Checks if file on given URL exists and optionally returns its sha256 sum
           Args:
               url - URL to check (starting with http or ftp, other protocols might not work)
               sha256 - set it to True to force downloading file and returning sha256 sum
                   (otherwise, for http[s] we use HEAD request)
           Returns:
               True, False, or sha256 of a linked file
        """
        log.debug('checking URL: ' + url)
        try:
            if not sha256 and url.startswith('http'):
                log.debug('testing with HEAD')
                r = requests.head(url)
                return r.status_code >= 200 and r.status_code < 300
            else:
                log.debug('getting whole file')
                m = hashlib.sha256()
                with urllib.request.urlopen(url) as f:
                    data = f.read(4096)
                    while data:
                        m.update(data)
                        data = f.read(4096)
                return m.hexdigest()
        except:
            return False

    def maybe_replace(self, string, match, old, new):
        """replaces `old` with `new` in `string` only if it contains `match`
            Does caseless compare by converting `string` to lowercase for comparison
            Args:
                string - string to work on
                match - string to look for, MUST BE lowercase
                old - string to replace
                new - string to replace with
        """
        if match not in string.lower():
            return string
        return string.replace(old, new)

    def extract_version_from_filename(self, dep, filename):
        if dep == 'openssl':
            # On different branches we use openssl from different sources
            # (this will be cleaned up soon). When downloading from github,
            # filename is OpenSSL_1_1_1.tar.gz, where 1_1_1 is version.
            # When downloading from openssl website, filename for same version
            # is openssl-1.1.1.tar.gz, and version is 1.1.1.
            # We first check for website-style version, and if it doesn't match
            # then we fallback to github-style version. If neither version is
            # found - match.group(1) will raise an exception.
            match = re.search('-([0-9a-z.]*).tar', filename)
            if not match:
                match = re.search('_([0-9a-z_]*).tar', filename)
            version = match.group(1)
            separator = 'char'
        elif dep == 'pthreads-w32':
            version = re.search('w32-([0-9-]*)-rel', filename).group(1)
            separator = '-'
        else:
            version = re.search('[-_]([0-9.]*)[\.-]', filename).group(1)
            separator = '.'
        return (version, separator)

    def find_new_version(self, old_url, old_version, separator):
        """Finds new version by iteratively increasing version in URL and
        checking if it's still possible to download a file.
        Returns highest version for which a file exists.
        Note that if old_version is 1.2.3, and somebody released version
        1.2.5 WITHOUT releasing 1.2.4 before that, then this function will NOT
        find it
        """
        increment = 0
        url_result = True
        while url_result:
            increment += 1
            new_version = self.increase_version(old_version, increment,
                                                separator)
            # note that we change version on URL level, not on filename level -
            # because sometimes version might be in directory name, too
            new_url = old_url.replace(old_version, new_version)
            url_result = self.checkfile(new_url)
            # note that url_result might be True, False, or string with sha256 hash
        # Loop ends when `increment` points to non-existing version -
        # so we need to decrease it to point to last existing one
        increment -= 1
        if increment == 0:
            return old_version
        return self.increase_version(old_version, increment, separator)

    def get_version_from_monitoring(self, dep):
        """Gets latest version of a dependency from release-monitoring.org site.
        Returns latest version (string), or False if dependency not found in
        release-monitoring.json file.
        """
        if dep not in self.monitoring_ids:
            return False
        id = self.monitoring_ids[dep]
        url = 'https://release-monitoring.org/api/project/{}'.format(id)
        try:
            data = requests.get(url).json()
        except:
            raise ReleaseMonitoringException(
                'Failed to do a request to release-monitoring.org website')
        try:
            return data['version']
        except:
            raise ReleaseMonitoringException(
                'Failed to get version from data received from release-monitoring.org website'
            )

    def get_current_version(self, dep):
        """Get current version of dependency dep"""
        # Note: this function partially duplicates next one.
        # It is done on purpose, since that one does some extra stuff.
        dist_file_path = 'deps-packaging/{}/distfiles'.format(dep)
        dist_file = self.buildscripts.get_file(dist_file_path)
        dist_file = dist_file.strip()
        old_filename = re.sub('.* ', '', dist_file)
        (old_version,
         separator) = self.extract_version_from_filename(dep, old_filename)
        return old_version

    def update_single_dep(self, dep):
        """Check if new version of dependency dep was released and create
        commit updating it in *.spec, dist, source, and README.md files
        """
        # Note: this function partially duplicates above one.
        # It is done on purpose, since it will need several other variables
        # afterwards: dist_file_path, old_filename, and separator.
        log.info('Checking new version of {}'.format(dep))
        dist_file_path = 'deps-packaging/{}/distfiles'.format(dep)
        dist_file = self.buildscripts.get_file(dist_file_path)
        dist_file = dist_file.strip()
        source_file_path = 'deps-packaging/{}/source'.format(dep)
        source_file = self.buildscripts.get_file(source_file_path)
        source_file = source_file.strip()
        old_filename = re.sub('.* ', '', dist_file)
        old_url = '{}{}'.format(source_file, old_filename)
        (old_version,
         separator) = self.extract_version_from_filename(dep, old_filename)
        new_version = self.get_version_from_monitoring(dep)
        if not new_version:
            log.warning(
                'Dependency {} not found in release-monitoring.org or in data file'
                .format(dep))
            new_version = self.find_new_version(old_url, old_version,
                                                separator)
        if new_version == old_version:
            # no update needed
            return False
        new_filename = old_filename.replace(old_version, new_version)
        new_url = old_url.replace(old_version, new_version)
        sha256sum = self.checkfile(new_url, True)
        if not sha256sum:
            message = 'Update {} from {} to {} FAILED to download {}'.format(
                dep, old_version, new_version, new_url)
            log.warn(message)
            self.slack.reply(message)
            return False
        message = 'Update {} from {} to {}'.format(dep, old_version,
                                                   new_version)
        log.info(message)
        dist_file = '{}  {}'.format(sha256sum, new_filename)
        self.buildscripts.put_file(dist_file_path, dist_file + '\n')
        source_file = source_file.replace(old_version, new_version)
        self.buildscripts.put_file(source_file_path, source_file + '\n')
        self.readme_lines = [
            self.maybe_replace(x, '* [{}]('.format(dep.replace('-hub', '')),
                               old_version, new_version)
            for x in self.readme_lines
        ]
        readme_file = '\n'.join(self.readme_lines)
        self.buildscripts.put_file(self.readme_file_path, readme_file)
        spec_file_path = 'deps-packaging/{}/cfbuild-{}.spec'.format(dep, dep)
        try:
            spec_file = self.buildscripts.get_file(spec_file_path)
        except:
            pass
        else:
            spec_file = spec_file.replace(old_version, new_version)
            self.buildscripts.put_file(spec_file_path, spec_file)
        self.buildscripts.commit(message)
        return message

    def collect_deps(self, branch):
        """List used dependencies for a branch, returns a dict like this:
            {"dep1": "version", "dep2": "version",...}
            """
        deps_versions = {}
        deps_list = self.get_deps_list(branch)
        for dep in deps_list:
            deps_versions[dep] = self.get_current_version(dep)
        return deps_versions

    def update_deps_version(self, branches):
        # prepare repo
        repo_name = 'buildscripts'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                    self.username, 'master')

        branches = branches.split(',')

        # fetch versions from all branches
        deps_table = {}
        # deps_table is a 2d dict: deps_table[dep][branch]=version
        branch_column_widths = {}
        for branch in branches:
            branch_column_widths[branch] = len(branch)
            self.buildscripts.checkout(branch)
            self.buildscripts.run_command('pull')
            deps_versions = self.collect_deps(branch)
            # deps_versions is a dict: deps_versions[dep]=version
            for dep in deps_versions:
                if not dep in deps_table:
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                deps_table[dep][branch] = deps_versions[dep]
                branch_column_widths[branch] = max(
                    branch_column_widths[branch], len(deps_versions[dep]))

        # patch the readme
        self.buildscripts.checkout('master')
        self.readme_file_path = 'README.md'
        readme_file = self.buildscripts.get_file(self.readme_file_path)
        readme_lines = readme_file.split('\n')
        has_notes = False  # flag to say that we're in a table that has "Notes" column
        in_hub = False  # flag that we're in Hub section
        for i, line in enumerate(readme_lines):
            if ' Hub ' in line:
                in_hub = True
            if not line.startswith('| '):
                continue
            if line.startswith('| CFEngine version '):
                has_notes = 'Notes' in line
                # Desired output row: ['CFEngine version', branches..., 'Notes']
                # Also note that list addition is concatenation: [1] + [2] == [1, 2]
                row = (["CFEngine version"] + [branch for branch in branches] +
                       (['Notes'] if has_notes else []))
                # Width of source columns
                column_widths = [len(x) for x in line.split('|')]
                # Note that first and last column widths are zero, since line
                # begins and ends with '|'. We're actually interested in widths
                # of first column (with words "CFEngine version" in it) and,
                # possibly, last ("Notes", which is now second-to-last).
                # Between them are branch column widths, calculated earlier.
                # Also we substract 2 to remove column "padding".
                column_widths = (
                    [column_widths[1] - 2] +  # "CFEngine version"
                    [branch_column_widths[branch] for branch in branches] +
                    ([column_widths[-2] - 2] if has_notes else []
                     )  # "Notes", if exists
                )
                line = '| ' + (" | ".join(
                    (val.ljust(width)
                     for val, width in zip(row, column_widths)))) + ' |'
            elif line.startswith('| --'):
                line = '| ' + (' | '.join(
                    ('-' * width for width in column_widths))) + ' |'
            else:
                # Sample line:
                # | [PHP](http://php.net/) ...
                # For it, in regexp below,
                # \[([a-z0-9-]*)\] will match [PHP]
                # \((.*?)\) will match (http://php.net/)
                match = re.match('\| \[([a-z0-9-]*)\]\((.*?)\) ',
                                 line,
                                 flags=re.IGNORECASE)
                if match:
                    dep_title = match.group(1)
                    dep = dep_title.lower()
                    url = match.group(2)
                else:
                    log.warn("didn't find dep in line [%s]", line)
                    continue
                if dep not in deps_table:
                    log.warn(
                        "unknown dependency in README: [%s] line [%s], will be EMPTY",
                        dep, line)
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                if has_notes:
                    note = re.search('\| ([^|]*) \|$', line)
                    if not note:
                        log.warn("didn't find note in line [%s]", line)
                        note = ''
                    else:
                        note = note.group(1)
                if in_hub:
                    dep = re.sub('-hub$', '', dep)
                row = (["[%s](%s)" % (dep_title, url)] +
                       [deps_table[dep][branch] for branch in branches] +
                       ([note] if has_notes else []))
                line = '| ' + (" | ".join(
                    (val.ljust(width)
                     for val, width in zip(row, column_widths)))) + ' |'
            readme_lines[i] = line

        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = 'deptables-{}'.format(timestamp)
        self.buildscripts.checkout(new_branchname, new=True)
        readme_file = '\n'.join(readme_lines)
        self.buildscripts.put_file(self.readme_file_path, readme_file)
        self.buildscripts.commit('Update dependency tables')
        self.buildscripts.push(new_branchname)
        pr_text = self.github.create_pr(target_repo='{}/{}'.format(
            upstream_name, repo_name),
                                        target_branch='master',
                                        source_user=self.username,
                                        source_branch=new_branchname,
                                        title='Update dependency tables',
                                        text='')
        self.slack.reply("Dependency tables:\n{}".format(pr_text), True)

    def run(self, branch):
        """Run the dependency update for a branch, creating PR in the end"""
        self.slack.reply("Running dependency updates for " + branch)
        # prepare repo
        repo_name = 'buildscripts'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                    self.username, branch)
        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = '{}-deps-{}'.format(branch, timestamp)
        self.buildscripts.checkout(new_branchname, new=True)
        self.readme_file_path = 'deps-packaging/README.md'
        readme_file = self.buildscripts.get_file(self.readme_file_path)
        self.readme_lines = readme_file.split('\n')
        self.monitoring_file_path = 'deps-packaging/release-monitoring.json'
        self.monitoring_ids = json.loads(
            self.buildscripts.get_file(self.monitoring_file_path))
        updates_summary = []
        only_deps = self.get_deps_list(branch)
        for dep in only_deps:
            single_result = self.update_single_dep(dep)
            if single_result:
                updates_summary.append(single_result)
                self.slack.reply(single_result)
        if len(updates_summary) == 0:
            self.slack.reply("Dependency checked, nothing to update")
            return
        self.buildscripts.push(new_branchname)
        updates_summary = '\n'.join(updates_summary)
        pr_text = self.github.create_pr(
            target_repo='{}/{}'.format(upstream_name, repo_name),
            target_branch=branch,
            source_user=self.username,
            source_branch=new_branchname,
            title='Dependency updates for ' + branch,
            text=updates_summary)
        self.slack.reply(
            "Dependency updates:\n```\n{}\n```\n{}".format(
                updates_summary, pr_text), True)
Example #7
0
    def update_deps_version(self, branches):
        # prepare repo
        repo_name = 'buildscripts'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                    self.username, 'master')

        branches = branches.split(',')

        # fetch versions from all branches
        deps_table = {}
        # deps_table is a 2d dict: deps_table[dep][branch]=version
        branch_column_widths = {}
        for branch in branches:
            branch_column_widths[branch] = len(branch)
            self.buildscripts.checkout(branch)
            self.buildscripts.run_command('pull')
            deps_versions = self.collect_deps(branch)
            # deps_versions is a dict: deps_versions[dep]=version
            for dep in deps_versions:
                if not dep in deps_table:
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                deps_table[dep][branch] = deps_versions[dep]
                branch_column_widths[branch] = max(
                    branch_column_widths[branch], len(deps_versions[dep]))

        # patch the readme
        self.buildscripts.checkout('master')
        self.readme_file_path = 'deps-packaging/README.md'
        readme_file = self.buildscripts.get_file(self.readme_file_path)
        readme_lines = readme_file.split('\n')
        has_notes = False  # flag to say that we're in a table that has "Notes" column
        in_hub = False  # flag that we're in Hub section
        for i, line in enumerate(readme_lines):
            if ' Hub ' in line:
                in_hub = True
            if not line.startswith('| '):
                continue
            if line.startswith('| CFEngine version '):
                has_notes = 'Notes' in line
                branches_text = ' | '.join([
                    branch.ljust(branch_column_widths[branch])
                    for branch in branches
                ])
                line = '| CFEngine version | ' + branches_text + (
                    ' | Notes |' if has_notes else ' |')
            elif line.startswith('|  --'):
                branches_text = ' | '.join([
                    '-' * branch_column_widths[branch] for branch in branches
                ])
                line = '|  --------------  | ' + branches_text + (
                    ' | ----- |' if has_notes else ' |')
            else:
                dep = re.match('\\|  ([a-z0-9-]*) ', line, flags=re.IGNORECASE)
                if dep:
                    dep = dep.group(1)
                else:
                    log.warn("didn't find dep in line [%s]", line)
                    continue
                if in_hub and dep == 'postgresql':
                    dep = 'postgresql-hub'
                if dep not in deps_table:
                    log.warn(
                        "unknown dependency in README: [%s] line [%s], will be EMPTY",
                        dep, line)
                    deps_table[dep] = collections.defaultdict(lambda: "-")
                if has_notes:
                    note = re.search('\| [^|]* \|$', line)
                    if not note:
                        log.warn("didn't find note in line [%s]", line)
                        note = '|  |'
                    note = note.group(0)  # group(0) is full matched string
                branches_text = ' | '.join(
                    deps_table[dep][branch].ljust(branch_column_widths[branch])
                    for branch in branches)
                if in_hub:
                    dep = re.sub('-hub$', '', dep)
                line = '|  %-15s | %s %s' % (dep, branches_text,
                                             (note if has_notes else '|'))
            readme_lines[i] = line

        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = 'deptables-{}'.format(timestamp)
        self.buildscripts.checkout(new_branchname, new=True)
        readme_file = '\n'.join(readme_lines)
        self.buildscripts.put_file(self.readme_file_path, readme_file)
        self.buildscripts.commit('Update dependency tables')
        self.buildscripts.push(new_branchname)
        pr_text = self.github.create_pr(target_repo='{}/{}'.format(
            upstream_name, repo_name),
                                        target_branch='master',
                                        source_user=self.username,
                                        source_branch=new_branchname,
                                        title='Update dependency tables',
                                        text='')
        self.slack.reply("Dependency tables:\n{}".format(pr_text), True)
Example #8
0
    def run(self, inputs):
        """Update the packages_mapping file and create PR"""
        # prepare repo
        repo_name = 'system-testing'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        repo = GitRepo(local_path, repo_name, upstream_name, self.username, "master")
        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = 'packages_mapping-{}'.format(timestamp)
        repo.checkout(new_branchname, new=True)
        # load current file
        packages_mapping_file_path = 'deployment_tests/packages_mapping.json'
        try:
            packages_mapping_contents = repo.get_file(packages_mapping_file_path)
        except FileNotFoundError as e:
            raise RepoFileNotFoundException('file %s not found in repo. Is repo broken?' % packages_mapping_file_path) from e
        try:
            packages_mapping = json.loads(packages_mapping_contents, object_pairs_hook=collections.OrderedDict)
        except json.decoder.JSONDecodeError as e:
            raise JSONParsingError('file %s is not a valid JSON' % packages_mapping_file_path) from e

        # sample of the file
        # {
        #   "packages": {
        #     "3.6.6": {
        #       "agent": {
        #         "community": {
        #           "PACKAGES_i386_linux_debian_4": {
        #             "url": "https://cfengine-package-repos.s3.amazonaws.com/community_binaries/cfengine-community_3.6.6-1_i386.deb"
        #           },
        for value in inputs.split(','):
            value_type = is_branch_or_version(value)
            assert value_type, 'couldn\'t decide if [%s] is branch or version' % value
            self.slack.reply("Updating packages mapping for %s %s " % ( value_type, value))
            result = {'agent':{}, 'hub':{}}
            for (product, codename) in [('community', 'community'), ('enterprise', 'nova')]:
                if value_type == 'version':
                    version = value
                    branch = re.sub('^(\\d+\\.\\d+\\.).*', '\\1x', version)
                    release_url = 'https://cfengine.com/release-data/%s/%s.json' % (product, version)
                else:
                    branch = value
                    releases_url = 'https://cfengine.com/release-data/%s/releases.json' % product
                    releases_request = requests.get(releases_url)
                    if not releases_request.ok:
                        raise URLDownloadFailureException('failed to download %s, return code %d' % (releases_url, releases_request.status_code))
                    try:
                        releases_data = releases_request.json()
                    except json.decoder.JSONDecodeError as e:
                        raise JSONParsingError('file %s is not a valid JSON' % releases_url) from e
                    if 'releases' not in releases_data:
                        raise JSONStructureError('no "releases" in %s JSON' % releases_url)
                    try:
                        release_url, version = next((release['URL'], release['version']) for release in releases_data['releases']
                                                    if "lts_branch" in release and release["lts_branch"] == branch
                                                    and "latest_on_branch" in release and release["latest_on_branch"])
                    except StopIteration as e:
                        raise JSONStructureError('no release with "lts_branch"=="%s" and "latest_on_branch"==true in %s JSON' % (branch, releases_url)) from e

                packages = self.collect_packages(release_url)
                # masterfiles hack #1: different parts of code expect the package to be called differently.
                # Instead of fixing it properly, we will just support both namings
                packages['VIRTUAL_PACKAGES_masterfiles'] = packages['PACKAGES_VIRTUAL_PACKAGES_masterfiles']
                # sort packages into agent/hub ones
                if product == 'community':
                    # for community, they are the same
                    hub_packages = agent_packages = packages
                else:
                    # masterfiles hack #2: different parts of code expect package to be in different groups
                    # ('hub' vs 'agent'). Instead of fixing it properly, we will just include it in both.
                    hub_packages = {platform: value for (platform, value) in packages.items() if
                                    ('_HUB_' in platform) or ("masterfiles" in platform)}
                    agent_packages = {platform: value for (platform, value) in packages.items()
                                      if ('_HUB_' not in platform)}
                result['agent'][codename] = agent_packages
                result['hub'][codename] = hub_packages
            # note that following code uses "branch" and "version" variables
            # they are left from last iteration of the above loop and should be
            # the same in all loop iterations
            packages_mapping['packages'][version] = result
            repo.put_file(packages_mapping_file_path, json.dumps(packages_mapping, indent=2))
            # update version in upgrade_from.json
            branch_nox = re.sub('\\.x$', '', branch)
            upgrade_from_file_path = 'deployment_tests/upgrade_from.json'
            try:
                upgrade_from_content = repo.get_file(upgrade_from_file_path)
            except FileNotFoundError as e:
                raise RepoFileNotFoundException('file %s not found in repo. Is repo broken?' % upgrade_from_file_path) from e
            try:
                upgrade_from = json.loads(upgrade_from_content, object_pairs_hook=collections.OrderedDict)
            except json.decoder.JSONDecodeError as e:
                raise JSONParsingError('file %s is not a valid JSON' % upgrade_from_file_path) from e
            upgrade_from['latest_version'][branch_nox] = version
            repo.put_file(upgrade_from_file_path, json.dumps(upgrade_from, indent=4))
            repo.commit('Add %s packages to packages_mapping' % version)
        repo.push(new_branchname)
        pr_text = self.github.create_pr(
            target_repo='{}/{}'.format(upstream_name, repo_name),
            target_branch='master',
            source_user=self.username,
            source_branch=new_branchname,
            title='Add %s packages to packages_mapping' % inputs,
            text='')
        self.slack.reply(pr_text, True)
Example #9
0
class UpdateChecker():
    """Class responsible for doing dependency updates
    Currently it's working only with cfengine/buildscripts repo, as described at
    https://github.com/mendersoftware/infra/blob/master/files/buildcache/release-scripts/RELEASE_PROCESS.org#minor-dependencies-update
    """
    def __init__(self, github, slack, dispatcher, username):
        self.github = github
        self.slack = slack
        self.username = username
        dispatcher.register_command(
            keyword='deps',
            callback=lambda branch: self.run(branch),
            parameter_name='branch',
            short_help='Run dependency updates',
            long_help=
            'Try to find new versions of dependencies on given branch and create PR with them'
        )

    def get_deps_list(self, branch='master'):
        """Get list of dependencies for given branch.
           Assumes proper branch checked out by `self.buildscripts` repo.
           Returns a list, like this: ["lcov", "pthreads-w32", "libgnurx"]
        """
        # TODO: get value of $EMBEDDED_DB from file
        embedded_db = 'lmdb'
        if branch == '3.7.x':
            options_file = self.buildscripts.get_file(
                'build-scripts/install-dependencies')
        else:
            options_file = self.buildscripts.get_file(
                'build-scripts/compile-options')
        options_lines = options_file.splitlines()
        if branch == '3.7.x':
            filtered_lines = (x for x in options_lines
                              if re.match('\s*DEPS=".*\\$DEPS', x))
            only_deps = (re.sub('\\$?DEPS', '', x) for x in filtered_lines)
            only_deps = (re.sub('[=";]', '', x) for x in only_deps)
            only_deps = (x.strip() for x in only_deps)
        else:
            filtered_lines = (x for x in options_lines
                              if 'var_append DEPS' in x)
            only_deps = (re.sub('.*DEPS "(.*)".*', "\\1", x)
                         for x in filtered_lines)
        # currently only_deps is generator of space-separated deps,
        # i.e. each item can contain several items, like this:
        # list(only_deps) = ["lcov", "pthreads-w32 libgnurx"]
        # to "flattern" it we first join using spaces and then split on spaces
        # in the middle we also do some clean-ups
        only_deps = ' '.join(only_deps)\
                .replace('$EMBEDDED_DB', embedded_db)\
                .replace('libgcc ','')\
                .split(' ')
        # now only_deps looks like this: ["lcov", "pthreads-w32", "libgnurx"]
        log.debug(pretty(only_deps))
        return only_deps

    def increase_version(self, version, increment, separator='.'):
        """increase last part of version - so 1.2.9 becomes 1.2.10
           Args:
               version - old version represented as string
               increment - by how much to increase
               separator - separator character between version parts. Typical
                   values are '.' and '-'. Special case: if separator is 'char'
                   string, then increase last character by 1 - so version
                   '1.2b' becomes '1.2c'
                   (we assume that we never meet version ending with 'z')
           Returns:
             new version as a string
        """
        if separator == 'char':
            return version[:-1] + chr(ord(version[-1]) + increment)
        version_components = version.split(separator)
        version_components[-1] = str(int(version_components[-1]) + increment)
        return separator.join(version_components)

    def checkfile(self, url, md5=False):
        """Checks if file on given URL exists and optionally returns its md5 sum
           Args:
               url - URL to check (starting with http or ftp, other protocols might not work)
               md5 - set it to True to force downloading file and returning md5 sum
                   (otherwise, for http[s] we use HEAD request)
           Returns:
               True, False, or md5 of a linked file
        """
        log.debug('checking URL: ' + url)
        try:
            if not md5 and url.startswith('http'):
                log.debug('testing with HEAD')
                r = requests.head(url)
                return r.status_code >= 200 and r.status_code < 300
            else:
                log.debug('getting whole file')
                m = hashlib.md5()
                with urllib.request.urlopen(url) as f:
                    data = f.read(4096)
                    while data:
                        m.update(data)
                        data = f.read(4096)
                return m.hexdigest()
        except:
            return False

    def maybe_replace(self, string, match, old, new):
        """replaces `old` with `new` in `string` only if it contains `match`
            Does caseless compare by converting `string` to lowercase for comparison
            Args:
                string - string to work on
                match - string to look for, MUST BE lowercase
                old - string to replace
                new - string to replace with
        """
        if match not in string.lower():
            return string
        return string.replace(old, new)

    def extract_version_from_filename(self, dep, filename):
        if dep == 'openssl':
            version = re.search('-([0-9a-z.]*).tar', filename).group(1)
            separator = 'char'
        elif dep == 'pthreads-w32':
            version = re.search('w32-([0-9-]*)-rel', filename).group(1)
            separator = '-'
        else:
            version = re.search('[-_]([0-9.]*)[\.-]', filename).group(1)
            separator = '.'
        return (version, separator)

    def find_new_version(self, old_url, old_version, separator):
        """Finds new version by iteratively increasing version in URL and
        checking if it's still possible to download a file.
        Returns highest version for which a file exists.
        Note that if old_version is 1.2.3, and somebody released version
        1.2.5 WITHOUT releasing 1.2.4 before that, then this function will NOT
        find it
        """
        increment = 0
        url_result = True
        while url_result:
            increment += 1
            new_version = self.increase_version(old_version, increment,
                                                separator)
            # note that we change version on URL level, not on filename level -
            # because sometimes version might be in directory name, too
            new_url = old_url.replace(old_version, new_version)
            url_result = self.checkfile(new_url)
            # note that url_result might be True, False, or string with md5 hash
        # Loop ends when `increment` points to non-existing version -
        # so we need to decrease it to point to last existing one
        increment -= 1
        if increment == 0:
            return old_version
        return self.increase_version(old_version, increment, separator)

    def get_version_from_monitoring(self, dep):
        """Gets latest version of a dependency from release-monitoring.org site.
        Returns latest version (string), or False if dependency not found in
        release-monitoring.json file.
        """
        if dep not in self.monitoring_ids:
            return False
        id = self.monitoring_ids[dep]
        url = 'https://release-monitoring.org/api/project/{}'.format(id)
        try:
            data = requests.get(url).json()
        except:
            raise ReleaseMonitoringException(
                'Failed to do a request to release-monitoring.org website')
        try:
            return data['version']
        except:
            raise ReleaseMonitoringException(
                'Failed to get version from data received from release-monitoring.org website'
            )

    def update_single_dep(self, dep):
        """Check if new version of dependency dep was released and create
        commit updating it in *.spec, dist, source, and README.md files
        """
        log.info('Checking new version of {}'.format(dep))
        dist_file_path = 'deps-packaging/{}/distfiles'.format(dep)
        dist_file = self.buildscripts.get_file(dist_file_path)
        dist_file = dist_file.strip()
        source_file_path = 'deps-packaging/{}/source'.format(dep)
        source_file = self.buildscripts.get_file(source_file_path)
        source_file = source_file.strip()
        old_filename = re.sub('.* ', '', dist_file)
        old_url = '{}{}'.format(source_file, old_filename)
        (old_version,
         separator) = self.extract_version_from_filename(dep, old_filename)
        new_version = self.get_version_from_monitoring(dep)
        if not new_version:
            log.warning(
                'Dependency {} not found in release-monitoring.org or in data file'
                .format(dep))
            new_version = self.find_new_version(old_url, old_version,
                                                separator)
        if new_version == old_version:
            # no update needed
            return False
        new_filename = old_filename.replace(old_version, new_version)
        new_url = old_url.replace(old_version, new_version)
        md5sum = self.checkfile(new_url, True)
        if not md5sum:
            message = 'Update {} from {} to {} FAILED to download {}'.format(
                dep, old_version, new_version, new_url)
            log.warn(message)
            self.slack.reply(message)
            return False
        message = 'Update {} from {} to {}'.format(dep, old_version,
                                                   new_version)
        log.info(message)
        dist_file = '{}  {}'.format(md5sum, new_filename)
        self.buildscripts.put_file(dist_file_path, dist_file + '\n')
        source_file = source_file.replace(old_version, new_version)
        self.buildscripts.put_file(source_file_path, source_file + '\n')
        self.readme_lines = [
            self.maybe_replace(x, '* [{}]('.format(dep.replace('-hub', '')),
                               old_version, new_version)
            for x in self.readme_lines
        ]
        readme_file = '\n'.join(self.readme_lines)
        self.buildscripts.put_file(self.readme_file_path, readme_file)
        spec_file_path = 'deps-packaging/{}/cfbuild-{}.spec'.format(dep, dep)
        try:
            spec_file = self.buildscripts.get_file(spec_file_path)
        except:
            pass
        else:
            spec_file = spec_file.replace(old_version, new_version)
            self.buildscripts.put_file(spec_file_path, spec_file + '\n')
        self.buildscripts.commit(message)
        return message

    def run(self, branch):
        """Run the dependency update for a branch, creating PR in the end"""
        self.slack.reply("Running dependency updates for " + branch)
        # prepare repo
        repo_name = 'buildscripts'
        upstream_name = 'cfengine'
        local_path = "../" + repo_name
        self.buildscripts = GitRepo(local_path, repo_name, upstream_name,
                                    self.username, branch)
        timestamp = re.sub('[^0-9-]', '_', str(datetime.datetime.today()))
        new_branchname = '{}-deps-{}'.format(branch, timestamp)
        self.buildscripts.checkout(new_branchname, True)
        self.readme_file_path = 'deps-packaging/README.md'
        readme_file = self.buildscripts.get_file(self.readme_file_path)
        self.readme_lines = readme_file.split('\n')
        self.monitoring_file_path = 'deps-packaging/release-monitoring.json'
        self.monitoring_ids = json.loads(
            self.buildscripts.get_file(self.monitoring_file_path))
        updates_summary = []
        only_deps = self.get_deps_list(branch)
        for dep in only_deps:
            single_result = self.update_single_dep(dep)
            if single_result:
                updates_summary.append(single_result)
                self.slack.reply(single_result)
        if len(updates_summary) == 0:
            self.slack.reply("Dependency checked, nothing to update")
            return
        self.buildscripts.push(new_branchname)
        updates_summary = '\n'.join(updates_summary)
        pr_text = self.github.create_pr(
            target_repo='{}/{}'.format(upstream_name, repo_name),
            target_branch=branch,
            source_user=self.username,
            source_branch=new_branchname,
            title='Dependency updates for ' + branch,
            text=updates_summary)
        self.slack.reply(
            "Dependency updates:\n```\n{}\n```\n{}".format(
                updates_summary, pr_text), True)