# -*- coding: utf-8 -*- import glob import logging import operator import os import re import resource import shlex import shutil import signal import subprocess import time from subprocess import CalledProcessError from ..common import dt2time, fqdn, now, grep, time2str, rfind, uniq_list, local_pgadmin_cursor, get_py_version from ..container import docker_build, docker_run, docker_stop, docker_is_running, docker_get_gateway_ip from odoo import models, fields, api from odoo.exceptions import UserError from odoo.http import request from odoo.tools import config, appdirs _re_error = r'^(?:\d{4}-\d\d-\d\d \d\d:\d\d:\d\d,\d{3} \d+ (?:ERROR|CRITICAL) )|(?:Traceback \(most recent call last\):)$' _re_warning = r'^\d{4}-\d\d-\d\d \d\d:\d\d:\d\d,\d{3} \d+ WARNING ' re_job = re.compile('_job_\d') _logger = logging.getLogger(__name__) def runbot_job(*accepted_job_types): """ Decorator for runbot_build _job_x methods to filter build jobs """ accepted_job_types += ('all', ) def job_decorator(func): def wrapper(self, build, log_path): if build.job_type == 'none' or build.job_type not in accepted_job_types: build._log(func.__name__, 'Skipping job') return -2 return func(self, build, log_path) return wrapper return job_decorator class runbot_build(models.Model): _name = "runbot.build" _order = 'id desc' branch_id = fields.Many2one('runbot.branch', 'Branch', required=True, ondelete='cascade', index=True) repo_id = fields.Many2one(related='branch_id.repo_id', readonly=True, store=True) name = fields.Char('Revno', required=True) host = fields.Char('Host') port = fields.Integer('Port') dest = fields.Char(compute='_get_dest', type='char', string='Dest', readonly=1, store=True) domain = fields.Char(compute='_get_domain', type='char', string='URL') date = fields.Datetime('Commit date') author = fields.Char('Author') author_email = fields.Char('Author Email') committer = fields.Char('Committer') committer_email = fields.Char('Committer Email') subject = fields.Text('Subject') sequence = fields.Integer('Sequence') modules = fields.Char("Modules to Install") result = fields.Char('Result', default='') # ok, ko, warn, skipped, killed, manually_killed guess_result = fields.Char(compute='_guess_result') pid = fields.Integer('Pid') state = fields.Char('Status', default='pending') # pending, testing, running, done, duplicate, deathrow job = fields.Char('Job') # job_* job_start = fields.Datetime('Job start') job_end = fields.Datetime('Job end') job_time = fields.Integer(compute='_get_time', string='Job time') job_age = fields.Integer(compute='_get_age', string='Job age') duplicate_id = fields.Many2one('runbot.build', 'Corresponding Build') server_match = fields.Selection([('builtin', 'This branch includes Odoo server'), ('exact', 'branch/PR exact name'), ('prefix', 'branch whose name is a prefix of current one'), ('fuzzy', 'Fuzzy - common ancestor found'), ('exact PR', 'Exact match between two PR'), ('no PR', 'PR matching a branch without PR'), ('default', 'No match found - defaults to master')], string='Server branch matching') revdep_build_ids = fields.Many2many('runbot.build', 'runbot_rev_dep_builds', column1='rev_dep_id', column2='dependent_id', string='Builds that depends on this build') extra_params = fields.Char('Extra cmd args') coverage = fields.Boolean('Enable code coverage') coverage_result = fields.Float('Coverage result', digits=(5, 2)) build_type = fields.Selection([('scheduled', 'This build was automatically scheduled'), ('rebuild', 'This build is a rebuild'), ('normal', 'normal build'), ('indirect', 'Automatic rebuild'), ], default='normal', string='Build type') job_type = fields.Selection([ ('testing', 'Testing jobs only'), ('running', 'Running job only'), ('all', 'All jobs'), ('none', 'Do not execute jobs'), ], ) def copy(self, values=None): raise UserError("Cannot duplicate build!") def create(self, vals): branch = self.env['runbot.branch'].search([('id', '=', vals.get('branch_id', False))]) if branch.job_type == 'none' or vals.get('job_type', '') == 'none': return self.env['runbot.build'] build_id = super(runbot_build, self).create(vals) extra_info = {'sequence': build_id.id if not build_id.sequence else build_id.sequence} job_type = vals['job_type'] if 'job_type' in vals else build_id.branch_id.job_type extra_info.update({'job_type': job_type}) context = self.env.context if not context.get('force_rebuild'): # detect duplicate duplicate_id = None domain = [ ('repo_id', '=', build_id.repo_id.duplicate_id.id), ('name', '=', build_id.name), ('duplicate_id', '=', False), '|', ('result', '=', False), ('result', '!=', 'skipped') ] # TODO xdo extract build_id._get_closest_branch_name(extra_repo.id) here for duplicate in self.search(domain, limit=10): duplicate_id = duplicate.id # Consider the duplicate if its closest branches are the same than the current build closest branches. for extra_repo in build_id.repo_id.dependency_ids: build_closest_name = build_id._get_closest_branch_name(extra_repo.id)[1] duplicate_closest_name = duplicate._get_closest_branch_name(extra_repo.id)[1] if build_closest_name != duplicate_closest_name: duplicate_id = None if duplicate_id: extra_info.update({'state': 'duplicate', 'duplicate_id': duplicate_id}) break build_id.write(extra_info) if build_id.state == 'duplicate' and build_id.duplicate_id.state in ('running', 'done'): build_id._github_status() return build_id def _reset(self): self.write({'state': 'pending'}) def _branch_exists(self, branch_id): Branch = self.env['runbot.branch'] branch = Branch.search([('id', '=', branch_id)]) if branch and branch[0]._is_on_remote(): return True return False def _get_closest_branch_name(self, target_repo_id): """Return (repo, branch name) of the closest common branch between build's branch and any branch of target_repo or its duplicated repos. Rules priority for choosing the branch from the other repo is: 1. Same branch name 2. A PR whose head name match 3. Match a branch which is the dashed-prefix of current branch name 4. Common ancestors (git merge-base) Note that PR numbers are replaced by the branch name of the PR target to prevent the above rules to mistakenly link PR of different repos together. """ self.ensure_one() Branch = self.env['runbot.branch'] build = self branch, repo = build.branch_id, build.repo_id name = branch.pull_head_name or branch.branch_name target_branch = branch.target_branch_name or 'master' target_repo = self.env['runbot.repo'].browse(target_repo_id) target_repo_ids = [target_repo.id] r = target_repo.duplicate_id while r: if r.id in target_repo_ids: break target_repo_ids.append(r.id) r = r.duplicate_id _logger.debug('Search closest of %s (%s) in repos %r', name, repo.name, target_repo_ids) sort_by_repo = lambda d: (not d['sticky'], # sticky first target_repo_ids.index(d['repo_id'][0]), -1 * len(d.get('branch_name', '')), -1 * d['id']) result_for = lambda d, match='exact': (d['repo_id'][0], d['name'], match) fields = ['name', 'repo_id', 'sticky'] # 1. same name, not a PR domain = [ ('repo_id', 'in', target_repo_ids), ('branch_name', '=', name), ('name', '=like', 'refs/heads/%'), ] targets = Branch.search_read(domain, fields, order='id DESC') targets = sorted(targets, key=sort_by_repo) if targets and self._branch_exists(targets[0]['id']): return result_for(targets[0]) # 2. PR with head name equals domain = [ ('repo_id', 'in', target_repo_ids), ('pull_head_name', '=', name), ('name', '=like', 'refs/pull/%'), ] pulls = Branch.search_read(domain, fields, order='id DESC') pulls = sorted(pulls, key=sort_by_repo) for pull in Branch.browse([pu['id'] for pu in pulls]): pi = pull._get_pull_info() if pi.get('state') == 'open': if ':' in name: # we assume that branch exists if we got pull info pr_branch_name = name.split(':')[1] return (pull.repo_id.duplicate_id.id, 'refs/heads/%s' % pr_branch_name, 'exact PR') else: return (pull.repo_id.id, pull.name, 'exact PR') # 3. Match a branch which is the dashed-prefix of current branch name branches = Branch.search_read( [('repo_id', 'in', target_repo_ids), ('name', '=like', 'refs/heads/%')], fields + ['branch_name'], order='id DESC', ) branches = sorted(branches, key=sort_by_repo) for branch in branches: if name.startswith(branch['branch_name'] + '-') and self._branch_exists(branch['id']): return result_for(branch, 'prefix') # 4.Match a PR in enterprise without community PR if build.branch_id.name.startswith('refs/pull') and ':' in name: pr_branch_name = name.split(':')[1] duplicate_branch_name = 'refs/heads/%s' % pr_branch_name domain = [ ('repo_id', 'in', target_repo_ids), # target_repo_ids should contain the target duplicate repo ('branch_name', '=', pr_branch_name), ('pull_head_name', '=', False), ] targets = Branch.search_read(domain, fields, order='id DESC') targets = sorted(targets, key=sort_by_repo) if targets and self._branch_exists(targets[0]['id']): return result_for(targets[0], 'no PR') # 5. last-resort value return target_repo_id, 'refs/heads/%s' % target_branch, 'default' @api.depends('name', 'branch_id.name') def _get_dest(self): for build in self: nickname = build.branch_id.name.split('/')[2] nickname = re.sub(r'"|\'|~|\:', '', nickname) nickname = re.sub(r'_|/|\.', '-', nickname) build.dest = ("%05d-%s-%s" % (build.id, nickname[:32], build.name[:6])).lower() def _get_domain(self): domain = self.env['ir.config_parameter'].sudo().get_param('runbot.runbot_domain', fqdn()) for build in self: if build.repo_id.nginx: build.domain = "%s.%s" % (build.dest, build.host) else: build.domain = "%s:%s" % (domain, build.port) def _guess_result(self): cr = self.env.cr cr.execute(""" SELECT b.id, CASE WHEN array_agg(l.level)::text[] && ARRAY['ERROR', 'CRITICAL'] THEN 'ko' WHEN array_agg(l.level)::text[] && ARRAY['WARNING'] THEN 'warn' ELSE 'ok' END FROM runbot_build b LEFT JOIN ir_logging l ON (l.build_id = b.id AND l.level != 'INFO') WHERE b.id = ANY(%s) GROUP BY b.id """, [list(self.filtered(lambda b: b.state == 'testing').ids)]) result = {row[0]: row[1] for row in cr.fetchall()} for build in self: build.guess_result = result.get(build.id, build.result) def _get_time(self): """Return the time taken by the tests""" for build in self: if build.job_end: build.job_time = int(dt2time(build.job_end) - dt2time(build.job_start)) elif build.job_start: build.job_time = int(time.time() - dt2time(build.job_start)) def _get_age(self): """Return the time between job start and now""" for build in self: if build.job_start: build.job_age = int(time.time() - dt2time(build.job_start)) def _force(self, message=None): """Force a rebuild and return a recordset of forced builds""" forced_builds = self.env['runbot.build'] for build in self: pending_ids = self.search([('state', '=', 'pending')], order='id', limit=1) if pending_ids: sequence = pending_ids[0].id else: sequence = self.search([], order='id desc', limit=1)[0].id # Force it now rebuild = True if build.state == 'done' and build.result == 'skipped': build.write({'state': 'pending', 'sequence': sequence, 'result': ''}) # or duplicate it elif build.state in ['running', 'done', 'duplicate', 'deathrow']: new_build = build.with_context(force_rebuild=True).create({ 'sequence': sequence, 'branch_id': build.branch_id.id, 'name': build.name, 'author': build.author, 'author_email': build.author_email, 'committer': build.committer, 'committer_email': build.committer_email, 'subject': build.subject, 'modules': build.modules, 'build_type': 'rebuild' }) build = new_build else: rebuild = False if rebuild: forced_builds |= build user = request.env.user if request else self.env.user build._log('rebuild', 'Rebuild initiated by %s' % user.name) if message: build._log('rebuild', message) return forced_builds def _skip(self, reason=None): """Mark builds ids as skipped""" if reason: self._logger('skip %s', reason) self.write({'state': 'done', 'result': 'skipped'}) to_unduplicate = self.search([('id', 'in', self.ids), ('duplicate_id', '!=', False)]) to_unduplicate._force() def _local_cleanup(self): for build in self: # Cleanup the *local* cluster with local_pgadmin_cursor() as local_cr: local_cr.execute(""" SELECT datname FROM pg_database WHERE pg_get_userbyid(datdba) = current_user AND datname LIKE %s """, [build.dest + '%']) to_delete = local_cr.fetchall() for db, in to_delete: self._local_pg_dropdb(db) # cleanup: find any build older than 7 days. root = self.env['runbot.repo']._root() build_dir = os.path.join(root, 'build') builds = os.listdir(build_dir) if builds: self.env.cr.execute(""" SELECT dest FROM runbot_build WHERE dest IN %s AND (state != 'done' OR job_end > (now() - interval '7 days')) """, [tuple(builds)]) actives = set(b[0] for b in self.env.cr.fetchall()) for b in builds: path = os.path.join(build_dir, b) if b not in actives and os.path.isdir(path) and os.path.isabs(path): shutil.rmtree(path) # cleanup old unused databases self.env.cr.execute("select id from runbot_build where state in ('testing', 'running')") db_ids = [id[0] for id in self.env.cr.fetchall()] if db_ids: with local_pgadmin_cursor() as local_cr: local_cr.execute(""" SELECT datname FROM pg_database WHERE pg_get_userbyid(datdba) = current_user AND datname ~ '^[0-9]+-.*' AND SUBSTRING(datname, '^([0-9]+)-.*')::int not in %s """, [tuple(db_ids)]) to_delete = local_cr.fetchall() for db, in to_delete: self._local_pg_dropdb(db) def _list_jobs(self): """List methods that starts with _job_[[:digit:]]""" return sorted(job[1:] for job in dir(self) if re_job.match(job)) def _find_port(self): # currently used port ids = self.search([('state', 'not in', ['pending', 'done'])]) ports = set(i['port'] for i in ids.read(['port'])) # starting port icp = self.env['ir.config_parameter'] port = int(icp.get_param('runbot.runbot_starting_port', default=2000)) # find next free port while port in ports: port += 3 return port def _logger(self, *l): l = list(l) for build in self: l[0] = "%s %s" % (build.dest, l[0]) _logger.debug(*l) def _get_docker_name(self): self.ensure_one() return '%s_%s' % (self.dest, self.job) def _schedule(self): """schedule the build""" jobs = self._list_jobs() icp = self.env['ir.config_parameter'] # For retro-compatibility, keep this parameter in seconds default_timeout = int(icp.get_param('runbot.runbot_timeout', default=1800)) / 60 for build in self: if build.state == 'deathrow': build._kill(result='manually_killed') continue elif build.state == 'pending': # allocate port and schedule first job port = self._find_port() values = { 'host': fqdn(), 'port': port, 'state': 'testing', 'job': jobs[0], 'job_start': now(), 'job_end': False, } build.write(values) else: # check if current job is finished if docker_is_running(build._get_docker_name()): # kill if overpassed timeout = (build.branch_id.job_timeout or default_timeout) * 60 * ( build.coverage and 1.5 or 1) if build.job != jobs[-1] and build.job_time > timeout: build._logger('%s time exceded (%ss)', build.job, build.job_time) build.write({'job_end': now()}) build._kill(result='killed') else: # failfast if not build.result and build.guess_result in ('ko', 'warn'): build.result = build.guess_result build._github_status() continue build._logger('%s finished', build.job) # schedule v = {} # testing -> running if build.job == jobs[-2]: v['state'] = 'running' v['job'] = jobs[-1] v['job_end'] = now(), # running -> done elif build.job == jobs[-1]: v['state'] = 'done' v['job'] = '' # testing else: v['job'] = jobs[jobs.index(build.job) + 1] build.write(v) # run job pid = None if build.state != 'done': build._logger('running %s', build.job) job_method = getattr(self, '_' + build.job) # compute the job method to run os.makedirs(build._path('logs'), exist_ok=True) os.makedirs(build._path('datadir'), exist_ok=True) log_path = build._path('logs', '%s.txt' % build.job) try: pid = job_method(build, log_path) build.write({'pid': pid}) except Exception: _logger.exception('%s failed running method %s', build.dest, build.job) build._log(build.job, "failed running job method, see runbot log") build._kill(result='ko') continue if pid == -2: # no process to wait, directly call next job # FIXME find a better way that this recursive call build._schedule() # cleanup only needed if it was not killed if build.state == 'done': build._local_cleanup() def _path(self, *l, **kw): """Return the repo build path""" self.ensure_one() build = self root = self.env['runbot.repo']._root() return os.path.join(root, 'build', build.dest, *l) def _server(self, *l, **kw): """Return the build server path""" self.ensure_one() build = self if os.path.exists(build._path('odoo')): return build._path('odoo', *l) return build._path('openerp', *l) def _filter_modules(self, modules, available_modules, explicit_modules): blacklist_modules = set(['auth_ldap', 'document_ftp', 'base_gengo', 'website_gengo', 'website_instantclick', 'pad', 'pad_project', 'note_pad', 'pos_cache', 'pos_blackbox_be']) mod_filter = lambda m: ( m in available_modules and (m in explicit_modules or (not m.startswith(('hw_', 'theme_', 'l10n_')) and m not in blacklist_modules)) ) return uniq_list(filter(mod_filter, modules)) def _checkout(self): for build in self: # starts from scratch if os.path.isdir(build._path()): shutil.rmtree(build._path()) # runbot log path os.makedirs(build._path("logs"), exist_ok=True) os.makedirs(build._server('addons'), exist_ok=True) # update repo if needed if not build.repo_id._hash_exists(build.name): build.repo_id._update(build.repo_id) # checkout branch build.branch_id.repo_id._git_export(build.name, build._path()) has_server = os.path.isfile(build._server('__init__.py')) server_match = 'builtin' # build complete set of modules to install modules_to_move = [] modules_to_test = ((build.branch_id.modules or '') + ',' + (build.repo_id.modules or '')) modules_to_test = list(filter(None, modules_to_test.split(','))) # ??? explicit_modules = set(modules_to_test) _logger.debug("manual modules_to_test for build %s: %s", build.dest, modules_to_test) if not has_server: if build.repo_id.modules_auto == 'repo': modules_to_test += [ os.path.basename(os.path.dirname(a)) for a in (glob.glob(build._path('*/__openerp__.py')) + glob.glob(build._path('*/__manifest__.py'))) ] _logger.debug("local modules_to_test for build %s: %s", build.dest, modules_to_test) for extra_repo in build.repo_id.dependency_ids: repo_id, closest_name, server_match = build._get_closest_branch_name(extra_repo.id) repo = self.env['runbot.repo'].browse(repo_id) _logger.debug('branch %s of %s: %s match branch %s of %s', build.branch_id.name, build.repo_id.name, server_match, closest_name, repo.name) build._log( 'Building environment', '%s match branch %s of %s' % (server_match, closest_name, repo.name) ) repo._update_git(force=True) latest_commit = repo._git(['rev-parse', closest_name]).strip() commit_oneline = repo._git(['show', '--pretty="%H -- %s"', '-s', latest_commit]).strip() build._log( 'Building environment', 'Server built based on commit %s from %s' % (commit_oneline, closest_name) ) repo._git_export(closest_name, build._path()) # Finally mark all addons to move to openerp/addons modules_to_move += [ os.path.dirname(module) for module in (glob.glob(build._path('*/__openerp__.py')) + glob.glob(build._path('*/__manifest__.py'))) ] # move all addons to server addons path for module in uniq_list(glob.glob(build._path('addons/*')) + modules_to_move): basename = os.path.basename(module) addon_path = build._server('addons', basename) if os.path.exists(addon_path): build._log( 'Building environment', 'You have duplicate modules in your branches "%s"' % basename ) if os.path.islink(addon_path) or os.path.isfile(addon_path): os.remove(addon_path) else: shutil.rmtree(addon_path) shutil.move(module, build._server('addons')) available_modules = [ os.path.basename(os.path.dirname(a)) for a in (glob.glob(build._server('addons/*/__openerp__.py')) + glob.glob(build._server('addons/*/__manifest__.py'))) ] if build.repo_id.modules_auto == 'all' or (build.repo_id.modules_auto != 'none' and has_server): modules_to_test += available_modules modules_to_test = self._filter_modules(modules_to_test, set(available_modules), explicit_modules) _logger.debug("modules_to_test for build %s: %s", build.dest, modules_to_test) build.write({'server_match': server_match, 'modules': ','.join(modules_to_test)}) def _local_pg_dropdb(self, dbname): with local_pgadmin_cursor() as local_cr: pid_col = 'pid' if local_cr.connection.server_version >= 90200 else 'procpid' query = 'SELECT pg_terminate_backend({}) FROM pg_stat_activity WHERE datname=%s'.format(pid_col) local_cr.execute(query, [dbname]) local_cr.execute('DROP DATABASE IF EXISTS "%s"' % dbname) # cleanup filestore datadir = appdirs.user_data_dir() paths = [os.path.join(datadir, pn, 'filestore', dbname) for pn in 'OpenERP Odoo'.split()] cmd = ['rm', '-rf'] + paths _logger.debug(' '.join(cmd)) subprocess.call(cmd) def _local_pg_createdb(self, dbname): self._local_pg_dropdb(dbname) _logger.debug("createdb %s", dbname) with local_pgadmin_cursor() as local_cr: local_cr.execute("""CREATE DATABASE "%s" TEMPLATE template0 LC_COLLATE 'C' ENCODING 'unicode'""" % dbname) def _log(self, func, message): self.ensure_one() _logger.debug("Build %s %s %s", self.id, func, message) self.env['ir.logging'].create({ 'build_id': self.id, 'level': 'INFO', 'type': 'runbot', 'name': 'odoo.runbot', 'message': message, 'path': 'runbot', 'func': func, 'line': '0', }) def reset(self): self.write({'state': 'pending'}) def _reap(self): while True: try: pid, status, rusage = os.wait3(os.WNOHANG) except OSError: break if pid == 0: break _logger.debug('reaping: pid: %s status: %s', pid, status) def _kill(self, result=None): host = fqdn() for build in self: if build.host != host: continue build._log('kill', 'Kill build %s' % build.dest) docker_stop(build._get_docker_name()) v = {'state': 'done', 'job': False} if result: v['result'] = result build.write(v) self.env.cr.commit() build._github_status() build._local_cleanup() def _ask_kill(self): self.ensure_one() user = request.env.user if request else self.env.user uid = user.id build = self.duplicate_id if self.state == 'duplicate' else self if build.state == 'pending': build._skip() build._log('_ask_kill', 'Skipping build %s, requested by %s (user #%s)' % (build.dest, user.name, uid)) elif build.state in ['testing', 'running']: build.write({'state': 'deathrow'}) build._log('_ask_kill', 'Killing build %s, requested by %s (user #%s)' % (build.dest, user.name, uid)) def _cmd(self): """Return a tuple describing the command to start the build First part is list with the command and parameters Second part is a list of Odoo modules """ self.ensure_one() build = self bins = [ 'odoo-bin', # >= 10.0 'openerp-server', # 9.0, 8.0 'openerp-server.py', # 7.0 'bin/openerp-server.py', # < 7.0 ] for odoo_bin in bins: if os.path.isfile(build._path(odoo_bin)): break # commandline cmd = [ os.path.join('/data/build', odoo_bin), ] # options if grep(build._server("tools/config.py"), "no-xmlrpcs"): cmd.append("--no-xmlrpcs") if grep(build._server("tools/config.py"), "no-netrpc"): cmd.append("--no-netrpc") if grep(build._server("tools/config.py"), "log-db"): logdb_uri = self.env['ir.config_parameter'].get_param('runbot.runbot_logdb_uri') logdb = self.env.cr.dbname if logdb_uri and grep(build._server('sql_db.py'), 'allow_uri'): logdb = '%s' % logdb_uri cmd += ["--log-db=%s" % logdb] if grep(build._server('tools/config.py'), 'log-db-level'): cmd += ["--log-db-level", '25'] if grep(build._server("tools/config.py"), "data-dir"): datadir = build._path('datadir') if not os.path.exists(datadir): os.mkdir(datadir) cmd += ["--data-dir", '/data/build/datadir'] # if build.branch_id.test_tags: # cmd.extend(['--test_tags', "'%s'" % build.branch_id.test_tags]) # keep for next version return cmd, build.modules def _github_status_notify_all(self, status): """Notify each repo with a status""" self.ensure_one() commits = {(b.repo_id, b.name) for b in self.search([('name', '=', self.name)])} for repo, commit_hash in commits: _logger.debug("github updating %s status %s to %s in repo %s", status['context'], commit_hash, status['state'], repo.name) repo._github('/repos/:owner/:repo/statuses/%s' % commit_hash, status, ignore_errors=True) def _github_status(self): """Notify github of failed/successful builds""" runbot_domain = self.env['runbot.repo']._domain() for build in self: desc = "runbot build %s" % (build.dest,) if build.state == 'testing': state = 'pending' elif build.state in ('running', 'done'): state = 'error' else: continue desc += " (runtime %ss)" % (build.job_time,) if build.result == 'ok': state = 'success' if build.result in ('ko', 'warn'): state = 'failure' status = { "state": state, "target_url": "http://%s/runbot/build/%s" % (runbot_domain, build.id), "description": desc, "context": "ci/runbot" } build._github_status_notify_all(status) # Jobs definitions # They all need "build log_path" parameters @runbot_job('testing', 'running') def _job_00_init(self, build, log_path): build._log('init', 'Init build environment') # notify pending build - avoid confusing users by saying nothing build._github_status() build._checkout() return -2 @runbot_job('testing', 'running') def _job_02_docker_build(self, build, log_path): """Build the docker image""" build._log('docker_build', 'Building docker image') docker_build(log_path, build._path()) return -2 @runbot_job('testing') def _job_10_test_base(self, build, log_path): build._log('test_base', 'Start test base module') self._local_pg_createdb("%s-base" % build.dest) cmd, mods = build._cmd() cmd += ['-d', '%s-base' % build.dest, '-i', 'base', '--stop-after-init', '--log-level=test', '--max-cron-threads=0'] if build.extra_params: cmd.extend(shlex.split(build.extra_params)) return docker_run(cmd, log_path, build._path(), build._get_docker_name(), cpu_limit=600) @runbot_job('testing', 'running') def _job_20_test_all(self, build, log_path): cpu_limit = 2400 self._local_pg_createdb("%s-all" % build.dest) cmd, mods = build._cmd() build._log('test_all', 'Start test all modules') if grep(build._server("tools/config.py"), "test-enable") and build.job_type in ('testing', 'all'): cmd.extend(['--test-enable', '--log-level=test']) else: build._log('test_all', 'Installing modules without testing') cmd += ['-d', '%s-all' % build.dest, '-i', mods, '--stop-after-init', '--max-cron-threads=0'] if build.extra_params: cmd.extend(build.extra_params.split(' ')) if build.coverage: cpu_limit *= 1.5 available_modules = [ os.path.basename(os.path.dirname(a)) for a in (glob.glob(build._server('addons/*/__openerp__.py')) + glob.glob(build._server('addons/*/__manifest__.py'))) ] bad_modules = set(available_modules) - set((mods or '').split(',')) omit = ['--omit', ','.join('*addons/%s/*' %m for m in bad_modules) + '*__manifest__.py'] cmd = [ get_py_version(build), '-m', 'coverage', 'run', '--branch', '--source', '/data/build'] + omit + cmd # reset job_start to an accurate job_20 job_time build.write({'job_start': now()}) return docker_run(cmd, log_path, build._path(), build._get_docker_name(), cpu_limit=cpu_limit) @runbot_job('testing') def _job_21_coverage_html(self, build, log_path): if not build.coverage: return -2 build._log('coverage_html', 'Start generating coverage html') cov_path = build._path('coverage') os.makedirs(cov_path, exist_ok=True) cmd = [ get_py_version(build), "-m", "coverage", "html", "-d", "/data/build/coverage", "--ignore-errors"] return docker_run(cmd, log_path, build._path(), build._get_docker_name()) @runbot_job('testing') def _job_22_coverage_result(self, build, log_path): if not build.coverage: return -2 build._log('coverage_result', 'Start getting coverage result') cov_path = build._path('coverage/index.html') if os.path.exists(cov_path): with open(cov_path,'r') as f: data = f.read() covgrep = re.search(r'pc_cov.>(?P\d+)%', data) build.coverage_result = covgrep and covgrep.group('coverage') or False else: build._log('coverage_result', 'Coverage file not found') return -2 # nothing to wait for @runbot_job('testing', 'running') def _job_29_results(self, build, log_path): build._log('run', 'Getting results for build %s' % build.dest) log_all = build._path('logs', 'job_20_test_all.txt') log_time = time.localtime(os.path.getmtime(log_all)) v = { 'job_end': time2str(log_time), } if grep(log_all, ".modules.loading: Modules loaded."): if rfind(log_all, _re_error): v['result'] = "ko" elif rfind(log_all, _re_warning): v['result'] = "warn" elif not grep(build._server("test/common.py"), "post_install") or grep(log_all, "Initiating shutdown."): v['result'] = "ok" else: v['result'] = "ko" build.write(v) build._github_status() return -2 @runbot_job('running') def _job_30_run(self, build, log_path): # adjust job_end to record an accurate job_20 job_time build._log('run', 'Start running build %s' % build.dest) # run server cmd, mods = build._cmd() if os.path.exists(build._server('addons/im_livechat')): cmd += ["--workers", "2"] cmd += ["--longpolling-port", "8070"] cmd += ["--max-cron-threads", "1"] else: # not sure, to avoid old server to check other dbs cmd += ["--max-cron-threads", "0"] cmd += ['-d', '%s-all' % build.dest] if grep(build._server("tools/config.py"), "db-filter"): if build.repo_id.nginx: cmd += ['--db-filter', '%d.*$'] else: cmd += ['--db-filter', '%s.*$' % build.dest] smtp_host = docker_get_gateway_ip() if smtp_host: cmd += ['--smtp', smtp_host] return docker_run(cmd, log_path, build._path(), build._get_docker_name(), exposed_ports = [build.port, build.port + 1])