mirror of
https://github.com/odoo/runbot.git
synced 2025-04-27 13:26:40 +07:00

The goal is to reduce maintenance and odd disk interactions & concurrency issues, by not creating concurrent clones, not having to push forks back in the repository, etc... it also removes the need to cleanup "scratch" working copies though that looks not to have been an issue in a while. The work is done on isolated objects without using or mutating refs, so even concurrent work should not be a problem. This turns out to not be any more verbose (less so if anything) than using `cherry-pick`, as that is not really designed for scripted / non-interactive use, or for squashing commits thereafter. Working directly with trees and commits is quite a bit cleaner even without a ton of helpers. Much of the credit goes to Julia Evans for [their investigation of 3-way merges as the underpinnings of cherry-picking][3-way merge], this would have been a lot more difficult if I'd had to rediscover the merge-base trick independently. A few things have been changed by this: - The old trace/stderr from cherrypick has disappeared as it's generated by cherrypick, but for a non-interactive use it's kinda useless anyway so I probably should have looked into removing it earlier (I think the main use was investigation of the inflateinit issue). - Error on emptied commits has to be hand-rolled as `merge-tree` couldn't care less, this is not hard but is a bit annoying. - `merge-tree`'s conflict information only references raw commits, which makes sense, but requires updating a bunch of tests. Then again so does the fact that it *usually* doesn't send anything to stderr, so that's usually disappearing. Conveniently `merge-tree` merges the conflict marker directly in the files / tree so we don't have to mess about moving them back out of the repository and into the working copy as I assume cherry-pick does, which means we don't have to try and commit them back in ether. That is a huge part of the gain over faffing about with the working copy. Fixes #847 [3-way merge]: https://jvns.ca/blog/2023/11/10/how-cherry-pick-and-revert-work/
576 lines
22 KiB
Python
576 lines
22 KiB
Python
import enum
|
|
import itertools
|
|
import json
|
|
import logging
|
|
from collections import Counter
|
|
from typing import Dict
|
|
|
|
from markupsafe import Markup
|
|
|
|
from odoo import models, fields, api, Command
|
|
from odoo.exceptions import UserError
|
|
from odoo.tools import drop_view_if_exists
|
|
|
|
from ... import git
|
|
from ..pull_requests import Repository
|
|
|
|
_logger = logging.getLogger(__name__)
|
|
class FreezeWizard(models.Model):
|
|
_name = 'runbot_merge.project.freeze'
|
|
_description = "Wizard for freezing a project('s master)"
|
|
|
|
project_id = fields.Many2one('runbot_merge.project', required=True)
|
|
errors = fields.Text(compute='_compute_errors')
|
|
branch_name = fields.Char(required=True, help="Name of the new branches to create")
|
|
|
|
required_pr_ids = fields.Many2many(
|
|
'runbot_merge.pull_requests', string="Required Pull Requests",
|
|
domain="[('state', 'not in', ('closed', 'merged'))]",
|
|
help="Pull requests which must have been merged before the freeze is allowed",
|
|
)
|
|
|
|
pr_state_key = fields.Html(string="Color Key", compute='_compute_state_key', readonly=True)
|
|
def _compute_state_key(self):
|
|
s = dict(self.env['runbot_merge.pull_requests']._fields['state'].selection)
|
|
self.pr_state_key = Markup("""
|
|
<p>%s</p>
|
|
""") % Markup(" ").join(
|
|
Markup('<span class="badge border-0 fucking_color_key_{}">{}</span>').format(v, s[k])
|
|
for k, v in STATE_COLORMAP.items()
|
|
if v
|
|
)
|
|
|
|
release_label = fields.Many2one('runbot_merge.freeze.labels', store=False, string="Release label", help="Find release PRs by label")
|
|
release_pr_ids = fields.One2many(
|
|
'runbot_merge.project.freeze.prs', 'wizard_id',
|
|
string="Release pull requests",
|
|
help="Pull requests used as tips for the freeze branches, "
|
|
"one per repository",
|
|
)
|
|
|
|
bump_label = fields.Many2one('runbot_merge.freeze.labels', store=False, string="Bump label", help="Find bump PRs by label")
|
|
bump_pr_ids = fields.One2many(
|
|
'runbot_merge.project.freeze.bumps', 'wizard_id',
|
|
string="Bump pull requests",
|
|
help="Pull requests used as tips of the frozen-off branches, "
|
|
"one per repository",
|
|
)
|
|
|
|
_sql_constraints = [
|
|
('unique_per_project', 'unique (project_id)',
|
|
"There should be only one ongoing freeze per project"),
|
|
]
|
|
|
|
@api.onchange('release_label')
|
|
def _onchange_release_label(self):
|
|
if not self.release_label:
|
|
return
|
|
|
|
prs = self.env['runbot_merge.pull_requests'].search([
|
|
('label', '=', self.release_label.label),
|
|
('state', 'not in', ('merged', 'closed')),
|
|
])
|
|
for release_pr in self.release_pr_ids:
|
|
p = prs.filtered(lambda p: p.repository == release_pr.repository_id)
|
|
if len(p) < 2:
|
|
release_pr.pr_id = p
|
|
|
|
@api.onchange('release_pr_ids')
|
|
def _onchange_release_prs(self):
|
|
labels = {p.pr_id.label for p in self.release_pr_ids if p.pr_id}
|
|
self.release_label = len(labels) == 1 and self.env['runbot_merge.freeze.labels'].search([
|
|
('label', '=', labels.pop()),
|
|
])
|
|
|
|
@api.onchange('bump_label')
|
|
def _onchange_bump_label(self):
|
|
if not self.bump_label:
|
|
return
|
|
prs = self.env['runbot_merge.pull_requests'].search([
|
|
('label', '=', self.bump_label.label),
|
|
('state', 'not in', ('merged', 'closed')),
|
|
])
|
|
commands = []
|
|
for bump_pr in self.bump_pr_ids:
|
|
p = prs.filtered(lambda p: p.repository == bump_pr.repository_id)
|
|
if len(p) == 1:
|
|
commands.append(Command.update(bump_pr.id, {'pr_id': p.id}))
|
|
else:
|
|
commands.append(Command.delete(bump_pr.id))
|
|
prs -= p
|
|
|
|
commands.extend(
|
|
Command.create({'repository_id': pr.repository.id, 'pr_id': pr.id})
|
|
for pr in prs
|
|
)
|
|
|
|
self.bump_pr_ids = commands
|
|
|
|
@api.onchange('bump_pr_ids')
|
|
def _onchange_bump_prs(self):
|
|
labels = {p.pr_id.label for p in self.bump_pr_ids if p.pr_id}
|
|
self.bump_label = len(labels) == 1 and self.env['runbot_merge.freeze.labels'].search([
|
|
('label', '=', labels.pop()),
|
|
])
|
|
|
|
@api.depends('release_pr_ids.pr_id.label', 'required_pr_ids.state')
|
|
def _compute_errors(self):
|
|
errors = []
|
|
|
|
release_repos = Counter(self.mapped('release_pr_ids.repository_id'))
|
|
release_repos.subtract(self.project_id.repo_ids)
|
|
excess = {k.name for k, v in release_repos.items() if v > 0}
|
|
if excess:
|
|
errors.append("* Every repository must have one release PR, found multiple for %s" % ', '.join(excess))
|
|
|
|
without = self.release_pr_ids.filtered(lambda p: not p.pr_id)
|
|
if without:
|
|
errors.append("* Every repository must have a release PR, missing release PRs for %s." % ', '.join(
|
|
without.mapped('repository_id.name')
|
|
))
|
|
|
|
labels = set(self.mapped('release_pr_ids.pr_id.label'))
|
|
if len(labels) != 1:
|
|
errors.append("* All release PRs must have the same label, found %r." % ', '.join(sorted(labels)))
|
|
non_squash = self.mapped('release_pr_ids.pr_id').filtered(lambda p: not p.squash)
|
|
if non_squash:
|
|
errors.append("* Release PRs should have a single commit, found more in %s." % ', '.join(p.display_name for p in non_squash))
|
|
|
|
bump_repos = Counter(self.mapped('bump_pr_ids.repository_id'))
|
|
excess = {k.name for k, v in bump_repos.items() if v > 1}
|
|
if excess:
|
|
errors.append("* Every repository may have one bump PR, found multiple for %s" % ', '.join(excess))
|
|
|
|
bump_labels = set(self.mapped('bump_pr_ids.pr_id.label'))
|
|
if len(bump_labels) > 1:
|
|
errors.append("* All bump PRs must have the same label, found %r" % ', '.join(sorted(bump_labels)))
|
|
non_squash = self.mapped('bump_pr_ids.pr_id').filtered(lambda p: not p.squash)
|
|
if non_squash:
|
|
errors.append("* Bump PRs should have a single commit, found more in %s." % ', '.join(p.display_name for p in non_squash))
|
|
|
|
unready = sum(p.state not in ('closed', 'merged') for p in self.required_pr_ids)
|
|
if unready:
|
|
errors.append(f"* {unready} required PRs not ready.")
|
|
|
|
self.errors = '\n'.join(errors) or False
|
|
|
|
def action_cancel(self):
|
|
self.project_id.check_access_rights('write')
|
|
self.project_id.check_access_rule('write')
|
|
self.sudo().unlink()
|
|
|
|
return {'type': 'ir.actions.act_window_close'}
|
|
|
|
def action_open(self):
|
|
return {
|
|
'type': 'ir.actions.act_window',
|
|
'target': 'new',
|
|
'name': f'Freeze project {self.project_id.name}',
|
|
'view_mode': 'form',
|
|
'res_model': self._name,
|
|
'res_id': self.id,
|
|
}
|
|
|
|
def action_freeze(self):
|
|
""" Attempts to perform the freeze.
|
|
"""
|
|
# if there are still errors, reopen the wizard
|
|
if self.errors:
|
|
return self.action_open()
|
|
|
|
conflict_crons = self.env.ref('runbot_merge.merge_cron')\
|
|
| self.env.ref('runbot_merge.staging_cron')\
|
|
| self.env.ref('runbot_merge.process_updated_commits')
|
|
# we don't want to run concurrently to the crons above, though we
|
|
# don't need to prevent read access to them
|
|
self.env.cr.execute(
|
|
'SELECT FROM ir_cron WHERE id =ANY(%s) FOR SHARE NOWAIT',
|
|
[conflict_crons.ids]
|
|
)
|
|
|
|
project_id = self.project_id
|
|
# need to create the new branch, but at the same time resequence
|
|
# everything so the new branch is the second one, just after the branch
|
|
# it "forks"
|
|
master, rest = project_id.branch_ids[0], project_id.branch_ids[1:]
|
|
if self.bump_pr_ids and master.active_staging_id:
|
|
self.env.cr.execute(
|
|
'SELECT FROM runbot_merge_stagings WHERE id = %s FOR UPDATE NOWAIT',
|
|
[master.active_staging_id.id]
|
|
)
|
|
|
|
seq = itertools.count(start=1) # start reseq at 1
|
|
commands = [
|
|
(1, master.id, {'sequence': next(seq)}),
|
|
(0, 0, {
|
|
'name': self.branch_name,
|
|
'sequence': next(seq),
|
|
})
|
|
]
|
|
commands.extend((1, b.id, {'sequence': s}) for s, b in zip(seq, rest))
|
|
project_id.branch_ids = commands
|
|
master_name = master.name
|
|
|
|
gh_sessions = {r: r.github() for r in self.project_id.repo_ids}
|
|
repos: Dict[Repository, git.Repo] = {
|
|
r: git.get_local(r).check(False)
|
|
for r in self.project_id.repo_ids
|
|
}
|
|
for repo, copy in repos.items():
|
|
copy.fetch(git.source_url(repo), '+refs/heads/*:refs/heads/*')
|
|
all_prs = self.release_pr_ids.pr_id | self.bump_pr_ids.pr_id
|
|
for pr in all_prs:
|
|
repos[pr.repository].fetch(
|
|
git.source_url(pr.repository),
|
|
pr.head,
|
|
)
|
|
|
|
# prep new branch (via tmp refs) on every repo
|
|
rel_heads: Dict[Repository, str] = {}
|
|
# store for master heads as odds are high the bump pr(s) will be on the
|
|
# same repo as one of the release PRs
|
|
prevs: Dict[Repository, str] = {}
|
|
for rel in self.release_pr_ids:
|
|
repo_id = rel.repository_id
|
|
gh = gh_sessions[repo_id]
|
|
try:
|
|
prev = prevs[repo_id] = gh.head(master_name)
|
|
except Exception as e:
|
|
raise UserError(f"Unable to resolve branch {master_name} of repository {repo_id.name} to a commit.") from e
|
|
|
|
try:
|
|
commits = gh.commits(rel.pr_id.number)
|
|
except Exception as e:
|
|
raise UserError(f"Unable to fetch commits of release PR {rel.pr_id.display_name}.") from e
|
|
|
|
_logger.debug("rebasing %s on %s (commits=%s)",
|
|
rel.pr_id.display_name, prev, len(commits))
|
|
rel_heads[repo_id] = repos[repo_id].rebase(prev, commits)[0]
|
|
|
|
# prep bump
|
|
bump_heads: Dict[Repository, str] = {}
|
|
for bump in self.bump_pr_ids:
|
|
repo_id = bump.repository_id
|
|
gh = gh_sessions[repo_id]
|
|
|
|
try:
|
|
prev = prevs[repo_id] = prevs.get(repo_id) or gh.head(master_name)
|
|
except Exception as e:
|
|
raise UserError(f"Unable to resolve branch {master_name} of repository {repo_id.name} to a commit.") from e
|
|
|
|
try:
|
|
commits = gh.commits(bump.pr_id.number)
|
|
except Exception as e:
|
|
raise UserError(f"Unable to fetch commits of bump PR {bump.pr_id.display_name}.") from e
|
|
|
|
_logger.debug("rebasing %s on %s (commits=%s)",
|
|
bump.pr_id.display_name, prev, len(commits))
|
|
bump_heads[repo_id] = repos[repo_id].rebase(prev, commits)[0]
|
|
|
|
# prevent concurrent updates to the commits table so we control the
|
|
# creation of commit objects from rebasing the release & bump PRs, do it
|
|
# only just before *pushing*
|
|
self.env.cr.execute("LOCK runbot_merge_commit IN ACCESS EXCLUSIVE MODE NOWAIT")
|
|
|
|
deployed = {}
|
|
# at this point we've got a bunch of tmp branches with merged release
|
|
# and bump PRs, it's time to update the corresponding targets
|
|
to_delete = [] # release prs go on new branches which we try to delete on failure
|
|
to_revert = [] # bump prs go on new branch which we try to revert on failure
|
|
failure = None
|
|
for rel in self.release_pr_ids:
|
|
repo_id = rel.repository_id
|
|
|
|
if repos[repo_id].push(
|
|
git.source_url(repo_id),
|
|
f'{rel_heads[repo_id]}:refs/heads/{self.branch_name}',
|
|
).returncode:
|
|
failure = ('create', repo_id.name, self.branch_name)
|
|
break
|
|
|
|
deployed[rel.pr_id.id] = rel_heads[repo_id]
|
|
to_delete.append(repo_id)
|
|
else: # all release deployments succeeded
|
|
for bump in self.bump_pr_ids:
|
|
repo_id = bump.repository_id
|
|
if repos[repo_id].push(
|
|
git.source_url(repo_id),
|
|
f'{bump_heads[repo_id]}:refs/heads/{master_name}'
|
|
).returncode:
|
|
failure = ('fast-forward', repo_id.name, master_name)
|
|
break
|
|
|
|
deployed[bump.pr_id.id] = bump_heads[repo_id]
|
|
to_revert.append(repo_id)
|
|
|
|
if failure:
|
|
addendums = []
|
|
# creating the branch failed, try to delete all previous branches
|
|
failures = []
|
|
for prev_id in to_revert:
|
|
if repos[prev_id].push(
|
|
'-f',
|
|
git.source_url(prev_id),
|
|
f'{prevs[prev_id]}:refs/heads/{master_name}',
|
|
).returncode:
|
|
failures.append(prev_id.name)
|
|
if failures:
|
|
addendums.append(
|
|
"Subsequently unable to revert branches created in %s." % \
|
|
', '.join(failures)
|
|
)
|
|
failures.clear()
|
|
|
|
for prev_id in to_delete:
|
|
if repos[prev_id].push(
|
|
git.source_url(prev_id),
|
|
f':refs/heads/{self.branch_name}'
|
|
).returncode:
|
|
failures.append(prev_id.name)
|
|
if failures:
|
|
addendums.append(
|
|
"Subsequently unable to delete branches created in %s." % \
|
|
", ".join(failures)
|
|
)
|
|
failures.clear()
|
|
|
|
if addendums:
|
|
addendum = '\n\n' + '\n'.join(addendums)
|
|
else:
|
|
addendum = ''
|
|
|
|
reason, repo, branch = failure
|
|
raise UserError(
|
|
f"Unable to {reason} branch {repo}:{branch}.{addendum}"
|
|
)
|
|
|
|
b = self.env['runbot_merge.branch'].search([('name', '=', self.branch_name)])
|
|
self.env.cr.execute(
|
|
"UPDATE runbot_merge_batch SET target=%s WHERE id = %s;"
|
|
"UPDATE runbot_merge_pull_requests SET target=%s WHERE id = any(%s)",
|
|
[
|
|
b.id, self.release_pr_ids.pr_id.batch_id.id,
|
|
b.id, self.release_pr_ids.pr_id.ids,
|
|
]
|
|
)
|
|
all_prs.batch_id.merge_date = fields.Datetime.now()
|
|
all_prs.reviewed_by = self.env.user.partner_id.id
|
|
for p in all_prs:
|
|
p.commits_map = json.dumps({
|
|
'': deployed[p.id],
|
|
p.head: deployed[p.id]
|
|
})
|
|
|
|
# stagings have to be created conditionally as otherwise we might not
|
|
# have a `target` to set and it's mandatory
|
|
laster = self.env['runbot_merge.stagings'].search(
|
|
[('target', '=', master.id), ('state', '=', 'success')],
|
|
order='id desc',
|
|
limit=1,
|
|
).commits.mapped(lambda c: (c.repository_id, c.commit_id))
|
|
if self.release_pr_ids:
|
|
rel_items = [(0, 0, {
|
|
'repository_id': repo.id,
|
|
'commit_id': self.env['runbot_merge.commit'].create({
|
|
'sha': sha,
|
|
'to_check': False,
|
|
}).id,
|
|
} if (sha := rel_heads.get(repo)) else {
|
|
'repository_id': repo.id,
|
|
'commit_id': commit.id,
|
|
})
|
|
for repo, commit in laster
|
|
]
|
|
self.env['runbot_merge.stagings'].create([{
|
|
'state': 'success',
|
|
'reason': 'release freeze staging',
|
|
'active': False,
|
|
'target': b.id,
|
|
'staging_batch_ids': [
|
|
(0, 0, {'runbot_merge_batch_id': batch.id})
|
|
for batch in self.release_pr_ids.pr_id.batch_id
|
|
],
|
|
'heads': rel_items,
|
|
'commits': rel_items,
|
|
}])
|
|
|
|
if self.bump_pr_ids:
|
|
bump_items = [(0, 0, {
|
|
'repository_id': repo.id,
|
|
'commit_id': self.env['runbot_merge.commit'].create({
|
|
'sha': sha,
|
|
'to_check': False,
|
|
}).id,
|
|
} if (sha := bump_heads.get(repo)) else {
|
|
'repository_id': repo.id,
|
|
'commit_id': commit.id,
|
|
})
|
|
for repo, commit in laster
|
|
]
|
|
self.env['runbot_merge.stagings'].create([{
|
|
'state': 'success',
|
|
'reason': 'bump freeze staging',
|
|
'active': False,
|
|
'target': master.id,
|
|
'staging_batch_ids': [
|
|
(0, 0, {'runbot_merge_batch_id': batch.id})
|
|
for batch in self.bump_pr_ids.pr_id.batch_id
|
|
],
|
|
'heads': bump_items,
|
|
'commits': bump_items,
|
|
}])
|
|
|
|
self.env['runbot_merge.pull_requests.feedback'].create([{
|
|
'repository': pr.repository.id,
|
|
'pull_request': pr.number,
|
|
'close': True,
|
|
'message': json.dumps({
|
|
'sha': deployed[pr.id],
|
|
'base': self.branch_name if pr in self.release_pr_ids.pr_id else None
|
|
})
|
|
} for pr in all_prs])
|
|
|
|
if self.bump_pr_ids:
|
|
master.active_staging_id.cancel("freeze by %s", self.env.user.login)
|
|
# delete wizard
|
|
self.sudo().unlink()
|
|
# managed to create all the things, show reminder text (or close)
|
|
if project_id.freeze_reminder:
|
|
return {
|
|
'type': 'ir.actions.act_window',
|
|
'target': 'new',
|
|
'name': f'Freeze reminder {project_id.name}',
|
|
'view_mode': 'form',
|
|
'res_model': project_id._name,
|
|
'res_id': project_id.id,
|
|
'view_id': self.env.ref('runbot_merge.project_freeze_reminder').id
|
|
}
|
|
|
|
return {'type': 'ir.actions.act_window_close'}
|
|
|
|
class ReleasePullRequest(models.Model):
|
|
_name = 'runbot_merge.project.freeze.prs'
|
|
_description = "links to pull requests used to \"cap\" freezes"
|
|
|
|
wizard_id = fields.Many2one('runbot_merge.project.freeze', required=True, index=True, ondelete='cascade')
|
|
repository_id = fields.Many2one('runbot_merge.repository', required=True)
|
|
pr_id = fields.Many2one(
|
|
'runbot_merge.pull_requests',
|
|
domain='[("repository", "=", repository_id), ("state", "not in", ("closed", "merged"))]',
|
|
string="Release Pull Request",
|
|
)
|
|
label = fields.Char(related='pr_id.label')
|
|
|
|
def write(self, vals):
|
|
# only the pr should be writeable after initial creation
|
|
assert 'wizard_id' not in vals
|
|
assert 'repository_id' not in vals
|
|
# and if the PR gets set, it should match the requested repository
|
|
if 'pr_id' in vals:
|
|
assert self.env['runbot_merge.pull_requests'].browse(vals['pr_id'])\
|
|
.repository == self.repository_id
|
|
|
|
return super().write(vals)
|
|
|
|
class BumpPullRequest(models.Model):
|
|
_name = 'runbot_merge.project.freeze.bumps'
|
|
_description = "links to pull requests used to \"bump\" the development branches"
|
|
|
|
wizard_id = fields.Many2one('runbot_merge.project.freeze', required=True, index=True, ondelete='cascade')
|
|
# FIXME: repo = wizard.repo?
|
|
repository_id = fields.Many2one('runbot_merge.repository', required=True)
|
|
pr_id = fields.Many2one(
|
|
'runbot_merge.pull_requests',
|
|
domain='[("repository", "=", repository_id), ("state", "not in", ("closed", "merged"))]',
|
|
string="Bump Pull Request",
|
|
)
|
|
label = fields.Char(related='pr_id.label')
|
|
|
|
@api.onchange('repository_id')
|
|
def _onchange_repository(self):
|
|
self.pr_id = False
|
|
|
|
def write(self, vals):
|
|
# only the pr should be writeable after initial creation
|
|
assert 'wizard_id' not in vals
|
|
# and if the PR gets set, it should match the requested repository
|
|
if vals.get('pr_id'):
|
|
assert self.env['runbot_merge.pull_requests'].browse(vals['pr_id'])\
|
|
.repository == self.repository_id
|
|
|
|
return super().write(vals)
|
|
|
|
class RepositoryFreeze(models.Model):
|
|
_inherit = 'runbot_merge.repository'
|
|
freeze = fields.Boolean(required=True, default=True,
|
|
help="Freeze this repository by default")
|
|
|
|
@enum.unique
|
|
class Colors(enum.IntEnum):
|
|
No = 0
|
|
Red = 1
|
|
Orange = 2
|
|
Yellow = 3
|
|
LightBlue = 4
|
|
DarkPurple = 5
|
|
Salmon = 6
|
|
MediumBlue = 7
|
|
DarkBlue = 8
|
|
Fuchsia = 9
|
|
Green = 10
|
|
Purple = 11
|
|
|
|
STATE_COLORMAP = {
|
|
'opened': Colors.No,
|
|
'closed': Colors.Orange,
|
|
'validated': Colors.No,
|
|
'approved': Colors.No,
|
|
'ready': Colors.LightBlue,
|
|
'merged': Colors.Green,
|
|
'error': Colors.Red,
|
|
}
|
|
class PullRequest(models.Model):
|
|
_inherit = 'runbot_merge.pull_requests'
|
|
|
|
state_color = fields.Integer(compute='_compute_state_color')
|
|
|
|
@api.depends('state')
|
|
def _compute_state_color(self):
|
|
for p in self:
|
|
p.state_color = STATE_COLORMAP[p.state]
|
|
|
|
class OpenPRLabels(models.Model):
|
|
"""Hacking around using contextual display_name to try and autocomplete
|
|
labels through PRs doesn't work because the client fucks up the display_name
|
|
(apparently they're not keyed on the context), therefore the behaviour
|
|
is inconsistent as the label shown in the autocomplete will result in the
|
|
PR being shown as its label in the o2m, and the other way around (if a PR
|
|
is selected directly in the o2m, then the PR's base display_name will be
|
|
shown in the label lookup field).
|
|
|
|
Therefore create a dumbshit view of label records.
|
|
|
|
Under the assumption that we'll have less than 256 repositories, the id of a
|
|
label record is the PR's id shifted as the high 24 bits, and the repo id as
|
|
the low 8.
|
|
"""
|
|
_name = 'runbot_merge.freeze.labels'
|
|
_description = "view representing labels for open PRs so they can autocomplete properly"
|
|
_rec_name = "label"
|
|
_auto = False
|
|
|
|
def init(self):
|
|
super().init()
|
|
drop_view_if_exists(self.env.cr, "runbot_merge_freeze_labels")
|
|
self.env.cr.execute("""
|
|
CREATE VIEW runbot_merge_freeze_labels AS (
|
|
SELECT DISTINCT ON (label)
|
|
id << 8 | repository as id,
|
|
label
|
|
FROM runbot_merge_pull_requests
|
|
WHERE state != 'merged' AND state != 'closed'
|
|
ORDER BY label, repository, id
|
|
)""")
|
|
|
|
label = fields.Char()
|