# # (C) 2011 coolo@suse.de, Novell Inc, openSUSE.org # Distribute under GPLv2 or GPLv3 # # Copy this script to ~/.osc-plugins/ or /var/lib/osc-plugins . # Then try to run 'osc check_repo --help' to see the usage. import os import re import shutil import subprocess import tempfile from urllib import quote_plus import urllib2 import sys from xml.etree import cElementTree as ET from osc import oscerr from osc import cmdln from osc.core import get_binary_file from osc.core import get_buildinfo from osc.core import http_GET from osc.core import http_POST from osc.core import makeurl from osc.core import Request # Expand sys.path to search modules inside the pluging directory _plugin_dir = os.path.expanduser('~/.osc-plugins') sys.path.append(_plugin_dir) from osclib.checkrepo import CheckRepo from osclib.cycle import CycleDetector from osclib.memoize import memoize, CACHEDIR # Directory where download binary packages. DOWNLOADS = os.path.expanduser('~/co/downloads') # # XXX - Ugly Hack. Because the way that osc import plugings we need to # declare some functions and objects used in the decorator as global # global tempfile global wraps global build global last_build_success global jobhistory @memoize() def build(apiurl, project, repo, arch, package): root = None try: url = makeurl(apiurl, ['build', project, repo, arch, package]) root = http_GET(url).read() except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) return root @memoize() def last_build_success(apiurl, src_project, tgt_project, src_package, rev): root = None try: url = makeurl(apiurl, ['build', src_project, '_result?lastsuccess&package=%s&pathproject=%s&srcmd5=%s' % ( quote_plus(src_package), quote_plus(tgt_project), rev)]) root = http_GET(url).read() except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) return root def get_project_repos(apiurl, src_project, tgt_project, src_package, rev): """Read the repositories of the project from _meta.""" # XXX TODO - Shitty logic here. A better proposal is refactorize # _check_repo_buildsuccess. repos = [] url = makeurl(apiurl, ['build', src_project, '_result?lastsuccess&package=%s&pathproject=%s&srcmd5=%s' % ( quote_plus(src_package), quote_plus(tgt_project), rev)]) try: root = ET.parse(http_GET(url)).getroot() for element in root.findall('repository'): archs = [(e.get('arch'), e.get('result')) for e in element.findall('arch')] repos.append((element.get('name'), archs)) except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) return repos def old_md5(apiurl, src_project, tgt_project, src_package, rev): """Recollect old MD5 for a package.""" # XXX TODO - instead of fixing the limit, use endtime to make # sure that we have the correct time frame. limit = 20 query = { 'package': src_package, # 'code': 'succeeded', 'limit': limit, } repositories = get_project_repos(apiurl, src_project, tgt_project, src_package, rev) srcmd5_list = [] for repository, archs in repositories: for arch, status in archs: if srcmd5_list: break if status not in ('succeeded', 'outdated'): continue url = makeurl(apiurl, ['build', src_project, repository, arch, '_jobhistory'], query=query) try: root = ET.parse(http_GET(url)).getroot() srcmd5_list = [e.get('srcmd5') for e in root.findall('jobhist')] except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) md5_set = set() for srcmd5 in srcmd5_list: query = { 'expand': 1, 'rev': srcmd5, } url = makeurl(apiurl, ['source', src_project, src_package], query=query) root = ET.parse(http_GET(url)).getroot() md5_set.add(root.find('linkinfo').get('srcmd5')) return md5_set def _check_repo_change_review_state(self, opts, id_, newstate, message='', supersed=None): """Taken from osc/osc/core.py, improved: - verbose option added, - empty by_user=& removed. - numeric id can be int(). """ query = { 'cmd': 'changereviewstate', 'newstate': newstate, 'by_user': 'factory-repo-checker', } if supersed: query['superseded_by'] = supersed # if message: # query['comment'] = message code = 404 url = makeurl(opts.apiurl, ['request', str(id_)], query=query) try: f = http_POST(url, data=message) root = ET.parse(f).getroot() code = root.attrib['code'] except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) return code def _check_repo_find_submit_request(self, opts, project, package): xpath = "(action/target/@project='%s' and "\ "action/target/@package='%s' and "\ "action/@type='submit' and "\ "(state/@name='new' or state/@name='review' or "\ "state/@name='accepted'))" % (project, package) try: url = makeurl(opts.apiurl, ['search', 'request'], 'match=%s' % quote_plus(xpath)) f = http_GET(url) collection = ET.parse(f).getroot() except urllib2.HTTPError, e: print('ERROR in URL %s [%s]' % (url, e)) return None for root in collection.findall('request'): r = Request() r.read(root) return int(r.reqid) return None def _check_repo_avoid_wrong_friends(self, prj, repo, arch, pkg, opts): xml = build(opts.apiurl, prj, repo, arch, pkg) if xml: root = ET.fromstring(xml) for binary in root.findall('binary'): # if there are binaries, we're out return False return True def _check_repo_one_request(self, rq, opts): class CheckRepoPackage: def __repr__(self): return '[%d:%s/%s]' % (int(self.request), self.sproject, self.spackage) def __init__(self): self.updated = False self.error = None self.build_excluded = False id_ = int(rq.get('id')) actions = rq.findall('action') if len(actions) > 1: msg = 'only one action per request is supported - create a group instead: '\ 'https://github.com/SUSE/hackweek/wiki/Improved-Factory-devel-project-submission-workflow' print('DECLINED', msg) self._check_repo_change_review_state(opts, id_, 'declined', message=msg) return [] act = actions[0] type_ = act.get('type') if type_ != 'submit': msg = 'Unchecked request type %s' % type_ print 'ACCEPTED', msg self._check_repo_change_review_state(opts, id_, 'accepted', message=msg) return [] pkg = act.find('source').get('package') prj = act.find('source').get('project') rev = act.find('source').get('rev') tprj = act.find('target').get('project') tpkg = act.find('target').get('package') subm_id = 'SUBMIT(%d):' % id_ print '%s %s/%s -> %s/%s' % (subm_id, prj, pkg, tprj, tpkg) packs = [] p = CheckRepoPackage() p.spackage = pkg p.sproject = prj p.tpackage = tpkg p.tproject = tprj p.group = opts.grouped.get(id_, id_) p.request = id_ # Get source information about the SR: # - Source MD5 # - Entries (.tar.gz, .changes, .spec ...) and MD5 try: url = makeurl(opts.apiurl, ['source', prj, pkg, '?expand=1&rev=%s' % rev]) root = ET.parse(http_GET(url)).getroot() except urllib2.HTTPError, e: print 'ERROR in URL %s [%s]' % (url, e) return [] p.rev = root.attrib['srcmd5'] # Recover the .spec files specs = [en.attrib['name'][:-5] for en in root.findall('entry') if en.attrib['name'].endswith('.spec')] # source checker validated it exists specs.remove(tpkg) packs.append(p) # Validate the rest of the spec files for spec in specs: lprj, lpkg, lmd5 = '', '', '' try: url = makeurl(opts.apiurl, ['source', prj, spec, '?expand=1']) root = ET.parse(http_GET(url)).getroot() link = root.find('linkinfo') if link is not None: lprj = link.attrib.get('project', '') lpkg = link.attrib.get('package', '') lmd5 = link.attrib['srcmd5'] except urllib2.HTTPError: pass # leave lprj if lprj != prj or lpkg != pkg and not p.updated: msg = '%s/%s should _link to %s/%s' % (prj, spec, prj, pkg) print 'DECLINED', msg self._check_repo_change_review_state(opts, id_, 'declined', message=msg) p.updated = True if lmd5 != p.rev and not p.updated: if lmd5 not in old_md5(opts.apiurl, lprj, p.tproject, spec, p.rev): msg = '%s/%s is a link but has a different md5sum than %s?' % (prj, spec, pkg) else: msg = '%s is no longer the submitted version, please resubmit HEAD' % spec print '[DECLINED] CHECK MANUALLY', msg # self._check_repo_change_review_state(opts, id_, 'declined', message=msg) p.updated = True sp = CheckRepoPackage() sp.spackage = spec sp.sproject = prj sp.tpackage = spec sp.tproject = tprj sp.group = p.group sp.request = id_ packs.append(sp) sp.rev = root.attrib['srcmd5'] return packs def _check_repo_buildsuccess(self, p, opts): root_xml = last_build_success(opts.apiurl, p.sproject, p.tproject, p.spackage, p.rev) root = ET.fromstring(root_xml) if not root: return False if 'code' in root.attrib: print ET.tostring(root) return False result = False p.goodrepos = [] missings = {} alldisabled = True foundbuilding = None foundfailed = None tocheckrepos = [] for repo in root.findall('repository'): archs = [a.attrib['arch'] for a in repo.findall('arch')] foundarchs = len([a for a in archs if a in ('i586', 'x86_64')]) if foundarchs == 2: tocheckrepos.append(repo) if not tocheckrepos: msg = 'Missing i586 and x86_64 in the repo list' print msg self._check_repo_change_review_state(opts, p.request, 'new', message=msg) # Next line not needed, but for documentation p.updated = True return False for repo in tocheckrepos: isgood = True founddisabled = False r_foundbuilding = None r_foundfailed = None r_missings = {} for arch in repo.findall('arch'): if arch.attrib['arch'] not in ('i586', 'x86_64'): continue if 'missing' in arch.attrib: for pkg in arch.attrib['missing'].split(','): if not self._check_repo_avoid_wrong_friends(p.sproject, repo.attrib['name'], arch.attrib['arch'], pkg, opts): missings[pkg] = 1 if not (arch.attrib['result'] in ['succeeded', 'excluded']): isgood = False if arch.attrib['result'] == 'excluded' and arch.attrib['arch'] == 'x86_64': p.build_excluded = True if arch.attrib['result'] == 'disabled': founddisabled = True if arch.attrib['result'] == 'failed' or arch.attrib['result'] == 'unknown': # Sometimes an unknown status is equivalent to # disabled, but we map it as failed to have a human # check (no autoreject) r_foundfailed = repo.attrib['name'] if arch.attrib['result'] == 'building': r_foundbuilding = repo.attrib['name'] if arch.attrib['result'] == 'outdated': msg = "%s's sources were changed after submissions and the old sources never built. Please resubmit" % p.spackage print 'DECLINED', msg self._check_repo_change_review_state(opts, p.request, 'declined', message=msg) # Next line is not needed, but for documentation p.updated = True return False r_missings = r_missings.keys() for pkg in r_missings: missings[pkg] = 1 if not founddisabled: alldisabled = False if isgood: p.goodrepos.append(repo.attrib['name']) result = True if r_foundbuilding: foundbuilding = r_foundbuilding if r_foundfailed: foundfailed = r_foundfailed p.missings = sorted(missings) if result: return True if alldisabled: msg = '%s is disabled or does not build against factory. Please fix and resubmit' % p.spackage print 'DECLINED', msg self._check_repo_change_review_state(opts, p.request, 'declined', message=msg) # Next line not needed, but for documentation p.updated = True return False if foundbuilding: msg = '%s is still building for repository %s' % (p.spackage, foundbuilding) print msg self._check_repo_change_review_state(opts, p.request, 'new', message=msg) # Next line not needed, but for documentation p.updated = True return False if foundfailed: msg = '%s failed to build in repository %s - not accepting' % (p.spackage, foundfailed) # failures might be temporary, so don't autoreject but wait for a human to check print msg self._check_repo_change_review_state(opts, p.request, 'new', message=msg) # Next line not needed, but for documentation p.updated = True return False return True def _check_repo_repo_list(self, prj, repo, arch, pkg, opts, ignore=False): url = makeurl(opts.apiurl, ['build', prj, repo, arch, pkg]) files = [] try: binaries = ET.parse(http_GET(url)).getroot() for bin_ in binaries.findall('binary'): fn = bin_.attrib['filename'] mt = int(bin_.attrib['mtime']) result = re.match(r'(.*)-([^-]*)-([^-]*)\.([^-\.]+)\.rpm', fn) if not result: if fn == 'rpmlint.log': files.append((fn, '', '', mt)) continue pname = result.group(1) if pname.endswith('-debuginfo') or pname.endswith('-debuginfo-32bit'): continue if pname.endswith('-debugsource'): continue if result.group(4) == 'src': continue files.append((fn, pname, result.group(4), mt)) except urllib2.HTTPError: pass # if not ignore: # print 'ERROR in URL %s [%s]' % (url, e) return files def _check_repo_get_binary(self, apiurl, prj, repo, arch, package, file, target, mtime): if os.path.exists(target): # we need to check the mtime too as the file might get updated cur = os.path.getmtime(target) if cur > mtime: return get_binary_file(apiurl, prj, repo, arch, file, package=package, target_filename=target) def _get_verifymd5(self, p, rev): try: url = makeurl(self.get_api_url(), ['source', p.sproject, p.spackage, '?view=info&rev=%s' % rev]) root = ET.parse(http_GET(url)).getroot() except urllib2.HTTPError, e: print 'ERROR in URL %s [%s]' % (url, e) return [] return root.attrib['verifymd5'] def _checker_compare_disturl(self, disturl, p): distmd5 = os.path.basename(disturl).split('-')[0] if distmd5 == p.rev: return True vrev1 = self._get_verifymd5(p, p.rev) vrev2 = self._get_verifymd5(p, distmd5) if vrev1 == vrev2: return True print 'ERROR Revision missmatch: %s, %s' % (vrev1, vrev2) return False def _check_repo_download(self, p, opts): p.downloads = dict() if p.build_excluded: return set() for repo in p.goodrepos: # we can assume x86_64 is there todownload = [] for fn in self._check_repo_repo_list(p.sproject, repo, 'x86_64', p.spackage, opts): todownload.append(('x86_64', fn[0], fn[3])) # now fetch -32bit packs # for fn in self._check_repo_repo_list(p.sproject, repo, 'i586', p.spackage, opts): # if fn[2] == 'x86_64': # todownload.append(('i586', fn[0], fn[3])) p.downloads[repo] = [] for arch, fn, mt in todownload: repodir = os.path.join(DOWNLOADS, p.spackage, repo) if not os.path.exists(repodir): os.makedirs(repodir) t = os.path.join(repodir, fn) self._check_repo_get_binary(opts.apiurl, p.sproject, repo, arch, p.spackage, fn, t, mt) p.downloads[repo].append(t) if fn.endswith('.rpm'): pid = subprocess.Popen(['rpm', '--nosignature', '--queryformat', '%{DISTURL}', '-qp', t], stdout=subprocess.PIPE, close_fds=True) os.waitpid(pid.pid, 0)[1] disturl = pid.stdout.readlines()[0] if not self._checker_compare_disturl(disturl, p): p.error = '[%s] %s does not match revision %s' % (p, disturl, p.rev) return set() toignore = set() for fn in self._check_repo_repo_list(p.tproject, 'standard', 'x86_64', p.tpackage, opts, ignore=True): toignore.add(fn[1]) # now fetch -32bit pack list for fn in self._check_repo_repo_list(p.tproject, 'standard', 'i586', p.tpackage, opts, ignore=True): if fn[2] == 'x86_64': toignore.add(fn[1]) return toignore def _get_buildinfo(self, opts, prj, repo, arch, pkg): """Get the build info for a package""" xml = get_buildinfo(opts.apiurl, prj, pkg, repo, arch) root = ET.fromstring(xml) return [e.attrib['name'] for e in root.findall('bdep')] def _check_repo_group(self, id_, reqs, opts): print '\nCheck group', reqs if not all(self._check_repo_buildsuccess(r, opts) for r in reqs): return # all succeeded toignore = set() destdir = os.path.expanduser('~/co/%s' % str(reqs[0].group)) fetched = dict((r, False) for r in opts.groups.get(id_, [])) packs = [] for p in reqs: i = self._check_repo_download(p, opts) if p.error: if not p.updated: print p.error self._check_repo_change_review_state(opts, p.request, 'new', message=p.error) p.updated = True else: print p.error return toignore.update(i) fetched[p.request] = True packs.append(p) for req, f in fetched.items(): if not f: packs.extend(self._check_repo_fetch_request(req, opts)) for p in packs: if fetched[p.request]: continue # we need to call it to fetch the good repos to download # but the return value is of no interest right now self._check_repo_buildsuccess(p, opts) i = self._check_repo_download(p, opts) if p.error: print 'ERROR (ALREADY ACEPTED?):', p.error p.updated = True toignore.update(i) # Detect cycles into the current Factory graph after we update the # links with the current list of request. cycle_detector = CycleDetector(opts.apiurl) cycle_detector.cycles(packs) for p in reqs: smissing = [] for package in p.missings: alreadyin = False # print package, packs for t in packs: if package == t.tpackage: alreadyin = True if alreadyin: continue # print package, packs, downloads, toignore request = self._check_repo_find_submit_request(opts, p.tproject, package) if request: greqs = opts.groups.get(p.group, []) if request in greqs: continue package = '%s(rq%s)' % (package, request) smissing.append(package) if len(smissing): msg = 'Please make sure to wait before these depencencies are in %s: %s' % (p.tproject, ', '.join(smissing)) if not p.updated: self._check_repo_change_review_state(opts, p.request, 'new', message=msg) print msg p.updated = True else: print msg return # Create a temporal file for the params params_file = tempfile.NamedTemporaryFile(delete=False) params_file.write('\n'.join(f for f in toignore if f.strip())) params_file.close() reposets = [] if len(packs) == 1: p = packs[0] for r in p.downloads.keys(): reposets.append([(p, r, p.downloads[r])]) else: # TODO: for groups we just pick the first repo - we'd need to create a smart # matrix dirstolink = [] for p in packs: keys = p.downloads.keys() if not keys: continue r = keys[0] dirstolink.append((p, r, p.downloads[r])) reposets.append(dirstolink) if len(reposets) == 0: print 'NO REPOS' return for dirstolink in reposets: if os.path.exists(destdir): shutil.rmtree(destdir) os.makedirs(destdir) for p, repo, downloads in dirstolink: dir = destdir + '/%s' % p.tpackage for d in downloads: if not os.path.exists(dir): os.mkdir(dir) os.symlink(d, os.path.join(dir, os.path.basename(d))) repochecker = os.path.join(self.repocheckerdir, 'repo-checker.pl') civs = "LC_ALL=C perl %s '%s' -r %s -f %s" % (repochecker, destdir, self.repodir, params_file.name) # print civs # continue # exit(1) p = subprocess.Popen(civs, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True) # ret = os.waitpid(p.pid, 0)[1] stdoutdata, stderrdata = p.communicate() ret = p.returncode # print ret, stdoutdata, stderrdata if not ret: # skip the others for p, repo, downloads in dirstolink: p.goodrepo = repo break os.unlink(params_file.name) updated = {} if ret: # print stdoutdata, set(map(lambda x: x.request, reqs)) for p in reqs: if updated.get(p.request, False) or p.updated: continue print stdoutdata self._check_repo_change_review_state(opts, p.request, 'new', message=stdoutdata) p.updated = True updated[p.request] = 1 return for p in reqs: if updated.get(p.request, False) or p.updated: continue msg = 'Builds for repo %s' % p.goodrepo print 'ACCEPTED', msg self._check_repo_change_review_state(opts, p.request, 'accepted', message=msg) p.updated = True updated[p.request] = 1 shutil.rmtree(destdir) def _check_repo_fetch_request(self, id_, opts): url = makeurl(opts.apiurl, ['request', str(id_)]) root = ET.parse(http_GET(url)).getroot() return self._check_repo_one_request(root, opts) @cmdln.alias('check', 'cr') @cmdln.option('-s', '--skip', action='store_true', help='skip review') def do_check_repo(self, subcmd, opts, *args): """${cmd_name}: Checker review of submit requests. Usage: ${cmd_name} [SRID]... Shows pending review requests and their current state. ${cmd_option_list} """ opts.mode = '' opts.verbose = False opts.apiurl = self.get_api_url() checkrepo = CheckRepo(opts.apiurl) # XXX TODO - Remove this the all access to opt.group[s|ed] comes # from checkrepo. opts.grouped = checkrepo.grouped opts.groups = checkrepo.groups if opts.skip: if not len(args): raise oscerr.WrongArgs('Provide #IDs to skip.') for id_ in args: msg = 'skip review' print 'ACCEPTED', msg self._check_repo_change_review_state(opts, id_, 'accepted', message=msg) return ids = [arg for arg in args if arg.isdigit()] packs = [] if not ids: # xpath query, using the -m, -r, -s options where = "@by_user='factory-repo-checker'+and+@state='new'" url = makeurl(opts.apiurl, ['search', 'request'], "match=state/@name='review'+and+review[%s]" % where) f = http_GET(url) root = ET.parse(f).getroot() for rq in root.findall('request'): packs.extend(self._check_repo_one_request(rq, opts)) else: # we have a list, use them. for id_ in ids: packs.extend(self._check_repo_fetch_request(id_, opts)) # Order the packs before grouping packs = sorted(packs, key=lambda p: p.request, reverse=True) groups = {} for p in packs: a = groups.get(p.group, []) a.append(p) groups[p.group] = a self.repocheckerdir = os.path.dirname(os.path.realpath(os.path.expanduser('~/.osc-plugins/osc-check_repo.py'))) self.repodir = "%s/repo-%s-%s-x86_64" % (CACHEDIR, 'openSUSE:Factory', 'standard') if not os.path.exists(self.repodir): os.mkdir(self.repodir) civs = 'LC_ALL=C perl %s/bs_mirrorfull --nodebug https://build.opensuse.org/build/%s/%s/x86_64 %s' % ( self.repocheckerdir, 'openSUSE:Factory', 'standard', self.repodir) os.system(civs) # Sort the groups, from high to low. This put first the stating # projects also for id_, reqs in sorted(groups.items(), reverse=True): self._check_repo_group(id_, reqs, opts)