Fix Flake E241
One more step towards white space sanity. Next group is E20x.
This commit is contained in:
parent
670c4ce8dc
commit
ec88760187
2
.flake8
2
.flake8
@ -1,4 +1,4 @@
|
||||
[flake8]
|
||||
exclude = .open-build-service/, abichecker, openqa, openqa-maintenance.py
|
||||
max-line-length = 100
|
||||
ignore = E501,F401,E302,E128,E251,E201,E202,E203,E305,F841,E265,E261,E266,E712,E401,E126,E502,E241,E711,E125,E123,W293,W391,E101,E713,E124,E127,E701,E714,W503,W504,E129,E741,E722,E731
|
||||
ignore = E501,F401,E302,E128,E251,E201,E202,E203,E305,F841,E265,E261,E266,E712,E401,E126,E502,E711,E125,E123,W293,W391,E101,E713,E124,E127,E701,E714,W503,W504,E129,E741,E722,E731
|
||||
|
@ -826,8 +826,8 @@ class CommandLineInterface(cmdln.Cmdln):
|
||||
def get_optparser(self):
|
||||
parser = cmdln.Cmdln.get_optparser(self)
|
||||
parser.add_option("--apiurl", '-A', metavar="URL", help="api url")
|
||||
parser.add_option("--user", metavar="USER", help="reviewer user name")
|
||||
parser.add_option("--group", metavar="GROUP", help="reviewer group name")
|
||||
parser.add_option("--user", metavar="USER", help="reviewer user name")
|
||||
parser.add_option("--group", metavar="GROUP", help="reviewer group name")
|
||||
parser.add_option("--dry", action="store_true", help="dry run")
|
||||
parser.add_option("--debug", action="store_true", help="debug output")
|
||||
parser.add_option("--osc-debug", action="store_true", help="osc debug output")
|
||||
|
@ -77,7 +77,7 @@ class ContainerCleaner(ToolBase.ToolBase):
|
||||
if srccontainer not in srccontainerarchs:
|
||||
srccontainerarchs[srccontainer] = []
|
||||
|
||||
logging.debug("%s provides binaries for %s", srccontainer, arch)
|
||||
logging.debug("%s provides binaries for %s", srccontainer, arch)
|
||||
srccontainerarchs[srccontainer] += [arch]
|
||||
|
||||
# Now go through each bucket and find out what doesn't contribute to the newest five
|
||||
@ -97,7 +97,7 @@ class ContainerCleaner(ToolBase.ToolBase):
|
||||
contributes = True
|
||||
|
||||
if contributes:
|
||||
logging.debug("%s contributes to %s", srccontainer, package)
|
||||
logging.debug("%s contributes to %s", srccontainer, package)
|
||||
else:
|
||||
logging.info("%s does not contribute", srccontainer)
|
||||
if len([count for count in archs_found.values() if count > 0]) == 0:
|
||||
|
@ -281,7 +281,7 @@ if __name__ == '__main__':
|
||||
|
||||
parser_notify = subparsers.add_parser('notify', help='notify maintainers of their packages')
|
||||
parser_notify.set_defaults(func=notify)
|
||||
parser_notify.add_argument('--dry', action='store_true', help='dry run emails')
|
||||
parser_notify.add_argument('--dry', action='store_true', help='dry run emails')
|
||||
parser_notify.add_argument("packages", nargs='*', help="packages to check")
|
||||
|
||||
parser_requests = subparsers.add_parser('requests', help='List open requests.')
|
||||
|
@ -103,8 +103,8 @@ class ChangeLogger(cmdln.Cmdln):
|
||||
raise Exception("Could not open %s as an ISO-9660 image." % arg)
|
||||
|
||||
# On Tumbleweed, there is no '/suse' prefix
|
||||
for path in ['/suse/x86_64', '/suse/noarch', '/suse/aarch64', '/suse/s390x',
|
||||
'/x86_64', '/noarch', '/aarch64', '/s390x' ]:
|
||||
for path in ['/suse/x86_64', '/suse/noarch', '/suse/aarch64',
|
||||
'/suse/s390x', '/x86_64', '/noarch', '/aarch64', '/s390x' ]:
|
||||
file_stats = iso.readdir(path)
|
||||
if file_stats is None:
|
||||
continue
|
||||
|
@ -211,7 +211,7 @@ def ingest_requests(api, project):
|
||||
tags['type'] = '_'.join(tags['type'])
|
||||
|
||||
point('review', {'open_for': (completed_at - opened_at).total_seconds()}, completed_at, tags)
|
||||
point('review_count', {'count': 1}, opened_at, tags, True)
|
||||
point('review_count', {'count': 1}, opened_at, tags, True)
|
||||
point('review_count', {'count': -1}, completed_at, tags, True)
|
||||
|
||||
found = []
|
||||
|
@ -249,7 +249,7 @@ class AcceptCommand(object):
|
||||
|
||||
if product_spec != new_product:
|
||||
update_version_attr = True
|
||||
url = self.api.makeurl(['source', project, product_pkg, product_name])
|
||||
url = self.api.makeurl(['source', project, product_pkg, product_name])
|
||||
http_PUT(url + '?comment=Update+version', data=new_product)
|
||||
|
||||
if update_version_attr:
|
||||
|
@ -73,7 +73,7 @@ class CleanupRings(object):
|
||||
self.links[mainpkg] = pkg
|
||||
|
||||
def fill_pkgdeps(self, prj, repo, arch):
|
||||
root = builddepinfo(self.api.apiurl, prj, repo, arch)
|
||||
root = builddepinfo(self.api.apiurl, prj, repo, arch)
|
||||
|
||||
for package in root.findall('package'):
|
||||
# use main package name for multibuild. We can't just ignore
|
||||
|
@ -314,7 +314,7 @@ def source_file_save(apiurl, project, package, filename, content, comment=None):
|
||||
url = makeurl(apiurl, ['source', project, package, filename], {'comment': comment})
|
||||
http_PUT(url, data=content)
|
||||
|
||||
def source_file_ensure(apiurl, project, package, filename, content, comment=None):
|
||||
def source_file_ensure(apiurl, project, package, filename, content, comment=None):
|
||||
if content != source_file_load(apiurl, project, package, filename):
|
||||
source_file_save(apiurl, project, package, filename, content, comment)
|
||||
|
||||
|
@ -187,7 +187,7 @@ class RepoChecker():
|
||||
for line in difflib.unified_diff(old_output, per_source[source]['output'], 'before', 'now'):
|
||||
self.logger.debug(line.strip())
|
||||
oldstate['check'][source] = {'problem': per_source[source]['output'],
|
||||
'rebuild': str(datetime.datetime.now())}
|
||||
'rebuild': str(datetime.datetime.now())}
|
||||
|
||||
for source in list(oldstate['check']):
|
||||
if not source.startswith('{}/{}/{}/'.format(project, repository, arch)):
|
||||
|
@ -312,7 +312,7 @@ class StagingWorkflow(object):
|
||||
source_package.name))
|
||||
return target_package
|
||||
|
||||
def create_project(self, name, reviewer={}, maintainer={}, project_links=[]):
|
||||
def create_project(self, name, reviewer={}, maintainer={}, project_links=[]):
|
||||
if isinstance(name, Project):
|
||||
return name
|
||||
if name in self.projects:
|
||||
|
Loading…
x
Reference in New Issue
Block a user