2006-04-20 16:26:50 +02:00
|
|
|
#!/usr/bin/python
|
|
|
|
|
|
|
|
# Copyright (C) 2006 Peter Poeml. All rights reserved.
|
|
|
|
# This program is free software; it may be used, copied, modified
|
|
|
|
# and distributed under the terms of the GNU General Public Licence,
|
|
|
|
# either version 2, or (at your option) any later version.
|
|
|
|
|
2006-05-30 12:31:25 +02:00
|
|
|
__version__ = '0.6'
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import urllib2
|
|
|
|
from urlparse import urlunsplit
|
2006-05-02 10:17:45 +02:00
|
|
|
import cElementTree as ET
|
|
|
|
from cStringIO import StringIO
|
2006-05-30 12:07:16 +02:00
|
|
|
import shutil
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
|
|
|
|
from xml.dom.ext.reader import Sax2
|
|
|
|
from xml.dom.ext import PrettyPrint
|
|
|
|
|
|
|
|
netloc = 'api.opensuse.org'
|
|
|
|
scheme = 'http'
|
|
|
|
|
|
|
|
BUFSIZE = 1024*1024
|
|
|
|
store = '.osc'
|
|
|
|
exclude_stuff = [store, '.svn', 'CVS']
|
|
|
|
|
|
|
|
|
2006-05-23 17:27:43 +02:00
|
|
|
new_project_templ = """\
|
2006-05-23 18:16:14 +02:00
|
|
|
<project name="%s">
|
2006-05-23 17:27:43 +02:00
|
|
|
<title>Short title of NewProject</title>
|
|
|
|
<description>This project aims at providing some foo and bar.
|
|
|
|
|
|
|
|
It also does some weird stuff.
|
|
|
|
</description>
|
|
|
|
<person role="maintainer" userid="%s" />
|
|
|
|
</project>
|
|
|
|
"""
|
|
|
|
|
|
|
|
new_package_templ = """\
|
|
|
|
<package name="%s">
|
|
|
|
<title>Title of NewPackage</title>
|
|
|
|
<description>DESCIPTION</description>
|
|
|
|
<person role="maintainer" userid="%s"/>
|
|
|
|
</package>
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
class File:
|
|
|
|
"""represent a file, including its metadata"""
|
|
|
|
def __init__(self, name, md5, size, mtime):
|
|
|
|
self.name = name
|
|
|
|
self.md5 = md5
|
|
|
|
self.size = size
|
|
|
|
self.mtime = mtime
|
|
|
|
def __str__(self):
|
|
|
|
return self.name
|
|
|
|
|
|
|
|
|
2006-05-22 12:50:37 +02:00
|
|
|
class Project:
|
|
|
|
"""represent a project directory, holding packages"""
|
|
|
|
def __init__(self, dir):
|
|
|
|
self.dir = dir
|
2006-06-06 12:50:40 +02:00
|
|
|
self.absdir = os.path.abspath(dir)
|
2006-05-22 12:50:37 +02:00
|
|
|
|
|
|
|
self.name = store_read_project(self.dir)
|
|
|
|
|
|
|
|
self.pacs_available = meta_get_packagelist(self.name)
|
|
|
|
|
|
|
|
self.pacs_have = []
|
|
|
|
for i in os.listdir(self.dir):
|
|
|
|
if i in self.pacs_available:
|
|
|
|
self.pacs_have.append(i)
|
|
|
|
|
|
|
|
self.pacs_missing = []
|
|
|
|
for i in self.pacs_available:
|
|
|
|
if i not in self.pacs_have:
|
|
|
|
self.pacs_missing.append(i)
|
|
|
|
|
|
|
|
def checkout_missing_pacs(self):
|
|
|
|
for pac in self.pacs_missing:
|
|
|
|
print 'checking out new package %s' % pac
|
|
|
|
olddir = os.getcwd()
|
|
|
|
os.chdir(os.pardir)
|
|
|
|
checkout_package(self.name, pac)
|
|
|
|
os.chdir(olddir)
|
|
|
|
|
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
r = []
|
|
|
|
r.append('*****************************************************')
|
|
|
|
r.append('Project %s (dir=%s, absdir=%s)' % (self.name, self.dir, self.absdir))
|
|
|
|
r.append('have pacs:\n%s' % ', '.join(self.pacs_have))
|
|
|
|
r.append('missing pacs:\n%s' % ', '.join(self.pacs_missing))
|
|
|
|
r.append('*****************************************************')
|
|
|
|
return '\n'.join(r)
|
|
|
|
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
class Package:
|
|
|
|
"""represent a package (its directory) and read/keep/write its metadata"""
|
|
|
|
def __init__(self, workingdir):
|
|
|
|
self.dir = workingdir
|
2006-06-06 12:50:40 +02:00
|
|
|
self.absdir = os.path.abspath(self.dir)
|
2006-05-19 22:13:29 +02:00
|
|
|
self.storedir = os.path.join(self.dir, store)
|
|
|
|
|
|
|
|
check_store_version(self.dir)
|
|
|
|
|
|
|
|
self.prjname = store_read_project(self.dir)
|
|
|
|
self.name = store_read_package(self.dir)
|
|
|
|
|
|
|
|
files_tree = read_filemeta(self.dir)
|
|
|
|
files_tree_root = files_tree.getroot()
|
|
|
|
|
|
|
|
self.rev = files_tree_root.get('rev')
|
|
|
|
|
|
|
|
self.filenamelist = []
|
|
|
|
self.filelist = []
|
|
|
|
for node in files_tree_root.findall('entry'):
|
2006-05-22 16:12:06 +02:00
|
|
|
try:
|
2006-05-23 10:43:04 +02:00
|
|
|
f = File(node.get('name'),
|
|
|
|
node.get('md5'),
|
|
|
|
int(node.get('size')),
|
|
|
|
int(node.get('mtime')))
|
2006-05-22 16:12:06 +02:00
|
|
|
except:
|
2006-05-23 10:43:04 +02:00
|
|
|
# okay, a very old version of _files, which didn't contain any metadata yet...
|
|
|
|
f = File(node.get('name'), '', 0, 0)
|
2006-05-19 22:13:29 +02:00
|
|
|
self.filelist.append(f)
|
|
|
|
self.filenamelist.append(f.name)
|
|
|
|
|
|
|
|
self.to_be_deleted = read_tobedeleted(self.dir)
|
2006-05-22 16:12:06 +02:00
|
|
|
self.in_conflict = read_inconflict(self.dir)
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
self.todo = []
|
|
|
|
self.todo_send = []
|
|
|
|
self.todo_delete = []
|
|
|
|
|
|
|
|
# gather unversioned files (the ones not listed in _meta)
|
|
|
|
self.filenamelist_unvers = []
|
|
|
|
for i in os.listdir(self.dir):
|
|
|
|
if i in exclude_stuff:
|
|
|
|
continue
|
|
|
|
if not i in self.filenamelist:
|
|
|
|
self.filenamelist_unvers.append(i)
|
|
|
|
|
|
|
|
def addfile(self, n):
|
|
|
|
st = os.stat(os.path.join(self.dir, n))
|
2006-06-16 14:19:02 +02:00
|
|
|
f = File(n, None, st.st_size, st.st_mtime)
|
2006-05-19 22:13:29 +02:00
|
|
|
self.filelist.append(f)
|
|
|
|
self.filenamelist.append(n)
|
|
|
|
self.filenamelist_unvers.remove(n)
|
2006-05-30 12:07:16 +02:00
|
|
|
shutil.copy2(os.path.join(self.dir, n), os.path.join(self.storedir, n))
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-05-30 12:07:16 +02:00
|
|
|
def delete_localfile(self, n):
|
|
|
|
try: os.unlink(os.path.join(self.dir, n))
|
|
|
|
except: pass
|
|
|
|
try: os.unlink(os.path.join(self.storedir, n))
|
|
|
|
except: pass
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
def put_on_deletelist(self, n):
|
|
|
|
if n not in self.to_be_deleted:
|
|
|
|
self.to_be_deleted.append(n)
|
|
|
|
|
2006-05-22 16:12:06 +02:00
|
|
|
def put_on_conflictlist(self, n):
|
|
|
|
if n not in self.in_conflict:
|
|
|
|
self.in_conflict.append(n)
|
|
|
|
|
|
|
|
def clear_from_conflictlist(self, n):
|
|
|
|
"""delete an entry from the file, and remove the file if it would be empty"""
|
|
|
|
if n in self.in_conflict:
|
|
|
|
|
|
|
|
filename = os.path.join(self.dir, n)
|
|
|
|
storefilename = os.path.join(self.storedir, n)
|
|
|
|
myfilename = os.path.join(self.dir, n + '.mine')
|
|
|
|
upfilename = os.path.join(self.dir, n + '.r' + self.rev)
|
|
|
|
|
2006-06-08 12:30:29 +02:00
|
|
|
try:
|
|
|
|
os.unlink(myfilename)
|
|
|
|
# the working copy may be updated, so the .r* ending may be obsolete...
|
|
|
|
# then we don't care
|
|
|
|
os.unlink(upfilename)
|
|
|
|
except:
|
|
|
|
pass
|
2006-05-22 16:12:06 +02:00
|
|
|
|
|
|
|
self.in_conflict.remove(n)
|
|
|
|
|
2006-05-23 15:59:29 +02:00
|
|
|
self.write_conflictlist()
|
2006-05-22 16:12:06 +02:00
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def write_deletelist(self):
|
2006-05-22 16:12:06 +02:00
|
|
|
if len(self.to_be_deleted) == 0:
|
2006-05-22 19:51:44 +02:00
|
|
|
try:
|
|
|
|
os.unlink(os.path.join(self.storedir, '_to_be_deleted'))
|
|
|
|
except:
|
|
|
|
pass
|
2006-05-22 16:12:06 +02:00
|
|
|
else:
|
|
|
|
fname = os.path.join(self.storedir, '_to_be_deleted')
|
|
|
|
f = open(fname, 'w')
|
|
|
|
f.write('\n'.join(self.to_be_deleted))
|
|
|
|
f.write('\n')
|
|
|
|
f.close()
|
|
|
|
|
2006-05-30 12:07:16 +02:00
|
|
|
def delete_source_file(self, n):
|
|
|
|
import othermethods
|
|
|
|
|
|
|
|
u = makeurl(['source', self.prjname, self.name, n])
|
|
|
|
othermethods.delfile(u, n, username, password)
|
|
|
|
|
|
|
|
self.delete_localfile(n)
|
|
|
|
|
|
|
|
def put_source_file(self, n):
|
|
|
|
import othermethods
|
|
|
|
|
2006-06-06 12:50:40 +02:00
|
|
|
# escaping '+' in the URL path (note: not in the URL query string) is
|
|
|
|
# only a workaround for ruby on rails, which swallows it otherwise
|
|
|
|
u = makeurl(['source', self.prjname, self.name, n.replace('+', '%2B')])
|
2006-05-30 12:07:16 +02:00
|
|
|
othermethods.putfile(u, os.path.join(self.dir, n), username, password)
|
|
|
|
|
|
|
|
shutil.copy2(os.path.join(self.dir, n), os.path.join(self.storedir, n))
|
|
|
|
|
2006-05-22 16:12:06 +02:00
|
|
|
def write_conflictlist(self):
|
|
|
|
if len(self.in_conflict) == 0:
|
|
|
|
os.unlink(os.path.join(self.storedir, '_in_conflict'))
|
|
|
|
else:
|
|
|
|
fname = os.path.join(self.storedir, '_in_conflict')
|
|
|
|
f = open(fname, 'w')
|
|
|
|
f.write('\n'.join(self.in_conflict))
|
|
|
|
f.write('\n')
|
|
|
|
f.close()
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
def updatefile(self, n):
|
|
|
|
filename = os.path.join(self.dir, n)
|
|
|
|
storefilename = os.path.join(self.storedir, n)
|
2006-05-22 10:16:31 +02:00
|
|
|
mtime = self.findfilebyname(n).mtime
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
get_source_file(self.prjname, self.name, n, targetfilename=filename)
|
|
|
|
os.utime(filename, (-1, mtime))
|
|
|
|
|
2006-05-30 12:07:16 +02:00
|
|
|
shutil.copy2(filename, storefilename)
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-05-22 16:12:06 +02:00
|
|
|
def mergefile(self, n):
|
|
|
|
filename = os.path.join(self.dir, n)
|
|
|
|
storefilename = os.path.join(self.storedir, n)
|
|
|
|
myfilename = os.path.join(self.dir, n + '.mine')
|
|
|
|
upfilename = os.path.join(self.dir, n + '.r' + self.rev)
|
|
|
|
os.rename(filename, myfilename)
|
|
|
|
|
|
|
|
get_source_file(self.prjname, self.name, n, targetfilename=upfilename)
|
|
|
|
|
2006-06-08 12:30:29 +02:00
|
|
|
ret = os.system('diff3 -m -E %s %s %s > %s' \
|
|
|
|
% (myfilename, storefilename, upfilename, filename))
|
2006-05-22 16:12:06 +02:00
|
|
|
if ret == 0:
|
|
|
|
# merge was successful... clean up
|
|
|
|
os.rename(upfilename, filename)
|
2006-05-30 12:07:16 +02:00
|
|
|
shutil.copy2(filename, storefilename)
|
2006-05-22 16:12:06 +02:00
|
|
|
os.unlink(myfilename)
|
2006-05-23 11:59:22 +02:00
|
|
|
return 'G'
|
2006-05-22 16:12:06 +02:00
|
|
|
else:
|
|
|
|
# unsuccessful merge
|
|
|
|
self.in_conflict.append(n)
|
|
|
|
self.write_conflictlist()
|
|
|
|
return 'C'
|
|
|
|
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def update_filesmeta(self):
|
|
|
|
meta = ''.join(show_files_meta(self.prjname, self.name))
|
|
|
|
f = open(os.path.join(self.storedir, '_files'), 'w')
|
|
|
|
f.write(meta)
|
|
|
|
f.close()
|
|
|
|
|
|
|
|
def update_pacmeta(self):
|
|
|
|
meta = ''.join(show_package_meta(self.prjname, self.name))
|
|
|
|
f = open(os.path.join(self.storedir, '_meta'), 'w')
|
|
|
|
f.write(meta)
|
|
|
|
f.close()
|
|
|
|
|
|
|
|
def findfilebyname(self, n):
|
|
|
|
for i in self.filelist:
|
|
|
|
if i.name == n:
|
|
|
|
return i
|
|
|
|
|
|
|
|
def status(self, n):
|
|
|
|
"""
|
|
|
|
status can be:
|
|
|
|
|
|
|
|
file storefile file present STATUS
|
|
|
|
exists exists in _files
|
|
|
|
|
|
|
|
x x - 'A'
|
2006-05-22 16:12:06 +02:00
|
|
|
x x x ' ' if digest differs: 'M'
|
|
|
|
and if in conflicts file: 'C'
|
2006-05-19 22:13:29 +02:00
|
|
|
x - - '?'
|
|
|
|
x - x 'D' and listed in _to_be_deleted
|
|
|
|
- x x '!'
|
|
|
|
- x - 'D' (when file in working copy is already deleted)
|
|
|
|
- - x 'F' (new in repo, but not yet in working copy)
|
|
|
|
- - - NOT DEFINED
|
|
|
|
|
|
|
|
"""
|
|
|
|
|
|
|
|
known_by_meta = False
|
|
|
|
exists = False
|
|
|
|
exists_in_store = False
|
|
|
|
if n in self.filenamelist:
|
|
|
|
known_by_meta = True
|
|
|
|
if os.path.exists(os.path.join(self.dir, n)):
|
|
|
|
exists = True
|
|
|
|
if os.path.exists(os.path.join(self.storedir, n)):
|
|
|
|
exists_in_store = True
|
|
|
|
|
|
|
|
|
|
|
|
if exists and not exists_in_store and known_by_meta:
|
|
|
|
state = 'D'
|
|
|
|
elif n in self.to_be_deleted:
|
|
|
|
state = 'D'
|
2006-05-22 16:12:06 +02:00
|
|
|
elif n in self.in_conflict:
|
|
|
|
state = 'C'
|
2006-05-19 22:13:29 +02:00
|
|
|
elif exists and exists_in_store and known_by_meta:
|
|
|
|
#print self.findfilebyname(n)
|
|
|
|
if dgst(os.path.join(self.dir, n)) != self.findfilebyname(n).md5:
|
|
|
|
state = 'M'
|
|
|
|
else:
|
|
|
|
state = ' '
|
|
|
|
elif exists and not exists_in_store and not known_by_meta:
|
|
|
|
state = '?'
|
|
|
|
elif exists and exists_in_store and not known_by_meta:
|
|
|
|
state = 'A'
|
|
|
|
elif not exists and exists_in_store and known_by_meta:
|
|
|
|
state = '!'
|
|
|
|
elif not exists and not exists_in_store and known_by_meta:
|
|
|
|
state = 'F'
|
|
|
|
elif not exists and exists_in_store and not known_by_meta:
|
|
|
|
state = 'D'
|
|
|
|
elif not exists and not exists_in_store and not known_by_meta:
|
|
|
|
print '%s: not exists and not exists_in_store and not nown_by_meta' % n
|
|
|
|
print 'this code path should never be reached!'
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
return state
|
|
|
|
|
|
|
|
|
|
|
|
def merge(self, otherpac):
|
|
|
|
self.todo += otherpac.todo
|
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
r = """
|
|
|
|
name: %s
|
|
|
|
prjname: %s
|
|
|
|
workingdir: %s
|
|
|
|
localfilelist: %s
|
|
|
|
rev: %s
|
|
|
|
'todo' files: %s
|
|
|
|
""" % (self.name,
|
|
|
|
self.prjname,
|
|
|
|
self.dir,
|
|
|
|
'\n '.join(self.filenamelist),
|
|
|
|
self.rev,
|
|
|
|
self.todo)
|
|
|
|
|
|
|
|
return r
|
|
|
|
|
|
|
|
|
2006-05-31 14:13:26 +02:00
|
|
|
def read_meta_from_spec(self):
|
|
|
|
specfile = os.path.join(self.dir, self.name + '.spec')
|
|
|
|
name, summary, descr = read_meta_from_spec(specfile)
|
|
|
|
|
|
|
|
if name != self.name:
|
|
|
|
print 'name from spec does not match name of package... this is probably a problem'
|
|
|
|
sys.exit(1)
|
|
|
|
self.summary = summary
|
|
|
|
self.descr = descr
|
|
|
|
|
|
|
|
|
|
|
|
def update_pac_meta(self):
|
|
|
|
import othermethods
|
|
|
|
import tempfile
|
|
|
|
|
|
|
|
(f, filename) = tempfile.mkstemp(prefix = 'osc_editmeta.', suffix = '.xml', dir = '/tmp')
|
|
|
|
|
|
|
|
try:
|
|
|
|
m = show_package_meta(self.prjname, self.name)
|
|
|
|
except urllib2.HTTPError, e:
|
|
|
|
if e.code == 404:
|
|
|
|
print 'package does not exist yet... creating it'
|
|
|
|
m = new_package_templ % (pac, username)
|
|
|
|
else:
|
|
|
|
print 'error getting package meta for project \'%s\' package \'%s\':' % (prj, pac)
|
|
|
|
print e
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
f = open(filename, 'w')
|
|
|
|
f.write(''.join(m))
|
|
|
|
f.close()
|
|
|
|
|
|
|
|
tree = ET.parse(filename)
|
|
|
|
tree.find('title').text = self.summary
|
|
|
|
tree.find('description').text = ''.join(self.descr)
|
|
|
|
tree.write(filename)
|
|
|
|
|
|
|
|
# FIXME: escape stuff for xml
|
|
|
|
print '*' * 36, 'old', '*' * 36
|
|
|
|
print ''.join(m)
|
|
|
|
print '*' * 36, 'new', '*' * 36
|
|
|
|
tree.write(sys.stdout)
|
|
|
|
print '*' * 72
|
|
|
|
|
|
|
|
# FIXME: for testing...
|
|
|
|
# open the new description in $EDITOR instead?
|
|
|
|
repl = raw_input('Write? (y/N) ')
|
|
|
|
if repl == 'y':
|
|
|
|
print 'Sending meta data...',
|
|
|
|
u = makeurl(['source', self.prjname, self.name, '_meta'])
|
|
|
|
othermethods.putfile(u, filename, username, password)
|
|
|
|
print 'Done.'
|
|
|
|
else:
|
|
|
|
print 'discarding', filename
|
|
|
|
|
|
|
|
os.unlink(filename)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2006-05-22 12:50:37 +02:00
|
|
|
def is_project_dir(d):
|
|
|
|
if os.path.exists(os.path.join(d, store, '_project')) and not \
|
|
|
|
os.path.exists(os.path.join(d, store, '_package')):
|
|
|
|
return True
|
|
|
|
else:
|
|
|
|
return False
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
def findpacs(files):
|
|
|
|
pacs = []
|
|
|
|
for f in files:
|
|
|
|
if f in exclude_stuff:
|
|
|
|
break
|
|
|
|
|
|
|
|
p = filedir_to_pac(f)
|
|
|
|
known = None
|
|
|
|
for i in pacs:
|
|
|
|
if i.name == p.name:
|
|
|
|
known = i
|
|
|
|
break
|
|
|
|
if known:
|
|
|
|
i.merge(p)
|
|
|
|
else:
|
|
|
|
pacs.append(p)
|
|
|
|
return pacs
|
|
|
|
|
|
|
|
|
|
|
|
def read_filemeta(dir):
|
|
|
|
return ET.parse(os.path.join(dir, store, '_files'))
|
|
|
|
|
|
|
|
|
|
|
|
def read_tobedeleted(dir):
|
|
|
|
r = []
|
|
|
|
fname = os.path.join(dir, store, '_to_be_deleted')
|
|
|
|
|
|
|
|
if os.path.exists(fname):
|
|
|
|
|
|
|
|
for i in open(fname, 'r').readlines():
|
|
|
|
r.append(i.strip())
|
|
|
|
|
|
|
|
return r
|
|
|
|
|
|
|
|
|
2006-05-22 16:12:06 +02:00
|
|
|
def read_inconflict(dir):
|
|
|
|
r = []
|
|
|
|
fname = os.path.join(dir, store, '_in_conflict')
|
|
|
|
|
|
|
|
if os.path.exists(fname):
|
|
|
|
|
|
|
|
for i in open(fname, 'r').readlines():
|
|
|
|
r.append(i.strip())
|
|
|
|
|
|
|
|
return r
|
|
|
|
|
|
|
|
|
2006-05-23 15:48:58 +02:00
|
|
|
def parseargs(list_of_args):
|
|
|
|
if list_of_args:
|
|
|
|
return list_of_args
|
2006-05-11 09:27:50 +02:00
|
|
|
else:
|
2006-05-23 15:48:58 +02:00
|
|
|
return [ os.curdir ]
|
2006-05-11 09:27:50 +02:00
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def filedir_to_pac(f):
|
|
|
|
|
|
|
|
if os.path.isdir(f):
|
|
|
|
wd = f
|
|
|
|
p = Package(wd)
|
|
|
|
|
|
|
|
elif os.path.isfile(f):
|
|
|
|
wd = os.path.dirname(f)
|
|
|
|
if wd == '':
|
|
|
|
wd = os.curdir
|
|
|
|
p = Package(wd)
|
|
|
|
p.todo = [ os.path.basename(f) ]
|
|
|
|
|
|
|
|
else:
|
|
|
|
wd = os.path.dirname(f)
|
|
|
|
if wd == '':
|
|
|
|
wd = os.curdir
|
|
|
|
p = Package(wd)
|
|
|
|
p.todo = [ os.path.basename(f) ]
|
|
|
|
|
|
|
|
|
|
|
|
#else:
|
|
|
|
# print
|
|
|
|
# print 'error: %s is neither a valid file or directory' % f
|
|
|
|
# sys.exit(1)
|
|
|
|
|
|
|
|
return p
|
|
|
|
|
|
|
|
|
|
|
|
def statfrmt(statusletter, filename):
|
|
|
|
return '%s %s' % (statusletter, filename)
|
|
|
|
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def makeurl(l):
|
|
|
|
"""given a list of path compoments, construct a complete URL"""
|
|
|
|
return urlunsplit((scheme, netloc, '/'.join(l), '', ''))
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def readauth():
|
|
|
|
"""look for the credentials. If there aren't any, ask and store them"""
|
|
|
|
|
|
|
|
#
|
|
|
|
# try .netrc first
|
|
|
|
#
|
|
|
|
|
|
|
|
# the needed entry in .netrc looks like this:
|
|
|
|
# machine api.opensuse.org login your_login password your_pass
|
|
|
|
# but it is not able for credentials containing spaces
|
|
|
|
import netrc
|
|
|
|
global username, password
|
|
|
|
|
|
|
|
try:
|
|
|
|
info = netrc.netrc()
|
|
|
|
username, account, password = info.authenticators(netloc)
|
|
|
|
return username, password
|
|
|
|
|
|
|
|
except (IOError, TypeError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
#
|
|
|
|
# try .oscrc next
|
|
|
|
#
|
|
|
|
import ConfigParser
|
|
|
|
conffile = os.path.expanduser('~/.oscrc')
|
|
|
|
if os.path.exists(conffile):
|
|
|
|
config = ConfigParser.ConfigParser()
|
|
|
|
config.read(conffile)
|
|
|
|
username = config.get(netloc, 'user')
|
|
|
|
password = config.get(netloc, 'pass')
|
|
|
|
return username, password
|
|
|
|
|
|
|
|
#
|
|
|
|
# create .oscrc
|
|
|
|
#
|
|
|
|
import getpass
|
|
|
|
print >>sys.stderr, \
|
|
|
|
"""your user account / password are not configured yet.
|
|
|
|
You will be asked for them below, and they will be stored in
|
|
|
|
%s for later use.
|
|
|
|
""" % conffile
|
|
|
|
|
|
|
|
username = raw_input('Username: ')
|
|
|
|
password = getpass.getpass()
|
|
|
|
|
|
|
|
fd = open(conffile, 'w')
|
|
|
|
os.chmod(conffile, 0600)
|
|
|
|
print >>fd, '[%s]\nuser: %s\npass: %s' % (netloc, username, password)
|
|
|
|
fd.close()
|
|
|
|
|
|
|
|
return username, password
|
|
|
|
|
|
|
|
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def init_basicauth():
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
username, password = readauth()
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
passmgr = urllib2.HTTPPasswordMgrWithDefaultRealm()
|
|
|
|
# this creates a password manager
|
|
|
|
passmgr.add_password(None, netloc, username, password)
|
|
|
|
# because we have put None at the start it will always
|
|
|
|
# use this username/password combination for urls
|
|
|
|
# for which `netloc` is a super-url
|
|
|
|
|
|
|
|
authhandler = urllib2.HTTPBasicAuthHandler(passmgr)
|
|
|
|
# create the AuthHandler
|
|
|
|
|
|
|
|
opener = urllib2.build_opener(authhandler)
|
2006-05-30 13:44:06 +02:00
|
|
|
opener.addheaders = [('User-agent', 'osc/%s' % __version__)]
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
urllib2.install_opener(opener)
|
|
|
|
# All calls to urllib2.urlopen will now use our handler
|
|
|
|
# Make sure not to include the protocol in with the URL, or
|
|
|
|
# HTTPPasswordMgrWithDefaultRealm will be very confused.
|
|
|
|
# You must (of course) use it when fetching the page though.
|
|
|
|
|
|
|
|
|
|
|
|
def init_package_dir(project, package, dir):
|
|
|
|
if not os.path.isdir(store):
|
|
|
|
os.mkdir(store)
|
|
|
|
os.chdir(store)
|
|
|
|
f = open('_project', 'w')
|
|
|
|
f.write(project + '\n')
|
|
|
|
f.close
|
|
|
|
f = open('_package', 'w')
|
|
|
|
f.write(package + '\n')
|
|
|
|
f.close
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
f = open('_files', 'w')
|
|
|
|
f.write(''.join(show_files_meta(project, package)))
|
|
|
|
f.close()
|
|
|
|
|
|
|
|
f = open('_osclib_version', 'w')
|
|
|
|
f.write(__version__ + '\n')
|
2006-04-20 16:26:50 +02:00
|
|
|
f.close()
|
|
|
|
|
2006-05-22 10:16:31 +02:00
|
|
|
os.chdir(os.pardir)
|
2006-04-20 16:26:50 +02:00
|
|
|
return
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def check_store_version(dir):
|
|
|
|
versionfile = os.path.join(dir, store, '_osclib_version')
|
2006-04-28 17:37:25 +02:00
|
|
|
try:
|
2006-05-19 22:13:29 +02:00
|
|
|
v = open(versionfile).read().strip()
|
2006-04-28 17:37:25 +02:00
|
|
|
except:
|
|
|
|
v = ''
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
if v == '':
|
|
|
|
print 'error: "%s" is not an osc working copy' % dir
|
|
|
|
sys.exit(1)
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
if v != __version__:
|
2006-05-30 12:31:25 +02:00
|
|
|
if v in ['0.2', '0.3', '0.4', '0.5']:
|
2006-05-22 16:23:13 +02:00
|
|
|
# version is fine, no migration needed
|
2006-05-19 22:13:29 +02:00
|
|
|
f = open(versionfile, 'w')
|
|
|
|
f.write(__version__ + '\n')
|
|
|
|
f.close()
|
|
|
|
return
|
2006-04-28 17:37:25 +02:00
|
|
|
print
|
2006-05-19 22:13:29 +02:00
|
|
|
print 'the osc metadata of your working copy "%s"' % dir
|
2006-04-28 17:37:25 +02:00
|
|
|
print 'has the wrong version (%s), should be %s' % (v, __version__)
|
|
|
|
print 'please do a fresh checkout'
|
|
|
|
print
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def meta_get_packagelist(prj):
|
|
|
|
|
|
|
|
u = makeurl(['source', prj, '_meta'])
|
2006-06-01 11:05:47 +02:00
|
|
|
|
|
|
|
try:
|
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
except urllib2.HTTPError, e:
|
|
|
|
if e.code == 404:
|
|
|
|
print 'project \'%s\' does not exist' % prj
|
|
|
|
sys.exit(1)
|
|
|
|
else:
|
|
|
|
print e
|
|
|
|
print 'url: \'%s\'' % u
|
|
|
|
sys.exit(1)
|
2006-05-02 10:17:45 +02:00
|
|
|
|
|
|
|
tree = ET.parse(f)
|
|
|
|
root = tree.getroot()
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in root.findall('package'):
|
|
|
|
r.append(node.get('name'))
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
|
|
|
|
|
|
|
def meta_get_filelist(prj, package):
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
u = makeurl(['source', prj, package])
|
2006-04-20 16:26:50 +02:00
|
|
|
f = urllib2.urlopen(u)
|
2006-05-02 10:17:45 +02:00
|
|
|
tree = ET.parse(f)
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in tree.getroot():
|
|
|
|
r.append(node.get('name'))
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
|
|
|
|
|
|
|
def localmeta_addfile(filename):
|
|
|
|
|
|
|
|
if filename in localmeta_get_filelist():
|
|
|
|
return
|
|
|
|
|
|
|
|
reader = Sax2.Reader()
|
2006-04-28 17:37:25 +02:00
|
|
|
f = open(os.path.join(store, '_files')).read()
|
2006-04-20 16:26:50 +02:00
|
|
|
doc = reader.fromString(f)
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
new = doc.createElement('entry')
|
|
|
|
#new.setAttribute('filetype', 'source')
|
|
|
|
new.setAttribute('name', filename)
|
2006-04-20 16:26:50 +02:00
|
|
|
doc.documentElement.appendChild(new)
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
o = open(os.path.join(store, '_files'), 'w')
|
2006-04-20 16:26:50 +02:00
|
|
|
PrettyPrint(doc, stream=o)
|
|
|
|
o.close()
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
def localmeta_removefile(filename):
|
|
|
|
|
|
|
|
reader = Sax2.Reader()
|
2006-04-28 17:37:25 +02:00
|
|
|
f = open(os.path.join(store, '_files')).read()
|
2006-04-20 16:26:50 +02:00
|
|
|
doc = reader.fromString(f)
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
for i in doc.getElementsByTagName('entry'):
|
|
|
|
if i.getAttribute('name') == filename:
|
2006-04-20 16:26:50 +02:00
|
|
|
i.parentNode.removeChild(i)
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
o = open(os.path.join(store, '_files'), 'w')
|
2006-04-20 16:26:50 +02:00
|
|
|
PrettyPrint(doc, stream=o)
|
|
|
|
o.close()
|
|
|
|
|
|
|
|
|
|
|
|
def localmeta_get_filelist():
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
tree = ET.parse(os.path.join(store, '_files'))
|
|
|
|
root = tree.getroot()
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in root.findall('entry'):
|
|
|
|
r.append(node.get('name'))
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
|
|
|
|
|
|
|
def get_slash_source():
|
|
|
|
u = makeurl(['source'])
|
2006-05-02 10:17:45 +02:00
|
|
|
tree = ET.parse(urllib2.urlopen(u))
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in tree.getroot():
|
|
|
|
r.append(node.get('name'))
|
|
|
|
r.sort()
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def show_project_meta(prj):
|
|
|
|
f = urllib2.urlopen(makeurl(['source', prj, '_meta']))
|
|
|
|
return f.readlines()
|
|
|
|
|
|
|
|
|
|
|
|
def show_package_meta(prj, pac):
|
|
|
|
f = urllib2.urlopen(makeurl(['source', prj, pac, '_meta']))
|
|
|
|
return f.readlines()
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-05-23 17:27:43 +02:00
|
|
|
def edit_meta(prj, pac):
|
|
|
|
import othermethods
|
|
|
|
import tempfile
|
|
|
|
|
|
|
|
(f, filename) = tempfile.mkstemp(prefix = 'osc_editmeta.', suffix = '.xml', dir = '/tmp')
|
|
|
|
|
|
|
|
if pac:
|
2006-05-31 14:13:26 +02:00
|
|
|
# package meta
|
2006-05-23 17:27:43 +02:00
|
|
|
u = makeurl(['source', prj, pac, '_meta'])
|
|
|
|
try:
|
|
|
|
m = show_package_meta(prj, pac)
|
|
|
|
except urllib2.HTTPError, e:
|
|
|
|
if e.code == 404:
|
|
|
|
m = new_package_templ % (pac, username)
|
2006-05-31 14:13:26 +02:00
|
|
|
else:
|
|
|
|
print 'error getting package meta for project \'%s\' package \'%s\':' % (prj, pac)
|
|
|
|
print e
|
|
|
|
sys.exit(1)
|
2006-05-23 17:27:43 +02:00
|
|
|
|
|
|
|
else:
|
2006-05-31 14:13:26 +02:00
|
|
|
# project meta
|
2006-05-23 17:27:43 +02:00
|
|
|
u = makeurl(['source', prj, '_meta'])
|
|
|
|
try:
|
|
|
|
m = show_project_meta(prj)
|
|
|
|
except urllib2.HTTPError, e:
|
|
|
|
if e.code == 404:
|
2006-05-23 18:16:14 +02:00
|
|
|
m = new_project_templ % (prj, username)
|
2006-05-31 14:13:26 +02:00
|
|
|
else:
|
|
|
|
print 'error getting package meta for project \'%s\':' % prj
|
|
|
|
print e
|
|
|
|
sys.exit(1)
|
2006-05-23 17:27:43 +02:00
|
|
|
|
|
|
|
f = open(filename, 'w')
|
|
|
|
f.write(''.join(m))
|
|
|
|
f.close()
|
|
|
|
|
2006-06-16 14:19:02 +02:00
|
|
|
timestamp = os.path.getmtime(filename)
|
2006-05-23 17:27:43 +02:00
|
|
|
|
|
|
|
editor = os.getenv('EDITOR', default='vim')
|
|
|
|
os.system('%s %s' % (editor, filename))
|
|
|
|
|
2006-06-16 14:19:02 +02:00
|
|
|
if os.path.getmtime(filename) == timestamp:
|
2006-05-23 17:27:43 +02:00
|
|
|
print 'File unchanged. Not saving.'
|
|
|
|
os.unlink(filename)
|
|
|
|
|
|
|
|
else:
|
|
|
|
print 'Sending meta data...',
|
|
|
|
othermethods.putfile(u, filename, username, password)
|
|
|
|
os.unlink(filename)
|
|
|
|
print 'Done.'
|
|
|
|
|
|
|
|
|
2006-04-28 17:37:25 +02:00
|
|
|
def show_files_meta(prj, pac):
|
|
|
|
f = urllib2.urlopen(makeurl(['source', prj, pac]))
|
|
|
|
return f.readlines()
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
|
2006-06-06 12:50:40 +02:00
|
|
|
def show_upstream_rev(prj, pac):
|
|
|
|
m = show_files_meta(prj, pac)
|
|
|
|
return ET.parse(StringIO(''.join(m))).getroot().get('rev')
|
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def read_meta_from_spec(specfile):
|
|
|
|
"""read Name, Summary and %description from spec file"""
|
|
|
|
in_descr = False
|
|
|
|
descr = []
|
|
|
|
|
|
|
|
if not os.path.isfile(specfile):
|
|
|
|
print 'file \'%s\' is not a readable file' % specfile
|
|
|
|
return None
|
|
|
|
|
|
|
|
for line in open(specfile, 'r'):
|
|
|
|
if line.startswith('Name:'):
|
|
|
|
name = line.split(':')[1].strip()
|
|
|
|
if line.startswith('Summary:'):
|
|
|
|
summary = line.split(':')[1].strip()
|
|
|
|
if line.startswith('%description'):
|
|
|
|
in_descr = True
|
|
|
|
continue
|
|
|
|
if in_descr and line.startswith('%'):
|
|
|
|
break
|
|
|
|
if in_descr:
|
|
|
|
descr.append(line)
|
|
|
|
|
|
|
|
return name, summary, descr
|
|
|
|
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def get_user_id(user):
|
2006-05-22 16:57:11 +02:00
|
|
|
u = makeurl(['person', user.replace(' ', '+')])
|
|
|
|
try:
|
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
return ''.join(f.readlines())
|
|
|
|
except urllib2.HTTPError:
|
|
|
|
print 'user \'%s\' not found' % user
|
|
|
|
return None
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def get_source_file(prj, package, filename, targetfilename=None):
|
2006-06-08 12:30:29 +02:00
|
|
|
u = makeurl(['source', prj, package, filename.replace('+', '%2B')])
|
2006-04-20 16:26:50 +02:00
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
o = open(targetfilename or filename, 'w')
|
2006-04-20 16:26:50 +02:00
|
|
|
while 1:
|
|
|
|
buf = f.read(BUFSIZE)
|
|
|
|
if not buf: break
|
|
|
|
o.write(buf)
|
|
|
|
o.close()
|
|
|
|
|
|
|
|
|
|
|
|
def dgst(file):
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
#if not os.path.exists(file):
|
|
|
|
#return None
|
2006-04-20 16:26:50 +02:00
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
import md5
|
|
|
|
s = md5.new()
|
2006-04-20 16:26:50 +02:00
|
|
|
f = open(file, 'r')
|
|
|
|
while 1:
|
|
|
|
buf = f.read(BUFSIZE)
|
|
|
|
if not buf: break
|
|
|
|
s.update(buf)
|
2006-05-19 22:13:29 +02:00
|
|
|
return s.hexdigest()
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
|
2006-06-08 12:30:29 +02:00
|
|
|
def binary(s):
|
|
|
|
"""return true if a string is binary data using diff's heuristic"""
|
|
|
|
if s and '\0' in s[:4096]:
|
|
|
|
return True
|
|
|
|
return False
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def get_source_file_diff(dir, filename, rev):
|
|
|
|
import difflib
|
|
|
|
|
2006-05-19 22:28:04 +02:00
|
|
|
file1 = os.path.join(dir, store, filename) # stored original
|
|
|
|
file2 = os.path.join(dir, filename) # working copy
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
f1 = open(file1, 'r')
|
|
|
|
f2 = open(file2, 'r')
|
|
|
|
|
2006-06-08 12:30:29 +02:00
|
|
|
s1 = f1.read()
|
|
|
|
s2 = f2.read()
|
|
|
|
|
|
|
|
if binary(s1) or binary (s2):
|
|
|
|
d = ['Binary file %s has changed\n' % filename]
|
|
|
|
|
|
|
|
else:
|
|
|
|
d = difflib.unified_diff(\
|
|
|
|
s1.splitlines(1), \
|
|
|
|
s2.splitlines(1), \
|
|
|
|
fromfile = '%s (revision %s)' % (filename, rev), \
|
|
|
|
tofile = '%s (working copy)' % filename)
|
2006-05-19 22:13:29 +02:00
|
|
|
|
|
|
|
f1.close()
|
|
|
|
f2.close()
|
|
|
|
|
|
|
|
return ''.join(d)
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
def make_dir(project, package):
|
|
|
|
#print "creating directory '%s'" % project
|
|
|
|
if not os.path.exists(project):
|
2006-05-22 10:16:31 +02:00
|
|
|
print statfrmt('A', project)
|
2006-04-20 16:26:50 +02:00
|
|
|
os.mkdir(project)
|
|
|
|
os.mkdir(os.path.join(project, store))
|
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
f = open(os.path.join(project, store, '_project'), 'w')
|
|
|
|
f.write(project + '\n')
|
|
|
|
f.close()
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
#print "creating directory '%s/%s'" % (project, package)
|
|
|
|
if not os.path.exists(os.path.join(project, package)):
|
2006-05-22 10:16:31 +02:00
|
|
|
print statfrmt('A', '%s/%s' % (project, package))
|
2006-04-20 16:26:50 +02:00
|
|
|
os.mkdir(os.path.join(project, package))
|
|
|
|
os.mkdir(os.path.join(project, package, store))
|
|
|
|
|
|
|
|
return(os.path.join(project, package))
|
|
|
|
|
|
|
|
|
|
|
|
def checkout_package(project, package):
|
|
|
|
olddir = os.getcwd()
|
|
|
|
|
|
|
|
os.chdir(make_dir(project, package))
|
|
|
|
init_package_dir(project, package, store)
|
2006-05-22 10:16:31 +02:00
|
|
|
p = Package(os.curdir)
|
|
|
|
|
|
|
|
for filename in p.filenamelist:
|
|
|
|
p.updatefile(filename)
|
|
|
|
print 'A ', os.path.join(project, package, filename)
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
os.chdir(olddir)
|
|
|
|
|
|
|
|
|
|
|
|
def get_platforms():
|
|
|
|
f = urllib2.urlopen(makeurl(['platform']))
|
2006-05-02 10:17:45 +02:00
|
|
|
tree = ET.parse(f)
|
2006-04-20 16:26:50 +02:00
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in tree.getroot():
|
|
|
|
r.append(node.get('name'))
|
|
|
|
r.sort()
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def get_platforms_of_project(prj):
|
|
|
|
f = show_project_meta(prj)
|
2006-05-02 10:17:45 +02:00
|
|
|
tree = ET.parse(StringIO(''.join(f)))
|
2006-04-20 16:26:50 +02:00
|
|
|
|
|
|
|
r = []
|
2006-05-02 10:17:45 +02:00
|
|
|
for node in tree.findall('repository'):
|
|
|
|
r.append(node.get('name'))
|
2006-04-20 16:26:50 +02:00
|
|
|
return r
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-04-29 22:44:09 +02:00
|
|
|
def show_results_meta(prj, package, platform):
|
2006-04-20 16:26:50 +02:00
|
|
|
u = makeurl(['result', prj, platform, package, 'result'])
|
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
return f.readlines()
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-04-29 22:44:09 +02:00
|
|
|
def get_results(prj, package, platform):
|
|
|
|
#print '----------------------------------------'
|
|
|
|
|
|
|
|
r = []
|
|
|
|
#result_line_templ = '%(prj)-15s %(pac)-15s %(rep)-15s %(arch)-10s %(status)s'
|
2006-05-11 09:27:50 +02:00
|
|
|
result_line_templ = '%(rep)-15s %(arch)-10s %(status)s'
|
2006-04-29 22:44:09 +02:00
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
f = show_results_meta(prj, package, platform)
|
|
|
|
tree = ET.parse(StringIO(''.join(f)))
|
|
|
|
|
|
|
|
root = tree.getroot()
|
|
|
|
|
|
|
|
rmap = {}
|
|
|
|
rmap['prj'] = root.get('project')
|
|
|
|
rmap['pac'] = root.get('package')
|
|
|
|
rmap['rep'] = root.get('repository')
|
|
|
|
|
|
|
|
for node in root.findall('archresult'):
|
|
|
|
rmap['arch'] = node.get('arch')
|
|
|
|
|
|
|
|
statusnode = node.find('status')
|
|
|
|
rmap['status'] = statusnode.get('code')
|
|
|
|
|
2006-06-01 11:05:47 +02:00
|
|
|
if rmap['status'] in ['expansion error', 'broken']:
|
2006-05-02 10:17:45 +02:00
|
|
|
rmap['status'] += ': ' + statusnode.find('summary').text
|
|
|
|
|
|
|
|
if rmap['status'] == 'failed':
|
2006-05-11 09:27:50 +02:00
|
|
|
rmap['status'] += ': %s://%s' % (scheme, netloc) + \
|
|
|
|
'/result/%(prj)s/%(rep)s/%(pac)s/%(arch)s/log' % rmap
|
2006-04-29 22:44:09 +02:00
|
|
|
|
|
|
|
r.append(result_line_templ % rmap)
|
|
|
|
return r
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def get_log(prj, package, platform, arch):
|
|
|
|
u = makeurl(['result', prj, platform, package, arch, 'log'])
|
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
return f.readlines()
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-05-19 22:13:29 +02:00
|
|
|
def get_history(prj, package):
|
|
|
|
# http://api.opensuse.org/rpm/Apache/factory/i586/apache2/history ?
|
|
|
|
# http://api.opensuse.org/package/Apache/apache2/history ?
|
|
|
|
u = makeurl(['package', prj, package, 'history'])
|
|
|
|
print u
|
|
|
|
f = urllib2.urlopen(u)
|
|
|
|
return f.readlines()
|
|
|
|
|
|
|
|
|
2006-06-08 12:30:29 +02:00
|
|
|
def cmd_rebuild(prj, package):
|
|
|
|
u = makeurl(['source', prj, package, '?cmd=rebuild'])
|
|
|
|
# adding data to the request makes it a POST
|
|
|
|
f = urllib2.urlopen(u, data=' ')
|
|
|
|
return f.readlines()
|
|
|
|
|
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def store_read_project(dir):
|
|
|
|
p = open(os.path.join(dir, store, '_project')).readlines()[0].strip()
|
|
|
|
return p
|
|
|
|
|
2006-05-02 10:17:45 +02:00
|
|
|
|
2006-04-20 16:26:50 +02:00
|
|
|
def store_read_package(dir):
|
|
|
|
p = open(os.path.join(dir, store, '_package')).readlines()[0].strip()
|
|
|
|
return p
|
|
|
|
|
2006-05-23 10:43:04 +02:00
|
|
|
def get_osc_version():
|
|
|
|
return __version__
|
2006-04-20 16:26:50 +02:00
|
|
|
|