#!/usr/bin/python # # Copyright (c) 2011, Novell Inc. # # This program is licensed under the BSD license, read LICENSE.BSD # for further information # # pysolv a little software installer demoing the sat solver library/bindings # things it does: # - understands globs for package names / dependencies # - understands .arch suffix # - repository data caching # - on demand loading of secondary repository data # - checksum verification # - deltarpm support # - installation of commandline packages # # things not yet ported: # - gpg verification # - file conflicts # - fastestmirror implementation # # things available in the library but missing from pysolv: # - vendor policy loading # - soft locks file handling # - multi version handling import sys import os import glob import solv import re import tempfile import time import subprocess import rpm from stat import * from solv import Pool, Repo, Dataiterator, Job, Solver, Transaction, Selection from iniparse import INIConfig from optparse import OptionParser #import gc #gc.set_debug(gc.DEBUG_LEAK) class repo_generic(dict): def __init__(self, name, type, attribs = {}): for k in attribs: self[k] = attribs[k] self.name = name self.type = type def calc_cookie_file(self, filename): chksum = solv.Chksum(solv.REPOKEY_TYPE_SHA256) chksum.add("1.1") chksum.add_stat(filename) return chksum.raw() def calc_cookie_fp(self, fp): chksum = solv.Chksum(solv.REPOKEY_TYPE_SHA256) chksum.add("1.1"); chksum.add_fp(fp) return chksum.raw() def calc_cookie_ext(self, f, cookie): chksum = solv.Chksum(solv.REPOKEY_TYPE_SHA256) chksum.add("1.1"); chksum.add(cookie) chksum.add_fstat(f.fileno()) extcookie = chksum.raw() # compatibility to c code if ord(extcookie[0]) == 0: extcookie[0] = chr(1) return extcookie def cachepath(self, ext = None): path = re.sub(r'^\.', '_', self.name) if ext: path += "_" + ext + ".solvx" else: path += ".solv" return "/var/cache/solv/" + re.sub(r'[/]', '_', path) def load(self, pool): self.handle = pool.add_repo(self.name) self.handle.appdata = self self.handle.priority = 99 - self['priority'] dorefresh = bool(int(self['autorefresh'])) if dorefresh: try: st = os.stat(self.cachepath()) if self['metadata_expire'] == -1 or time.time() - st[ST_MTIME] < self['metadata_expire']: dorefresh = False except OSError, e: pass self['cookie'] = '' if not dorefresh and self.usecachedrepo(None): print "repo: '%s': cached" % self.name return True return False def load_ext(self, repodata): return False def setfromurls(self, urls): if not urls: return url = urls[0] print "[using mirror %s]" % re.sub(r'^(.*?/...*?)/.*$', r'\1', url) self['baseurl'] = url def setfrommetalink(self, metalink): f = self.download(metalink, False, None) if not f: return None f = os.fdopen(f.dup(), 'r') urls = [] chksum = None for l in f.readlines(): l = l.strip() m = re.match(r'^([0-9a-fA-F]{64})', l) if m: chksum = solv.Chksum(solv.REPOKEY_TYPE_SHA256, m.group(1)) m = re.match(r'^(https?://.+)repodata/repomd.xml', l) if m: urls.append(m.group(1)) if not urls: chksum = None # in case the metalink is about a different file f.close() self.setfromurls(urls) return chksum def setfrommirrorlist(self, mirrorlist): f = self.download(mirrorlist, False, None) if not f: return f = os.fdopen(f.dup(), 'r') urls = [] for l in f.readline(): l = l.strip() if l[0:6] == 'http://' or l[0:7] == 'https://': urls.append(l) self.setfromurls(urls) f.close() def download(self, file, uncompress, chksum, markincomplete=False): url = None if 'baseurl' not in self: if 'metalink' in self: if file != self['metalink']: metalinkchksum = self.setfrommetalink(self['metalink']) if file == 'repodata/repomd.xml' and metalinkchksum and not chksum: chksum = metalinkchksum else: url = file elif 'mirrorlist' in self: if file != self['mirrorlist']: self.setfrommirrorlist(self['mirrorlist']) else: url = file if not url: if 'baseurl' not in self: print "%s: no baseurl" % self.name return None url = re.sub(r'/$', '', self['baseurl']) + '/' + file f = tempfile.TemporaryFile() st = subprocess.call(['curl', '-f', '-s', '-L', url], stdout=f.fileno()) if os.lseek(f.fileno(), 0, os.SEEK_CUR) == 0 and (st == 0 or not chksum): return None os.lseek(f.fileno(), 0, os.SEEK_SET) if st: print "%s: download error %d" % (file, st) if markincomplete: self['incomplete'] = True return None if chksum: fchksum = solv.Chksum(chksum.type) if not fchksum: print "%s: unknown checksum type" % file if markincomplete: self['incomplete'] = True return None fchksum.add_fd(f.fileno()) if fchksum != chksum: print "%s: checksum mismatch" % file if markincomplete: self['incomplete'] = True return None if uncompress: return solv.xfopen_fd(file, f.fileno()) return solv.xfopen_fd(None, f.fileno()) def usecachedrepo(self, ext, mark=False): try: repopath = self.cachepath(ext) f = open(repopath, 'r') f.seek(-32, os.SEEK_END) fcookie = f.read(32) if len(fcookie) != 32: return False if not ext: cookie = self['cookie'] else: cookie = self['extcookie'] if cookie and fcookie != cookie: return False if self.type != 'system' and not ext: f.seek(-32 * 2, os.SEEK_END) fextcookie = f.read(32) if len(fextcookie) != 32: return False f.seek(0) flags = 0 if ext: flags = Repo.REPO_USE_LOADING|Repo.REPO_EXTEND_SOLVABLES if ext != 'DL': flags |= Repo.REPO_LOCALPOOL if not self.handle.add_solv(f, flags): return False if self.type != 'system' and not ext: self['cookie'] = fcookie self['extcookie'] = fextcookie if mark: # no futimes in python? try: os.utime(repopath, None) except Exception, e: pass except IOError, e: return False return True def writecachedrepo(self, ext, info=None): if 'incomplete' in self: return try: if not os.path.isdir("/var/cache/solv"): os.mkdir("/var/cache/solv", 0755) (fd, tmpname) = tempfile.mkstemp(prefix='.newsolv-', dir='/var/cache/solv') os.fchmod(fd, 0444) f = os.fdopen(fd, 'w+') if not info: self.handle.write(f) elif ext: info.write(f) else: # rewrite_repos case self.handle.write_first_repodata(f) if self.type != 'system' and not ext: if 'extcookie' not in self: self['extcookie'] = self.calc_cookie_ext(f, self['cookie']) f.write(self['extcookie']) if not ext: f.write(self['cookie']) else: f.write(self['extcookie']) f.close() if self.handle.iscontiguous(): # switch to saved repo to activate paging and save memory nf = solv.xfopen(tmpname) if not ext: # main repo self.handle.empty() if not self.handle.add_solv(nf, Repo.SOLV_ADD_NO_STUBS): sys.exit("internal error, cannot reload solv file") else: # extension repodata # need to extend to repo boundaries, as this is how # info.write() has written the data info.extend_to_repo() flags = Repo.REPO_EXTEND_SOLVABLES if ext != 'DL': flags |= Repo.REPO_LOCALPOOL info.add_solv(nf, flags) os.rename(tmpname, self.cachepath(ext)) except IOError, e: if tmpname: os.unlink(tmpname) def updateaddedprovides(self, addedprovides): if 'incomplete' in self: return if not hasattr(self, 'handle'): return if self.handle.isempty(): return # make sure there's just one real repodata with extensions repodata = self.handle.first_repodata() if not repodata: return oldaddedprovides = repodata.lookup_idarray(solv.SOLVID_META, solv.REPOSITORY_ADDEDFILEPROVIDES) if not set(addedprovides) <= set(oldaddedprovides): for id in addedprovides: repodata.add_idarray(solv.SOLVID_META, solv.REPOSITORY_ADDEDFILEPROVIDES, id) repodata.internalize() self.writecachedrepo(None, repodata) def packagespath(self): return '' class repo_repomd(repo_generic): def load(self, pool): if super(repo_repomd, self).load(pool): return True print "rpmmd repo '%s':" % self.name, sys.stdout.flush() f = self.download("repodata/repomd.xml", False, None, None) if not f: print "no repomd.xml file, skipped" self.handle.free(True) del self.handle return False self['cookie'] = self.calc_cookie_fp(f) if self.usecachedrepo(None, True): print "cached" return True self.handle.add_repomdxml(f, 0) print "fetching" (filename, filechksum) = self.find('primary') if filename: f = self.download(filename, True, filechksum, True) if f: self.handle.add_rpmmd(f, None, 0) if 'incomplete' in self: return False # hopeless, need good primary (filename, filechksum) = self.find('updateinfo') if filename: f = self.download(filename, True, filechksum, True) if f: self.handle.add_updateinfoxml(f, 0) self.add_exts() self.writecachedrepo(None) # must be called after writing the repo self.handle.create_stubs() return True def find(self, what): di = self.handle.Dataiterator(solv.SOLVID_META, solv.REPOSITORY_REPOMD_TYPE, what, Dataiterator.SEARCH_STRING) di.prepend_keyname(solv.REPOSITORY_REPOMD) for d in di: dp = d.parentpos() filename = dp.lookup_str(solv.REPOSITORY_REPOMD_LOCATION) chksum = dp.lookup_checksum(solv.REPOSITORY_REPOMD_CHECKSUM) if filename and not chksum: print "no %s file checksum!" % filename filename = None chksum = None if filename: return (filename, chksum) return (None, None) def add_ext(self, repodata, what, ext): filename, chksum = self.find(what) if not filename and what == 'deltainfo': filename, chksum = self.find('prestodelta') if not filename: return handle = repodata.new_handle() repodata.set_poolstr(handle, solv.REPOSITORY_REPOMD_TYPE, what) repodata.set_str(handle, solv.REPOSITORY_REPOMD_LOCATION, filename) repodata.set_checksum(handle, solv.REPOSITORY_REPOMD_CHECKSUM, chksum) if ext == 'DL': repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.REPOSITORY_DELTAINFO) repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.REPOKEY_TYPE_FLEXARRAY) elif ext == 'FL': repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.SOLVABLE_FILELIST) repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.REPOKEY_TYPE_DIRSTRARRAY) repodata.add_flexarray(solv.SOLVID_META, solv.REPOSITORY_EXTERNAL, handle) def add_exts(self): repodata = self.handle.add_repodata(0) self.add_ext(repodata, 'deltainfo', 'DL') self.add_ext(repodata, 'filelists', 'FL') repodata.internalize() def load_ext(self, repodata): repomdtype = repodata.lookup_str(solv.SOLVID_META, solv.REPOSITORY_REPOMD_TYPE) if repomdtype == 'filelists': ext = 'FL' elif repomdtype == 'deltainfo': ext = 'DL' else: return False sys.stdout.write("[%s:%s: " % (self.name, ext)) if self.usecachedrepo(ext): sys.stdout.write("cached]\n") sys.stdout.flush() return True sys.stdout.write("fetching]\n") sys.stdout.flush() filename = repodata.lookup_str(solv.SOLVID_META, solv.REPOSITORY_REPOMD_LOCATION) filechksum = repodata.lookup_checksum(solv.SOLVID_META, solv.REPOSITORY_REPOMD_CHECKSUM) f = self.download(filename, True, filechksum) if not f: return False if ext == 'FL': self.handle.add_rpmmd(f, 'FL', Repo.REPO_USE_LOADING|Repo.REPO_EXTEND_SOLVABLES|Repo.REPO_LOCALPOOL) elif ext == 'DL': self.handle.add_deltainfoxml(f, Repo.REPO_USE_LOADING) self.writecachedrepo(ext, repodata) return True class repo_susetags(repo_generic): def load(self, pool): if super(repo_susetags, self).load(pool): return True print "susetags repo '%s':" % self.name, sys.stdout.flush() f = self.download("content", False, None, None) if not f: print "no content file, skipped" self.handle.free(True) del self.handle return False self['cookie'] = self.calc_cookie_fp(f) if self.usecachedrepo(None, True): print "cached" return True self.handle.add_content(f, 0) print "fetching" defvendorid = self.handle.lookup_id(solv.SOLVID_META, solv.SUSETAGS_DEFAULTVENDOR) descrdir = self.handle.lookup_str(solv.SOLVID_META, solv.SUSETAGS_DESCRDIR) if not descrdir: descrdir = "suse/setup/descr" (filename, filechksum) = self.find('packages.gz') if not filename: (filename, filechksum) = self.find('packages') if filename: f = self.download(descrdir + '/' + filename, True, filechksum, True) if f: self.handle.add_susetags(f, defvendorid, None, Repo.REPO_NO_INTERNALIZE|Repo.SUSETAGS_RECORD_SHARES) (filename, filechksum) = self.find('packages.en.gz') if not filename: (filename, filechksum) = self.find('packages.en') if filename: f = self.download(descrdir + '/' + filename, True, filechksum, True) if f: self.handle.add_susetags(f, defvendorid, None, Repo.REPO_NO_INTERNALIZE|Repo.REPO_REUSE_REPODATA|Repo.REPO_EXTEND_SOLVABLES) self.handle.internalize() self.add_exts() self.writecachedrepo(None) # must be called after writing the repo self.handle.create_stubs() return True def find(self, what): di = self.handle.Dataiterator(solv.SOLVID_META, solv.SUSETAGS_FILE_NAME, what, Dataiterator.SEARCH_STRING) di.prepend_keyname(solv.SUSETAGS_FILE) for d in di: dp = d.parentpos() chksum = dp.lookup_checksum(solv.SUSETAGS_FILE_CHECKSUM) return (what, chksum) return (None, None) def add_ext(self, repodata, what, ext): (filename, chksum) = self.find(what) if not filename: return handle = repodata.new_handle() repodata.set_str(handle, solv.SUSETAGS_FILE_NAME, filename) if chksum: repodata.set_checksum(handle, solv.SUSETAGS_FILE_CHECKSUM, chksum) if ext == 'DU': repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.SOLVABLE_DISKUSAGE) repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.REPOKEY_TYPE_DIRNUMNUMARRAY) elif ext == 'FL': repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.SOLVABLE_FILELIST) repodata.add_idarray(handle, solv.REPOSITORY_KEYS, solv.REPOKEY_TYPE_DIRSTRARRAY) else: for langtag, langtagtype in [ (solv.SOLVABLE_SUMMARY, solv.REPOKEY_TYPE_STR), (solv.SOLVABLE_DESCRIPTION, solv.REPOKEY_TYPE_STR), (solv.SOLVABLE_EULA, solv.REPOKEY_TYPE_STR), (solv.SOLVABLE_MESSAGEINS, solv.REPOKEY_TYPE_STR), (solv.SOLVABLE_MESSAGEDEL, solv.REPOKEY_TYPE_STR), (solv.SOLVABLE_CATEGORY, solv.REPOKEY_TYPE_ID) ]: repodata.add_idarray(handle, solv.REPOSITORY_KEYS, self.handle.pool.id2langid(langtag, ext, 1)) repodata.add_idarray(handle, solv.REPOSITORY_KEYS, langtagtype) repodata.add_flexarray(solv.SOLVID_META, solv.REPOSITORY_EXTERNAL, handle) def add_exts(self): repodata = self.handle.add_repodata(0) di = self.handle.Dataiterator(solv.SOLVID_META, solv.SUSETAGS_FILE_NAME, None, 0) di.prepend_keyname(solv.SUSETAGS_FILE) for d in di: filename = d.str() if not filename: continue if filename[0:9] != "packages.": continue if len(filename) == 11 and filename != "packages.gz": ext = filename[9:11] elif filename[11:12] == ".": ext = filename[9:11] else: continue if ext == "en": continue self.add_ext(repodata, filename, ext) repodata.internalize() def load_ext(self, repodata): filename = repodata.lookup_str(solv.SOLVID_META, solv.SUSETAGS_FILE_NAME) ext = filename[9:11] sys.stdout.write("[%s:%s: " % (self.name, ext)) if self.usecachedrepo(ext): sys.stdout.write("cached]\n") sys.stdout.flush() return True sys.stdout.write("fetching]\n") sys.stdout.flush() defvendorid = self.handle.lookup_id(solv.SOLVID_META, solv.SUSETAGS_DEFAULTVENDOR) descrdir = self.handle.lookup_str(solv.SOLVID_META, solv.SUSETAGS_DESCRDIR) if not descrdir: descrdir = "suse/setup/descr" filechksum = repodata.lookup_checksum(solv.SOLVID_META, solv.SUSETAGS_FILE_CHECKSUM) f = self.download(descrdir + '/' + filename, True, filechksum) if not f: return False flags = Repo.REPO_USE_LOADING|Repo.REPO_EXTEND_SOLVABLES if ext != 'DL': flags |= Repo.REPO_LOCALPOOL self.handle.add_susetags(f, defvendorid, ext, flags) self.writecachedrepo(ext, repodata) return True def packagespath(self): datadir = repo.handle.lookup_str(solv.SOLVID_META, solv.SUSETAGS_DATADIR) if not datadir: datadir = 'suse' return datadir + '/' class repo_unknown(repo_generic): def load(self, pool): print "unsupported repo '%s': skipped" % self.name return False class repo_system(repo_generic): def load(self, pool): self.handle = pool.add_repo(self.name) self.handle.appdata = self pool.installed = self.handle print "rpm database:", self['cookie'] = self.calc_cookie_file("/var/lib/rpm/Packages") if self.usecachedrepo(None): print "cached" return True print "reading" if hasattr(self.handle.__class__, 'add_products'): self.handle.add_products("/etc/products.d", Repo.REPO_NO_INTERNALIZE) self.handle.add_rpmdb(None, Repo.REPO_REUSE_REPODATA) self.writecachedrepo(None) return True class repo_cmdline(repo_generic): def load(self, pool): self.handle = pool.add_repo(self.name) self.handle.appdata = self return True def load_stub(repodata): repo = repodata.repo.appdata if repo: return repo.load_ext(repodata) return False parser = OptionParser(usage="usage: solv.py [options] COMMAND") parser.add_option('-r', '--repo', action="append", type="string", dest="repos") (options, args) = parser.parse_args() if not args: parser.print_help(sys.stderr) sys.exit(1) cmd = args[0] args = args[1:] if cmd == 'li': cmd = 'list' if cmd == 'in': cmd = 'install' if cmd == 'rm': cmd = 'erase' if cmd == 've': cmd = 'verify' if cmd == 'se': cmd = 'search' # read all repo configs repos = [] reposdirs = [] if os.path.isdir("/etc/zypp/repos.d"): reposdirs = [ "/etc/zypp/repos.d" ] else: reposdirs = [ "/etc/yum/repos.d" ] for reposdir in reposdirs: if not os.path.isdir(reposdir): continue for reponame in sorted(glob.glob('%s/*.repo' % reposdir)): cfg = INIConfig(open(reponame)) for alias in cfg: repoattr = {'enabled': 0, 'priority': 99, 'autorefresh': 1, 'type': 'rpm-md', 'metadata_expire': 900} for k in cfg[alias]: repoattr[k] = cfg[alias][k] if 'mirrorlist' in repoattr and 'metalink' not in repoattr: if repoattr['mirrorlist'].find('/metalink'): repoattr['metalink'] = repoattr['mirrorlist'] del repoattr['mirrorlist'] if repoattr['type'] == 'rpm-md': repo = repo_repomd(alias, 'repomd', repoattr) elif repoattr['type'] == 'yast2': repo = repo_susetags(alias, 'susetags', repoattr) else: repo = repo_unknown(alias, 'unknown', repoattr) repos.append(repo) pool = solv.Pool() pool.setarch() pool.set_loadcallback(load_stub) # now load all enabled repos into the pool sysrepo = repo_system('@System', 'system') sysrepo.load(pool) for repo in repos: if int(repo['enabled']): repo.load(pool) repolimiter = None if options.repos: for reponame in options.repos: mrepos = [ repo for repo in repos if repo.name == reponame ] if not mrepos: print "no repository matches '%s'" % reponame sys.exit(1) repo = mrepos[0] if hasattr(repo, 'handle'): if not repolimiter: repolimiter = pool.Selection() repolimiter.add_raw(Job.SOLVER_SOLVABLE_REPO|Job.SOLVER_SETREPO|Job.SOLVER_SETVENDOR, repo.handle.id) if cmd == 'search': matches = {} di = pool.Dataiterator(0, solv.SOLVABLE_NAME, args[0], Dataiterator.SEARCH_SUBSTRING|Dataiterator.SEARCH_NOCASE) for d in di: matches[d.solvid] = True for solvid in sorted(matches.keys()): print " - %s [%s]: %s" % (pool.solvid2str(solvid), pool.solvables[solvid].repo.name, pool.lookup_str(solvid, solv.SOLVABLE_SUMMARY)) sys.exit(0) cmdlinerepo = None if cmd == 'list' or cmd == 'info' or cmd == 'install': for arg in args: if arg.endswith(".rpm") and os.access(arg, os.R_OK): if not cmdlinerepo: cmdlinerepo = repo_cmdline('@commandline', 'cmdline') cmdlinerepo.load(pool) cmdlinerepo['packages'] = {} cmdlinerepo['packages'][arg] = cmdlinerepo.handle.add_rpm(arg, Repo.REPO_REUSE_REPODATA|Repo.REPO_NO_INTERNALIZE) if cmdlinerepo: cmdlinerepo.handle.internalize() addedprovides = pool.addfileprovides_queue() if addedprovides: sysrepo.updateaddedprovides(addedprovides) for repo in repos: repo.updateaddedprovides(addedprovides) pool.createwhatprovides() # convert arguments into jobs jobs = [] for arg in args: if cmdlinerepo and arg in cmdlinerepo['packages']: jobs.append(pool.Job(Job.SOLVER_SOLVABLE, cmdlinerepo['packages'][arg])) else: flags = Selection.SELECTION_NAME|Selection.SELECTION_PROVIDES|Selection.SELECTION_GLOB if len(arg) and arg[0] == '/': flags |= Selection.SELECTION_FILELIST if cmd == 'erase': flags |= Selection.SELECTION_INSTALLED_ONLY sel = pool.select(arg, flags) if repolimiter: sel.limit(repolimiter) if sel.isempty(): sel = pool.select(arg, flags | Selection.SELECTION_NOCASE) if repolimiter: sel.limit(repolimiter) if not sel.isempty(): print "[ignoring case for '%s']" % arg if sel.isempty(): print "nothing matches '%s'" % arg sys.exit(1) if sel.flags() & Selection.SELECTION_FILELIST: print "[using file list match for '%s']" % arg if sel.flags() & Selection.SELECTION_PROVIDES: print "[using capability match for '%s']" % arg jobs += sel.jobs(0) if not jobs and (cmd == 'up' or cmd == 'dup' or cmd == 'verify' or repolimiter): sel = pool.Selection() sel.add_raw(Job.SOLVER_SOLVABLE_ALL, 0) if repolimiter: sel.limit(repolimiter) jobs += sel.jobs(0) if cmd == 'list' or cmd == 'info': if not jobs: print "no package matched." sys.exit(1) for job in jobs: for s in job.solvables(): if cmd == 'info': print "Name: %s" % s print "Repo: %s" % s.repo print "Summary: %s" % s.lookup_str(solv.SOLVABLE_SUMMARY) str = s.lookup_str(solv.SOLVABLE_URL) if str: print "Url: %s" % str str = s.lookup_str(solv.SOLVABLE_LICENSE) if str: print "License: %s" % str print "Description:\n%s" % s.lookup_str(solv.SOLVABLE_DESCRIPTION) print else: print " - %s [%s]" % (s, s.repo) print " %s" % s.lookup_str(solv.SOLVABLE_SUMMARY) sys.exit(0) if cmd == 'install' or cmd == 'erase' or cmd == 'up' or cmd == 'dup' or cmd == 'verify': if not jobs: print "no package matched." sys.exit(1) for job in jobs: if cmd == 'up': job.how |= Job.SOLVER_UPDATE # up magic: use install instead of update if no installed package matches if job.isemptyupdate(): job.how ^= Job.SOLVER_UPDATE ^ Job.SOLVER_INSTALL elif cmd == 'install': job.how |= Job.SOLVER_INSTALL elif cmd == 'erase': job.how |= Job.SOLVER_ERASE elif cmd == 'dup': job.how |= Job.SOLVER_DISTUPGRADE elif cmd == 'verify': job.how |= Job.SOLVER_VERIFY #pool.set_debuglevel(2) solver = None while True: solver = pool.Solver() solver.set_flag(Solver.SOLVER_FLAG_SPLITPROVIDES, 1); if cmd == 'erase': solver.set_flag(Solver.SOLVER_FLAG_ALLOW_UNINSTALL, 1); problems = solver.solve(jobs) if not problems: break for problem in problems: print "Problem %d:" % problem.id r = problem.findproblemrule() ri = r.info() print ri.problemstr() solutions = problem.solutions() for solution in solutions: print " Solution %d:" % solution.id elements = solution.elements(True) for element in elements: print " - %s" % element.str() print sol = '' while not (sol == 's' or sol == 'q' or (sol.isdigit() and int(sol) >= 1 and int(sol) <= len(solutions))): sys.stdout.write("Please choose a solution: ") sys.stdout.flush() sol = sys.stdin.readline().strip() if sol == 's': continue # skip problem if sol == 'q': sys.exit(1) solution = solutions[int(sol) - 1] for element in solution.elements(): newjob = element.Job() if element.type == Solver.SOLVER_SOLUTION_JOB: jobs[element.jobidx] = newjob else: if newjob and newjob not in jobs: jobs.append(newjob) # no problems, show transaction trans = solver.transaction() del solver if trans.isempty(): print "Nothing to do." sys.exit(0) print print "Transaction summary:" print for cl in trans.classify(): if cl.type == Transaction.SOLVER_TRANSACTION_ERASE: print "%d erased packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_INSTALL: print "%d installed packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_REINSTALLED: print "%d reinstalled packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_DOWNGRADED: print "%d downgraded packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_CHANGED: print "%d changed packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_UPGRADED: print "%d upgraded packages:" % cl.count elif cl.type == Transaction.SOLVER_TRANSACTION_VENDORCHANGE: print "%d vendor changes from '%s' to '%s':" % (cl.count, cl.fromdep(), cl.todep()) elif cl.type == Transaction.SOLVER_TRANSACTION_ARCHCHANGE: print "%d arch changes from '%s' to '%s':" % (cl.count, cl.fromdep(), cl.todep()) else: continue for p in cl.solvables(): if cl.type == Transaction.SOLVER_TRANSACTION_UPGRADED or cl.type == Transaction.SOLVER_TRANSACTION_DOWNGRADED: op = trans.othersolvable(p) print " - %s -> %s" % (p, op) else: print " - %s" % p print print "install size change: %d K" % trans.calc_installsizechange() print while True: sys.stdout.write("OK to continue (y/n)? ") sys.stdout.flush() yn = sys.stdin.readline().strip() if yn == 'y': break if yn == 'n': sys.exit(1) newpkgs = trans.newpackages() newpkgsfp = {} if newpkgs: downloadsize = 0 for p in newpkgs: downloadsize += p.lookup_num(solv.SOLVABLE_DOWNLOADSIZE) print "Downloading %d packages, %d K" % (len(newpkgs), downloadsize) for p in newpkgs: repo = p.repo.appdata location, medianr = p.lookup_location() if not location: continue if repo.type == 'commandline': f = solv.xfopen(location) if not f: sys.exit("\n%s: %s not found" % location) newpkgsfp[p.id] = f continue if not sysrepo.handle.isempty() and os.access('/usr/bin/applydeltarpm', os.X_OK): pname = p.name di = p.repo.Dataiterator(solv.SOLVID_META, solv.DELTA_PACKAGE_NAME, pname, Dataiterator.SEARCH_STRING) di.prepend_keyname(solv.REPOSITORY_DELTAINFO) for d in di: dp = d.parentpos() if dp.lookup_id(solv.DELTA_PACKAGE_EVR) != p.evrid or dp.lookup_id(solv.DELTA_PACKAGE_ARCH) != p.archid: continue baseevrid = dp.lookup_id(solv.DELTA_BASE_EVR) candidate = None for installedp in pool.whatprovides(p.nameid): if installedp.isinstalled() and installedp.nameid == p.nameid and installedp.archid == p.archid and installedp.evrid == baseevrid: candidate = installedp if not candidate: continue seq = dp.lookup_deltaseq() st = subprocess.call(['/usr/bin/applydeltarpm', '-a', p.arch, '-c', '-s', seq]) if st: continue chksum = dp.lookup_checksum(solv.DELTA_CHECKSUM) if not chksum: continue dloc, dmedianr = dp.lookup_deltalocation() dloc = repo.packagespath() + dloc f = repo.download(dloc, False, chksum) if not f: continue nf = tempfile.TemporaryFile() nf = os.dup(nf.fileno()) # get rid of CLOEXEC st = subprocess.call(['/usr/bin/applydeltarpm', '-a', p.arch, "/dev/fd/%d" % f.fileno(), "/dev/fd/%d" % nf]) os.lseek(nf, 0, os.SEEK_SET) newpkgsfp[p.id] = solv.xfopen_fd("", nf) os.close(nf) break if p.id in newpkgsfp: sys.stdout.write("d") sys.stdout.flush() continue chksum = p.lookup_checksum(solv.SOLVABLE_CHECKSUM) location = repo.packagespath() + location f = repo.download(location, False, chksum) if not f: sys.exit("\n%s: %s not found in repository" % (repo.name, location)) newpkgsfp[p.id] = f sys.stdout.write(".") sys.stdout.flush() print print "Committing transaction:" print ts = rpm.TransactionSet('/') ts.setVSFlags(rpm._RPMVSF_NOSIGNATURES) erasenamehelper = {} for p in trans.steps(): type = trans.steptype(p, Transaction.SOLVER_TRANSACTION_RPM_ONLY) if type == Transaction.SOLVER_TRANSACTION_ERASE: rpmdbid = p.lookup_num(solv.RPM_RPMDBID) erasenamehelper[p.name] = p if not rpmdbid: sys.exit("\ninternal error: installed package %s has no rpmdbid\n" % p) ts.addErase(rpmdbid) elif type == Transaction.SOLVER_TRANSACTION_INSTALL: f = newpkgsfp[p.id] h = ts.hdrFromFdno(solv.xfileno(f)) os.lseek(solv.xfileno(f), 0, os.SEEK_SET) ts.addInstall(h, p, 'u') elif type == Transaction.SOLVER_TRANSACTION_MULTIINSTALL: f = newpkgsfp[p.id] h = ts.hdrFromFdno(solv.xfileno(f)) os.lseek(solv.xfileno(f), 0, os.SEEK_SET) ts.addInstall(h, p, 'i') checkproblems = ts.check() if checkproblems: print checkproblems sys.exit("Sorry.") ts.order() def runCallback(reason, amount, total, p, d): if reason == rpm.RPMCALLBACK_INST_OPEN_FILE: return solv.xfileno(newpkgsfp[p.id]) if reason == rpm.RPMCALLBACK_INST_START: print "install", p if reason == rpm.RPMCALLBACK_UNINST_START: # argh, p is just the name of the package if p in erasenamehelper: p = erasenamehelper[p] print "erase", p runproblems = ts.run(runCallback, '') if runproblems: print runproblems sys.exit(1) sys.exit(0) print "unknown command", cmd sys.exit(1) # vim: sw=4 et