Robustify package upgrades and dependency checking.

- Unmerge appropriate packages during upgrades and downgrades.
- Calculate time spent in dependency generation to the tenth of a second.
- Only track dependencies of packages that are actually being installed.
- Ignore PDEPEND, as it has no impact on dependency ordering.
- Only break dependency chains that go against Portage's install order.
- Rename Failed -> Retrying.
- Print emerge command lines as they are run.

TEST=Emerged hard-host-depends and ran build_packages with parallel_emerge
BUG=none

Review URL: http://codereview.chromium.org/2886010
This commit is contained in:
David James 2010-07-01 20:52:59 -07:00
parent a0fcf30e54
commit a27ae994b5

View File

@ -138,58 +138,49 @@ def GetDepsFromPortage(package):
Args: Args:
package: string containing the packages to build. package: string containing the packages to build.
Returns: Returns:
text output of emege -p --debug, which can be processed elsewhere. text output of emerge -p --debug, which can be processed elsewhere.
""" """
print "Calculating deps for package %s" % package print "Calculating deps for package %s" % package
cmdline = EmergeCommand() + " -p --debug " + package cmdline = EmergeCommand() + " -p --debug --color=n " + package
print "+ %s" % cmdline print "+ %s" % cmdline
# Store output in a temp file as it is too big for a unix pipe. # Store output in a temp file as it is too big for a unix pipe.
stderr_buffer = tempfile.TemporaryFile() stderr_buffer = tempfile.TemporaryFile()
stdout_buffer = tempfile.TemporaryFile() stdout_buffer = tempfile.TemporaryFile()
# Launch the subprocess. # Launch the subprocess.
start = time.time()
depsproc = subprocess.Popen(shlex.split(cmdline), stderr=stderr_buffer, depsproc = subprocess.Popen(shlex.split(cmdline), stderr=stderr_buffer,
stdout=stdout_buffer, bufsize=64*1024) stdout=stdout_buffer, bufsize=64*1024)
# Wait for this to complete.
seconds = 0
while depsproc.poll() is not None:
seconds += 1
time.sleep(1)
if seconds % 5 == 0:
print ".",
print " done"
print "Deps calculated in %d:%02ds" % (seconds / 60, seconds % 60)
depsproc.wait() depsproc.wait()
seconds = time.time() - start
print "Deps calculated in %d:%04.1fs" % (seconds / 60, seconds % 60)
stderr_buffer.seek(0) stderr_buffer.seek(0)
stderr_raw = stderr_buffer.read() stderr_raw = stderr_buffer.read()
info_start = stderr_raw.find("digraph") info_start = stderr_raw.find("digraph")
stdout_buffer.seek(0)
stdout_raw = stdout_buffer.read()
lines = []
if info_start != -1: if info_start != -1:
stdout = stderr_raw[info_start:] lines = stderr_raw[info_start:].split("\n")
else: lines.extend(stdout_raw.split("\n"))
stdout_buffer.seek(0)
stdout_raw = stdout_buffer.read()
stdout = stderr_raw + stdout_raw
if VERBOSE or depsproc.returncode != 0: if VERBOSE or depsproc.returncode != 0:
print stdout output = stderr_raw + stdout_raw
print output
if depsproc.returncode != 0: if depsproc.returncode != 0:
print "Failed to generate deps" print "Failed to generate deps"
sys.exit(1) sys.exit(1)
lines = stdout.split("\n")
return lines return lines
def DepsToTree(lines): def DepsToTree(lines):
"""Regex the emerge --tree output to generate a nested dict of dependencies. """Regex the output from 'emerge --debug' to generate a nested dict of deps.
Args: Args:
lines: text dump from 'emerge -p --tree package' lines: output from 'emerge -p --debug package'
Returns: Returns:
dep_tree: nested dict of dependencies, as specified by emerge. dep_tree: nested dict of dependencies, as specified by emerge.
there may be dupes, or circular deps. there may be dupes, or circular deps.
We need to regex lines as follows: We need to regex lines as follows:
hard-host-depends depends on hard-host-depends depends on
@ -205,13 +196,21 @@ def DepsToTree(lines):
r"(?P<version>\d+[\w\.-]*)\', \'(?P<action>\w+)\'\) " r"(?P<version>\d+[\w\.-]*)\', \'(?P<action>\w+)\'\) "
r"(?P<deptype>(depends on|\(.*\)))") r"(?P<deptype>(depends on|\(.*\)))")
re_origdeps = re.compile(r"(?P<pkgname>[\w\+/-]+) depends on") re_origdeps = re.compile(r"(?P<pkgname>[\w\+/-]+) depends on")
re_installed_package = re.compile(
r"\[(?P<desc>[^\]]*)\] "
r"(?P<pkgdir>[\w\+-]+)/"
r"(?P<pkgname>[\w\+-]+)-"
r"(?P<version>\d+[\w\.-]*)( \["
r"(?P<oldversion>\d+[\w\.-]*)\])?"
)
re_failed = re.compile(r".*depends on.*") re_failed = re.compile(r".*depends on.*")
deps_tree = {} deps_tree = {}
deps_stack = [] deps_stack = []
deps_info = {}
for line in lines: for line in lines:
m = re_deps.match(line) m = re_deps.match(line)
m_orig = re_origdeps.match(line) m_orig = re_origdeps.match(line)
m_installed = re_installed_package.match(line)
if m: if m:
pkgname = m.group("pkgname") pkgname = m.group("pkgname")
pkgdir = m.group("pkgdir") pkgdir = m.group("pkgdir")
@ -283,15 +282,30 @@ def DepsToTree(lines):
deps_stack = deps_stack[0:depth] deps_stack = deps_stack[0:depth]
# Add ourselves to the end of the stack. # Add ourselves to the end of the stack.
deps_stack.append(pkgname) deps_stack.append(pkgname)
elif m_installed:
pkgname = m_installed.group("pkgname")
pkgdir = m_installed.group("pkgdir")
version = m_installed.group("version")
oldversion = m_installed.group("oldversion")
desc = m_installed.group("desc")
uninstall = False
if oldversion and (desc.find("U") != -1 or desc.find("D") != -1):
uninstall = True
fullpkg = "%s/%s-%s" % (pkgdir, pkgname, version)
deps_info[fullpkg] = {"idx": len(deps_info),
"pkgdir": pkgdir,
"pkgname": pkgname,
"oldversion": oldversion,
"uninstall": uninstall}
else: else:
# Is this a package that failed to match uor huge regex? # Is this a package that failed to match our huge regex?
m = re_failed.match(line) m = re_failed.match(line)
if m: if m:
print "FAIL: Couldn't understand line:" print "FAIL: Couldn't understand line:"
print line print line
sys.exit(1) sys.exit(1)
return deps_tree return deps_tree, deps_info
def PrintTree(deps, depth=""): def PrintTree(deps, depth=""):
@ -307,11 +321,12 @@ def PrintTree(deps, depth=""):
PrintTree(deps[entry]["deps"], depth=depth + " ") PrintTree(deps[entry]["deps"], depth=depth + " ")
def GenDependencyGraph(deps_tree): def GenDependencyGraph(deps_tree, deps_info):
"""Generate a doubly linked dependency graph. """Generate a doubly linked dependency graph.
Args: Args:
deps_tree: dependency tree structure. deps_tree: dependency tree structure.
deps_info: more info on the dependencies.
Returns: Returns:
Deps graph in the form of a dict of packages, with each package Deps graph in the form of a dict of packages, with each package
specifying a "needs" list and "provides" list. specifying a "needs" list and "provides" list.
@ -331,48 +346,78 @@ def GenDependencyGraph(deps_tree):
for pkg in packages: for pkg in packages:
action = packages[pkg]["action"] action = packages[pkg]["action"]
this_pkg = deps_map.setdefault( this_pkg = deps_map.setdefault(
pkg, {"needs": {}, "provides": set(), "action": "nomerge"}) pkg, {"needs": set(), "provides": set(), "action": "nomerge"})
if action != "nomerge": if action != "nomerge":
this_pkg["action"] = action this_pkg["action"] = action
this_pkg["deps_info"] = deps_info.get(pkg)
ReverseTree(packages[pkg]["deps"]) ReverseTree(packages[pkg]["deps"])
for dep, dep_item in packages[pkg]["deps"].items(): for dep, dep_item in packages[pkg]["deps"].items():
dep_pkg = deps_map[dep] dep_pkg = deps_map[dep]
dep_type = dep_item["deptype"] dep_type = dep_item["deptype"]
if dep_type == "(runtime_post)": if dep_type != "(runtime_post)":
dep_pkg["needs"][pkg] = dep_type
this_pkg["provides"].add(dep)
else:
dep_pkg["provides"].add(pkg) dep_pkg["provides"].add(pkg)
this_pkg["needs"][dep] = dep_type this_pkg["needs"].add(dep)
def RemoveInstalledPackages():
"""Remove installed packages, propagating dependencies"""
rm_pkgs = set(deps_map.keys()) - set(deps_info.keys())
for pkg in rm_pkgs:
this_pkg = deps_map[pkg]
needs = this_pkg["needs"]
provides = this_pkg["provides"]
for dep in needs:
dep_provides = deps_map[dep]["provides"]
dep_provides.update(provides)
dep_provides.discard(pkg)
dep_provides.discard(dep)
for target in provides:
target_needs = deps_map[target]["needs"]
target_needs.update(needs)
target_needs.discard(pkg)
target_needs.discard(target)
del deps_map[pkg]
def SanitizeDep(basedep, currdep, oldstack, limit): def SanitizeDep(basedep, currdep, oldstack, limit):
"""Remove any circular dependencies between basedep, currdep, then recurse. """Search for circular deps between basedep and currdep, then recurse.
Args: Args:
basedep: original dependency, top of stack. basedep: original dependency, top of stack.
currdep: bottom of our current recursion, bottom of stack. currdep: bottom of our current recursion, bottom of stack.
oldstack: current dependency chain. oldstack: current dependency chain.
limit: how many more levels of recusion to go through, max. limit: how many more levels of recusion to go through, max.
TODO(): Break PDEPEND preferentially, then RDEPEND. Also extract emerge TODO(): Break RDEPEND preferentially.
linear ordering and break cycles on default emerge linear order. Returns:
True iff circular dependencies are found.
""" """
if limit == 0: if limit == 0:
return return
for dep in deps_map[currdep]["needs"]: for dep in deps_map[currdep]["needs"]:
stack = oldstack + [dep] stack = oldstack + [dep]
if basedep in deps_map[dep]["needs"]: if basedep in deps_map[dep]["needs"] or dep == basedep:
if dep != basedep:
stack += [basedep]
print "Remove cyclic dependency from:" print "Remove cyclic dependency from:"
for i in xrange(0, len(stack) - 1): for i in xrange(0, len(stack) - 1):
print " %s (%s)-> %s " % ( print " %s -> %s " % (stack[i], stack[i+1])
stack[i], deps_map[stack[i]]["needs"][stack[i+1]], stack[i+1]) return True
del deps_map[dep]["needs"][basedep] if dep not in oldstack and SanitizeDep(basedep, dep, stack, limit - 1):
deps_map[basedep]["provides"].remove(dep) return True
SanitizeDep(basedep, dep, stack, limit - 1) return
def SanitizeTree(): def SanitizeTree():
"""Remove circular dependencies up to cycle length 8.""" """Remove circular dependencies up to cycle length 32."""
for dep in deps_map: start = time.time()
SanitizeDep(dep, dep, [dep], 8) for basedep in deps_map:
for dep in deps_map[basedep]["needs"].copy():
if deps_info[basedep]["idx"] <= deps_info[dep]["idx"]:
if SanitizeDep(basedep, dep, [basedep, dep], 31):
print "Breaking", basedep, " -> ", dep
deps_map[basedep]["needs"].remove(dep)
deps_map[dep]["provides"].remove(basedep)
seconds = time.time() - start
print "Tree sanitized in %d:%04.1fs" % (seconds / 60, seconds % 60)
def AddSecretDeps(): def AddSecretDeps():
"""Find these tagged packages and add extra dependencies. """Find these tagged packages and add extra dependencies.
@ -390,10 +435,11 @@ def GenDependencyGraph(deps_tree):
needed_pkg = dep needed_pkg = dep
if bad_pkg and needed_pkg: if bad_pkg and needed_pkg:
deps_map[needed_pkg]["provides"].add(bad_pkg) deps_map[needed_pkg]["provides"].add(bad_pkg)
deps_map[bad_pkg]["needs"][needed_pkg] = "(manually forced)" deps_map[bad_pkg]["needs"].add(needed_pkg)
ReverseTree(deps_tree) ReverseTree(deps_tree)
AddSecretDeps() AddSecretDeps()
RemoveInstalledPackages()
SanitizeTree() SanitizeTree()
return deps_map return deps_map
@ -402,8 +448,8 @@ def PrintDepsMap(deps_map):
"""Print dependency graph, for each package list it's prerequisites.""" """Print dependency graph, for each package list it's prerequisites."""
for i in deps_map: for i in deps_map:
print "%s: (%s) needs" % (i, deps_map[i]["action"]) print "%s: (%s) needs" % (i, deps_map[i]["action"])
for j, dep_type in deps_map[i]["needs"].items(): for j in deps_map[i]["needs"]:
print " %s ( %s )" % (j, dep_type) print " %s" % (j)
class EmergeQueue(object): class EmergeQueue(object):
@ -426,9 +472,9 @@ class EmergeQueue(object):
def _Status(self): def _Status(self):
"""Print status.""" """Print status."""
print "Pending %s, Ready %s, Running %s, Failed %s, Total %s" % ( print "Pending %s, Ready %s, Running %s, Retrying %s, Total %s" % (
len(self._deps_map), len(self._emerge_queue), len(self._deps_map), len(self._emerge_queue),
len(self._jobs), len(self._failed), self._total_jobs) len(self._jobs), len(self._retry_queue), self._total_jobs)
def _LaunchOneEmerge(self, target): def _LaunchOneEmerge(self, target):
"""Run emerge --nodeps to do a single package install. """Run emerge --nodeps to do a single package install.
@ -455,14 +501,19 @@ class EmergeQueue(object):
# in the "world" file, which represents explicit intalls. # in the "world" file, which represents explicit intalls.
# "--oneshot" here will prevent it from being tagged in world. # "--oneshot" here will prevent it from being tagged in world.
cmdline = EmergeCommand() + " --nodeps --oneshot =" + target cmdline = EmergeCommand() + " --nodeps --oneshot =" + target
if VERBOSE: deps_info = self._deps_map[target]["deps_info"]
print "running %s" % cmdline if deps_info["uninstall"]:
package = "%(pkgdir)s/%(pkgname)s-%(oldversion)s" % deps_info
cmdline += " && %s -1C =%s" % (EmergeCommand(), package)
print "+ %s" % cmdline
# Store output in a temp file as it is too big for a unix pipe. # Store output in a temp file as it is too big for a unix pipe.
stdout_buffer = tempfile.TemporaryFile() stdout_buffer = tempfile.TemporaryFile()
# Modify the environment to disable locking. # Modify the environment to disable locking.
portage_env = os.environ.copy() portage_env = os.environ.copy()
portage_env["PORTAGE_LOCKS"] = "false" portage_env["PORTAGE_LOCKS"] = "false"
portage_env["UNMERGE_DELAY"] = "0"
# Autoclean rummages around in the portage database and uninstalls # Autoclean rummages around in the portage database and uninstalls
# old packages. Definitely not necessary for build_image. However # old packages. Definitely not necessary for build_image. However
# it may be necessary for incremental build_packages. It may also # it may be necessary for incremental build_packages. It may also
@ -471,7 +522,7 @@ class EmergeQueue(object):
portage_env["AUTOCLEAN"] = "no" portage_env["AUTOCLEAN"] = "no"
# Launch the subprocess. # Launch the subprocess.
emerge_proc = subprocess.Popen( emerge_proc = subprocess.Popen(
shlex.split(cmdline), stdout=stdout_buffer, cmdline, shell=True, stdout=stdout_buffer,
stderr=subprocess.STDOUT, bufsize=64*1024, env=portage_env) stderr=subprocess.STDOUT, bufsize=64*1024, env=portage_env)
return (target, emerge_proc, stdout_buffer) return (target, emerge_proc, stdout_buffer)
@ -479,7 +530,7 @@ class EmergeQueue(object):
def _Finish(self, target): def _Finish(self, target):
"""Mark a target as completed and unblock dependecies.""" """Mark a target as completed and unblock dependecies."""
for dep in self._deps_map[target]["provides"]: for dep in self._deps_map[target]["provides"]:
del self._deps_map[dep]["needs"][target] self._deps_map[dep]["needs"].remove(target)
if not self._deps_map[dep]["needs"]: if not self._deps_map[dep]["needs"]:
if VERBOSE: if VERBOSE:
print "Unblocking %s" % dep print "Unblocking %s" % dep
@ -591,13 +642,13 @@ print " Building package %s on %s (%s)" % (PACKAGE, EMERGE_ARGS, BOARD)
print "Running emerge to generate deps" print "Running emerge to generate deps"
deps_output = GetDepsFromPortage(PACKAGE) deps_output = GetDepsFromPortage(PACKAGE)
print "Processing emerge output" print "Processing emerge output"
dependency_tree = DepsToTree(deps_output) dependency_tree, dependency_info = DepsToTree(deps_output)
if VERBOSE: if VERBOSE:
print "Print tree" print "Print tree"
PrintTree(dependency_tree) PrintTree(dependency_tree)
print "Generate dependency graph." print "Generate dependency graph."
dependency_graph = GenDependencyGraph(dependency_tree) dependency_graph = GenDependencyGraph(dependency_tree, dependency_info)
if VERBOSE: if VERBOSE:
PrintDepsMap(dependency_graph) PrintDepsMap(dependency_graph)