| # discovery.py - protocol changeset discovery functions |
| # |
| # Copyright 2010 Matt Mackall <mpm@selenic.com> |
| # |
| # This software may be used and distributed according to the terms of the |
| # GNU General Public License version 2 or any later version. |
| |
| from node import nullid, short |
| from i18n import _ |
| import util, setdiscovery, treediscovery, phases, obsolete, bookmarks |
| import branchmap |
| |
| def findcommonincoming(repo, remote, heads=None, force=False): |
| """Return a tuple (common, anyincoming, heads) used to identify the common |
| subset of nodes between repo and remote. |
| |
| "common" is a list of (at least) the heads of the common subset. |
| "anyincoming" is testable as a boolean indicating if any nodes are missing |
| locally. If remote does not support getbundle, this actually is a list of |
| roots of the nodes that would be incoming, to be supplied to |
| changegroupsubset. No code except for pull should be relying on this fact |
| any longer. |
| "heads" is either the supplied heads, or else the remote's heads. |
| |
| If you pass heads and they are all known locally, the response lists just |
| these heads in "common" and in "heads". |
| |
| Please use findcommonoutgoing to compute the set of outgoing nodes to give |
| extensions a good hook into outgoing. |
| """ |
| |
| if not remote.capable('getbundle'): |
| return treediscovery.findcommonincoming(repo, remote, heads, force) |
| |
| if heads: |
| allknown = True |
| nm = repo.changelog.nodemap |
| for h in heads: |
| if nm.get(h) is None: |
| allknown = False |
| break |
| if allknown: |
| return (heads, False, heads) |
| |
| res = setdiscovery.findcommonheads(repo.ui, repo, remote, |
| abortwhenunrelated=not force) |
| common, anyinc, srvheads = res |
| return (list(common), anyinc, heads or list(srvheads)) |
| |
| class outgoing(object): |
| '''Represents the set of nodes present in a local repo but not in a |
| (possibly) remote one. |
| |
| Members: |
| |
| missing is a list of all nodes present in local but not in remote. |
| common is a list of all nodes shared between the two repos. |
| excluded is the list of missing changeset that shouldn't be sent remotely. |
| missingheads is the list of heads of missing. |
| commonheads is the list of heads of common. |
| |
| The sets are computed on demand from the heads, unless provided upfront |
| by discovery.''' |
| |
| def __init__(self, revlog, commonheads, missingheads): |
| self.commonheads = commonheads |
| self.missingheads = missingheads |
| self._revlog = revlog |
| self._common = None |
| self._missing = None |
| self.excluded = [] |
| |
| def _computecommonmissing(self): |
| sets = self._revlog.findcommonmissing(self.commonheads, |
| self.missingheads) |
| self._common, self._missing = sets |
| |
| @util.propertycache |
| def common(self): |
| if self._common is None: |
| self._computecommonmissing() |
| return self._common |
| |
| @util.propertycache |
| def missing(self): |
| if self._missing is None: |
| self._computecommonmissing() |
| return self._missing |
| |
| def findcommonoutgoing(repo, other, onlyheads=None, force=False, |
| commoninc=None, portable=False): |
| '''Return an outgoing instance to identify the nodes present in repo but |
| not in other. |
| |
| If onlyheads is given, only nodes ancestral to nodes in onlyheads |
| (inclusive) are included. If you already know the local repo's heads, |
| passing them in onlyheads is faster than letting them be recomputed here. |
| |
| If commoninc is given, it must be the result of a prior call to |
| findcommonincoming(repo, other, force) to avoid recomputing it here. |
| |
| If portable is given, compute more conservative common and missingheads, |
| to make bundles created from the instance more portable.''' |
| # declare an empty outgoing object to be filled later |
| og = outgoing(repo.changelog, None, None) |
| |
| # get common set if not provided |
| if commoninc is None: |
| commoninc = findcommonincoming(repo, other, force=force) |
| og.commonheads, _any, _hds = commoninc |
| |
| # compute outgoing |
| mayexclude = (repo._phasecache.phaseroots[phases.secret] or repo.obsstore) |
| if not mayexclude: |
| og.missingheads = onlyheads or repo.heads() |
| elif onlyheads is None: |
| # use visible heads as it should be cached |
| og.missingheads = repo.filtered("served").heads() |
| og.excluded = [ctx.node() for ctx in repo.set('secret() or extinct()')] |
| else: |
| # compute common, missing and exclude secret stuff |
| sets = repo.changelog.findcommonmissing(og.commonheads, onlyheads) |
| og._common, allmissing = sets |
| og._missing = missing = [] |
| og.excluded = excluded = [] |
| for node in allmissing: |
| ctx = repo[node] |
| if ctx.phase() >= phases.secret or ctx.extinct(): |
| excluded.append(node) |
| else: |
| missing.append(node) |
| if len(missing) == len(allmissing): |
| missingheads = onlyheads |
| else: # update missing heads |
| missingheads = phases.newheads(repo, onlyheads, excluded) |
| og.missingheads = missingheads |
| if portable: |
| # recompute common and missingheads as if -r<rev> had been given for |
| # each head of missing, and --base <rev> for each head of the proper |
| # ancestors of missing |
| og._computecommonmissing() |
| cl = repo.changelog |
| missingrevs = set(cl.rev(n) for n in og._missing) |
| og._common = set(cl.ancestors(missingrevs)) - missingrevs |
| commonheads = set(og.commonheads) |
| og.missingheads = [h for h in og.missingheads if h not in commonheads] |
| |
| return og |
| |
| def _headssummary(repo, remote, outgoing): |
| """compute a summary of branch and heads status before and after push |
| |
| return {'branch': ([remoteheads], [newheads], [unsyncedheads])} mapping |
| |
| - branch: the branch name |
| - remoteheads: the list of remote heads known locally |
| None is the branch is new |
| - newheads: the new remote heads (known locally) with outgoing pushed |
| - unsyncedheads: the list of remote heads unknown locally. |
| """ |
| cl = repo.changelog |
| headssum = {} |
| # A. Create set of branches involved in the push. |
| branches = set(repo[n].branch() for n in outgoing.missing) |
| remotemap = remote.branchmap() |
| newbranches = branches - set(remotemap) |
| branches.difference_update(newbranches) |
| |
| # A. register remote heads |
| remotebranches = set() |
| for branch, heads in remote.branchmap().iteritems(): |
| remotebranches.add(branch) |
| known = [] |
| unsynced = [] |
| for h in heads: |
| if h in cl.nodemap: |
| known.append(h) |
| else: |
| unsynced.append(h) |
| headssum[branch] = (known, list(known), unsynced) |
| # B. add new branch data |
| missingctx = list(repo[n] for n in outgoing.missing) |
| touchedbranches = set() |
| for ctx in missingctx: |
| branch = ctx.branch() |
| touchedbranches.add(branch) |
| if branch not in headssum: |
| headssum[branch] = (None, [], []) |
| |
| # C drop data about untouched branches: |
| for branch in remotebranches - touchedbranches: |
| del headssum[branch] |
| |
| # D. Update newmap with outgoing changes. |
| # This will possibly add new heads and remove existing ones. |
| newmap = branchmap.branchcache((branch, heads[1]) |
| for branch, heads in headssum.iteritems() |
| if heads[0] is not None) |
| newmap.update(repo, (ctx.rev() for ctx in missingctx)) |
| for branch, newheads in newmap.iteritems(): |
| headssum[branch][1][:] = newheads |
| return headssum |
| |
| def _oldheadssummary(repo, remoteheads, outgoing, inc=False): |
| """Compute branchmapsummary for repo without branchmap support""" |
| |
| cl = repo.changelog |
| # 1-4b. old servers: Check for new topological heads. |
| # Construct {old,new}map with branch = None (topological branch). |
| # (code based on update) |
| oldheads = set(h for h in remoteheads if h in cl.nodemap) |
| # all nodes in outgoing.missing are children of either: |
| # - an element of oldheads |
| # - another element of outgoing.missing |
| # - nullrev |
| # This explains why the new head are very simple to compute. |
| r = repo.set('heads(%ln + %ln)', oldheads, outgoing.missing) |
| newheads = list(c.node() for c in r) |
| unsynced = inc and set([None]) or set() |
| return {None: (oldheads, newheads, unsynced)} |
| |
| def checkheads(repo, remote, outgoing, remoteheads, newbranch=False, inc=False): |
| """Check that a push won't add any outgoing head |
| |
| raise Abort error and display ui message as needed. |
| """ |
| # Check for each named branch if we're creating new remote heads. |
| # To be a remote head after push, node must be either: |
| # - unknown locally |
| # - a local outgoing head descended from update |
| # - a remote head that's known locally and not |
| # ancestral to an outgoing head |
| if remoteheads == [nullid]: |
| # remote is empty, nothing to check. |
| return |
| |
| if remote.capable('branchmap'): |
| headssum = _headssummary(repo, remote, outgoing) |
| else: |
| headssum = _oldheadssummary(repo, remoteheads, outgoing, inc) |
| newbranches = [branch for branch, heads in headssum.iteritems() |
| if heads[0] is None] |
| # 1. Check for new branches on the remote. |
| if newbranches and not newbranch: # new branch requires --new-branch |
| branchnames = ', '.join(sorted(newbranches)) |
| raise util.Abort(_("push creates new remote branches: %s!") |
| % branchnames, |
| hint=_("use 'hg push --new-branch' to create" |
| " new remote branches")) |
| |
| # 2 compute newly pushed bookmarks. We |
| # we don't warned about bookmarked heads. |
| localbookmarks = repo._bookmarks |
| remotebookmarks = remote.listkeys('bookmarks') |
| bookmarkedheads = set() |
| for bm in localbookmarks: |
| rnode = remotebookmarks.get(bm) |
| if rnode and rnode in repo: |
| lctx, rctx = repo[bm], repo[rnode] |
| if bookmarks.validdest(repo, rctx, lctx): |
| bookmarkedheads.add(lctx.node()) |
| |
| # 3. Check for new heads. |
| # If there are more heads after the push than before, a suitable |
| # error message, depending on unsynced status, is displayed. |
| error = None |
| unsynced = False |
| allmissing = set(outgoing.missing) |
| allfuturecommon = set(c.node() for c in repo.set('%ld', outgoing.common)) |
| allfuturecommon.update(allmissing) |
| for branch, heads in sorted(headssum.iteritems()): |
| if heads[0] is None: |
| # Maybe we should abort if we push more that one head |
| # for new branches ? |
| continue |
| candidate_newhs = set(heads[1]) |
| # add unsynced data |
| oldhs = set(heads[0]) |
| oldhs.update(heads[2]) |
| candidate_newhs.update(heads[2]) |
| dhs = None |
| discardedheads = set() |
| if repo.obsstore: |
| # remove future heads which are actually obsolete by another |
| # pushed element: |
| # |
| # XXX as above, There are several cases this case does not handle |
| # XXX properly |
| # |
| # (1) if <nh> is public, it won't be affected by obsolete marker |
| # and a new is created |
| # |
| # (2) if the new heads have ancestors which are not obsolete and |
| # not ancestors of any other heads we will have a new head too. |
| # |
| # This two case will be easy to handle for know changeset but much |
| # more tricky for unsynced changes. |
| newhs = set() |
| for nh in candidate_newhs: |
| if nh in repo and repo[nh].phase() <= phases.public: |
| newhs.add(nh) |
| else: |
| for suc in obsolete.allsuccessors(repo.obsstore, [nh]): |
| if suc != nh and suc in allfuturecommon: |
| discardedheads.add(nh) |
| break |
| else: |
| newhs.add(nh) |
| else: |
| newhs = candidate_newhs |
| if [h for h in heads[2] if h not in discardedheads]: |
| unsynced = True |
| if len(newhs) > len(oldhs): |
| # strip updates to existing remote heads from the new heads list |
| dhs = sorted(newhs - bookmarkedheads - oldhs) |
| if dhs: |
| if error is None: |
| if branch not in ('default', None): |
| error = _("push creates new remote head %s " |
| "on branch '%s'!") % (short(dhs[0]), branch) |
| else: |
| error = _("push creates new remote head %s!" |
| ) % short(dhs[0]) |
| if heads[2]: # unsynced |
| hint = _("you should pull and merge or " |
| "use push -f to force") |
| else: |
| hint = _("did you forget to merge? " |
| "use push -f to force") |
| if branch is not None: |
| repo.ui.note(_("new remote heads on branch '%s'\n") % branch) |
| for h in dhs: |
| repo.ui.note(_("new remote head %s\n") % short(h)) |
| if error: |
| raise util.Abort(error, hint=hint) |
| |
| # 6. Check for unsynced changes on involved branches. |
| if unsynced: |
| repo.ui.warn(_("note: unsynced remote changes!\n")) |