35
46
WorkingTreeNotRevision,
49
from bzrlib.graph import Graph
38
50
from bzrlib.merge3 import Merge3
40
from bzrlib.osutils import rename, pathjoin, rmtree
51
from bzrlib.osutils import rename, pathjoin
41
52
from progress import DummyProgress, ProgressPhase
42
from bzrlib.revision import common_ancestor, is_ancestor, NULL_REVISION
53
from bzrlib.revision import (NULL_REVISION, ensure_null)
43
54
from bzrlib.textfile import check_text_lines
44
from bzrlib.trace import mutter, warning, note
45
from bzrlib.transform import (TreeTransform, resolve_conflicts, cook_conflicts,
46
FinalPaths, create_by_entry, unique_add)
47
from bzrlib.versionedfile import WeaveMerge
55
from bzrlib.trace import mutter, warning, note, is_quiet
56
from bzrlib.transform import (TransformPreview, TreeTransform,
57
resolve_conflicts, cook_conflicts,
58
conflict_pass, FinalPaths, create_from_tree,
59
unique_add, ROOT_PARENT)
60
from bzrlib.versionedfile import PlanWeaveMerge
48
61
from bzrlib import ui
50
63
# TODO: Report back as changes are merged in
52
def _get_tree(treespec, local_branch=None):
53
from bzrlib import workingtree
54
location, revno = treespec
56
tree = workingtree.WorkingTree.open_containing(location)[0]
57
return tree.branch, tree
58
branch = Branch.open_containing(location)[0]
60
revision = branch.last_revision()
62
revision = branch.get_rev_id(revno)
64
revision = NULL_REVISION
65
return branch, _get_revid_tree(branch, revision, local_branch)
68
def _get_revid_tree(branch, revision, local_branch):
70
base_tree = branch.bzrdir.open_workingtree()
72
if local_branch is not None:
73
if local_branch.base != branch.base:
74
local_branch.fetch(branch, revision)
75
base_tree = local_branch.repository.revision_tree(revision)
77
base_tree = branch.repository.revision_tree(revision)
81
66
def transform_tree(from_tree, to_tree, interesting_ids=None):
82
merge_inner(from_tree.branch, to_tree, from_tree, ignore_zero=True,
83
interesting_ids=interesting_ids, this_tree=from_tree)
67
from_tree.lock_tree_write()
69
merge_inner(from_tree.branch, to_tree, from_tree, ignore_zero=True,
70
interesting_ids=interesting_ids, this_tree=from_tree)
86
75
class Merger(object):
87
def __init__(self, this_branch, other_tree=None, base_tree=None,
88
this_tree=None, pb=DummyProgress()):
76
def __init__(self, this_branch, other_tree=None, base_tree=None,
77
this_tree=None, pb=None, change_reporter=None,
78
recurse='down', revision_graph=None):
89
79
object.__init__(self)
90
assert this_tree is not None, "this_tree is required"
91
80
self.this_branch = this_branch
92
self.this_basis = this_branch.last_revision()
81
self.this_basis = _mod_revision.ensure_null(
82
this_branch.last_revision())
93
83
self.this_rev_id = None
94
84
self.this_tree = this_tree
95
85
self.this_revision_tree = None
96
86
self.this_basis_tree = None
97
87
self.other_tree = other_tree
88
self.other_branch = None
98
89
self.base_tree = base_tree
99
90
self.ignore_zero = False
100
91
self.backup_files = False
101
92
self.interesting_ids = None
93
self.interesting_files = None
102
94
self.show_base = False
103
95
self.reprocess = False
108
def revision_tree(self, revision_id):
109
return self.this_branch.repository.revision_tree(revision_id)
100
self.recurse = recurse
101
self.change_reporter = change_reporter
102
self._cached_trees = {}
103
self._revision_graph = revision_graph
104
self._base_is_ancestor = None
105
self._base_is_other_ancestor = None
106
self._is_criss_cross = None
107
self._lca_trees = None
109
def cache_trees_with_revision_ids(self, trees):
110
"""Cache any tree in trees if it has a revision_id."""
111
for maybe_tree in trees:
112
if maybe_tree is None:
115
rev_id = maybe_tree.get_revision_id()
116
except AttributeError:
118
self._cached_trees[rev_id] = maybe_tree
121
def revision_graph(self):
122
if self._revision_graph is None:
123
self._revision_graph = self.this_branch.repository.get_graph()
124
return self._revision_graph
126
def _set_base_is_ancestor(self, value):
127
self._base_is_ancestor = value
129
def _get_base_is_ancestor(self):
130
if self._base_is_ancestor is None:
131
self._base_is_ancestor = self.revision_graph.is_ancestor(
132
self.base_rev_id, self.this_basis)
133
return self._base_is_ancestor
135
base_is_ancestor = property(_get_base_is_ancestor, _set_base_is_ancestor)
137
def _set_base_is_other_ancestor(self, value):
138
self._base_is_other_ancestor = value
140
def _get_base_is_other_ancestor(self):
141
if self._base_is_other_ancestor is None:
142
if self.other_basis is None:
144
self._base_is_other_ancestor = self.revision_graph.is_ancestor(
145
self.base_rev_id, self.other_basis)
146
return self._base_is_other_ancestor
148
base_is_other_ancestor = property(_get_base_is_other_ancestor,
149
_set_base_is_other_ancestor)
152
def from_uncommitted(tree, other_tree, pb=None, base_tree=None):
153
"""Return a Merger for uncommitted changes in other_tree.
155
:param tree: The tree to merge into
156
:param other_tree: The tree to get uncommitted changes from
157
:param pb: A progress indicator
158
:param base_tree: The basis to use for the merge. If unspecified,
159
other_tree.basis_tree() will be used.
161
if base_tree is None:
162
base_tree = other_tree.basis_tree()
163
merger = Merger(tree.branch, other_tree, base_tree, tree, pb)
164
merger.base_rev_id = merger.base_tree.get_revision_id()
165
merger.other_rev_id = None
166
merger.other_basis = merger.base_rev_id
170
def from_mergeable(klass, tree, mergeable, pb):
171
"""Return a Merger for a bundle or merge directive.
173
:param tree: The tree to merge changes into
174
:param mergeable: A merge directive or bundle
175
:param pb: A progress indicator
177
mergeable.install_revisions(tree.branch.repository)
178
base_revision_id, other_revision_id, verified =\
179
mergeable.get_merge_request(tree.branch.repository)
180
revision_graph = tree.branch.repository.get_graph()
181
if base_revision_id is not None:
182
if (base_revision_id != _mod_revision.NULL_REVISION and
183
revision_graph.is_ancestor(
184
base_revision_id, tree.branch.last_revision())):
185
base_revision_id = None
187
warning('Performing cherrypick')
188
merger = klass.from_revision_ids(pb, tree, other_revision_id,
189
base_revision_id, revision_graph=
191
return merger, verified
194
def from_revision_ids(pb, tree, other, base=None, other_branch=None,
195
base_branch=None, revision_graph=None,
197
"""Return a Merger for revision-ids.
199
:param pb: A progress indicator
200
:param tree: The tree to merge changes into
201
:param other: The revision-id to use as OTHER
202
:param base: The revision-id to use as BASE. If not specified, will
204
:param other_branch: A branch containing the other revision-id. If
205
not supplied, tree.branch is used.
206
:param base_branch: A branch containing the base revision-id. If
207
not supplied, other_branch or tree.branch will be used.
208
:param revision_graph: If you have a revision_graph precomputed, pass
209
it in, otherwise it will be created for you.
210
:param tree_branch: The branch associated with tree. If not supplied,
211
tree.branch will be used.
213
if tree_branch is None:
214
tree_branch = tree.branch
215
merger = Merger(tree_branch, this_tree=tree, pb=pb,
216
revision_graph=revision_graph)
217
if other_branch is None:
218
other_branch = tree.branch
219
merger.set_other_revision(other, other_branch)
223
if base_branch is None:
224
base_branch = other_branch
225
merger.set_base_revision(base, base_branch)
228
def revision_tree(self, revision_id, branch=None):
229
if revision_id not in self._cached_trees:
231
branch = self.this_branch
233
tree = self.this_tree.revision_tree(revision_id)
234
except errors.NoSuchRevisionInTree:
235
tree = branch.repository.revision_tree(revision_id)
236
self._cached_trees[revision_id] = tree
237
return self._cached_trees[revision_id]
239
def _get_tree(self, treespec, possible_transports=None):
240
from bzrlib import workingtree
241
location, revno = treespec
243
tree = workingtree.WorkingTree.open_containing(location)[0]
244
return tree.branch, tree
245
branch = Branch.open_containing(location, possible_transports)[0]
247
revision_id = branch.last_revision()
249
revision_id = branch.get_rev_id(revno)
250
revision_id = ensure_null(revision_id)
251
return branch, self.revision_tree(revision_id, branch)
111
253
def ensure_revision_trees(self):
112
254
if self.this_revision_tree is None:
113
self.this_basis_tree = self.this_branch.repository.revision_tree(
255
self.this_basis_tree = self.revision_tree(self.this_basis)
115
256
if self.this_basis == self.this_rev_id:
116
257
self.this_revision_tree = self.this_basis_tree
118
259
if self.other_rev_id is None:
119
260
other_basis_tree = self.revision_tree(self.other_basis)
120
changes = other_basis_tree.changes_from(self.other_tree)
121
if changes.has_changed():
261
if other_basis_tree.has_changes(self.other_tree):
122
262
raise WorkingTreeNotRevision(self.this_tree)
123
263
other_rev_id = self.other_basis
124
264
self.other_tree = other_basis_tree
140
279
def check_basis(self, check_clean, require_commits=True):
141
280
if self.this_basis is None and require_commits is True:
142
raise BzrCommandError("This branch has no commits")
281
raise BzrCommandError("This branch has no commits."
282
" (perhaps you would prefer 'bzr pull')")
144
284
self.compare_basis()
145
285
if self.this_basis != self.this_rev_id:
146
raise BzrCommandError("Working tree has uncommitted changes.")
286
raise errors.UncommittedChanges(self.this_tree)
148
288
def compare_basis(self):
149
changes = self.this_tree.changes_from(self.this_tree.basis_tree())
150
if not changes.has_changed():
290
basis_tree = self.revision_tree(self.this_tree.last_revision())
291
except errors.NoSuchRevision:
292
basis_tree = self.this_tree.basis_tree()
293
if not self.this_tree.has_changes(basis_tree):
151
294
self.this_rev_id = self.this_basis
153
296
def set_interesting_files(self, file_list):
155
self._set_interesting_files(file_list)
156
except NotVersionedError, e:
157
raise BzrCommandError("%s is not a source file in any"
160
def _set_interesting_files(self, file_list):
161
"""Set the list of interesting ids from a list of files."""
162
if file_list is None:
163
self.interesting_ids = None
166
interesting_ids = set()
167
for path in file_list:
169
for tree in (self.this_tree, self.base_tree, self.other_tree):
170
file_id = tree.inventory.path2id(path)
171
if file_id is not None:
172
interesting_ids.add(file_id)
175
raise NotVersionedError(path=path)
176
self.interesting_ids = interesting_ids
297
self.interesting_files = file_list
178
299
def set_pending(self):
179
if not self.base_is_ancestor:
181
if self.other_rev_id is None:
183
ancestry = self.this_branch.repository.get_ancestry(self.this_basis)
184
if self.other_rev_id in ancestry:
186
self.this_tree.add_parent_tree((self.other_rev_id, self.other_tree))
188
def set_other(self, other_revision):
300
if not self.base_is_ancestor or not self.base_is_other_ancestor or self.other_rev_id is None:
304
def _add_parent(self):
305
new_parents = self.this_tree.get_parent_ids() + [self.other_rev_id]
306
new_parent_trees = []
307
for revision_id in new_parents:
309
tree = self.revision_tree(revision_id)
310
except errors.NoSuchRevision:
314
new_parent_trees.append((revision_id, tree))
316
self.this_tree.set_parent_trees(new_parent_trees,
317
allow_leftmost_as_ghost=True)
319
for _revision_id, tree in new_parent_trees:
323
def set_other(self, other_revision, possible_transports=None):
189
324
"""Set the revision and tree to merge from.
191
326
This sets the other_tree, other_rev_id, other_basis attributes.
193
328
:param other_revision: The [path, revision] list to merge from.
195
other_branch, self.other_tree = _get_tree(other_revision,
330
self.other_branch, self.other_tree = self._get_tree(other_revision,
197
332
if other_revision[1] == -1:
198
self.other_rev_id = other_branch.last_revision()
199
if self.other_rev_id is None:
200
raise NoCommits(other_branch)
333
self.other_rev_id = _mod_revision.ensure_null(
334
self.other_branch.last_revision())
335
if _mod_revision.is_null(self.other_rev_id):
336
raise NoCommits(self.other_branch)
201
337
self.other_basis = self.other_rev_id
202
338
elif other_revision[1] is not None:
203
self.other_rev_id = other_branch.get_rev_id(other_revision[1])
339
self.other_rev_id = self.other_branch.get_rev_id(other_revision[1])
204
340
self.other_basis = self.other_rev_id
206
342
self.other_rev_id = None
207
self.other_basis = other_branch.last_revision()
343
self.other_basis = self.other_branch.last_revision()
208
344
if self.other_basis is None:
209
raise NoCommits(other_branch)
210
if other_branch.base != self.this_branch.base:
211
self.this_branch.fetch(other_branch, last_revision=self.other_basis)
345
raise NoCommits(self.other_branch)
346
if self.other_rev_id is not None:
347
self._cached_trees[self.other_rev_id] = self.other_tree
348
self._maybe_fetch(self.other_branch,self.this_branch, self.other_basis)
350
def set_other_revision(self, revision_id, other_branch):
351
"""Set 'other' based on a branch and revision id
353
:param revision_id: The revision to use for a tree
354
:param other_branch: The branch containing this tree
356
self.other_rev_id = revision_id
357
self.other_branch = other_branch
358
self._maybe_fetch(other_branch, self.this_branch, self.other_rev_id)
359
self.other_tree = self.revision_tree(revision_id)
360
self.other_basis = revision_id
362
def set_base_revision(self, revision_id, branch):
363
"""Set 'base' based on a branch and revision id
365
:param revision_id: The revision to use for a tree
366
:param branch: The branch containing this tree
368
self.base_rev_id = revision_id
369
self.base_branch = branch
370
self._maybe_fetch(branch, self.this_branch, revision_id)
371
self.base_tree = self.revision_tree(revision_id)
373
def _maybe_fetch(self, source, target, revision_id):
374
if not source.repository.has_same_location(target.repository):
375
target.fetch(source, revision_id)
213
377
def find_base(self):
214
self.set_base([None, None])
378
revisions = [ensure_null(self.this_basis),
379
ensure_null(self.other_basis)]
380
if NULL_REVISION in revisions:
381
self.base_rev_id = NULL_REVISION
382
self.base_tree = self.revision_tree(self.base_rev_id)
383
self._is_criss_cross = False
385
lcas = self.revision_graph.find_lca(revisions[0], revisions[1])
386
self._is_criss_cross = False
388
self.base_rev_id = NULL_REVISION
390
self.base_rev_id = list(lcas)[0]
391
else: # len(lcas) > 1
393
# find_unique_lca can only handle 2 nodes, so we have to
394
# start back at the beginning. It is a shame to traverse
395
# the graph again, but better than re-implementing
397
self.base_rev_id = self.revision_graph.find_unique_lca(
398
revisions[0], revisions[1])
400
self.base_rev_id = self.revision_graph.find_unique_lca(
402
self._is_criss_cross = True
403
if self.base_rev_id == NULL_REVISION:
404
raise UnrelatedBranches()
405
if self._is_criss_cross:
406
warning('Warning: criss-cross merge encountered. See bzr'
407
' help criss-cross.')
408
mutter('Criss-cross lcas: %r' % lcas)
409
interesting_revision_ids = [self.base_rev_id]
410
interesting_revision_ids.extend(lcas)
411
interesting_trees = dict((t.get_revision_id(), t)
412
for t in self.this_branch.repository.revision_trees(
413
interesting_revision_ids))
414
self._cached_trees.update(interesting_trees)
415
self.base_tree = interesting_trees.pop(self.base_rev_id)
416
sorted_lca_keys = self.revision_graph.find_merge_order(
418
self._lca_trees = [interesting_trees[key]
419
for key in sorted_lca_keys]
421
self.base_tree = self.revision_tree(self.base_rev_id)
422
self.base_is_ancestor = True
423
self.base_is_other_ancestor = True
424
mutter('Base revid: %r' % self.base_rev_id)
216
426
def set_base(self, base_revision):
217
427
"""Set the base revision to use for the merge.
264
460
kwargs['show_base'] = self.show_base
265
461
elif self.show_base:
266
462
raise BzrError("Showing base is not supported for this"
267
" merge type. %s" % self.merge_type)
268
merge = self.merge_type(pb=self._pb, **kwargs)
463
" merge type. %s" % self.merge_type)
464
if (not getattr(self.merge_type, 'supports_reverse_cherrypick', True)
465
and not self.base_is_other_ancestor):
466
raise errors.CannotReverseCherrypick()
467
if self.merge_type.supports_cherrypick:
468
kwargs['cherrypick'] = (not self.base_is_ancestor or
469
not self.base_is_other_ancestor)
470
if self._is_criss_cross and getattr(self.merge_type,
471
'supports_lca_trees', False):
472
kwargs['lca_trees'] = self._lca_trees
473
return self.merge_type(pb=self._pb,
474
change_reporter=self.change_reporter,
477
def _do_merge_to(self, merge):
478
if self.other_branch is not None:
479
self.other_branch.update_references(self.this_branch)
481
if self.recurse == 'down':
482
for relpath, file_id in self.this_tree.iter_references():
483
sub_tree = self.this_tree.get_nested_tree(file_id, relpath)
484
other_revision = self.other_tree.get_reference_revision(
486
if other_revision == sub_tree.last_revision():
488
sub_merge = Merger(sub_tree.branch, this_tree=sub_tree)
489
sub_merge.merge_type = self.merge_type
490
other_branch = self.other_branch.reference_parent(file_id, relpath)
491
sub_merge.set_other_revision(other_revision, other_branch)
492
base_revision = self.base_tree.get_reference_revision(file_id)
493
sub_merge.base_tree = \
494
sub_tree.branch.repository.revision_tree(base_revision)
495
sub_merge.base_rev_id = base_revision
499
self.this_tree.lock_tree_write()
501
if self.base_tree is not None:
502
self.base_tree.lock_read()
504
if self.other_tree is not None:
505
self.other_tree.lock_read()
507
merge = self.make_merger()
508
self._do_merge_to(merge)
510
if self.other_tree is not None:
511
self.other_tree.unlock()
513
if self.base_tree is not None:
514
self.base_tree.unlock()
516
self.this_tree.unlock()
269
517
if len(merge.cooked_conflicts) == 0:
270
if not self.ignore_zero:
518
if not self.ignore_zero and not is_quiet():
271
519
note("All changes applied successfully.")
273
521
note("%d conflicts encountered." % len(merge.cooked_conflicts))
275
523
return len(merge.cooked_conflicts)
277
def regen_inventory(self, new_entries):
278
old_entries = self.this_tree.read_working_inventory()
282
for path, file_id in new_entries:
285
new_entries_map[file_id] = path
287
def id2path(file_id):
288
path = new_entries_map.get(file_id)
291
entry = old_entries[file_id]
292
if entry.parent_id is None:
294
return pathjoin(id2path(entry.parent_id), entry.name)
296
for file_id in old_entries:
297
entry = old_entries[file_id]
298
path = id2path(file_id)
299
if file_id in self.base_tree.inventory:
300
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
302
executable = getattr(entry, 'executable', False)
303
new_inventory[file_id] = (path, file_id, entry.parent_id,
304
entry.kind, executable)
306
by_path[path] = file_id
311
for path, file_id in new_entries:
313
del new_inventory[file_id]
316
new_path_list.append((path, file_id))
317
if file_id not in old_entries:
319
# Ensure no file is added before its parent
321
for path, file_id in new_path_list:
325
parent = by_path[os.path.dirname(path)]
326
abspath = pathjoin(self.this_tree.basedir, path)
327
kind = bzrlib.osutils.file_kind(abspath)
328
if file_id in self.base_tree.inventory:
329
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
332
new_inventory[file_id] = (path, file_id, parent, kind, executable)
333
by_path[path] = file_id
335
# Get a list in insertion order
336
new_inventory_list = new_inventory.values()
337
mutter ("""Inventory regeneration:
338
old length: %i insertions: %i deletions: %i new_length: %i"""\
339
% (len(old_entries), insertions, deletions,
340
len(new_inventory_list)))
341
assert len(new_inventory_list) == len(old_entries) + insertions\
343
new_inventory_list.sort()
344
return new_inventory_list
526
class _InventoryNoneEntry(object):
527
"""This represents an inventory entry which *isn't there*.
529
It simplifies the merging logic if we always have an InventoryEntry, even
530
if it isn't actually present
537
symlink_target = None
540
_none_entry = _InventoryNoneEntry()
347
543
class Merge3Merger(object):
363
594
self.cooked_conflicts = []
364
595
self.reprocess = reprocess
365
596
self.show_base = show_base
597
self._lca_trees = lca_trees
598
# Uncommenting this will change the default algorithm to always use
599
# _entries_lca. This can be useful for running the test suite and
600
# making sure we haven't missed any corner cases.
601
# if lca_trees is None:
602
# self._lca_trees = [self.base_tree]
605
self.change_reporter = change_reporter
606
self.cherrypick = cherrypick
368
607
if self.pp is None:
369
608
self.pp = ProgressPhase("Merge phase", 3, self.pb)
371
if interesting_ids is not None:
372
all_ids = interesting_ids
613
self.this_tree.lock_tree_write()
614
self.base_tree.lock_read()
615
self.other_tree.lock_read()
617
self.tt = TreeTransform(self.this_tree, self.pb)
620
self._compute_transform()
622
results = self.tt.apply(no_conflicts=True)
623
self.write_modified(results)
625
self.this_tree.add_conflicts(self.cooked_conflicts)
626
except UnsupportedOperation:
631
self.other_tree.unlock()
632
self.base_tree.unlock()
633
self.this_tree.unlock()
636
def make_preview_transform(self):
637
self.base_tree.lock_read()
638
self.other_tree.lock_read()
639
self.tt = TransformPreview(self.this_tree)
642
self._compute_transform()
645
self.other_tree.unlock()
646
self.base_tree.unlock()
650
def _compute_transform(self):
651
if self._lca_trees is None:
652
entries = self._entries3()
653
resolver = self._three_way
374
all_ids = set(base_tree)
375
all_ids.update(other_tree)
376
working_tree.lock_write()
377
self.tt = TreeTransform(working_tree, self.pb)
655
entries = self._entries_lca()
656
resolver = self._lca_multi_way
657
child_pb = ui.ui_factory.nested_progress_bar()
380
child_pb = ui.ui_factory.nested_progress_bar()
382
for num, file_id in enumerate(all_ids):
383
child_pb.update('Preparing file merge', num, len(all_ids))
384
self.merge_names(file_id)
659
for num, (file_id, changed, parents3, names3,
660
executable3) in enumerate(entries):
661
child_pb.update('Preparing file merge', num, len(entries))
662
self._merge_names(file_id, parents3, names3, resolver=resolver)
385
664
file_status = self.merge_contents(file_id)
386
self.merge_executable(file_id, file_status)
391
child_pb = ui.ui_factory.nested_progress_bar()
393
fs_conflicts = resolve_conflicts(self.tt, child_pb)
396
self.cook_conflicts(fs_conflicts)
397
for conflict in self.cooked_conflicts:
400
results = self.tt.apply()
401
self.write_modified(results)
403
working_tree.add_conflicts(self.cooked_conflicts)
404
except UnsupportedOperation:
408
working_tree.unlock()
666
file_status = 'unmodified'
667
self._merge_executable(file_id,
668
executable3, file_status, resolver=resolver)
673
child_pb = ui.ui_factory.nested_progress_bar()
675
fs_conflicts = resolve_conflicts(self.tt, child_pb,
676
lambda t, c: conflict_pass(t, c, self.other_tree))
679
if self.change_reporter is not None:
680
from bzrlib import delta
681
delta.report_changes(
682
self.tt.iter_changes(), self.change_reporter)
683
self.cook_conflicts(fs_conflicts)
684
for conflict in self.cooked_conflicts:
688
"""Gather data about files modified between three trees.
690
Return a list of tuples of file_id, changed, parents3, names3,
691
executable3. changed is a boolean indicating whether the file contents
692
or kind were changed. parents3 is a tuple of parent ids for base,
693
other and this. names3 is a tuple of names for base, other and this.
694
executable3 is a tuple of execute-bit values for base, other and this.
697
iterator = self.other_tree.iter_changes(self.base_tree,
698
include_unchanged=True, specific_files=self.interesting_files,
699
extra_trees=[self.this_tree])
700
this_entries = dict((e.file_id, e) for p, e in
701
self.this_tree.iter_entries_by_dir(
702
self.interesting_ids))
703
for (file_id, paths, changed, versioned, parents, names, kind,
704
executable) in iterator:
705
if (self.interesting_ids is not None and
706
file_id not in self.interesting_ids):
708
entry = this_entries.get(file_id)
709
if entry is not None:
710
this_name = entry.name
711
this_parent = entry.parent_id
712
this_executable = entry.executable
716
this_executable = None
717
parents3 = parents + (this_parent,)
718
names3 = names + (this_name,)
719
executable3 = executable + (this_executable,)
720
result.append((file_id, changed, parents3, names3, executable3))
723
def _entries_lca(self):
724
"""Gather data about files modified between multiple trees.
726
This compares OTHER versus all LCA trees, and for interesting entries,
727
it then compares with THIS and BASE.
729
For the multi-valued entries, the format will be (BASE, [lca1, lca2])
730
:return: [(file_id, changed, parents, names, executable)]
731
file_id Simple file_id of the entry
732
changed Boolean, True if the kind or contents changed
734
parents ((base, [parent_id, in, lcas]), parent_id_other,
736
names ((base, [name, in, lcas]), name_in_other, name_in_this)
737
executable ((base, [exec, in, lcas]), exec_in_other, exec_in_this)
739
if self.interesting_files is not None:
740
lookup_trees = [self.this_tree, self.base_tree]
741
lookup_trees.extend(self._lca_trees)
742
# I think we should include the lca trees as well
743
interesting_ids = self.other_tree.paths2ids(self.interesting_files,
746
interesting_ids = self.interesting_ids
748
walker = _mod_tree.MultiWalker(self.other_tree, self._lca_trees)
750
base_inventory = self.base_tree.inventory
751
this_inventory = self.this_tree.inventory
752
for path, file_id, other_ie, lca_values in walker.iter_all():
753
# Is this modified at all from any of the other trees?
755
other_ie = _none_entry
756
if interesting_ids is not None and file_id not in interesting_ids:
759
# If other_revision is found in any of the lcas, that means this
760
# node is uninteresting. This is because when merging, if there are
761
# multiple heads(), we have to create a new node. So if we didn't,
762
# we know that the ancestry is linear, and that OTHER did not
764
# See doc/developers/lca_merge_resolution.txt for details
765
other_revision = other_ie.revision
766
if other_revision is not None:
767
# We can't use this shortcut when other_revision is None,
768
# because it may be None because things are WorkingTrees, and
769
# not because it is *actually* None.
770
is_unmodified = False
771
for lca_path, ie in lca_values:
772
if ie is not None and ie.revision == other_revision:
779
for lca_path, lca_ie in lca_values:
781
lca_entries.append(_none_entry)
783
lca_entries.append(lca_ie)
785
if file_id in base_inventory:
786
base_ie = base_inventory[file_id]
788
base_ie = _none_entry
790
if file_id in this_inventory:
791
this_ie = this_inventory[file_id]
793
this_ie = _none_entry
799
for lca_ie in lca_entries:
800
lca_kinds.append(lca_ie.kind)
801
lca_parent_ids.append(lca_ie.parent_id)
802
lca_names.append(lca_ie.name)
803
lca_executable.append(lca_ie.executable)
805
kind_winner = self._lca_multi_way(
806
(base_ie.kind, lca_kinds),
807
other_ie.kind, this_ie.kind)
808
parent_id_winner = self._lca_multi_way(
809
(base_ie.parent_id, lca_parent_ids),
810
other_ie.parent_id, this_ie.parent_id)
811
name_winner = self._lca_multi_way(
812
(base_ie.name, lca_names),
813
other_ie.name, this_ie.name)
815
content_changed = True
816
if kind_winner == 'this':
817
# No kind change in OTHER, see if there are *any* changes
818
if other_ie.kind == 'directory':
819
if parent_id_winner == 'this' and name_winner == 'this':
820
# No change for this directory in OTHER, skip
822
content_changed = False
823
elif other_ie.kind is None or other_ie.kind == 'file':
824
def get_sha1(ie, tree):
825
if ie.kind != 'file':
827
return tree.get_file_sha1(file_id)
828
base_sha1 = get_sha1(base_ie, self.base_tree)
829
lca_sha1s = [get_sha1(ie, tree) for ie, tree
830
in zip(lca_entries, self._lca_trees)]
831
this_sha1 = get_sha1(this_ie, self.this_tree)
832
other_sha1 = get_sha1(other_ie, self.other_tree)
833
sha1_winner = self._lca_multi_way(
834
(base_sha1, lca_sha1s), other_sha1, this_sha1,
835
allow_overriding_lca=False)
836
exec_winner = self._lca_multi_way(
837
(base_ie.executable, lca_executable),
838
other_ie.executable, this_ie.executable)
839
if (parent_id_winner == 'this' and name_winner == 'this'
840
and sha1_winner == 'this' and exec_winner == 'this'):
841
# No kind, parent, name, exec, or content change for
842
# OTHER, so this node is not considered interesting
844
if sha1_winner == 'this':
845
content_changed = False
846
elif other_ie.kind == 'symlink':
847
def get_target(ie, tree):
848
if ie.kind != 'symlink':
850
return tree.get_symlink_target(file_id)
851
base_target = get_target(base_ie, self.base_tree)
852
lca_targets = [get_target(ie, tree) for ie, tree
853
in zip(lca_entries, self._lca_trees)]
854
this_target = get_target(this_ie, self.this_tree)
855
other_target = get_target(other_ie, self.other_tree)
856
target_winner = self._lca_multi_way(
857
(base_target, lca_targets),
858
other_target, this_target)
859
if (parent_id_winner == 'this' and name_winner == 'this'
860
and target_winner == 'this'):
861
# No kind, parent, name, or symlink target change
864
if target_winner == 'this':
865
content_changed = False
866
elif other_ie.kind == 'tree-reference':
867
# The 'changed' information seems to be handled at a higher
868
# level. At least, _entries3 returns False for content
869
# changed, even when at a new revision_id.
870
content_changed = False
871
if (parent_id_winner == 'this' and name_winner == 'this'):
872
# Nothing interesting
875
raise AssertionError('unhandled kind: %s' % other_ie.kind)
876
# XXX: We need to handle kind == 'symlink'
878
# If we have gotten this far, that means something has changed
879
result.append((file_id, content_changed,
880
((base_ie.parent_id, lca_parent_ids),
881
other_ie.parent_id, this_ie.parent_id),
882
((base_ie.name, lca_names),
883
other_ie.name, this_ie.name),
884
((base_ie.executable, lca_executable),
885
other_ie.executable, this_ie.executable)
892
self.tt.final_kind(self.tt.root)
894
self.tt.cancel_deletion(self.tt.root)
895
if self.tt.final_file_id(self.tt.root) is None:
896
self.tt.version_file(self.tt.tree_file_id(self.tt.root),
898
other_root_file_id = self.other_tree.get_root_id()
899
if other_root_file_id is None:
901
other_root = self.tt.trans_id_file_id(other_root_file_id)
902
if other_root == self.tt.root:
905
self.tt.final_kind(other_root)
908
if self.other_tree.inventory.root.file_id in self.this_tree.inventory:
909
# the other tree's root is a non-root in the current tree
911
self.reparent_children(self.other_tree.inventory.root, self.tt.root)
912
self.tt.cancel_creation(other_root)
913
self.tt.cancel_versioning(other_root)
915
def reparent_children(self, ie, target):
916
for thing, child in ie.children.iteritems():
917
trans_id = self.tt.trans_id_file_id(child.file_id)
918
self.tt.adjust_path(self.tt.final_name(trans_id), target, trans_id)
411
920
def write_modified(self, results):
412
921
modified_hashes = {}
886
1490
name = self.tt.final_name(trans_id)
887
1491
parent_id = self.tt.final_parent(trans_id)
888
1492
self._dump_conflicts(name, parent_id, file_id)
889
self._raw_conflicts.append(('text conflict', trans_id))
1493
self._raw_conflicts.append(('text conflict', trans_id))
1495
osutils.rmtree(temp_dir)
894
1498
def merge_inner(this_branch, other_tree, base_tree, ignore_zero=False,
896
merge_type=Merge3Merger,
897
interesting_ids=None,
1500
merge_type=Merge3Merger,
1501
interesting_ids=None,
900
1504
other_rev_id=None,
901
1505
interesting_files=None,
904
"""Primary interface for merging.
1508
change_reporter=None):
1509
"""Primary interface for merging.
906
typical use is probably
1511
typical use is probably
907
1512
'merge_inner(branch, branch.get_revision_tree(other_revision),
908
1513
branch.get_revision_tree(base_revision))'
910
1515
if this_tree is None:
911
warnings.warn("bzrlib.merge.merge_inner requires a this_tree parameter as of "
912
"bzrlib version 0.8.",
915
this_tree = this_branch.bzrdir.open_workingtree()
916
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1516
raise BzrError("bzrlib.merge.merge_inner requires a this_tree "
1517
"parameter as of bzrlib version 0.8.")
1518
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1519
pb=pb, change_reporter=change_reporter)
918
1520
merger.backup_files = backup_files
919
1521
merger.merge_type = merge_type
920
1522
merger.interesting_ids = interesting_ids
921
1523
merger.ignore_zero = ignore_zero
922
1524
if interesting_files:
923
assert not interesting_ids, ('Only supply interesting_ids'
924
' or interesting_files')
925
merger._set_interesting_files(interesting_files)
1526
raise ValueError('Only supply interesting_ids'
1527
' or interesting_files')
1528
merger.interesting_files = interesting_files
926
1529
merger.show_base = show_base
927
1530
merger.reprocess = reprocess
928
1531
merger.other_rev_id = other_rev_id
929
1532
merger.other_basis = other_rev_id
1533
get_revision_id = getattr(base_tree, 'get_revision_id', None)
1534
if get_revision_id is None:
1535
get_revision_id = base_tree.last_revision
1536
merger.cache_trees_with_revision_ids([other_tree, base_tree, this_tree])
1537
merger.set_base_revision(get_revision_id(), this_branch)
930
1538
return merger.do_merge()
933
merge_types = { "merge3": (Merge3Merger, "Native diff3-style merge"),
934
"diff3": (Diff3Merger, "Merge using external diff3"),
935
'weave': (WeaveMerger, "Weave-based merge")
939
def merge_type_help():
940
templ = '%s%%7s: %%s' % (' '*12)
941
lines = [templ % (f[0], f[1][1]) for f in merge_types.iteritems()]
942
return '\n'.join(lines)
1540
def get_merge_type_registry():
1541
"""Merge type registry is in bzrlib.option to avoid circular imports.
1543
This method provides a sanctioned way to retrieve it.
1545
from bzrlib import option
1546
return option._merge_type_registry
1549
def _plan_annotate_merge(annotated_a, annotated_b, ancestors_a, ancestors_b):
1550
def status_a(revision, text):
1551
if revision in ancestors_b:
1552
return 'killed-b', text
1554
return 'new-a', text
1556
def status_b(revision, text):
1557
if revision in ancestors_a:
1558
return 'killed-a', text
1560
return 'new-b', text
1562
plain_a = [t for (a, t) in annotated_a]
1563
plain_b = [t for (a, t) in annotated_b]
1564
matcher = patiencediff.PatienceSequenceMatcher(None, plain_a, plain_b)
1565
blocks = matcher.get_matching_blocks()
1568
for ai, bi, l in blocks:
1569
# process all mismatched sections
1570
# (last mismatched section is handled because blocks always
1571
# includes a 0-length last block)
1572
for revision, text in annotated_a[a_cur:ai]:
1573
yield status_a(revision, text)
1574
for revision, text in annotated_b[b_cur:bi]:
1575
yield status_b(revision, text)
1576
# and now the matched section
1579
for text_a in plain_a[ai:a_cur]:
1580
yield "unchanged", text_a
1583
class _PlanMergeBase(object):
1585
def __init__(self, a_rev, b_rev, vf, key_prefix):
1588
:param a_rev: Revision-id of one revision to merge
1589
:param b_rev: Revision-id of the other revision to merge
1590
:param vf: A VersionedFiles containing both revisions
1591
:param key_prefix: A prefix for accessing keys in vf, typically
1597
self._last_lines = None
1598
self._last_lines_revision_id = None
1599
self._cached_matching_blocks = {}
1600
self._key_prefix = key_prefix
1601
self._precache_tip_lines()
1603
def _precache_tip_lines(self):
1604
lines = self.get_lines([self.a_rev, self.b_rev])
1605
self.lines_a = lines[self.a_rev]
1606
self.lines_b = lines[self.b_rev]
1608
def get_lines(self, revisions):
1609
"""Get lines for revisions from the backing VersionedFiles.
1611
:raises RevisionNotPresent: on absent texts.
1613
keys = [(self._key_prefix + (rev,)) for rev in revisions]
1615
for record in self.vf.get_record_stream(keys, 'unordered', True):
1616
if record.storage_kind == 'absent':
1617
raise errors.RevisionNotPresent(record.key, self.vf)
1618
result[record.key[-1]] = osutils.chunks_to_lines(
1619
record.get_bytes_as('chunked'))
1622
def plan_merge(self):
1623
"""Generate a 'plan' for merging the two revisions.
1625
This involves comparing their texts and determining the cause of
1626
differences. If text A has a line and text B does not, then either the
1627
line was added to text A, or it was deleted from B. Once the causes
1628
are combined, they are written out in the format described in
1629
VersionedFile.plan_merge
1631
blocks = self._get_matching_blocks(self.a_rev, self.b_rev)
1632
unique_a, unique_b = self._unique_lines(blocks)
1633
new_a, killed_b = self._determine_status(self.a_rev, unique_a)
1634
new_b, killed_a = self._determine_status(self.b_rev, unique_b)
1635
return self._iter_plan(blocks, new_a, killed_b, new_b, killed_a)
1637
def _iter_plan(self, blocks, new_a, killed_b, new_b, killed_a):
1640
for i, j, n in blocks:
1641
for a_index in range(last_i, i):
1642
if a_index in new_a:
1643
if a_index in killed_b:
1644
yield 'conflicted-a', self.lines_a[a_index]
1646
yield 'new-a', self.lines_a[a_index]
1648
yield 'killed-b', self.lines_a[a_index]
1649
for b_index in range(last_j, j):
1650
if b_index in new_b:
1651
if b_index in killed_a:
1652
yield 'conflicted-b', self.lines_b[b_index]
1654
yield 'new-b', self.lines_b[b_index]
1656
yield 'killed-a', self.lines_b[b_index]
1657
# handle common lines
1658
for a_index in range(i, i+n):
1659
yield 'unchanged', self.lines_a[a_index]
1663
def _get_matching_blocks(self, left_revision, right_revision):
1664
"""Return a description of which sections of two revisions match.
1666
See SequenceMatcher.get_matching_blocks
1668
cached = self._cached_matching_blocks.get((left_revision,
1670
if cached is not None:
1672
if self._last_lines_revision_id == left_revision:
1673
left_lines = self._last_lines
1674
right_lines = self.get_lines([right_revision])[right_revision]
1676
lines = self.get_lines([left_revision, right_revision])
1677
left_lines = lines[left_revision]
1678
right_lines = lines[right_revision]
1679
self._last_lines = right_lines
1680
self._last_lines_revision_id = right_revision
1681
matcher = patiencediff.PatienceSequenceMatcher(None, left_lines,
1683
return matcher.get_matching_blocks()
1685
def _unique_lines(self, matching_blocks):
1686
"""Analyse matching_blocks to determine which lines are unique
1688
:return: a tuple of (unique_left, unique_right), where the values are
1689
sets of line numbers of unique lines.
1695
for i, j, n in matching_blocks:
1696
unique_left.extend(range(last_i, i))
1697
unique_right.extend(range(last_j, j))
1700
return unique_left, unique_right
1703
def _subtract_plans(old_plan, new_plan):
1704
"""Remove changes from new_plan that came from old_plan.
1706
It is assumed that the difference between the old_plan and new_plan
1707
is their choice of 'b' text.
1709
All lines from new_plan that differ from old_plan are emitted
1710
verbatim. All lines from new_plan that match old_plan but are
1711
not about the 'b' revision are emitted verbatim.
1713
Lines that match and are about the 'b' revision are the lines we
1714
don't want, so we convert 'killed-b' -> 'unchanged', and 'new-b'
1715
is skipped entirely.
1717
matcher = patiencediff.PatienceSequenceMatcher(None, old_plan,
1720
for i, j, n in matcher.get_matching_blocks():
1721
for jj in range(last_j, j):
1723
for jj in range(j, j+n):
1724
plan_line = new_plan[jj]
1725
if plan_line[0] == 'new-b':
1727
elif plan_line[0] == 'killed-b':
1728
yield 'unchanged', plan_line[1]
1734
class _PlanMerge(_PlanMergeBase):
1735
"""Plan an annotate merge using on-the-fly annotation"""
1737
def __init__(self, a_rev, b_rev, vf, key_prefix):
1738
super(_PlanMerge, self).__init__(a_rev, b_rev, vf, key_prefix)
1739
self.a_key = self._key_prefix + (self.a_rev,)
1740
self.b_key = self._key_prefix + (self.b_rev,)
1741
self.graph = Graph(self.vf)
1742
heads = self.graph.heads((self.a_key, self.b_key))
1744
# one side dominates, so we can just return its values, yay for
1746
# Ideally we would know that before we get this far
1747
self._head_key = heads.pop()
1748
if self._head_key == self.a_key:
1752
mutter('found dominating revision for %s\n%s > %s', self.vf,
1753
self._head_key[-1], other)
1756
self._head_key = None
1759
def _precache_tip_lines(self):
1760
# Turn this into a no-op, because we will do this later
1763
def _find_recursive_lcas(self):
1764
"""Find all the ancestors back to a unique lca"""
1765
cur_ancestors = (self.a_key, self.b_key)
1766
# graph.find_lca(uncommon, keys) now returns plain NULL_REVISION,
1767
# rather than a key tuple. We will just map that directly to no common
1771
next_lcas = self.graph.find_lca(*cur_ancestors)
1772
# Map a plain NULL_REVISION to a simple no-ancestors
1773
if next_lcas == set([NULL_REVISION]):
1775
# Order the lca's based on when they were merged into the tip
1776
# While the actual merge portion of weave merge uses a set() of
1777
# active revisions, the order of insertion *does* effect the
1778
# implicit ordering of the texts.
1779
for rev_key in cur_ancestors:
1780
ordered_parents = tuple(self.graph.find_merge_order(rev_key,
1782
parent_map[rev_key] = ordered_parents
1783
if len(next_lcas) == 0:
1785
elif len(next_lcas) == 1:
1786
parent_map[list(next_lcas)[0]] = ()
1788
elif len(next_lcas) > 2:
1789
# More than 2 lca's, fall back to grabbing all nodes between
1790
# this and the unique lca.
1791
mutter('More than 2 LCAs, falling back to all nodes for:'
1792
' %s, %s\n=> %s', self.a_key, self.b_key, cur_ancestors)
1793
cur_lcas = next_lcas
1794
while len(cur_lcas) > 1:
1795
cur_lcas = self.graph.find_lca(*cur_lcas)
1796
if len(cur_lcas) == 0:
1797
# No common base to find, use the full ancestry
1800
unique_lca = list(cur_lcas)[0]
1801
if unique_lca == NULL_REVISION:
1802
# find_lca will return a plain 'NULL_REVISION' rather
1803
# than a key tuple when there is no common ancestor, we
1804
# prefer to just use None, because it doesn't confuse
1805
# _get_interesting_texts()
1807
parent_map.update(self._find_unique_parents(next_lcas,
1810
cur_ancestors = next_lcas
1813
def _find_unique_parents(self, tip_keys, base_key):
1814
"""Find ancestors of tip that aren't ancestors of base.
1816
:param tip_keys: Nodes that are interesting
1817
:param base_key: Cull all ancestors of this node
1818
:return: The parent map for all revisions between tip_keys and
1819
base_key. base_key will be included. References to nodes outside of
1820
the ancestor set will also be removed.
1822
# TODO: this would be simpler if find_unique_ancestors took a list
1823
# instead of a single tip, internally it supports it, but it
1824
# isn't a "backwards compatible" api change.
1825
if base_key is None:
1826
parent_map = dict(self.graph.iter_ancestry(tip_keys))
1827
# We remove NULL_REVISION because it isn't a proper tuple key, and
1828
# thus confuses things like _get_interesting_texts, and our logic
1829
# to add the texts into the memory weave.
1830
if NULL_REVISION in parent_map:
1831
parent_map.pop(NULL_REVISION)
1834
for tip in tip_keys:
1836
self.graph.find_unique_ancestors(tip, [base_key]))
1837
parent_map = self.graph.get_parent_map(interesting)
1838
parent_map[base_key] = ()
1839
culled_parent_map, child_map, tails = self._remove_external_references(
1841
# Remove all the tails but base_key
1842
if base_key is not None:
1843
tails.remove(base_key)
1844
self._prune_tails(culled_parent_map, child_map, tails)
1845
# Now remove all the uninteresting 'linear' regions
1846
simple_map = _mod_graph.collapse_linear_regions(culled_parent_map)
1850
def _remove_external_references(parent_map):
1851
"""Remove references that go outside of the parent map.
1853
:param parent_map: Something returned from Graph.get_parent_map(keys)
1854
:return: (filtered_parent_map, child_map, tails)
1855
filtered_parent_map is parent_map without external references
1856
child_map is the {parent_key: [child_keys]} mapping
1857
tails is a list of nodes that do not have any parents in the map
1859
# TODO: The basic effect of this function seems more generic than
1860
# _PlanMerge. But the specific details of building a child_map,
1861
# and computing tails seems very specific to _PlanMerge.
1862
# Still, should this be in Graph land?
1863
filtered_parent_map = {}
1866
for key, parent_keys in parent_map.iteritems():
1867
culled_parent_keys = [p for p in parent_keys if p in parent_map]
1868
if not culled_parent_keys:
1870
for parent_key in culled_parent_keys:
1871
child_map.setdefault(parent_key, []).append(key)
1872
# TODO: Do we want to do this, it adds overhead for every node,
1873
# just to say that the node has no children
1874
child_map.setdefault(key, [])
1875
filtered_parent_map[key] = culled_parent_keys
1876
return filtered_parent_map, child_map, tails
1879
def _prune_tails(parent_map, child_map, tails_to_remove):
1880
"""Remove tails from the parent map.
1882
This will remove the supplied revisions until no more children have 0
1885
:param parent_map: A dict of {child: [parents]}, this dictionary will
1886
be modified in place.
1887
:param tails_to_remove: A list of tips that should be removed,
1888
this list will be consumed
1889
:param child_map: The reverse dict of parent_map ({parent: [children]})
1890
this dict will be modified
1891
:return: None, parent_map will be modified in place.
1893
while tails_to_remove:
1894
next = tails_to_remove.pop()
1895
parent_map.pop(next)
1896
children = child_map.pop(next)
1897
for child in children:
1898
child_parents = parent_map[child]
1899
child_parents.remove(next)
1900
if len(child_parents) == 0:
1901
tails_to_remove.append(child)
1903
def _get_interesting_texts(self, parent_map):
1904
"""Return a dict of texts we are interested in.
1906
Note that the input is in key tuples, but the output is in plain
1909
:param parent_map: The output from _find_recursive_lcas
1910
:return: A dict of {'revision_id':lines} as returned by
1911
_PlanMergeBase.get_lines()
1913
all_revision_keys = set(parent_map)
1914
all_revision_keys.add(self.a_key)
1915
all_revision_keys.add(self.b_key)
1917
# Everything else is in 'keys' but get_lines is in 'revision_ids'
1918
all_texts = self.get_lines([k[-1] for k in all_revision_keys])
1921
def _build_weave(self):
1922
from bzrlib import weave
1923
self._weave = weave.Weave(weave_name='in_memory_weave',
1924
allow_reserved=True)
1925
parent_map = self._find_recursive_lcas()
1927
all_texts = self._get_interesting_texts(parent_map)
1929
# Note: Unfortunately, the order given by topo_sort will effect the
1930
# ordering resolution in the output. Specifically, if you add A then B,
1931
# then in the output text A lines will show up before B lines. And, of
1932
# course, topo_sort doesn't guarantee any real ordering.
1933
# So we use merge_sort, and add a fake node on the tip.
1934
# This ensures that left-hand parents will always be inserted into the
1935
# weave before right-hand parents.
1936
tip_key = self._key_prefix + (_mod_revision.CURRENT_REVISION,)
1937
parent_map[tip_key] = (self.a_key, self.b_key)
1939
for seq_num, key, depth, eom in reversed(tsort.merge_sort(parent_map,
1943
# for key in tsort.topo_sort(parent_map):
1944
parent_keys = parent_map[key]
1945
revision_id = key[-1]
1946
parent_ids = [k[-1] for k in parent_keys]
1947
self._weave.add_lines(revision_id, parent_ids,
1948
all_texts[revision_id])
1950
def plan_merge(self):
1951
"""Generate a 'plan' for merging the two revisions.
1953
This involves comparing their texts and determining the cause of
1954
differences. If text A has a line and text B does not, then either the
1955
line was added to text A, or it was deleted from B. Once the causes
1956
are combined, they are written out in the format described in
1957
VersionedFile.plan_merge
1959
if self._head_key is not None: # There was a single head
1960
if self._head_key == self.a_key:
1963
if self._head_key != self.b_key:
1964
raise AssertionError('There was an invalid head: %s != %s'
1965
% (self.b_key, self._head_key))
1967
head_rev = self._head_key[-1]
1968
lines = self.get_lines([head_rev])[head_rev]
1969
return ((plan, line) for line in lines)
1970
return self._weave.plan_merge(self.a_rev, self.b_rev)
1973
class _PlanLCAMerge(_PlanMergeBase):
1975
This merge algorithm differs from _PlanMerge in that:
1976
1. comparisons are done against LCAs only
1977
2. cases where a contested line is new versus one LCA but old versus
1978
another are marked as conflicts, by emitting the line as conflicted-a
1981
This is faster, and hopefully produces more useful output.
1984
def __init__(self, a_rev, b_rev, vf, key_prefix, graph):
1985
_PlanMergeBase.__init__(self, a_rev, b_rev, vf, key_prefix)
1986
lcas = graph.find_lca(key_prefix + (a_rev,), key_prefix + (b_rev,))
1989
if lca == NULL_REVISION:
1992
self.lcas.add(lca[-1])
1993
for lca in self.lcas:
1994
if _mod_revision.is_null(lca):
1997
lca_lines = self.get_lines([lca])[lca]
1998
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_a,
2000
blocks = list(matcher.get_matching_blocks())
2001
self._cached_matching_blocks[(a_rev, lca)] = blocks
2002
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_b,
2004
blocks = list(matcher.get_matching_blocks())
2005
self._cached_matching_blocks[(b_rev, lca)] = blocks
2007
def _determine_status(self, revision_id, unique_line_numbers):
2008
"""Determines the status unique lines versus all lcas.
2010
Basically, determines why the line is unique to this revision.
2012
A line may be determined new, killed, or both.
2014
If a line is determined new, that means it was not present in at least
2015
one LCA, and is not present in the other merge revision.
2017
If a line is determined killed, that means the line was present in
2020
If a line is killed and new, this indicates that the two merge
2021
revisions contain differing conflict resolutions.
2022
:param revision_id: The id of the revision in which the lines are
2024
:param unique_line_numbers: The line numbers of unique lines.
2025
:return a tuple of (new_this, killed_other):
2029
unique_line_numbers = set(unique_line_numbers)
2030
for lca in self.lcas:
2031
blocks = self._get_matching_blocks(revision_id, lca)
2032
unique_vs_lca, _ignored = self._unique_lines(blocks)
2033
new.update(unique_line_numbers.intersection(unique_vs_lca))
2034
killed.update(unique_line_numbers.difference(unique_vs_lca))