17
17
"""Tree classes, representing directory at point in time.
21
from collections import deque
21
from collections.abc import deque
22
except ImportError: # python < 3.7
23
from collections import deque
25
from .lazy_import import lazy_import
26
lazy_import(globals(), """
25
29
conflicts as _mod_conflicts,
30
33
revision as _mod_revision,
33
from bzrlib.decorators import needs_read_lock
34
from bzrlib.errors import BzrError, NoSuchId
35
from bzrlib import errors
36
from bzrlib.inventory import InventoryFile
37
from bzrlib.inter import InterObject
38
from bzrlib.osutils import fingerprint_file
39
from bzrlib.symbol_versioning import deprecated_function, deprecated_in
40
from bzrlib.trace import note
37
from breezy.i18n import gettext
45
from .inter import InterObject
48
class FileTimestampUnavailable(errors.BzrError):
50
_fmt = "The filestamp for %(path)s is not available."
54
def __init__(self, path):
58
class MissingNestedTree(errors.BzrError):
60
_fmt = "The nested tree for %(path)s can not be resolved."""
62
def __init__(self, path):
66
class TreeEntry(object):
67
"""An entry that implements the minimum interface used by commands.
72
def __eq__(self, other):
73
# yes, this is ugly, TODO: best practice __eq__ style.
74
return (isinstance(other, TreeEntry)
75
and other.__class__ == self.__class__)
79
def kind_character(self):
82
def is_unmodified(self, other):
83
"""Does this entry reference the same entry?
85
This is mostly the same as __eq__, but returns False
86
for entries without enough information (i.e. revision is None)
91
class TreeDirectory(TreeEntry):
92
"""See TreeEntry. This is a directory in a working tree."""
98
def kind_character(self):
102
class TreeFile(TreeEntry):
103
"""See TreeEntry. This is a regular file in a working tree."""
109
def kind_character(self):
113
class TreeLink(TreeEntry):
114
"""See TreeEntry. This is a symlink in a working tree."""
120
def kind_character(self):
124
class TreeReference(TreeEntry):
125
"""See TreeEntry. This is a reference to a nested tree in a working tree."""
129
kind = 'tree-reference'
131
def kind_character(self):
135
class TreeChange(object):
136
"""Describes the changes between the same item in two different trees."""
138
__slots__ = ['file_id', 'path', 'changed_content', 'versioned', 'parent_id',
139
'name', 'kind', 'executable', 'copied']
141
def __init__(self, file_id, path, changed_content, versioned, parent_id,
142
name, kind, executable, copied=False):
143
self.file_id = file_id
145
self.changed_content = changed_content
146
self.versioned = versioned
147
self.parent_id = parent_id
150
self.executable = executable
154
return "%s%r" % (self.__class__.__name__, self._as_tuple())
157
return len(self.__slots__)
160
return (self.file_id, self.path, self.changed_content, self.versioned,
161
self.parent_id, self.name, self.kind, self.executable, self.copied)
163
def __eq__(self, other):
164
if isinstance(other, TreeChange):
165
return self._as_tuple() == other._as_tuple()
166
if isinstance(other, tuple):
167
return self._as_tuple() == other
170
def __lt__(self, other):
171
return self._as_tuple() < other._as_tuple()
173
def meta_modified(self):
174
if self.versioned == (True, True):
175
return (self.executable[0] != self.executable[1])
178
def is_reparented(self):
179
return self.parent_id[0] != self.parent_id[1]
181
def discard_new(self):
182
return self.__class__(
183
self.file_id, (self.path[0], None), self.changed_content,
184
(self.versioned[0], None), (self.parent_id[0], None),
185
(self.name[0], None), (self.kind[0], None),
186
(self.executable[0], None),
43
190
class Tree(object):
194
335
The yield order (ignoring root) would be::
195
337
a, f, a/b, a/d, a/b/c, a/d/e, f/g
197
:param yield_parents: If True, yield the parents from the root leading
198
down to specific_file_ids that have been requested. This has no
199
impact if specific_file_ids is None.
201
return self.inventory.iter_entries_by_dir(
202
specific_file_ids=specific_file_ids, yield_parents=yield_parents)
339
If recurse_nested is enabled then nested trees are included as if
340
they were a part of the tree. If is disabled then TreeReference
341
objects (without any children) are yielded.
343
raise NotImplementedError(self.iter_entries_by_dir)
345
def iter_child_entries(self, path):
346
"""Iterate over the children of a directory or tree reference.
348
:param path: Path of the directory
349
:raise NoSuchFile: When the path does not exist
350
:return: Iterator over entries in the directory
352
raise NotImplementedError(self.iter_child_entries)
354
def list_files(self, include_root=False, from_dir=None, recursive=True,
355
recurse_nested=False):
356
"""List all files in this tree.
358
:param include_root: Whether to include the entry for the tree root
359
:param from_dir: Directory under which to list files
360
:param recursive: Whether to list files recursively
361
:param recurse_nested: enter nested trees
362
:return: iterator over tuples of
363
(path, versioned, kind, inventory entry)
365
raise NotImplementedError(self.list_files)
204
367
def iter_references(self):
205
368
if self.supports_tree_reference():
206
369
for path, entry in self.iter_entries_by_dir():
207
370
if entry.kind == 'tree-reference':
208
yield path, entry.file_id
210
def kind(self, file_id):
373
def get_containing_nested_tree(self, path):
374
"""Find the nested tree that contains a path.
376
:return: tuple with (nested tree and path inside the nested tree)
378
for nested_path in self.iter_references():
380
if path.startswith(nested_path):
381
nested_tree = self.get_nested_tree(nested_path)
382
return nested_tree, path[len(nested_path):]
386
def get_nested_tree(self, path):
387
"""Open the nested tree at the specified path.
389
:param path: Path from which to resolve tree reference.
390
:return: A Tree object for the nested tree
391
:raise MissingNestedTree: If the nested tree can not be resolved
393
raise NotImplementedError(self.get_nested_tree)
395
def kind(self, path):
211
396
raise NotImplementedError("Tree subclass %s must implement kind"
212
% self.__class__.__name__)
397
% self.__class__.__name__)
214
def stored_kind(self, file_id):
215
"""File kind stored for this file_id.
399
def stored_kind(self, path):
400
"""File kind stored for this path.
217
402
May not match kind on disk for working trees. Always available
218
403
for versioned files, even when the file itself is missing.
220
return self.kind(file_id)
405
return self.kind(path)
222
407
def path_content_summary(self, path):
223
408
"""Get a summary of the information about path.
254
439
raise NotImplementedError(self._comparison_data)
256
def _file_size(self, entry, stat_value):
257
raise NotImplementedError(self._file_size)
259
def _get_inventory(self):
260
return self._inventory
262
def get_file(self, file_id, path=None):
263
"""Return a file object for the file file_id in the tree.
265
If both file_id and path are defined, it is implementation defined as
266
to which one is used.
441
def get_file(self, path):
442
"""Return a file object for the file path in the tree.
268
444
raise NotImplementedError(self.get_file)
270
def get_file_with_stat(self, file_id, path=None):
271
"""Get a file handle and stat object for file_id.
446
def get_file_with_stat(self, path):
447
"""Get a file handle and stat object for path.
273
449
The default implementation returns (self.get_file, None) for backwards
276
:param file_id: The file id to read.
277
:param path: The path of the file, if it is known.
452
:param path: The path of the file.
278
453
:return: A tuple (file_handle, stat_value_or_None). If the tree has
279
454
no stat facility, or need for a stat cache feedback during commit,
280
455
it may return None for the second element of the tuple.
282
return (self.get_file(file_id, path), None)
457
return (self.get_file(path), None)
284
def get_file_text(self, file_id, path=None):
459
def get_file_text(self, path):
285
460
"""Return the byte content of a file.
287
:param file_id: The file_id of the file.
288
462
:param path: The path of the file.
289
If both file_id and path are supplied, an implementation may use
464
:returns: A single byte string for the whole file.
292
my_file = self.get_file(file_id, path)
466
with self.get_file(path) as my_file:
294
467
return my_file.read()
298
def get_file_lines(self, file_id, path=None):
469
def get_file_lines(self, path):
299
470
"""Return the content of a file, as lines.
301
:param file_id: The file_id of the file.
302
472
:param path: The path of the file.
303
If both file_id and path are supplied, an implementation may use
306
return osutils.split_lines(self.get_file_text(file_id, path))
308
def get_file_mtime(self, file_id, path=None):
474
return osutils.split_lines(self.get_file_text(path))
476
def get_file_verifier(self, path, stat_value=None):
477
"""Return a verifier for a file.
479
The default implementation returns a sha1.
481
:param path: The path that this file can be found at.
482
These must point to the same object.
483
:param stat_value: Optional stat value for the object
484
:return: Tuple with verifier name and verifier data
486
return ("SHA1", self.get_file_sha1(path, stat_value=stat_value))
488
def get_file_sha1(self, path, stat_value=None):
489
"""Return the SHA1 file for a file.
491
:note: callers should use get_file_verifier instead
492
where possible, as the underlying repository implementation may
493
have quicker access to a non-sha1 verifier.
495
:param path: The path that this file can be found at.
496
:param stat_value: Optional stat value for the object
498
raise NotImplementedError(self.get_file_sha1)
500
def get_file_mtime(self, path):
309
501
"""Return the modification time for a file.
311
:param file_id: The handle for this file.
312
503
:param path: The path that this file can be found at.
313
These must point to the same object.
315
505
raise NotImplementedError(self.get_file_mtime)
317
def get_file_size(self, file_id):
507
def get_file_size(self, path):
318
508
"""Return the size of a file in bytes.
320
510
This applies only to regular files. If invoked on directories or
321
511
symlinks, it will return None.
322
:param file_id: The file-id of the file
324
513
raise NotImplementedError(self.get_file_size)
326
def get_file_by_path(self, path):
327
return self.get_file(self._inventory.path2id(path), path)
515
def is_executable(self, path):
516
"""Check if a file is executable.
518
:param path: The path that this file can be found at.
520
raise NotImplementedError(self.is_executable)
329
522
def iter_files_bytes(self, desired_files):
330
523
"""Iterate through file contents.
342
535
this implementation, it is a tuple containing a single bytestring with
343
536
the complete text of the file.
345
:param desired_files: a list of (file_id, identifier) pairs
538
:param desired_files: a list of (path, identifier) pairs
347
for file_id, identifier in desired_files:
540
for path, identifier in desired_files:
348
541
# We wrap the string in a tuple so that we can return an iterable
349
542
# of bytestrings. (Technically, a bytestring is also an iterable
350
543
# of bytestrings, but iterating through each character is not
352
cur_file = (self.get_file_text(file_id),)
545
cur_file = (self.get_file_text(path),)
353
546
yield identifier, cur_file
355
def get_symlink_target(self, file_id):
356
"""Get the target for a given file_id.
548
def get_symlink_target(self, path):
549
"""Get the target for a given path.
358
It is assumed that the caller already knows that file_id is referencing
551
It is assumed that the caller already knows that path is referencing
360
:param file_id: Handle for the symlink entry.
553
:param path: The path of the file.
361
554
:return: The path the symlink points to.
363
556
raise NotImplementedError(self.get_symlink_target)
365
def get_canonical_inventory_paths(self, paths):
366
"""Like get_canonical_inventory_path() but works on multiple items.
368
:param paths: A sequence of paths relative to the root of the tree.
369
:return: A list of paths, with each item the corresponding input path
370
adjusted to account for existing elements that match case
373
return list(self._yield_canonical_inventory_paths(paths))
375
def get_canonical_inventory_path(self, path):
376
"""Returns the first inventory item that case-insensitively matches path.
378
If a path matches exactly, it is returned. If no path matches exactly
379
but more than one path matches case-insensitively, it is implementation
380
defined which is returned.
382
If no path matches case-insensitively, the input path is returned, but
383
with as many path entries that do exist changed to their canonical
386
If you need to resolve many names from the same tree, you should
387
use get_canonical_inventory_paths() to avoid O(N) behaviour.
389
:param path: A paths relative to the root of the tree.
390
:return: The input path adjusted to account for existing elements
391
that match case insensitively.
393
return self._yield_canonical_inventory_paths([path]).next()
395
def _yield_canonical_inventory_paths(self, paths):
397
# First, if the path as specified exists exactly, just use it.
398
if self.path2id(path) is not None:
402
cur_id = self.get_root_id()
404
bit_iter = iter(path.split("/"))
408
for child in self.iter_children(cur_id):
410
# XXX: it seem like if the child is known to be in the
411
# tree, we shouldn't need to go from its id back to
412
# its path -- mbp 2010-02-11
414
# XXX: it seems like we could be more efficient
415
# by just directly looking up the original name and
416
# only then searching all children; also by not
417
# chopping paths so much. -- mbp 2010-02-11
418
child_base = os.path.basename(self.id2path(child))
419
if (child_base == elt):
420
# if we found an exact match, we can stop now; if
421
# we found an approximate match we need to keep
422
# searching because there might be an exact match
425
new_path = osutils.pathjoin(cur_path, child_base)
427
elif child_base.lower() == lelt:
429
new_path = osutils.pathjoin(cur_path, child_base)
431
# before a change is committed we can see this error...
436
# got to the end of this directory and no entries matched.
437
# Return what matched so far, plus the rest as specified.
438
cur_path = osutils.pathjoin(cur_path, elt, *list(bit_iter))
443
def get_root_id(self):
444
"""Return the file_id for the root of this tree."""
445
raise NotImplementedError(self.get_root_id)
447
def annotate_iter(self, file_id,
558
def annotate_iter(self, path,
448
559
default_revision=_mod_revision.CURRENT_REVISION):
449
560
"""Return an iterator of revision_id, line tuples.
451
562
For working trees (and mutable trees in general), the special
452
563
revision_id 'current:' will be used for lines that are new in this
453
564
tree, e.g. uncommitted changes.
454
:param file_id: The file to produce an annotated version from
565
:param path: The file to produce an annotated version from
455
566
:param default_revision: For lines that don't match a basis, mark them
456
567
with this revision id. Not all implementations will make use of
459
570
raise NotImplementedError(self.annotate_iter)
461
def _get_plan_merge_data(self, file_id, other, base):
462
from bzrlib import versionedfile
463
vf = versionedfile._PlanMergeVersionedFile(file_id)
464
last_revision_a = self._get_file_revision(file_id, vf, 'this:')
465
last_revision_b = other._get_file_revision(file_id, vf, 'other:')
467
last_revision_base = None
469
last_revision_base = base._get_file_revision(file_id, vf, 'base:')
470
return vf, last_revision_a, last_revision_b, last_revision_base
472
def plan_file_merge(self, file_id, other, base=None):
473
"""Generate a merge plan based on annotations.
475
If the file contains uncommitted changes in this tree, they will be
476
attributed to the 'current:' pseudo-revision. If the file contains
477
uncommitted changes in the other tree, they will be assigned to the
478
'other:' pseudo-revision.
480
data = self._get_plan_merge_data(file_id, other, base)
481
vf, last_revision_a, last_revision_b, last_revision_base = data
482
return vf.plan_merge(last_revision_a, last_revision_b,
485
def plan_file_lca_merge(self, file_id, other, base=None):
486
"""Generate a merge plan based lca-newness.
488
If the file contains uncommitted changes in this tree, they will be
489
attributed to the 'current:' pseudo-revision. If the file contains
490
uncommitted changes in the other tree, they will be assigned to the
491
'other:' pseudo-revision.
493
data = self._get_plan_merge_data(file_id, other, base)
494
vf, last_revision_a, last_revision_b, last_revision_base = data
495
return vf.plan_lca_merge(last_revision_a, last_revision_b,
498
def _iter_parent_trees(self):
499
"""Iterate through parent trees, defaulting to Tree.revision_tree."""
500
for revision_id in self.get_parent_ids():
502
yield self.revision_tree(revision_id)
503
except errors.NoSuchRevisionInTree:
504
yield self.repository.revision_tree(revision_id)
507
def _file_revision(revision_tree, file_id):
508
"""Determine the revision associated with a file in a given tree."""
509
revision_tree.lock_read()
511
return revision_tree.inventory[file_id].revision
513
revision_tree.unlock()
515
def _get_file_revision(self, file_id, vf, tree_revision):
516
"""Ensure that file_id, tree_revision is in vf to plan the merge."""
518
if getattr(self, '_repository', None) is None:
519
last_revision = tree_revision
520
parent_keys = [(file_id, self._file_revision(t, file_id)) for t in
521
self._iter_parent_trees()]
522
vf.add_lines((file_id, last_revision), parent_keys,
523
self.get_file(file_id).readlines())
524
repo = self.branch.repository
527
last_revision = self._file_revision(self, file_id)
528
base_vf = self._repository.texts
529
if base_vf not in vf.fallback_versionedfiles:
530
vf.fallback_versionedfiles.append(base_vf)
533
inventory = property(_get_inventory,
534
doc="Inventory of this Tree")
536
def _check_retrieved(self, ie, f):
539
fp = fingerprint_file(f)
542
if ie.text_size is not None:
543
if ie.text_size != fp['size']:
544
raise BzrError("mismatched size for file %r in %r" % (ie.file_id, self._store),
545
["inventory expects %d bytes" % ie.text_size,
546
"file is actually %d bytes" % fp['size'],
547
"store is probably damaged/corrupt"])
549
if ie.text_sha1 != fp['sha1']:
550
raise BzrError("wrong SHA-1 for file %r in %r" % (ie.file_id, self._store),
551
["inventory expects %s" % ie.text_sha1,
552
"file is actually %s" % fp['sha1'],
553
"store is probably damaged/corrupt"])
556
572
def path2id(self, path):
557
573
"""Return the id for path in this tree."""
558
return self._inventory.path2id(path)
560
def paths2ids(self, paths, trees=[], require_versioned=True):
561
"""Return all the ids that can be reached by walking from paths.
563
Each path is looked up in this tree and any extras provided in
564
trees, and this is repeated recursively: the children in an extra tree
565
of a directory that has been renamed under a provided path in this tree
566
are all returned, even if none exist under a provided path in this
567
tree, and vice versa.
569
:param paths: An iterable of paths to start converting to ids from.
570
Alternatively, if paths is None, no ids should be calculated and None
571
will be returned. This is offered to make calling the api unconditional
572
for code that *might* take a list of files.
573
:param trees: Additional trees to consider.
574
:param require_versioned: If False, do not raise NotVersionedError if
575
an element of paths is not versioned in this tree and all of trees.
577
return find_ids_across_trees(paths, [self] + list(trees), require_versioned)
579
def iter_children(self, file_id):
580
entry = self.iter_entries_by_dir([file_id]).next()[1]
581
for child in getattr(entry, 'children', {}).itervalues():
574
raise NotImplementedError(self.path2id)
576
def is_versioned(self, path):
577
"""Check whether path is versioned.
579
:param path: Path to check
582
return self.path2id(path) is not None
584
def find_related_paths_across_trees(self, paths, trees=[],
585
require_versioned=True):
586
"""Find related paths in tree corresponding to specified filenames in any
589
All matches in all trees will be used, and all children of matched
590
directories will be used.
592
:param paths: The filenames to find related paths for (if None, returns
594
:param trees: The trees to find file_ids within
595
:param require_versioned: if true, all specified filenames must occur in
597
:return: a set of paths for the specified filenames and their children
600
raise NotImplementedError(self.find_related_paths_across_trees)
584
602
def lock_read(self):
585
603
"""Lock this tree for multiple read only operations.
587
:return: A bzrlib.lock.LogicalLockResult.
605
:return: A breezy.lock.LogicalLockResult.
607
return lock.LogicalLockResult(self.unlock)
591
609
def revision_tree(self, revision_id):
592
610
"""Obtain a revision tree for the revision revision_id.
724
738
searcher = default_searcher
728
######################################################################
731
# TODO: Merge these two functions into a single one that can operate
732
# on either a whole tree or a set of files.
734
# TODO: Return the diff in order by filename, not by category or in
735
# random order. Can probably be done by lock-stepping through the
736
# filenames from both trees.
739
def file_status(filename, old_tree, new_tree):
740
"""Return single-letter status, old and new names for a file.
742
The complexity here is in deciding how to represent renames;
743
many complex cases are possible.
745
old_inv = old_tree.inventory
746
new_inv = new_tree.inventory
747
new_id = new_inv.path2id(filename)
748
old_id = old_inv.path2id(filename)
750
if not new_id and not old_id:
751
# easy: doesn't exist in either; not versioned at all
752
if new_tree.is_ignored(filename):
753
return 'I', None, None
755
return '?', None, None
757
# There is now a file of this name, great.
760
# There is no longer a file of this name, but we can describe
761
# what happened to the file that used to have
762
# this name. There are two possibilities: either it was
763
# deleted entirely, or renamed.
764
if new_inv.has_id(old_id):
765
return 'X', old_inv.id2path(old_id), new_inv.id2path(old_id)
767
return 'D', old_inv.id2path(old_id), None
769
# if the file_id is new in this revision, it is added
770
if new_id and not old_inv.has_id(new_id):
773
# if there used to be a file of this name, but that ID has now
774
# disappeared, it is deleted
775
if old_id and not new_inv.has_id(old_id):
781
@deprecated_function(deprecated_in((1, 9, 0)))
782
def find_renames(old_inv, new_inv):
783
for file_id in old_inv:
784
if file_id not in new_inv:
786
old_name = old_inv.id2path(file_id)
787
new_name = new_inv.id2path(file_id)
788
if old_name != new_name:
789
yield (old_name, new_name)
792
def find_ids_across_trees(filenames, trees, require_versioned=True):
793
"""Find the ids corresponding to specified filenames.
795
All matches in all trees will be used, and all children of matched
796
directories will be used.
798
:param filenames: The filenames to find file_ids for (if None, returns
800
:param trees: The trees to find file_ids within
801
:param require_versioned: if true, all specified filenames must occur in
803
:return: a set of file ids for the specified filenames and their children.
807
specified_path_ids = _find_ids_across_trees(filenames, trees,
809
return _find_children_across_trees(specified_path_ids, trees)
812
def _find_ids_across_trees(filenames, trees, require_versioned):
813
"""Find the ids corresponding to specified filenames.
815
All matches in all trees will be used, but subdirectories are not scanned.
817
:param filenames: The filenames to find file_ids for
818
:param trees: The trees to find file_ids within
819
:param require_versioned: if true, all specified filenames must occur in
821
:return: a set of file ids for the specified filenames
824
interesting_ids = set()
825
for tree_path in filenames:
828
file_id = tree.path2id(tree_path)
829
if file_id is not None:
830
interesting_ids.add(file_id)
833
not_versioned.append(tree_path)
834
if len(not_versioned) > 0 and require_versioned:
835
raise errors.PathsNotVersionedError(not_versioned)
836
return interesting_ids
839
def _find_children_across_trees(specified_ids, trees):
840
"""Return a set including specified ids and their children.
842
All matches in all trees will be used.
844
:param trees: The trees to find file_ids within
845
:return: a set containing all specified ids and their children
847
interesting_ids = set(specified_ids)
848
pending = interesting_ids
849
# now handle children of interesting ids
850
# we loop so that we handle all children of each id in both trees
851
while len(pending) > 0:
853
for file_id in pending:
855
if not tree.has_or_had_id(file_id):
857
for child_id in tree.iter_children(file_id):
858
if child_id not in interesting_ids:
859
new_pending.add(child_id)
860
interesting_ids.update(new_pending)
861
pending = new_pending
862
return interesting_ids
741
def archive(self, format, name, root='', subdir=None,
743
"""Create an archive of this tree.
745
:param format: Format name (e.g. 'tar')
746
:param name: target file name
747
:param root: Root directory name (or None)
748
:param subdir: Subdirectory to export (or None)
749
:return: Iterator over archive chunks
751
from .archive import create_archive
752
with self.lock_read():
753
return create_archive(format, self, name, root,
754
subdir, force_mtime=force_mtime)
757
def versionable_kind(cls, kind):
758
"""Check if this tree support versioning a specific file kind."""
759
return (kind in ('file', 'directory', 'symlink', 'tree-reference'))
865
762
class InterTree(InterObject):
1059
948
seen_dirs = set()
1060
949
if want_unversioned:
1061
950
all_unversioned = sorted([(p.split('/'), p) for p in
1062
self.target.extras()
1063
if specific_files is None or
1064
osutils.is_inside_any(specific_files, p)])
952
if specific_files is None or
953
osutils.is_inside_any(specific_files, p)])
1065
954
all_unversioned = deque(all_unversioned)
1067
956
all_unversioned = deque()
1069
958
from_entries_by_dir = list(self.source.iter_entries_by_dir(
1070
specific_file_ids=specific_file_ids))
1071
from_data = dict((e.file_id, (p, e)) for p, e in from_entries_by_dir)
959
specific_files=source_specific_files))
960
from_data = dict(from_entries_by_dir)
1072
961
to_entries_by_dir = list(self.target.iter_entries_by_dir(
1073
specific_file_ids=specific_file_ids))
962
specific_files=target_specific_files))
963
path_equivs = self.find_source_paths([p for p, e in to_entries_by_dir])
1074
964
num_entries = len(from_entries_by_dir) + len(to_entries_by_dir)
1076
966
# the unversioned path lookup only occurs on real trees - where there
1077
967
# can be extras. So the fake_entry is solely used to look up
1078
968
# executable it values when execute is not supported.
1079
fake_entry = InventoryFile('unused', 'unused', 'unused')
969
fake_entry = TreeFile()
1080
970
for target_path, target_entry in to_entries_by_dir:
1081
971
while (all_unversioned and
1082
all_unversioned[0][0] < target_path.split('/')):
972
all_unversioned[0][0] < target_path.split('/')):
1083
973
unversioned_path = all_unversioned.popleft()
1084
974
target_kind, target_executable, target_stat = \
1085
self.target._comparison_data(fake_entry, unversioned_path[1])
1086
yield (None, (None, unversioned_path[1]), True, (False, False),
975
self.target._comparison_data(
976
fake_entry, unversioned_path[1])
978
None, (None, unversioned_path[1]), True, (False, False),
1088
980
(None, unversioned_path[0][-1]),
1089
981
(None, target_kind),
1090
982
(None, target_executable))
1091
source_path, source_entry = from_data.get(target_entry.file_id,
1093
result, changes = self._changes_from_entries(source_entry,
1094
target_entry, source_path=source_path, target_path=target_path)
1095
to_paths[result[0]] = result[1][1]
983
source_path = path_equivs[target_path]
984
if source_path is not None:
985
source_entry = from_data.get(source_path)
988
result, changes = self._changes_from_entries(
989
source_entry, target_entry, source_path=source_path, target_path=target_path)
990
to_paths[result.file_id] = result.path[1]
1096
991
entry_count += 1
992
if result.versioned[0]:
1098
993
entry_count += 1
1099
994
if pb is not None:
1100
995
pb.update('comparing files', entry_count, num_entries)
1101
996
if changes or include_unchanged:
1102
if specific_file_ids is not None:
1103
new_parent_id = result[4][1]
1104
precise_file_ids.add(new_parent_id)
1105
changed_file_ids.append(result[0])
997
if specific_files is not None:
998
precise_file_ids.add(result.parent_id[1])
999
changed_file_ids.append(result.file_id)
1107
1001
# Ensure correct behaviour for reparented/added specific files.
1108
1002
if specific_files is not None:
1109
1003
# Record output dirs
1110
if result[6][1] == 'directory':
1111
seen_dirs.add(result[0])
1004
if result.kind[1] == 'directory':
1005
seen_dirs.add(result.file_id)
1112
1006
# Record parents of reparented/added entries.
1113
versioned = result[3]
1115
if not versioned[0] or parents[0] != parents[1]:
1116
seen_parents.add(parents[1])
1007
if not result.versioned[0] or result.is_reparented():
1008
seen_parents.add(result.parent_id[1])
1117
1009
while all_unversioned:
1118
1010
# yield any trailing unversioned paths
1119
1011
unversioned_path = all_unversioned.popleft()
1120
1012
to_kind, to_executable, to_stat = \
1121
1013
self.target._comparison_data(fake_entry, unversioned_path[1])
1122
yield (None, (None, unversioned_path[1]), True, (False, False),
1015
None, (None, unversioned_path[1]), True, (False, False),
1124
1017
(None, unversioned_path[0][-1]),
1125
1018
(None, to_kind),
1228
1110
# Examine file_id
1229
1111
if discarded_changes:
1230
1112
result = discarded_changes.get(file_id)
1234
1116
if result is None:
1235
old_entry = self._get_entry(self.source, file_id)
1236
new_entry = self._get_entry(self.target, file_id)
1118
source_path = self.source.id2path(file_id)
1119
except errors.NoSuchId:
1123
source_entry = self._get_entry(
1124
self.source, source_path)
1126
target_path = self.target.id2path(file_id)
1127
except errors.NoSuchId:
1131
target_entry = self._get_entry(
1132
self.target, target_path)
1237
1133
result, changes = self._changes_from_entries(
1238
old_entry, new_entry)
1134
source_entry, target_entry, source_path, target_path)
1241
1137
# Get this parents parent to examine.
1242
new_parent_id = result[4][1]
1138
new_parent_id = result.parent_id[1]
1243
1139
precise_file_ids.add(new_parent_id)
1245
if (result[6][0] == 'directory' and
1246
result[6][1] != 'directory'):
1141
if (result.kind[0] == 'directory' and
1142
result.kind[1] != 'directory'):
1247
1143
# This stopped being a directory, the old children have
1248
1144
# to be included.
1249
if old_entry is None:
1145
if source_entry is None:
1250
1146
# Reusing a discarded change.
1251
old_entry = self._get_entry(self.source, file_id)
1252
for child in old_entry.children.values():
1253
precise_file_ids.add(child.file_id)
1254
changed_file_ids.add(result[0])
1147
source_entry = self._get_entry(
1148
self.source, result.path[0])
1149
precise_file_ids.update(
1151
for child in self.source.iter_child_entries(result.path[0]))
1152
changed_file_ids.add(result.file_id)
1258
class MultiWalker(object):
1259
"""Walk multiple trees simultaneously, getting combined results."""
1261
# Note: This could be written to not assume you can do out-of-order
1262
# lookups. Instead any nodes that don't match in all trees could be
1263
# marked as 'deferred', and then returned in the final cleanup loop.
1264
# For now, I think it is "nicer" to return things as close to the
1265
# "master_tree" order as we can.
1267
def __init__(self, master_tree, other_trees):
1268
"""Create a new MultiWalker.
1270
All trees being walked must implement "iter_entries_by_dir()", such
1271
that they yield (path, object) tuples, where that object will have a
1272
'.file_id' member, that can be used to check equality.
1274
:param master_tree: All trees will be 'slaved' to the master_tree such
1275
that nodes in master_tree will be used as 'first-pass' sync points.
1276
Any nodes that aren't in master_tree will be merged in a second
1278
:param other_trees: A list of other trees to walk simultaneously.
1280
self._master_tree = master_tree
1281
self._other_trees = other_trees
1283
# Keep track of any nodes that were properly processed just out of
1284
# order, that way we don't return them at the end, we don't have to
1285
# track *all* processed file_ids, just the out-of-order ones
1286
self._out_of_order_processed = set()
1289
def _step_one(iterator):
1290
"""Step an iter_entries_by_dir iterator.
1292
:return: (has_more, path, ie)
1293
If has_more is False, path and ie will be None.
1296
path, ie = iterator.next()
1297
except StopIteration:
1298
return False, None, None
1300
return True, path, ie
1303
def _cmp_path_by_dirblock(path1, path2):
1304
"""Compare two paths based on what directory they are in.
1306
This generates a sort order, such that all children of a directory are
1307
sorted together, and grandchildren are in the same order as the
1308
children appear. But all grandchildren come after all children.
1310
:param path1: first path
1311
:param path2: the second path
1312
:return: negative number if ``path1`` comes first,
1313
0 if paths are equal
1314
and a positive number if ``path2`` sorts first
1316
# Shortcut this special case
1319
# This is stolen from _dirstate_helpers_py.py, only switching it to
1320
# Unicode objects. Consider using encode_utf8() and then using the
1321
# optimized versions, or maybe writing optimized unicode versions.
1322
if not isinstance(path1, unicode):
1323
raise TypeError("'path1' must be a unicode string, not %s: %r"
1324
% (type(path1), path1))
1325
if not isinstance(path2, unicode):
1326
raise TypeError("'path2' must be a unicode string, not %s: %r"
1327
% (type(path2), path2))
1328
return cmp(MultiWalker._path_to_key(path1),
1329
MultiWalker._path_to_key(path2))
1332
def _path_to_key(path):
1333
dirname, basename = osutils.split(path)
1334
return (dirname.split(u'/'), basename)
1336
def _lookup_by_file_id(self, extra_entries, other_tree, file_id):
1337
"""Lookup an inventory entry by file_id.
1339
This is called when an entry is missing in the normal order.
1340
Generally this is because a file was either renamed, or it was
1341
deleted/added. If the entry was found in the inventory and not in
1342
extra_entries, it will be added to self._out_of_order_processed
1344
:param extra_entries: A dictionary of {file_id: (path, ie)}. This
1345
should be filled with entries that were found before they were
1346
used. If file_id is present, it will be removed from the
1348
:param other_tree: The Tree to search, in case we didn't find the entry
1350
:param file_id: The file_id to look for
1351
:return: (path, ie) if found or (None, None) if not present.
1353
if file_id in extra_entries:
1354
return extra_entries.pop(file_id)
1355
# TODO: Is id2path better as the first call, or is
1356
# inventory[file_id] better as a first check?
1358
cur_path = other_tree.id2path(file_id)
1359
except errors.NoSuchId:
1361
if cur_path is None:
1364
self._out_of_order_processed.add(file_id)
1365
cur_ie = other_tree.inventory[file_id]
1366
return (cur_path, cur_ie)
1369
"""Match up the values in the different trees."""
1370
for result in self._walk_master_tree():
1372
self._finish_others()
1373
for result in self._walk_others():
1376
def _walk_master_tree(self):
1377
"""First pass, walk all trees in lock-step.
1379
When we are done, all nodes in the master_tree will have been
1380
processed. _other_walkers, _other_entries, and _others_extra will be
1381
set on 'self' for future processing.
1383
# This iterator has the most "inlining" done, because it tends to touch
1384
# every file in the tree, while the others only hit nodes that don't
1386
master_iterator = self._master_tree.iter_entries_by_dir()
1388
other_walkers = [other.iter_entries_by_dir()
1389
for other in self._other_trees]
1390
other_entries = [self._step_one(walker) for walker in other_walkers]
1391
# Track extra nodes in the other trees
1392
others_extra = [{} for i in xrange(len(self._other_trees))]
1394
master_has_more = True
1395
step_one = self._step_one
1396
lookup_by_file_id = self._lookup_by_file_id
1397
out_of_order_processed = self._out_of_order_processed
1399
while master_has_more:
1400
(master_has_more, path, master_ie) = step_one(master_iterator)
1401
if not master_has_more:
1404
file_id = master_ie.file_id
1406
other_values_append = other_values.append
1407
next_other_entries = []
1408
next_other_entries_append = next_other_entries.append
1409
for idx, (other_has_more, other_path, other_ie) in enumerate(other_entries):
1410
if not other_has_more:
1411
other_values_append(lookup_by_file_id(
1412
others_extra[idx], self._other_trees[idx], file_id))
1413
next_other_entries_append((False, None, None))
1414
elif file_id == other_ie.file_id:
1415
# This is the critical code path, as most of the entries
1416
# should match between most trees.
1417
other_values_append((other_path, other_ie))
1418
next_other_entries_append(step_one(other_walkers[idx]))
1420
# This walker did not match, step it until it either
1421
# matches, or we know we are past the current walker.
1422
other_walker = other_walkers[idx]
1423
other_extra = others_extra[idx]
1424
while (other_has_more and
1425
self._cmp_path_by_dirblock(other_path, path) < 0):
1426
other_file_id = other_ie.file_id
1427
if other_file_id not in out_of_order_processed:
1428
other_extra[other_file_id] = (other_path, other_ie)
1429
other_has_more, other_path, other_ie = \
1430
step_one(other_walker)
1431
if other_has_more and other_ie.file_id == file_id:
1432
# We ended up walking to this point, match and step
1434
other_values_append((other_path, other_ie))
1435
other_has_more, other_path, other_ie = \
1436
step_one(other_walker)
1438
# This record isn't in the normal order, see if it
1440
other_values_append(lookup_by_file_id(
1441
other_extra, self._other_trees[idx], file_id))
1442
next_other_entries_append((other_has_more, other_path,
1444
other_entries = next_other_entries
1446
# We've matched all the walkers, yield this datapoint
1447
yield path, file_id, master_ie, other_values
1448
self._other_walkers = other_walkers
1449
self._other_entries = other_entries
1450
self._others_extra = others_extra
1452
def _finish_others(self):
1453
"""Finish walking the other iterators, so we get all entries."""
1454
for idx, info in enumerate(self._other_entries):
1455
other_extra = self._others_extra[idx]
1456
(other_has_more, other_path, other_ie) = info
1457
while other_has_more:
1458
other_file_id = other_ie.file_id
1459
if other_file_id not in self._out_of_order_processed:
1460
other_extra[other_file_id] = (other_path, other_ie)
1461
other_has_more, other_path, other_ie = \
1462
self._step_one(self._other_walkers[idx])
1463
del self._other_entries
1465
def _walk_others(self):
1466
"""Finish up by walking all the 'deferred' nodes."""
1467
# TODO: One alternative would be to grab all possible unprocessed
1468
# file_ids, and then sort by path, and then yield them. That
1469
# might ensure better ordering, in case a caller strictly
1470
# requires parents before children.
1471
for idx, other_extra in enumerate(self._others_extra):
1472
others = sorted(other_extra.itervalues(),
1473
key=lambda x: self._path_to_key(x[0]))
1474
for other_path, other_ie in others:
1475
file_id = other_ie.file_id
1476
# We don't need to check out_of_order_processed here, because
1477
# the lookup_by_file_id will be removing anything processed
1478
# from the extras cache
1479
other_extra.pop(file_id)
1480
other_values = [(None, None) for i in xrange(idx)]
1481
other_values.append((other_path, other_ie))
1482
for alt_idx, alt_extra in enumerate(self._others_extra[idx+1:]):
1483
alt_idx = alt_idx + idx + 1
1484
alt_extra = self._others_extra[alt_idx]
1485
alt_tree = self._other_trees[alt_idx]
1486
other_values.append(self._lookup_by_file_id(
1487
alt_extra, alt_tree, file_id))
1488
yield other_path, file_id, None, other_values
1155
def file_content_matches(
1156
self, source_path, target_path,
1157
source_stat=None, target_stat=None):
1158
"""Check if two files are the same in the source and target trees.
1160
This only checks that the contents of the files are the same,
1161
it does not touch anything else.
1163
:param source_path: Path of the file in the source tree
1164
:param target_path: Path of the file in the target tree
1165
:param source_stat: Optional stat value of the file in the source tree
1166
:param target_stat: Optional stat value of the file in the target tree
1167
:return: Boolean indicating whether the files have the same contents
1169
with self.lock_read():
1170
source_verifier_kind, source_verifier_data = (
1171
self.source.get_file_verifier(source_path, source_stat))
1172
target_verifier_kind, target_verifier_data = (
1173
self.target.get_file_verifier(
1174
target_path, target_stat))
1175
if source_verifier_kind == target_verifier_kind:
1176
return (source_verifier_data == target_verifier_data)
1177
# Fall back to SHA1 for now
1178
if source_verifier_kind != "SHA1":
1179
source_sha1 = self.source.get_file_sha1(
1180
source_path, source_stat)
1182
source_sha1 = source_verifier_data
1183
if target_verifier_kind != "SHA1":
1184
target_sha1 = self.target.get_file_sha1(
1185
target_path, target_stat)
1187
target_sha1 = target_verifier_data
1188
return (source_sha1 == target_sha1)
1190
def find_target_path(self, path, recurse='none'):
1191
"""Find target tree path.
1193
:param path: Path to search for (exists in source)
1194
:return: path in target, or None if there is no equivalent path.
1195
:raise NoSuchFile: If the path doesn't exist in source
1197
file_id = self.source.path2id(path)
1199
raise errors.NoSuchFile(path)
1201
return self.target.id2path(file_id, recurse=recurse)
1202
except errors.NoSuchId:
1205
def find_source_path(self, path, recurse='none'):
1206
"""Find the source tree path.
1208
:param path: Path to search for (exists in target)
1209
:return: path in source, or None if there is no equivalent path.
1210
:raise NoSuchFile: if the path doesn't exist in target
1212
file_id = self.target.path2id(path)
1214
raise errors.NoSuchFile(path)
1216
return self.source.id2path(file_id, recurse=recurse)
1217
except errors.NoSuchId:
1220
def find_target_paths(self, paths, recurse='none'):
1221
"""Find target tree paths.
1223
:param paths: Iterable over paths in target to search for
1224
:return: Dictionary mapping from source paths to paths in target , or
1225
None if there is no equivalent path.
1229
ret[path] = self.find_target_path(path, recurse=recurse)
1232
def find_source_paths(self, paths, recurse='none'):
1233
"""Find source tree paths.
1235
:param paths: Iterable over paths in target to search for
1236
:return: Dictionary mapping from target paths to paths in source, or
1237
None if there is no equivalent path.
1241
ret[path] = self.find_source_path(path, recurse=recurse)
1245
InterTree.register_optimiser(InterTree)
1248
def find_previous_paths(from_tree, to_tree, paths, recurse='none'):
1249
"""Find previous tree paths.
1251
:param from_tree: From tree
1252
:param to_tree: To tree
1253
:param paths: Iterable over paths in from_tree to search for
1254
:return: Dictionary mapping from from_tree paths to paths in to_tree, or
1255
None if there is no equivalent path.
1257
return InterTree.get(to_tree, from_tree).find_source_paths(paths, recurse=recurse)
1260
def find_previous_path(from_tree, to_tree, path, recurse='none'):
1261
"""Find previous tree path.
1263
:param from_tree: From tree
1264
:param to_tree: To tree
1265
:param path: Path to search for (exists in from_tree)
1266
:return: path in to_tree, or None if there is no equivalent path.
1267
:raise NoSuchFile: If the path doesn't exist in from_tree
1269
return InterTree.get(to_tree, from_tree).find_source_path(
1270
path, recurse=recurse)
1273
def get_canonical_path(tree, path, normalize):
1274
"""Find the canonical path of an item, ignoring case.
1276
:param tree: Tree to traverse
1277
:param path: Case-insensitive path to look up
1278
:param normalize: Function to normalize a filename for comparison
1279
:return: The canonical path
1283
bit_iter = iter(path.split("/"))
1284
for elt in bit_iter:
1285
lelt = normalize(elt)
1288
for child in tree.iter_child_entries(cur_path):
1290
if child.name == elt:
1291
# if we found an exact match, we can stop now; if
1292
# we found an approximate match we need to keep
1293
# searching because there might be an exact match
1295
new_path = osutils.pathjoin(cur_path, child.name)
1297
elif normalize(child.name) == lelt:
1298
new_path = osutils.pathjoin(cur_path, child.name)
1299
except errors.NoSuchId:
1300
# before a change is committed we can see this error...
1302
except errors.NotADirectory:
1307
# got to the end of this directory and no entries matched.
1308
# Return what matched so far, plus the rest as specified.
1309
cur_path = osutils.pathjoin(cur_path, elt, *list(bit_iter))