496
589
return os.path.getsize(self.id2abspath(file_id))
499
def get_file_sha1(self, file_id):
500
path = self._inventory.id2path(file_id)
501
return self._hashcache.get_sha1(path)
503
def is_executable(self, file_id):
504
if not supports_executable():
592
def get_file_sha1(self, file_id, path=None, stat_value=None):
594
path = self._inventory.id2path(file_id)
595
return self._hashcache.get_sha1(path, stat_value)
597
def get_file_mtime(self, file_id, path=None):
599
path = self.inventory.id2path(file_id)
600
return os.lstat(self.abspath(path)).st_mtime
602
if not supports_executable():
603
def is_executable(self, file_id, path=None):
505
604
return self._inventory[file_id].executable
507
path = self._inventory.id2path(file_id)
606
def is_executable(self, file_id, path=None):
608
path = self.id2path(file_id)
508
609
mode = os.lstat(self.abspath(path)).st_mode
509
return bool(stat.S_ISREG(mode) and stat.S_IEXEC&mode)
512
def add(self, files, ids=None):
513
"""Make files versioned.
515
Note that the command line normally calls smart_add instead,
516
which can automatically recurse.
518
This adds the files to the inventory, so that they will be
519
recorded by the next commit.
522
List of paths to add, relative to the base of the tree.
525
If set, use these instead of automatically generated ids.
526
Must be the same length as the list of files, but may
527
contain None for ids that are to be autogenerated.
529
TODO: Perhaps have an option to add the ids even if the files do
532
TODO: Perhaps callback with the ids and paths as they're added.
610
return bool(stat.S_ISREG(mode) and stat.S_IEXEC & mode)
612
@needs_tree_write_lock
613
def _add(self, files, ids, kinds):
614
"""See MutableTree._add."""
534
615
# TODO: Re-adding a file that is removed in the working copy
535
616
# should probably put it back with the previous ID.
536
if isinstance(files, basestring):
537
assert(ids is None or isinstance(ids, basestring))
543
ids = [None] * len(files)
545
assert(len(ids) == len(files))
547
inv = self.read_working_inventory()
548
for f,file_id in zip(files, ids):
549
if self.is_control_filename(f):
550
raise BzrError("cannot add control file %s" % quotefn(f))
555
raise BzrError("cannot add top-level %r" % f)
557
fullpath = normpath(self.abspath(f))
560
kind = file_kind(fullpath)
562
if e.errno == errno.ENOENT:
563
raise NoSuchFile(fullpath)
564
# maybe something better?
565
raise BzrError('cannot add: not a regular file, symlink or directory: %s' % quotefn(f))
567
if not InventoryEntry.versionable_kind(kind):
568
raise BzrError('cannot add: not a versionable file ('
569
'i.e. regular file, symlink or directory): %s' % quotefn(f))
617
# the read and write working inventory should not occur in this
618
# function - they should be part of lock_write and unlock.
620
for f, file_id, kind in zip(files, ids, kinds):
621
assert kind is not None
571
622
if file_id is None:
572
file_id = gen_file_id(f)
573
inv.add_path(f, kind=kind, file_id=file_id)
623
inv.add_path(f, kind=kind)
625
inv.add_path(f, kind=kind, file_id=file_id)
626
self._inventory_is_modified = True
575
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
576
self._write_inventory(inv)
628
@needs_tree_write_lock
629
def _gather_kinds(self, files, kinds):
630
"""See MutableTree._gather_kinds."""
631
for pos, f in enumerate(files):
632
if kinds[pos] is None:
633
fullpath = normpath(self.abspath(f))
635
kinds[pos] = file_kind(fullpath)
637
if e.errno == errno.ENOENT:
638
raise errors.NoSuchFile(fullpath)
578
640
@needs_write_lock
641
def add_parent_tree_id(self, revision_id, allow_leftmost_as_ghost=False):
642
"""Add revision_id as a parent.
644
This is equivalent to retrieving the current list of parent ids
645
and setting the list to its value plus revision_id.
647
:param revision_id: The revision id to add to the parent list. It may
648
be a ghost revision as long as its not the first parent to be added,
649
or the allow_leftmost_as_ghost parameter is set True.
650
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
652
parents = self.get_parent_ids() + [revision_id]
653
self.set_parent_ids(parents, allow_leftmost_as_ghost=len(parents) > 1
654
or allow_leftmost_as_ghost)
656
@needs_tree_write_lock
657
def add_parent_tree(self, parent_tuple, allow_leftmost_as_ghost=False):
658
"""Add revision_id, tree tuple as a parent.
660
This is equivalent to retrieving the current list of parent trees
661
and setting the list to its value plus parent_tuple. See also
662
add_parent_tree_id - if you only have a parent id available it will be
663
simpler to use that api. If you have the parent already available, using
664
this api is preferred.
666
:param parent_tuple: The (revision id, tree) to add to the parent list.
667
If the revision_id is a ghost, pass None for the tree.
668
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
670
parent_ids = self.get_parent_ids() + [parent_tuple[0]]
671
if len(parent_ids) > 1:
672
# the leftmost may have already been a ghost, preserve that if it
674
allow_leftmost_as_ghost = True
675
self.set_parent_ids(parent_ids,
676
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
678
@needs_tree_write_lock
579
679
def add_pending_merge(self, *revision_ids):
580
680
# TODO: Perhaps should check at this point that the
581
681
# history of the revision is actually present?
582
p = self.pending_merges()
682
parents = self.get_parent_ids()
584
684
for rev_id in revision_ids:
685
if rev_id in parents:
687
parents.append(rev_id)
590
self.set_pending_merges(p)
690
self.set_parent_ids(parents, allow_leftmost_as_ghost=True)
692
def path_content_summary(self, path, _lstat=osutils.lstat,
693
_mapper=osutils.file_kind_from_stat_mode):
694
"""See Tree.path_content_summary."""
695
abspath = self.abspath(path)
697
stat_result = _lstat(abspath)
699
if getattr(e, 'errno', None) == errno.ENOENT:
701
return ('missing', None, None, None)
702
# propagate other errors
704
kind = _mapper(stat_result.st_mode)
706
size = stat_result.st_size
707
# try for a stat cache lookup
708
if not supports_executable():
709
executable = None # caller can decide policy.
711
mode = stat_result.st_mode
712
executable = bool(stat.S_ISREG(mode) and stat.S_IEXEC & mode)
713
return (kind, size, executable, self._sha_from_stat(
715
elif kind == 'directory':
716
# perhaps it looks like a plain directory, but it's really a
718
if self._directory_is_tree_reference(path):
719
kind = 'tree-reference'
720
return kind, None, None, None
721
elif kind == 'symlink':
722
return ('symlink', None, None, os.readlink(abspath))
724
return (kind, None, None, None)
726
@deprecated_method(zero_eleven)
593
728
def pending_merges(self):
594
729
"""Return a list of pending merges.
596
731
These are revisions that have been merged into the working
597
732
directory but not yet committed.
600
merges_file = self._control_files.get_utf8('pending-merges')
602
if e.errno != errno.ENOENT:
606
for l in merges_file.readlines():
607
p.append(l.rstrip('\n'))
734
As of 0.11 this is deprecated. Please see WorkingTree.get_parent_ids()
735
instead - which is available on all tree objects.
737
return self.get_parent_ids()[1:]
739
def _check_parents_for_ghosts(self, revision_ids, allow_leftmost_as_ghost):
740
"""Common ghost checking functionality from set_parent_*.
742
This checks that the left hand-parent exists if there are any
745
if len(revision_ids) > 0:
746
leftmost_id = revision_ids[0]
747
if (not allow_leftmost_as_ghost and not
748
self.branch.repository.has_revision(leftmost_id)):
749
raise errors.GhostRevisionUnusableHere(leftmost_id)
751
def _set_merges_from_parent_ids(self, parent_ids):
752
merges = parent_ids[1:]
753
self._control_files.put_bytes('pending-merges', '\n'.join(merges))
755
@needs_tree_write_lock
756
def set_parent_ids(self, revision_ids, allow_leftmost_as_ghost=False):
757
"""Set the parent ids to revision_ids.
759
See also set_parent_trees. This api will try to retrieve the tree data
760
for each element of revision_ids from the trees repository. If you have
761
tree data already available, it is more efficient to use
762
set_parent_trees rather than set_parent_ids. set_parent_ids is however
763
an easier API to use.
765
:param revision_ids: The revision_ids to set as the parent ids of this
766
working tree. Any of these may be ghosts.
768
self._check_parents_for_ghosts(revision_ids,
769
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
770
for revision_id in revision_ids:
771
_mod_revision.check_not_reserved_id(revision_id)
773
if len(revision_ids) > 0:
774
self.set_last_revision(revision_ids[0])
776
self.set_last_revision(_mod_revision.NULL_REVISION)
778
self._set_merges_from_parent_ids(revision_ids)
780
@needs_tree_write_lock
781
def set_parent_trees(self, parents_list, allow_leftmost_as_ghost=False):
782
"""See MutableTree.set_parent_trees."""
783
parent_ids = [rev for (rev, tree) in parents_list]
784
for revision_id in parent_ids:
785
_mod_revision.check_not_reserved_id(revision_id)
787
self._check_parents_for_ghosts(parent_ids,
788
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
790
if len(parent_ids) == 0:
791
leftmost_parent_id = _mod_revision.NULL_REVISION
792
leftmost_parent_tree = None
794
leftmost_parent_id, leftmost_parent_tree = parents_list[0]
796
if self._change_last_revision(leftmost_parent_id):
797
if leftmost_parent_tree is None:
798
# If we don't have a tree, fall back to reading the
799
# parent tree from the repository.
800
self._cache_basis_inventory(leftmost_parent_id)
802
inv = leftmost_parent_tree.inventory
803
xml = self._create_basis_xml_from_inventory(
804
leftmost_parent_id, inv)
805
self._write_basis_inventory(xml)
806
self._set_merges_from_parent_ids(parent_ids)
808
@needs_tree_write_lock
611
809
def set_pending_merges(self, rev_list):
612
self._control_files.put_utf8('pending-merges', '\n'.join(rev_list))
810
parents = self.get_parent_ids()
811
leftmost = parents[:1]
812
new_parents = leftmost + rev_list
813
self.set_parent_ids(new_parents)
815
@needs_tree_write_lock
615
816
def set_merge_modified(self, modified_hashes):
616
817
def iter_stanzas():
617
818
for file_id, hash in modified_hashes.iteritems():
618
yield Stanza(file_id=file_id, hash=hash)
819
yield Stanza(file_id=file_id.decode('utf8'), hash=hash)
619
820
self._put_rio('merge-hashes', iter_stanzas(), MERGE_MODIFIED_HEADER_1)
822
def _sha_from_stat(self, path, stat_result):
823
"""Get a sha digest from the tree's stat cache.
825
The default implementation assumes no stat cache is present.
827
:param path: The path.
828
:param stat_result: The stat result being looked up.
622
832
def _put_rio(self, filename, stanzas, header):
833
self._must_be_locked()
623
834
my_file = rio_file(stanzas, header)
624
835
self._control_files.put(filename, my_file)
837
@needs_write_lock # because merge pulls data into the branch.
838
def merge_from_branch(self, branch, to_revision=None, from_revision=None,
840
"""Merge from a branch into this working tree.
842
:param branch: The branch to merge from.
843
:param to_revision: If non-None, the merge will merge to to_revision,
844
but not beyond it. to_revision does not need to be in the history
845
of the branch when it is supplied. If None, to_revision defaults to
846
branch.last_revision().
848
from bzrlib.merge import Merger, Merge3Merger
849
pb = bzrlib.ui.ui_factory.nested_progress_bar()
851
merger = Merger(self.branch, this_tree=self, pb=pb)
852
merger.pp = ProgressPhase("Merge phase", 5, pb)
853
merger.pp.next_phase()
854
# check that there are no
856
merger.check_basis(check_clean=True, require_commits=False)
857
if to_revision is None:
858
to_revision = _mod_revision.ensure_null(branch.last_revision())
859
merger.other_rev_id = to_revision
860
if _mod_revision.is_null(merger.other_rev_id):
861
raise errors.NoCommits(branch)
862
self.branch.fetch(branch, last_revision=merger.other_rev_id)
863
merger.other_basis = merger.other_rev_id
864
merger.other_tree = self.branch.repository.revision_tree(
866
merger.other_branch = branch
867
merger.pp.next_phase()
868
if from_revision is None:
871
merger.set_base_revision(from_revision, branch)
872
if merger.base_rev_id == merger.other_rev_id:
873
raise errors.PointlessMerge
874
merger.backup_files = False
875
if merge_type is None:
876
merger.merge_type = Merge3Merger
878
merger.merge_type = merge_type
879
merger.set_interesting_files(None)
880
merger.show_base = False
881
merger.reprocess = False
882
conflicts = merger.do_merge()
627
889
def merge_modified(self):
890
"""Return a dictionary of files modified by a merge.
892
The list is initialized by WorkingTree.set_merge_modified, which is
893
typically called after we make some automatic updates to the tree
896
This returns a map of file_id->sha1, containing only files which are
897
still in the working inventory and have that text hash.
629
900
hashfile = self._control_files.get('merge-hashes')
901
except errors.NoSuchFile:
632
903
merge_hashes = {}
634
905
if hashfile.next() != MERGE_MODIFIED_HEADER_1 + '\n':
635
raise MergeModifiedFormatError()
906
raise errors.MergeModifiedFormatError()
636
907
except StopIteration:
637
raise MergeModifiedFormatError()
908
raise errors.MergeModifiedFormatError()
638
909
for s in RioReader(hashfile):
639
file_id = s.get("file_id")
910
# RioReader reads in Unicode, so convert file_ids back to utf8
911
file_id = osutils.safe_file_id(s.get("file_id"), warn=False)
640
912
if file_id not in self.inventory:
643
if hash == self.get_file_sha1(file_id):
644
merge_hashes[file_id] = hash
914
text_hash = s.get("hash")
915
if text_hash == self.get_file_sha1(file_id):
916
merge_hashes[file_id] = text_hash
645
917
return merge_hashes
920
def mkdir(self, path, file_id=None):
921
"""See MutableTree.mkdir()."""
923
file_id = generate_ids.gen_file_id(os.path.basename(path))
924
os.mkdir(self.abspath(path))
925
self.add(path, file_id, 'directory')
647
928
def get_symlink_target(self, file_id):
648
929
return os.readlink(self.id2abspath(file_id))
650
def file_class(self, filename):
651
if self.path2id(filename):
653
elif self.is_ignored(filename):
658
def list_files(self):
659
"""Recursively list all files as (path, class, kind, id).
932
def subsume(self, other_tree):
933
def add_children(inventory, entry):
934
for child_entry in entry.children.values():
935
inventory._byid[child_entry.file_id] = child_entry
936
if child_entry.kind == 'directory':
937
add_children(inventory, child_entry)
938
if other_tree.get_root_id() == self.get_root_id():
939
raise errors.BadSubsumeSource(self, other_tree,
940
'Trees have the same root')
942
other_tree_path = self.relpath(other_tree.basedir)
943
except errors.PathNotChild:
944
raise errors.BadSubsumeSource(self, other_tree,
945
'Tree is not contained by the other')
946
new_root_parent = self.path2id(osutils.dirname(other_tree_path))
947
if new_root_parent is None:
948
raise errors.BadSubsumeSource(self, other_tree,
949
'Parent directory is not versioned.')
950
# We need to ensure that the result of a fetch will have a
951
# versionedfile for the other_tree root, and only fetching into
952
# RepositoryKnit2 guarantees that.
953
if not self.branch.repository.supports_rich_root():
954
raise errors.SubsumeTargetNeedsUpgrade(other_tree)
955
other_tree.lock_tree_write()
957
new_parents = other_tree.get_parent_ids()
958
other_root = other_tree.inventory.root
959
other_root.parent_id = new_root_parent
960
other_root.name = osutils.basename(other_tree_path)
961
self.inventory.add(other_root)
962
add_children(self.inventory, other_root)
963
self._write_inventory(self.inventory)
964
# normally we don't want to fetch whole repositories, but i think
965
# here we really do want to consolidate the whole thing.
966
for parent_id in other_tree.get_parent_ids():
967
self.branch.fetch(other_tree.branch, parent_id)
968
self.add_parent_tree_id(parent_id)
971
other_tree.bzrdir.retire_bzrdir()
973
def _directory_is_tree_reference(self, relpath):
974
# as a special case, if a directory contains control files then
975
# it's a tree reference, except that the root of the tree is not
976
return relpath and osutils.isdir(self.abspath(relpath) + u"/.bzr")
977
# TODO: We could ask all the control formats whether they
978
# recognize this directory, but at the moment there's no cheap api
979
# to do that. Since we probably can only nest bzr checkouts and
980
# they always use this name it's ok for now. -- mbp 20060306
982
# FIXME: There is an unhandled case here of a subdirectory
983
# containing .bzr but not a branch; that will probably blow up
984
# when you try to commit it. It might happen if there is a
985
# checkout in a subdirectory. This can be avoided by not adding
988
@needs_tree_write_lock
989
def extract(self, file_id, format=None):
990
"""Extract a subtree from this tree.
992
A new branch will be created, relative to the path for this tree.
996
segments = osutils.splitpath(path)
997
transport = self.branch.bzrdir.root_transport
998
for name in segments:
999
transport = transport.clone(name)
1000
transport.ensure_base()
1003
sub_path = self.id2path(file_id)
1004
branch_transport = mkdirs(sub_path)
1006
format = bzrdir.format_registry.make_bzrdir('dirstate-with-subtree')
1007
branch_transport.ensure_base()
1008
branch_bzrdir = format.initialize_on_transport(branch_transport)
1010
repo = branch_bzrdir.find_repository()
1011
except errors.NoRepositoryPresent:
1012
repo = branch_bzrdir.create_repository()
1013
assert repo.supports_rich_root()
1015
if not repo.supports_rich_root():
1016
raise errors.RootNotRich()
1017
new_branch = branch_bzrdir.create_branch()
1018
new_branch.pull(self.branch)
1019
for parent_id in self.get_parent_ids():
1020
new_branch.fetch(self.branch, parent_id)
1021
tree_transport = self.bzrdir.root_transport.clone(sub_path)
1022
if tree_transport.base != branch_transport.base:
1023
tree_bzrdir = format.initialize_on_transport(tree_transport)
1024
branch.BranchReferenceFormat().initialize(tree_bzrdir, new_branch)
1026
tree_bzrdir = branch_bzrdir
1027
wt = tree_bzrdir.create_workingtree(NULL_REVISION)
1028
wt.set_parent_ids(self.get_parent_ids())
1029
my_inv = self.inventory
1030
child_inv = Inventory(root_id=None)
1031
new_root = my_inv[file_id]
1032
my_inv.remove_recursive_id(file_id)
1033
new_root.parent_id = None
1034
child_inv.add(new_root)
1035
self._write_inventory(my_inv)
1036
wt._write_inventory(child_inv)
1039
def _serialize(self, inventory, out_file):
1040
xml5.serializer_v5.write_inventory(self._inventory, out_file,
1043
def _deserialize(selt, in_file):
1044
return xml5.serializer_v5.read_inventory(in_file)
1047
"""Write the in memory inventory to disk."""
1048
# TODO: Maybe this should only write on dirty ?
1049
if self._control_files._lock_mode != 'w':
1050
raise errors.NotWriteLocked(self)
1052
self._serialize(self._inventory, sio)
1054
self._control_files.put('inventory', sio)
1055
self._inventory_is_modified = False
1057
def _kind(self, relpath):
1058
return osutils.file_kind(self.abspath(relpath))
1060
def list_files(self, include_root=False):
1061
"""Recursively list all files as (path, class, kind, id, entry).
661
1063
Lists, but does not descend into unversioned directories.
666
1068
Skips the control directory.
668
inv = self._inventory
670
def descend(from_dir_relpath, from_dir_id, dp):
1070
# list_files is an iterator, so @needs_read_lock doesn't work properly
1071
# with it. So callers should be careful to always read_lock the tree.
1072
if not self.is_locked():
1073
raise errors.ObjectNotLocked(self)
1075
inv = self.inventory
1076
if include_root is True:
1077
yield ('', 'V', 'directory', inv.root.file_id, inv.root)
1078
# Convert these into local objects to save lookup times
1079
pathjoin = osutils.pathjoin
1080
file_kind = self._kind
1082
# transport.base ends in a slash, we want the piece
1083
# between the last two slashes
1084
transport_base_dir = self.bzrdir.transport.base.rsplit('/', 2)[1]
1086
fk_entries = {'directory':TreeDirectory, 'file':TreeFile, 'symlink':TreeLink}
1088
# directory file_id, relative path, absolute path, reverse sorted children
1089
children = os.listdir(self.basedir)
1091
# jam 20060527 The kernel sized tree seems equivalent whether we
1092
# use a deque and popleft to keep them sorted, or if we use a plain
1093
# list and just reverse() them.
1094
children = collections.deque(children)
1095
stack = [(inv.root.file_id, u'', self.basedir, children)]
1097
from_dir_id, from_dir_relpath, from_dir_abspath, children = stack[-1]
1100
f = children.popleft()
674
1101
## TODO: If we find a subdirectory with its own .bzr
675
1102
## directory, then that is a separate tree and we
676
1103
## should exclude it.
678
1105
# the bzrdir for this tree
679
if self.bzrdir.transport.base.endswith(f + '/'):
1106
if transport_base_dir == f:
683
fp = appendpath(from_dir_relpath, f)
1109
# we know that from_dir_relpath and from_dir_abspath never end in a slash
1110
# and 'f' doesn't begin with one, we can do a string op, rather
1111
# than the checks of pathjoin(), all relative paths will have an extra slash
1113
fp = from_dir_relpath + '/' + f
686
fap = appendpath(dp, f)
1116
fap = from_dir_abspath + '/' + f
688
1118
f_ie = inv.get_child(from_dir_id, f)
691
elif self.is_ignored(fp):
1121
elif self.is_ignored(fp[1:]):
1124
# we may not have found this file, because of a unicode issue
1125
f_norm, can_access = osutils.normalized_filename(f)
1126
if f == f_norm or not can_access:
1127
# No change, so treat this file normally
1130
# this file can be accessed by a normalized path
1131
# check again if it is versioned
1132
# these lines are repeated here for performance
1134
fp = from_dir_relpath + '/' + f
1135
fap = from_dir_abspath + '/' + f
1136
f_ie = inv.get_child(from_dir_id, f)
1139
elif self.is_ignored(fp[1:]):
696
1144
fk = file_kind(fap)
700
raise BzrCheckError("file %r entered as kind %r id %r, "
702
% (fap, f_ie.kind, f_ie.file_id, fk))
704
1146
# make a last minute entry
1148
yield fp[1:], c, fk, f_ie.file_id, f_ie
708
if fk == 'directory':
709
entry = TreeDirectory()
712
elif fk == 'symlink':
1151
yield fp[1:], c, fk, None, fk_entries[fk]()
1153
yield fp[1:], c, fk, None, TreeEntry()
717
yield fp, c, fk, (f_ie and f_ie.file_id), entry
719
1156
if fk != 'directory':
723
# don't descend unversioned directories
726
for ff in descend(fp, f_ie.file_id, fap):
729
for f in descend(u'', inv.root.file_id, self.basedir):
733
def move(self, from_paths, to_name):
1159
# But do this child first
1160
new_children = os.listdir(fap)
1162
new_children = collections.deque(new_children)
1163
stack.append((f_ie.file_id, fp, fap, new_children))
1164
# Break out of inner loop,
1165
# so that we start outer loop with child
1168
# if we finished all children, pop it off the stack
1171
@needs_tree_write_lock
1172
def move(self, from_paths, to_dir=None, after=False, **kwargs):
734
1173
"""Rename files.
736
to_name must exist in the inventory.
1175
to_dir must exist in the inventory.
738
If to_name exists and is a directory, the files are moved into
1177
If to_dir exists and is a directory, the files are moved into
739
1178
it, keeping their old names.
741
Note that to_name is only the last component of the new name;
1180
Note that to_dir is only the last component of the new name;
742
1181
this doesn't change the directory.
1183
For each entry in from_paths the move mode will be determined
1186
The first mode moves the file in the filesystem and updates the
1187
inventory. The second mode only updates the inventory without
1188
touching the file on the filesystem. This is the new mode introduced
1191
move uses the second mode if 'after == True' and the target is not
1192
versioned but present in the working tree.
1194
move uses the second mode if 'after == False' and the source is
1195
versioned but no longer in the working tree, and the target is not
1196
versioned but present in the working tree.
1198
move uses the first mode if 'after == False' and the source is
1199
versioned and present in the working tree, and the target is not
1200
versioned and not present in the working tree.
1202
Everything else results in an error.
744
1204
This returns a list of (from_path, to_path) pairs for each
745
1205
entry that is moved.
748
## TODO: Option to move IDs only
1210
# check for deprecated use of signature
1212
to_dir = kwargs.get('to_name', None)
1214
raise TypeError('You must supply a target directory')
1216
symbol_versioning.warn('The parameter to_name was deprecated'
1217
' in version 0.13. Use to_dir instead',
1220
# check destination directory
749
1221
assert not isinstance(from_paths, basestring)
750
1222
inv = self.inventory
751
to_abs = self.abspath(to_name)
1223
to_abs = self.abspath(to_dir)
752
1224
if not isdir(to_abs):
753
raise BzrError("destination %r is not a directory" % to_abs)
754
if not self.has_filename(to_name):
755
raise BzrError("destination %r not in working directory" % to_abs)
756
to_dir_id = inv.path2id(to_name)
757
if to_dir_id == None and to_name != '':
758
raise BzrError("destination %r is not a versioned directory" % to_name)
1225
raise errors.BzrMoveFailedError('',to_dir,
1226
errors.NotADirectory(to_abs))
1227
if not self.has_filename(to_dir):
1228
raise errors.BzrMoveFailedError('',to_dir,
1229
errors.NotInWorkingDirectory(to_dir))
1230
to_dir_id = inv.path2id(to_dir)
1231
if to_dir_id is None:
1232
raise errors.BzrMoveFailedError('',to_dir,
1233
errors.NotVersionedError(path=str(to_dir)))
759
1235
to_dir_ie = inv[to_dir_id]
760
if to_dir_ie.kind not in ('directory', 'root_directory'):
761
raise BzrError("destination %r is not a directory" % to_abs)
763
to_idpath = inv.get_idpath(to_dir_id)
766
if not self.has_filename(f):
767
raise BzrError("%r does not exist in working tree" % f)
768
f_id = inv.path2id(f)
770
raise BzrError("%r is not versioned" % f)
771
name_tail = splitpath(f)[-1]
772
dest_path = appendpath(to_name, name_tail)
773
if self.has_filename(dest_path):
774
raise BzrError("destination %r already exists" % dest_path)
775
if f_id in to_idpath:
776
raise BzrError("can't move %r to a subdirectory of itself" % f)
778
# OK, so there's a race here, it's possible that someone will
779
# create a file in this interval and then the rename might be
780
# left half-done. But we should have caught most problems.
781
orig_inv = deepcopy(self.inventory)
1236
if to_dir_ie.kind != 'directory':
1237
raise errors.BzrMoveFailedError('',to_dir,
1238
errors.NotADirectory(to_abs))
1240
# create rename entries and tuples
1241
for from_rel in from_paths:
1242
from_tail = splitpath(from_rel)[-1]
1243
from_id = inv.path2id(from_rel)
1245
raise errors.BzrMoveFailedError(from_rel,to_dir,
1246
errors.NotVersionedError(path=str(from_rel)))
1248
from_entry = inv[from_id]
1249
from_parent_id = from_entry.parent_id
1250
to_rel = pathjoin(to_dir, from_tail)
1251
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1253
from_tail=from_tail,
1254
from_parent_id=from_parent_id,
1255
to_rel=to_rel, to_tail=from_tail,
1256
to_parent_id=to_dir_id)
1257
rename_entries.append(rename_entry)
1258
rename_tuples.append((from_rel, to_rel))
1260
# determine which move mode to use. checks also for movability
1261
rename_entries = self._determine_mv_mode(rename_entries, after)
1263
original_modified = self._inventory_is_modified
784
name_tail = splitpath(f)[-1]
785
dest_path = appendpath(to_name, name_tail)
786
result.append((f, dest_path))
787
inv.rename(inv.path2id(f), to_dir_id, name_tail)
789
rename(self.abspath(f), self.abspath(dest_path))
791
raise BzrError("failed to rename %r to %r: %s" %
792
(f, dest_path, e[1]),
793
["rename rolled back"])
1266
self._inventory_is_modified = True
1267
self._move(rename_entries)
795
1269
# restore the inventory on error
796
self._set_inventory(orig_inv)
1270
self._inventory_is_modified = original_modified
798
1272
self._write_inventory(inv)
802
def rename_one(self, from_rel, to_rel):
1273
return rename_tuples
1275
def _determine_mv_mode(self, rename_entries, after=False):
1276
"""Determines for each from-to pair if both inventory and working tree
1277
or only the inventory has to be changed.
1279
Also does basic plausability tests.
1281
inv = self.inventory
1283
for rename_entry in rename_entries:
1284
# store to local variables for easier reference
1285
from_rel = rename_entry.from_rel
1286
from_id = rename_entry.from_id
1287
to_rel = rename_entry.to_rel
1288
to_id = inv.path2id(to_rel)
1289
only_change_inv = False
1291
# check the inventory for source and destination
1293
raise errors.BzrMoveFailedError(from_rel,to_rel,
1294
errors.NotVersionedError(path=str(from_rel)))
1295
if to_id is not None:
1296
raise errors.BzrMoveFailedError(from_rel,to_rel,
1297
errors.AlreadyVersionedError(path=str(to_rel)))
1299
# try to determine the mode for rename (only change inv or change
1300
# inv and file system)
1302
if not self.has_filename(to_rel):
1303
raise errors.BzrMoveFailedError(from_id,to_rel,
1304
errors.NoSuchFile(path=str(to_rel),
1305
extra="New file has not been created yet"))
1306
only_change_inv = True
1307
elif not self.has_filename(from_rel) and self.has_filename(to_rel):
1308
only_change_inv = True
1309
elif self.has_filename(from_rel) and not self.has_filename(to_rel):
1310
only_change_inv = False
1312
# something is wrong, so lets determine what exactly
1313
if not self.has_filename(from_rel) and \
1314
not self.has_filename(to_rel):
1315
raise errors.BzrRenameFailedError(from_rel,to_rel,
1316
errors.PathsDoNotExist(paths=(str(from_rel),
1319
raise errors.RenameFailedFilesExist(from_rel, to_rel,
1320
extra="(Use --after to update the Bazaar id)")
1321
rename_entry.only_change_inv = only_change_inv
1322
return rename_entries
1324
def _move(self, rename_entries):
1325
"""Moves a list of files.
1327
Depending on the value of the flag 'only_change_inv', the
1328
file will be moved on the file system or not.
1330
inv = self.inventory
1333
for entry in rename_entries:
1335
self._move_entry(entry)
1337
self._rollback_move(moved)
1341
def _rollback_move(self, moved):
1342
"""Try to rollback a previous move in case of an filesystem error."""
1343
inv = self.inventory
1346
self._move_entry(_RenameEntry(entry.to_rel, entry.from_id,
1347
entry.to_tail, entry.to_parent_id, entry.from_rel,
1348
entry.from_tail, entry.from_parent_id,
1349
entry.only_change_inv))
1350
except errors.BzrMoveFailedError, e:
1351
raise errors.BzrMoveFailedError( '', '', "Rollback failed."
1352
" The working tree is in an inconsistent state."
1353
" Please consider doing a 'bzr revert'."
1354
" Error message is: %s" % e)
1356
def _move_entry(self, entry):
1357
inv = self.inventory
1358
from_rel_abs = self.abspath(entry.from_rel)
1359
to_rel_abs = self.abspath(entry.to_rel)
1360
if from_rel_abs == to_rel_abs:
1361
raise errors.BzrMoveFailedError(entry.from_rel, entry.to_rel,
1362
"Source and target are identical.")
1364
if not entry.only_change_inv:
1366
osutils.rename(from_rel_abs, to_rel_abs)
1368
raise errors.BzrMoveFailedError(entry.from_rel,
1370
inv.rename(entry.from_id, entry.to_parent_id, entry.to_tail)
1372
@needs_tree_write_lock
1373
def rename_one(self, from_rel, to_rel, after=False):
803
1374
"""Rename one file.
805
1376
This can change the directory or the filename or both.
1378
rename_one has several 'modes' to work. First, it can rename a physical
1379
file and change the file_id. That is the normal mode. Second, it can
1380
only change the file_id without touching any physical file. This is
1381
the new mode introduced in version 0.15.
1383
rename_one uses the second mode if 'after == True' and 'to_rel' is not
1384
versioned but present in the working tree.
1386
rename_one uses the second mode if 'after == False' and 'from_rel' is
1387
versioned but no longer in the working tree, and 'to_rel' is not
1388
versioned but present in the working tree.
1390
rename_one uses the first mode if 'after == False' and 'from_rel' is
1391
versioned and present in the working tree, and 'to_rel' is not
1392
versioned and not present in the working tree.
1394
Everything else results in an error.
807
1396
inv = self.inventory
808
if not self.has_filename(from_rel):
809
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
810
if self.has_filename(to_rel):
811
raise BzrError("can't rename: new working file %r already exists" % to_rel)
813
file_id = inv.path2id(from_rel)
815
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
818
from_parent = entry.parent_id
819
from_name = entry.name
821
if inv.path2id(to_rel):
822
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
1399
# create rename entries and tuples
1400
from_tail = splitpath(from_rel)[-1]
1401
from_id = inv.path2id(from_rel)
1403
raise errors.BzrRenameFailedError(from_rel,to_rel,
1404
errors.NotVersionedError(path=str(from_rel)))
1405
from_entry = inv[from_id]
1406
from_parent_id = from_entry.parent_id
824
1407
to_dir, to_tail = os.path.split(to_rel)
825
1408
to_dir_id = inv.path2id(to_dir)
826
if to_dir_id == None and to_dir != '':
827
raise BzrError("can't determine destination directory id for %r" % to_dir)
829
mutter("rename_one:")
830
mutter(" file_id {%s}" % file_id)
831
mutter(" from_rel %r" % from_rel)
832
mutter(" to_rel %r" % to_rel)
833
mutter(" to_dir %r" % to_dir)
834
mutter(" to_dir_id {%s}" % to_dir_id)
836
inv.rename(file_id, to_dir_id, to_tail)
838
from_abs = self.abspath(from_rel)
839
to_abs = self.abspath(to_rel)
841
rename(from_abs, to_abs)
843
inv.rename(file_id, from_parent, from_name)
844
raise BzrError("failed to rename %r to %r: %s"
845
% (from_abs, to_abs, e[1]),
846
["rename rolled back"])
1409
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1411
from_tail=from_tail,
1412
from_parent_id=from_parent_id,
1413
to_rel=to_rel, to_tail=to_tail,
1414
to_parent_id=to_dir_id)
1415
rename_entries.append(rename_entry)
1417
# determine which move mode to use. checks also for movability
1418
rename_entries = self._determine_mv_mode(rename_entries, after)
1420
# check if the target changed directory and if the target directory is
1422
if to_dir_id is None:
1423
raise errors.BzrMoveFailedError(from_rel,to_rel,
1424
errors.NotVersionedError(path=str(to_dir)))
1426
# all checks done. now we can continue with our actual work
1427
mutter('rename_one:\n'
1432
' to_dir_id {%s}\n',
1433
from_id, from_rel, to_rel, to_dir, to_dir_id)
1435
self._move(rename_entries)
847
1436
self._write_inventory(inv)
1438
class _RenameEntry(object):
1439
def __init__(self, from_rel, from_id, from_tail, from_parent_id,
1440
to_rel, to_tail, to_parent_id, only_change_inv=False):
1441
self.from_rel = from_rel
1442
self.from_id = from_id
1443
self.from_tail = from_tail
1444
self.from_parent_id = from_parent_id
1445
self.to_rel = to_rel
1446
self.to_tail = to_tail
1447
self.to_parent_id = to_parent_id
1448
self.only_change_inv = only_change_inv
849
1450
@needs_read_lock
850
1451
def unknowns(self):
851
1452
"""Return all unknown files.
853
1454
These are files in the working directory that are not versioned or
854
1455
control files or ignored.
856
>>> from bzrlib.bzrdir import ScratchDir
857
>>> d = ScratchDir(files=['foo', 'foo~'])
858
>>> b = d.open_branch()
859
>>> tree = d.open_workingtree()
860
>>> map(str, tree.unknowns())
863
>>> list(b.unknowns())
865
>>> tree.remove('foo')
866
>>> list(b.unknowns())
869
for subp in self.extras():
870
if not self.is_ignored(subp):
1457
# force the extras method to be fully executed before returning, to
1458
# prevent race conditions with the lock
1460
[subp for subp in self.extras() if not self.is_ignored(subp)])
1462
@needs_tree_write_lock
1463
def unversion(self, file_ids):
1464
"""Remove the file ids in file_ids from the current versioned set.
1466
When a file_id is unversioned, all of its children are automatically
1469
:param file_ids: The file ids to stop versioning.
1470
:raises: NoSuchId if any fileid is not currently versioned.
1472
for file_id in file_ids:
1473
if self._inventory.has_id(file_id):
1474
self._inventory.remove_recursive_id(file_id)
1476
raise errors.NoSuchId(self, file_id)
1478
# in the future this should just set a dirty bit to wait for the
1479
# final unlock. However, until all methods of workingtree start
1480
# with the current in -memory inventory rather than triggering
1481
# a read, it is more complex - we need to teach read_inventory
1482
# to know when to read, and when to not read first... and possibly
1483
# to save first when the in memory one may be corrupted.
1484
# so for now, we just only write it if it is indeed dirty.
1486
self._write_inventory(self._inventory)
873
1488
@deprecated_method(zero_eight)
874
1489
def iter_conflicts(self):
875
1490
"""List all files in the tree that have text or content conflicts.
1056
1750
This is used to allow WorkingTree3 instances to not affect branch
1057
1751
when their last revision is set.
1059
if new_revision is None:
1753
if _mod_revision.is_null(new_revision):
1060
1754
self.branch.set_revision_history([])
1062
# current format is locked in with the branch
1063
revision_history = self.branch.revision_history()
1065
position = revision_history.index(new_revision)
1067
raise errors.NoSuchRevision(self.branch, new_revision)
1068
self.branch.set_revision_history(revision_history[:position + 1])
1757
self.branch.generate_revision_history(new_revision)
1758
except errors.NoSuchRevision:
1759
# not present in the repo - dont try to set it deeper than the tip
1760
self.branch.set_revision_history([new_revision])
1763
def _write_basis_inventory(self, xml):
1764
"""Write the basis inventory XML to the basis-inventory file"""
1765
assert isinstance(xml, str), 'serialised xml must be bytestring.'
1766
path = self._basis_inventory_name()
1768
self._control_files.put(path, sio)
1770
def _create_basis_xml_from_inventory(self, revision_id, inventory):
1771
"""Create the text that will be saved in basis-inventory"""
1772
inventory.revision_id = revision_id
1773
return xml7.serializer_v7.write_inventory_to_string(inventory)
1071
1775
def _cache_basis_inventory(self, new_revision):
1072
1776
"""Cache new_revision as the basis inventory."""
1777
# TODO: this should allow the ready-to-use inventory to be passed in,
1778
# as commit already has that ready-to-use [while the format is the
1074
1781
# this double handles the inventory - unpack and repack -
1075
1782
# but is easier to understand. We can/should put a conditional
1076
1783
# in here based on whether the inventory is in the latest format
1077
1784
# - perhaps we should repack all inventories on a repository
1079
inv = self.branch.repository.get_inventory(new_revision)
1080
inv.revision_id = new_revision
1081
xml = bzrlib.xml5.serializer_v5.write_inventory_to_string(inv)
1083
path = self._basis_inventory_name()
1084
self._control_files.put_utf8(path, xml)
1085
except WeaveRevisionNotPresent:
1786
# the fast path is to copy the raw xml from the repository. If the
1787
# xml contains 'revision_id="', then we assume the right
1788
# revision_id is set. We must check for this full string, because a
1789
# root node id can legitimately look like 'revision_id' but cannot
1791
xml = self.branch.repository.get_inventory_xml(new_revision)
1792
firstline = xml.split('\n', 1)[0]
1793
if (not 'revision_id="' in firstline or
1794
'format="7"' not in firstline):
1795
inv = self.branch.repository.deserialise_inventory(
1797
xml = self._create_basis_xml_from_inventory(new_revision, inv)
1798
self._write_basis_inventory(xml)
1799
except (errors.NoSuchRevision, errors.RevisionNotPresent):
1088
1802
def read_basis_inventory(self):
1089
1803
"""Read the cached basis inventory."""
1090
1804
path = self._basis_inventory_name()
1091
return self._control_files.get_utf8(path).read()
1805
return self._control_files.get(path).read()
1093
1807
@needs_read_lock
1094
1808
def read_working_inventory(self):
1095
"""Read the working inventory."""
1809
"""Read the working inventory.
1811
:raises errors.InventoryModified: read_working_inventory will fail
1812
when the current in memory inventory has been modified.
1814
# conceptually this should be an implementation detail of the tree.
1815
# XXX: Deprecate this.
1096
1816
# ElementTree does its own conversion from UTF-8, so open in
1098
result = bzrlib.xml5.serializer_v5.read_inventory(
1099
self._control_files.get('inventory'))
1100
self._set_inventory(result)
1818
if self._inventory_is_modified:
1819
raise errors.InventoryModified(self)
1820
result = self._deserialize(self._control_files.get('inventory'))
1821
self._set_inventory(result, dirty=False)
1104
def remove(self, files, verbose=False):
1105
"""Remove nominated files from the working inventory..
1107
This does not remove their text. This does not run on XXX on what? RBC
1109
TODO: Refuse to remove modified files unless --force is given?
1111
TODO: Do something useful with directories.
1113
TODO: Should this remove the text or not? Tough call; not
1114
removing may be useful and the user can just use use rm, and
1115
is the opposite of add. Removing it is consistent with most
1116
other tools. Maybe an option.
1824
@needs_tree_write_lock
1825
def remove(self, files, verbose=False, to_file=None, keep_files=True,
1827
"""Remove nominated files from the working inventory.
1829
:files: File paths relative to the basedir.
1830
:keep_files: If true, the files will also be kept.
1831
:force: Delete files and directories, even if they are changed and
1832
even if the directories are not empty.
1118
## TODO: Normalize names
1119
## TODO: Remove nested loops; better scalability
1120
1834
if isinstance(files, basestring):
1121
1835
files = [files]
1123
inv = self.inventory
1125
# do this before any modifications
1840
unknown_nested_files=set()
1842
def recurse_directory_to_add_files(directory):
1843
# Recurse directory and add all files
1844
# so we can check if they have changed.
1845
for parent_info, file_infos in\
1846
osutils.walkdirs(self.abspath(directory),
1848
for relpath, basename, kind, lstat, abspath in file_infos:
1849
# Is it versioned or ignored?
1850
if self.path2id(relpath) or self.is_ignored(relpath):
1851
# Add nested content for deletion.
1852
new_files.add(relpath)
1854
# Files which are not versioned and not ignored
1855
# should be treated as unknown.
1856
unknown_nested_files.add((relpath, None, kind))
1858
for filename in files:
1859
# Get file name into canonical form.
1860
abspath = self.abspath(filename)
1861
filename = self.relpath(abspath)
1862
if len(filename) > 0:
1863
new_files.add(filename)
1864
if osutils.isdir(abspath):
1865
recurse_directory_to_add_files(filename)
1867
files = list(new_files)
1870
return # nothing to do
1872
# Sort needed to first handle directory content before the directory
1873
files.sort(reverse=True)
1875
# Bail out if we are going to delete files we shouldn't
1876
if not keep_files and not force:
1877
has_changed_files = len(unknown_nested_files) > 0
1878
if not has_changed_files:
1879
for (file_id, path, content_change, versioned, parent_id, name,
1880
kind, executable) in self._iter_changes(self.basis_tree(),
1881
include_unchanged=True, require_versioned=False,
1882
want_unversioned=True, specific_files=files):
1883
# Check if it's an unknown (but not ignored) OR
1884
# changed (but not deleted) :
1885
if not self.is_ignored(path[1]) and (
1886
versioned == (False, False) or
1887
content_change and kind[1] != None):
1888
has_changed_files = True
1891
if has_changed_files:
1892
# Make delta show ALL applicable changes in error message.
1893
tree_delta = self.changes_from(self.basis_tree(),
1894
require_versioned=False, want_unversioned=True,
1895
specific_files=files)
1896
for unknown_file in unknown_nested_files:
1897
if unknown_file not in tree_delta.unversioned:
1898
tree_delta.unversioned.extend((unknown_file,))
1899
raise errors.BzrRemoveChangedFilesError(tree_delta)
1901
# Build inv_delta and delete files where applicaple,
1902
# do this before any modifications to inventory.
1126
1903
for f in files:
1127
fid = inv.path2id(f)
1904
fid = self.path2id(f)
1129
# TODO: Perhaps make this just a warning, and continue?
1130
# This tends to happen when
1131
raise NotVersionedError(path=f)
1132
mutter("remove inventory entry %s {%s}", quotefn(f), fid)
1134
# having remove it, it must be either ignored or unknown
1135
if self.is_ignored(f):
1139
show_status(new_status, inv[fid].kind, quotefn(f))
1142
self._write_inventory(inv)
1145
def revert(self, filenames, old_tree=None, backups=True,
1146
pb=DummyProgress()):
1147
from transform import revert
1148
from conflicts import resolve
1907
message = "%s is not versioned." % (f,)
1910
# having removed it, it must be either ignored or unknown
1911
if self.is_ignored(f):
1915
textui.show_status(new_status, self.kind(fid), f,
1918
inv_delta.append((f, None, fid, None))
1919
message = "removed %s" % (f,)
1922
abs_path = self.abspath(f)
1923
if osutils.lexists(abs_path):
1924
if (osutils.isdir(abs_path) and
1925
len(os.listdir(abs_path)) > 0):
1927
osutils.rmtree(abs_path)
1929
message = "%s is not an empty directory "\
1930
"and won't be deleted." % (f,)
1932
osutils.delete_any(abs_path)
1933
message = "deleted %s" % (f,)
1934
elif message is not None:
1935
# Only care if we haven't done anything yet.
1936
message = "%s does not exist." % (f,)
1938
# Print only one message (if any) per file.
1939
if message is not None:
1941
self.apply_inventory_delta(inv_delta)
1943
@needs_tree_write_lock
1944
def revert(self, filenames=None, old_tree=None, backups=True,
1945
pb=DummyProgress(), report_changes=False):
1946
from bzrlib.conflicts import resolve
1949
symbol_versioning.warn('Using [] to revert all files is deprecated'
1950
' as of bzr 0.91. Please use None (the default) instead.',
1951
DeprecationWarning, stacklevel=2)
1149
1952
if old_tree is None:
1150
1953
old_tree = self.basis_tree()
1151
conflicts = revert(self, old_tree, filenames, backups, pb)
1152
if not len(filenames):
1153
self.set_pending_merges([])
1954
conflicts = transform.revert(self, old_tree, filenames, backups, pb,
1956
if filenames is None:
1957
self.set_parent_ids(self.get_parent_ids()[:1])
1156
1960
resolve(self, filenames, ignore_misses=True)
1157
1961
return conflicts
1963
def revision_tree(self, revision_id):
1964
"""See Tree.revision_tree.
1966
WorkingTree can supply revision_trees for the basis revision only
1967
because there is only one cached inventory in the bzr directory.
1969
if revision_id == self.last_revision():
1971
xml = self.read_basis_inventory()
1972
except errors.NoSuchFile:
1976
inv = xml7.serializer_v7.read_inventory_from_string(xml)
1977
# dont use the repository revision_tree api because we want
1978
# to supply the inventory.
1979
if inv.revision_id == revision_id:
1980
return revisiontree.RevisionTree(self.branch.repository,
1982
except errors.BadInventoryFormat:
1984
# raise if there was no inventory, or if we read the wrong inventory.
1985
raise errors.NoSuchRevisionInTree(self, revision_id)
1159
1987
# XXX: This method should be deprecated in favour of taking in a proper
1160
1988
# new Inventory object.
1989
@needs_tree_write_lock
1162
1990
def set_inventory(self, new_inventory_list):
1163
1991
from bzrlib.inventory import (Inventory,
1164
1992
InventoryDirectory,
1204
2058
between multiple working trees, i.e. via shared storage, then we
1205
2059
would probably want to lock both the local tree, and the branch.
1207
# FIXME: We want to write out the hashcache only when the last lock on
1208
# this working copy is released. Peeking at the lock count is a bit
1209
# of a nasty hack; probably it's better to have a transaction object,
1210
# which can do some finalization when it's either successfully or
1211
# unsuccessfully completed. (Denys's original patch did that.)
1212
# RBC 20060206 hookinhg into transaction will couple lock and transaction
1213
# wrongly. Hookinh into unllock on the control files object is fine though.
1215
# TODO: split this per format so there is no ugly if block
1216
if self._hashcache.needs_write and (
1217
# dedicated lock files
1218
self._control_files._lock_count==1 or
1220
(self._control_files is self.branch.control_files and
1221
self._control_files._lock_count==3)):
1222
self._hashcache.write()
1223
# reverse order of locking.
1224
result = self._control_files.unlock()
1226
self.branch.unlock()
2061
raise NotImplementedError(self.unlock)
2063
def update(self, change_reporter=None, possible_transports=None):
1232
2064
"""Update a working tree along its branch.
1234
This will update the branch if its bound too, which means we have multiple trees involved:
1235
The new basis tree of the master.
1236
The old basis tree of the branch.
1237
The old basis tree of the working tree.
1238
The current working tree state.
1239
pathologically all three may be different, and non ancestors of each other.
1240
Conceptually we want to:
1241
Preserve the wt.basis->wt.state changes
1242
Transform the wt.basis to the new master basis.
1243
Apply a merge of the old branch basis to get any 'local' changes from it into the tree.
1244
Restore the wt.basis->wt.state changes.
2066
This will update the branch if its bound too, which means we have
2067
multiple trees involved:
2069
- The new basis tree of the master.
2070
- The old basis tree of the branch.
2071
- The old basis tree of the working tree.
2072
- The current working tree state.
2074
Pathologically, all three may be different, and non-ancestors of each
2075
other. Conceptually we want to:
2077
- Preserve the wt.basis->wt.state changes
2078
- Transform the wt.basis to the new master basis.
2079
- Apply a merge of the old branch basis to get any 'local' changes from
2081
- Restore the wt.basis->wt.state changes.
1246
2083
There isn't a single operation at the moment to do that, so we:
1247
Merge current state -> basis tree of the master w.r.t. the old tree basis.
1248
Do a 'normal' merge of the old branch basis if it is relevant.
1250
old_tip = self.branch.update()
1251
if old_tip is not None:
1252
self.add_pending_merge(old_tip)
1253
self.branch.lock_read()
1256
if self.last_revision() != self.branch.last_revision():
1257
# merge tree state up to new branch tip.
1258
basis = self.basis_tree()
2084
- Merge current state -> basis tree of the master w.r.t. the old tree
2086
- Do a 'normal' merge of the old branch basis if it is relevant.
2088
if self.branch.get_master_branch(possible_transports) is not None:
2090
update_branch = True
2092
self.lock_tree_write()
2093
update_branch = False
2096
old_tip = self.branch.update(possible_transports)
2099
return self._update_tree(old_tip, change_reporter)
2103
@needs_tree_write_lock
2104
def _update_tree(self, old_tip=None, change_reporter=None):
2105
"""Update a tree to the master branch.
2107
:param old_tip: if supplied, the previous tip revision the branch,
2108
before it was changed to the master branch's tip.
2110
# here if old_tip is not None, it is the old tip of the branch before
2111
# it was updated from the master branch. This should become a pending
2112
# merge in the working tree to preserve the user existing work. we
2113
# cant set that until we update the working trees last revision to be
2114
# one from the new branch, because it will just get absorbed by the
2115
# parent de-duplication logic.
2117
# We MUST save it even if an error occurs, because otherwise the users
2118
# local work is unreferenced and will appear to have been lost.
2122
last_rev = self.get_parent_ids()[0]
2124
last_rev = _mod_revision.NULL_REVISION
2125
if last_rev != _mod_revision.ensure_null(self.branch.last_revision()):
2126
# merge tree state up to new branch tip.
2127
basis = self.basis_tree()
1259
2130
to_tree = self.branch.basis_tree()
1260
result += merge_inner(self.branch,
2131
if basis.inventory.root is None:
2132
self.set_root_id(to_tree.inventory.root.file_id)
2134
result += merge.merge_inner(
1264
self.set_last_revision(self.branch.last_revision())
1265
if old_tip and old_tip != self.last_revision():
1266
# our last revision was not the prior branch last reivison
1267
# and we have converted that last revision to a pending merge.
1268
# base is somewhere between the branch tip now
1269
# and the now pending merge
1270
from bzrlib.revision import common_ancestor
1272
base_rev_id = common_ancestor(self.branch.last_revision(),
1274
self.branch.repository)
1275
except errors.NoCommonAncestor:
1277
base_tree = self.branch.repository.revision_tree(base_rev_id)
1278
other_tree = self.branch.repository.revision_tree(old_tip)
1279
result += merge_inner(self.branch,
1285
self.branch.unlock()
2139
change_reporter=change_reporter)
2142
# TODO - dedup parents list with things merged by pull ?
2143
# reuse the tree we've updated to to set the basis:
2144
parent_trees = [(self.branch.last_revision(), to_tree)]
2145
merges = self.get_parent_ids()[1:]
2146
# Ideally we ask the tree for the trees here, that way the working
2147
# tree can decide whether to give us teh entire tree or give us a
2148
# lazy initialised tree. dirstate for instance will have the trees
2149
# in ram already, whereas a last-revision + basis-inventory tree
2150
# will not, but also does not need them when setting parents.
2151
for parent in merges:
2152
parent_trees.append(
2153
(parent, self.branch.repository.revision_tree(parent)))
2154
if (old_tip is not None and not _mod_revision.is_null(old_tip)):
2155
parent_trees.append(
2156
(old_tip, self.branch.repository.revision_tree(old_tip)))
2157
self.set_parent_trees(parent_trees)
2158
last_rev = parent_trees[0][0]
2160
# the working tree had the same last-revision as the master
2161
# branch did. We may still have pivot local work from the local
2162
# branch into old_tip:
2163
if (old_tip is not None and not _mod_revision.is_null(old_tip)):
2164
self.add_parent_tree_id(old_tip)
2165
if (old_tip is not None and not _mod_revision.is_null(old_tip)
2166
and old_tip != last_rev):
2167
# our last revision was not the prior branch last revision
2168
# and we have converted that last revision to a pending merge.
2169
# base is somewhere between the branch tip now
2170
# and the now pending merge
2172
# Since we just modified the working tree and inventory, flush out
2173
# the current state, before we modify it again.
2174
# TODO: jam 20070214 WorkingTree3 doesn't require this, dirstate
2175
# requires it only because TreeTransform directly munges the
2176
# inventory and calls tree._write_inventory(). Ultimately we
2177
# should be able to remove this extra flush.
2179
graph = self.branch.repository.get_graph()
2180
base_rev_id = graph.find_unique_lca(self.branch.last_revision(),
2182
base_tree = self.branch.repository.revision_tree(base_rev_id)
2183
other_tree = self.branch.repository.revision_tree(old_tip)
2184
result += merge.merge_inner(
2189
change_reporter=change_reporter)
2192
def _write_hashcache_if_dirty(self):
2193
"""Write out the hashcache if it is dirty."""
2194
if self._hashcache.needs_write:
2196
self._hashcache.write()
2198
if e.errno not in (errno.EPERM, errno.EACCES):
2200
# TODO: jam 20061219 Should this be a warning? A single line
2201
# warning might be sufficient to let the user know what
2203
mutter('Could not write hashcache for %s\nError: %s',
2204
self._hashcache.cache_file_name(), e)
2206
@needs_tree_write_lock
1288
2207
def _write_inventory(self, inv):
1289
2208
"""Write inventory as the current inventory."""
1291
bzrlib.xml5.serializer_v5.write_inventory(inv, sio)
1293
self._control_files.put('inventory', sio)
1294
self._set_inventory(inv)
1295
mutter('wrote working inventory')
2209
self._set_inventory(inv, dirty=True)
1297
2212
def set_conflicts(self, arg):
1298
raise UnsupportedOperation(self.set_conflicts, self)
2213
raise errors.UnsupportedOperation(self.set_conflicts, self)
2215
def add_conflicts(self, arg):
2216
raise errors.UnsupportedOperation(self.add_conflicts, self)
1300
2218
@needs_read_lock
1301
2219
def conflicts(self):
1302
conflicts = ConflictList()
2220
conflicts = _mod_conflicts.ConflictList()
1303
2221
for conflicted in self._iter_conflicts():
1306
2224
if file_kind(self.abspath(conflicted)) != "file":
1309
if e.errno == errno.ENOENT:
2226
except errors.NoSuchFile:
1313
2228
if text is True:
1314
2229
for suffix in ('.THIS', '.OTHER'):
1316
2231
kind = file_kind(self.abspath(conflicted+suffix))
1318
if e.errno == errno.ENOENT:
2234
except errors.NoSuchFile:
1326
2238
ctype = {True: 'text conflict', False: 'contents conflict'}[text]
1327
conflicts.append(Conflict.factory(ctype, path=conflicted,
2239
conflicts.append(_mod_conflicts.Conflict.factory(ctype,
1328
2241
file_id=self.path2id(conflicted)))
1329
2242
return conflicts
2244
def walkdirs(self, prefix=""):
2245
"""Walk the directories of this tree.
2247
returns a generator which yields items in the form:
2248
((curren_directory_path, fileid),
2249
[(file1_path, file1_name, file1_kind, (lstat), file1_id,
2252
This API returns a generator, which is only valid during the current
2253
tree transaction - within a single lock_read or lock_write duration.
2255
If the tree is not locked, it may cause an error to be raised,
2256
depending on the tree implementation.
2258
disk_top = self.abspath(prefix)
2259
if disk_top.endswith('/'):
2260
disk_top = disk_top[:-1]
2261
top_strip_len = len(disk_top) + 1
2262
inventory_iterator = self._walkdirs(prefix)
2263
disk_iterator = osutils.walkdirs(disk_top, prefix)
2265
current_disk = disk_iterator.next()
2266
disk_finished = False
2268
if not (e.errno == errno.ENOENT or
2269
(sys.platform == 'win32' and e.errno == ERROR_PATH_NOT_FOUND)):
2272
disk_finished = True
2274
current_inv = inventory_iterator.next()
2275
inv_finished = False
2276
except StopIteration:
2279
while not inv_finished or not disk_finished:
2280
if not disk_finished:
2281
# strip out .bzr dirs
2282
if current_disk[0][1][top_strip_len:] == '':
2283
# osutils.walkdirs can be made nicer -
2284
# yield the path-from-prefix rather than the pathjoined
2286
bzrdir_loc = bisect_left(current_disk[1], ('.bzr', '.bzr'))
2287
if current_disk[1][bzrdir_loc][0] == '.bzr':
2288
# we dont yield the contents of, or, .bzr itself.
2289
del current_disk[1][bzrdir_loc]
2291
# everything is unknown
2294
# everything is missing
2297
direction = cmp(current_inv[0][0], current_disk[0][0])
2299
# disk is before inventory - unknown
2300
dirblock = [(relpath, basename, kind, stat, None, None) for
2301
relpath, basename, kind, stat, top_path in current_disk[1]]
2302
yield (current_disk[0][0], None), dirblock
2304
current_disk = disk_iterator.next()
2305
except StopIteration:
2306
disk_finished = True
2308
# inventory is before disk - missing.
2309
dirblock = [(relpath, basename, 'unknown', None, fileid, kind)
2310
for relpath, basename, dkind, stat, fileid, kind in
2312
yield (current_inv[0][0], current_inv[0][1]), dirblock
2314
current_inv = inventory_iterator.next()
2315
except StopIteration:
2318
# versioned present directory
2319
# merge the inventory and disk data together
2321
for relpath, subiterator in itertools.groupby(sorted(
2322
current_inv[1] + current_disk[1], key=operator.itemgetter(0)), operator.itemgetter(1)):
2323
path_elements = list(subiterator)
2324
if len(path_elements) == 2:
2325
inv_row, disk_row = path_elements
2326
# versioned, present file
2327
dirblock.append((inv_row[0],
2328
inv_row[1], disk_row[2],
2329
disk_row[3], inv_row[4],
2331
elif len(path_elements[0]) == 5:
2333
dirblock.append((path_elements[0][0],
2334
path_elements[0][1], path_elements[0][2],
2335
path_elements[0][3], None, None))
2336
elif len(path_elements[0]) == 6:
2337
# versioned, absent file.
2338
dirblock.append((path_elements[0][0],
2339
path_elements[0][1], 'unknown', None,
2340
path_elements[0][4], path_elements[0][5]))
2342
raise NotImplementedError('unreachable code')
2343
yield current_inv[0], dirblock
2345
current_inv = inventory_iterator.next()
2346
except StopIteration:
2349
current_disk = disk_iterator.next()
2350
except StopIteration:
2351
disk_finished = True
2353
def _walkdirs(self, prefix=""):
2354
"""Walk the directories of this tree.
2356
:prefix: is used as the directrory to start with.
2357
returns a generator which yields items in the form:
2358
((curren_directory_path, fileid),
2359
[(file1_path, file1_name, file1_kind, None, file1_id,
2362
_directory = 'directory'
2363
# get the root in the inventory
2364
inv = self.inventory
2365
top_id = inv.path2id(prefix)
2369
pending = [(prefix, '', _directory, None, top_id, None)]
2372
currentdir = pending.pop()
2373
# 0 - relpath, 1- basename, 2- kind, 3- stat, 4-id, 5-kind
2374
top_id = currentdir[4]
2376
relroot = currentdir[0] + '/'
2379
# FIXME: stash the node in pending
2381
for name, child in entry.sorted_children():
2382
dirblock.append((relroot + name, name, child.kind, None,
2383
child.file_id, child.kind
2385
yield (currentdir[0], entry.file_id), dirblock
2386
# push the user specified dirs from dirblock
2387
for dir in reversed(dirblock):
2388
if dir[2] == _directory:
2391
@needs_tree_write_lock
2392
def auto_resolve(self):
2393
"""Automatically resolve text conflicts according to contents.
2395
Only text conflicts are auto_resolvable. Files with no conflict markers
2396
are considered 'resolved', because bzr always puts conflict markers
2397
into files that have text conflicts. The corresponding .THIS .BASE and
2398
.OTHER files are deleted, as per 'resolve'.
2399
:return: a tuple of ConflictLists: (un_resolved, resolved).
2401
un_resolved = _mod_conflicts.ConflictList()
2402
resolved = _mod_conflicts.ConflictList()
2403
conflict_re = re.compile('^(<{7}|={7}|>{7})')
2404
for conflict in self.conflicts():
2405
if (conflict.typestring != 'text conflict' or
2406
self.kind(conflict.file_id) != 'file'):
2407
un_resolved.append(conflict)
2409
my_file = open(self.id2abspath(conflict.file_id), 'rb')
2411
for line in my_file:
2412
if conflict_re.search(line):
2413
un_resolved.append(conflict)
2416
resolved.append(conflict)
2419
resolved.remove_files(self)
2420
self.set_conflicts(un_resolved)
2421
return un_resolved, resolved
2423
def _validate(self):
2424
"""Validate internal structures.
2426
This is meant mostly for the test suite. To give it a chance to detect
2427
corruption after actions have occurred. The default implementation is a
2430
:return: None. An exception should be raised if there is an error.
2435
class WorkingTree2(WorkingTree):
2436
"""This is the Format 2 working tree.
2438
This was the first weave based working tree.
2439
- uses os locks for locking.
2440
- uses the branch last-revision.
2443
def __init__(self, *args, **kwargs):
2444
super(WorkingTree2, self).__init__(*args, **kwargs)
2445
# WorkingTree2 has more of a constraint that self._inventory must
2446
# exist. Because this is an older format, we don't mind the overhead
2447
# caused by the extra computation here.
2449
# Newer WorkingTree's should only have self._inventory set when they
2451
if self._inventory is None:
2452
self.read_working_inventory()
2454
def lock_tree_write(self):
2455
"""See WorkingTree.lock_tree_write().
2457
In Format2 WorkingTrees we have a single lock for the branch and tree
2458
so lock_tree_write() degrades to lock_write().
2460
self.branch.lock_write()
2462
return self._control_files.lock_write()
2464
self.branch.unlock()
2468
# do non-implementation specific cleanup
2471
# we share control files:
2472
if self._control_files._lock_count == 3:
2473
# _inventory_is_modified is always False during a read lock.
2474
if self._inventory_is_modified:
2476
self._write_hashcache_if_dirty()
2478
# reverse order of locking.
2480
return self._control_files.unlock()
2482
self.branch.unlock()
1332
2485
class WorkingTree3(WorkingTree):
1333
2486
"""This is the Format 3 working tree.