1
1
# Copyright (C) 2005, 2006 Canonical Ltd
3
3
# This program is free software; you can redistribute it and/or modify
4
4
# it under the terms of the GNU General Public License as published by
5
5
# the Free Software Foundation; either version 2 of the License, or
6
6
# (at your option) any later version.
8
8
# This program is distributed in the hope that it will be useful,
9
9
# but WITHOUT ANY WARRANTY; without even the implied warranty of
10
10
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11
11
# GNU General Public License for more details.
13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
39
39
# At the moment they may alias the inventory and have old copies of it in
40
40
# memory. (Now done? -- mbp 20060309)
42
from binascii import hexlify
42
44
from copy import deepcopy
43
45
from cStringIO import StringIO
55
from bzrlib import bzrdir, errors, ignores, osutils, urlutils
50
56
from bzrlib.atomicfile import AtomicFile
51
from bzrlib.branch import (Branch,
53
58
from bzrlib.conflicts import Conflict, ConflictList, CONFLICT_SUFFIXES
54
import bzrlib.bzrdir as bzrdir
55
59
from bzrlib.decorators import needs_read_lock, needs_write_lock
56
import bzrlib.errors as errors
57
60
from bzrlib.errors import (BzrCheckError,
59
62
ConflictFormatError,
61
63
WeaveRevisionNotPresent,
90
91
from bzrlib.progress import DummyProgress, ProgressPhase
91
92
from bzrlib.revision import NULL_REVISION
92
93
from bzrlib.rio import RioReader, rio_file, Stanza
93
from bzrlib.symbol_versioning import *
94
from bzrlib.textui import show_status
94
from bzrlib.symbol_versioning import (deprecated_passed,
100
from bzrlib.trace import mutter, note
96
101
from bzrlib.transform import build_tree
97
from bzrlib.trace import mutter, note
98
102
from bzrlib.transport import get_transport
99
103
from bzrlib.transport.local import LocalTransport
104
from bzrlib.textui import show_status
101
107
import bzrlib.xml5
110
# the regex removes any weird characters; we don't escape them
111
# but rather just pull them out
112
_gen_file_id_re = re.compile(r'[^\w.]')
113
_gen_id_suffix = None
117
def _next_id_suffix():
118
"""Create a new file id suffix that is reasonably unique.
120
On the first call we combine the current time with 64 bits of randomness
121
to give a highly probably globally unique number. Then each call in the same
122
process adds 1 to a serial number we append to that unique value.
124
# XXX TODO: change bzrlib.add.smart_add to call workingtree.add() rather
125
# than having to move the id randomness out of the inner loop like this.
126
# XXX TODO: for the global randomness this uses we should add the thread-id
127
# before the serial #.
128
global _gen_id_suffix, _gen_id_serial
129
if _gen_id_suffix is None:
130
_gen_id_suffix = "-%s-%s-" % (compact_date(time()), rand_chars(16))
132
return _gen_id_suffix + str(_gen_id_serial)
104
135
def gen_file_id(name):
105
"""Return new file id.
107
This should probably generate proper UUIDs, but for the moment we
108
cope with just randomness because running uuidgen every time is
111
from binascii import hexlify
112
from time import time
115
idx = name.rfind('/')
117
name = name[idx+1 : ]
118
idx = name.rfind('\\')
120
name = name[idx+1 : ]
122
# make it not a hidden file
123
name = name.lstrip('.')
125
# remove any wierd characters; we don't escape them but rather
127
name = re.sub(r'[^\w.]', '', name)
129
s = hexlify(rand_bytes(8))
130
return '-'.join((name, compact_date(time()), s))
136
"""Return new file id for the basename 'name'.
138
The uniqueness is supplied from _next_id_suffix.
140
# The real randomness is in the _next_id_suffix, the
141
# rest of the identifier is just to be nice.
143
# 1) Remove non-ascii word characters to keep the ids portable
144
# 2) squash to lowercase, so the file id doesn't have to
145
# be escaped (case insensitive filesystems would bork for ids
146
# that only differred in case without escaping).
147
# 3) truncate the filename to 20 chars. Long filenames also bork on some
149
# 4) Removing starting '.' characters to prevent the file ids from
150
# being considered hidden.
151
ascii_word_only = _gen_file_id_re.sub('', name.lower())
152
short_no_dots = ascii_word_only.lstrip('.')[:20]
153
return short_no_dots + _next_id_suffix()
133
156
def gen_root_id():
238
261
mutter("opening working tree %r", basedir)
239
262
if deprecated_passed(branch):
240
263
if not _internal:
241
warn("WorkingTree(..., branch=XXX) is deprecated as of bzr 0.8."
264
warnings.warn("WorkingTree(..., branch=XXX) is deprecated as of bzr 0.8."
242
265
" Please use bzrdir.open_workingtree() or"
243
266
" WorkingTree.open().",
244
267
DeprecationWarning,
270
self._branch = branch
249
self.branch = self.bzrdir.open_branch()
250
assert isinstance(self.branch, Branch), \
251
"branch %r is not a Branch" % self.branch
272
self._branch = self.bzrdir.open_branch()
252
273
self.basedir = realpath(basedir)
253
274
# if branch is at our basedir and is a format 6 or less
254
275
if isinstance(self._format, WorkingTreeFormat2):
255
276
# share control object
256
277
self._control_files = self.branch.control_files
258
# only ready for format 3
259
assert isinstance(self._format, WorkingTreeFormat3)
279
# assume all other formats have their own control files.
260
280
assert isinstance(_control_files, LockableFiles), \
261
281
"_control_files must be a LockableFiles, not %r" \
283
303
self._set_inventory(_inventory)
306
fget=lambda self: self._branch,
307
doc="""The branch this WorkingTree is connected to.
309
This cannot be set - it is reflective of the actual disk structure
310
the working tree has been constructed from.
313
def break_lock(self):
314
"""Break a lock if one is present from another instance.
316
Uses the ui factory to ask for confirmation if the lock may be from
319
This will probe the repository for its lock as well.
321
self._control_files.break_lock()
322
self.branch.break_lock()
285
324
def _set_inventory(self, inv):
325
assert inv.root is not None
286
326
self._inventory = inv
287
327
self.path2id = self._inventory.path2id
289
329
def is_control_filename(self, filename):
290
330
"""True if filename is the name of a control file in this tree.
332
:param filename: A filename within the tree. This is a relative path
333
from the root of this tree.
292
335
This is true IF and ONLY IF the filename is part of the meta data
293
336
that bzr controls in this tree. I.E. a random .bzr directory placed
294
337
on disk will not be a control file for this tree.
297
self.bzrdir.transport.relpath(self.abspath(filename))
299
except errors.PathNotChild:
339
return self.bzrdir.is_control_filename(filename)
303
342
def open(path=None, _unsupported=False):
352
394
return pathjoin(self.basedir, filename)
354
396
def basis_tree(self):
355
"""Return RevisionTree for the current last revision."""
397
"""Return RevisionTree for the current last revision.
399
If the left most parent is a ghost then the returned tree will be an
400
empty tree - one obtained by calling repository.revision_tree(None).
356
402
revision_id = self.last_revision()
357
403
if revision_id is not None:
359
405
xml = self.read_basis_inventory()
360
406
inv = bzrlib.xml5.serializer_v5.read_inventory_from_string(xml)
407
inv.root.revision = revision_id
361
408
except NoSuchFile:
363
410
if inv is not None and inv.revision_id == revision_id:
364
411
return bzrlib.tree.RevisionTree(self.branch.repository, inv,
366
413
# FIXME? RBC 20060403 should we cache the inventory here ?
367
return self.branch.repository.revision_tree(revision_id)
415
return self.branch.repository.revision_tree(revision_id)
416
except errors.RevisionNotPresent:
417
# the basis tree *may* be a ghost or a low level error may have
418
# occured. If the revision is present, its a problem, if its not
420
if self.branch.repository.has_revision(revision_id):
422
# the basis tree is a ghost so return an empty tree.
423
return self.branch.repository.revision_tree(None)
370
426
@deprecated_method(zero_eight)
405
461
return bzrdir.BzrDir.create_standalone_workingtree(directory)
407
def relpath(self, abs):
408
"""Return the local path portion from a given absolute path."""
409
return relpath(self.basedir, abs)
463
def relpath(self, path):
464
"""Return the local path portion from a given path.
466
The path may be absolute or relative. If its a relative path it is
467
interpreted relative to the python current working directory.
469
return relpath(self.basedir, path)
411
471
def has_filename(self, filename):
412
return bzrlib.osutils.lexists(self.abspath(filename))
472
return osutils.lexists(self.abspath(filename))
414
474
def get_file(self, file_id):
415
475
return self.get_file_byname(self.id2path(file_id))
477
def get_file_text(self, file_id):
478
return self.get_file(file_id).read()
417
480
def get_file_byname(self, filename):
418
481
return file(self.abspath(filename), 'rb')
483
def get_parent_ids(self):
484
"""See Tree.get_parent_ids.
486
This implementation reads the pending merges list and last_revision
487
value and uses that to decide what the parents list should be.
489
last_rev = self.last_revision()
495
merges_file = self._control_files.get_utf8('pending-merges')
499
for l in merges_file.readlines():
500
parents.append(l.rstrip('\n'))
420
503
def get_root_id(self):
421
504
"""Return the id of this trees root"""
422
505
inv = self.read_working_inventory()
471
555
# but with branch a kwarg now, passing in args as is results in the
472
556
#message being used for the branch
473
557
args = (DEPRECATED_PARAMETER, message, ) + args
474
Commit().commit(working_tree=self, revprops=revprops, *args, **kwargs)
558
committed_id = Commit().commit( working_tree=self, revprops=revprops,
475
560
self._set_inventory(self.read_working_inventory())
477
563
def id2abspath(self, file_id):
478
564
return self.abspath(self.id2path(file_id))
496
582
return os.path.getsize(self.id2abspath(file_id))
499
def get_file_sha1(self, file_id):
500
path = self._inventory.id2path(file_id)
585
def get_file_sha1(self, file_id, path=None):
587
path = self._inventory.id2path(file_id)
501
588
return self._hashcache.get_sha1(path)
503
def is_executable(self, file_id):
504
if not supports_executable():
590
def get_file_mtime(self, file_id, path=None):
592
path = self._inventory.id2path(file_id)
593
return os.lstat(self.abspath(path)).st_mtime
595
if not supports_executable():
596
def is_executable(self, file_id, path=None):
505
597
return self._inventory[file_id].executable
507
path = self._inventory.id2path(file_id)
599
def is_executable(self, file_id, path=None):
601
path = self._inventory.id2path(file_id)
508
602
mode = os.lstat(self.abspath(path)).st_mode
509
return bool(stat.S_ISREG(mode) and stat.S_IEXEC&mode)
603
return bool(stat.S_ISREG(mode) and stat.S_IEXEC & mode)
511
605
@needs_write_lock
512
606
def add(self, files, ids=None):
555
649
raise BzrError("cannot add top-level %r" % f)
557
651
fullpath = normpath(self.abspath(f))
560
653
kind = file_kind(fullpath)
561
654
except OSError, e:
562
655
if e.errno == errno.ENOENT:
563
656
raise NoSuchFile(fullpath)
564
# maybe something better?
565
raise BzrError('cannot add: not a regular file, symlink or directory: %s' % quotefn(f))
567
657
if not InventoryEntry.versionable_kind(kind):
568
raise BzrError('cannot add: not a versionable file ('
569
'i.e. regular file, symlink or directory): %s' % quotefn(f))
658
raise errors.BadFileKindError(filename=f, kind=kind)
571
659
if file_id is None:
572
file_id = gen_file_id(f)
573
inv.add_path(f, kind=kind, file_id=file_id)
660
inv.add_path(f, kind=kind)
662
inv.add_path(f, kind=kind, file_id=file_id)
575
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
576
664
self._write_inventory(inv)
578
666
@needs_write_lock
667
def add_parent_tree_id(self, revision_id, allow_leftmost_as_ghost=False):
668
"""Add revision_id as a parent.
670
This is equivalent to retrieving the current list of parent ids
671
and setting the list to its value plus revision_id.
673
:param revision_id: The revision id to add to the parent list. It may
674
be a ghost revision as long as its not the first parent to be added,
675
or the allow_leftmost_as_ghost parameter is set True.
676
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
678
parents = self.get_parent_ids() + [revision_id]
679
self.set_parent_ids(parents,
680
allow_leftmost_as_ghost=len(parents) > 1 or allow_leftmost_as_ghost)
683
def add_parent_tree(self, parent_tuple, allow_leftmost_as_ghost=False):
684
"""Add revision_id, tree tuple as a parent.
686
This is equivalent to retrieving the current list of parent trees
687
and setting the list to its value plus parent_tuple. See also
688
add_parent_tree_id - if you only have a parent id available it will be
689
simpler to use that api. If you have the parent already available, using
690
this api is preferred.
692
:param parent_tuple: The (revision id, tree) to add to the parent list.
693
If the revision_id is a ghost, pass None for the tree.
694
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
696
parent_ids = self.get_parent_ids() + [parent_tuple[0]]
697
if len(parent_ids) > 1:
698
# the leftmost may have already been a ghost, preserve that if it
700
allow_leftmost_as_ghost = True
701
self.set_parent_ids(parent_ids,
702
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
579
705
def add_pending_merge(self, *revision_ids):
580
706
# TODO: Perhaps should check at this point that the
581
707
# history of the revision is actually present?
582
p = self.pending_merges()
708
parents = self.get_parent_ids()
584
710
for rev_id in revision_ids:
711
if rev_id in parents:
713
parents.append(rev_id)
590
self.set_pending_merges(p)
716
self.set_parent_ids(parents, allow_leftmost_as_ghost=True)
593
719
def pending_merges(self):
596
722
These are revisions that have been merged into the working
597
723
directory but not yet committed.
600
merges_file = self._control_files.get_utf8('pending-merges')
602
if e.errno != errno.ENOENT:
606
for l in merges_file.readlines():
607
p.append(l.rstrip('\n'))
725
return self.get_parent_ids()[1:]
728
def set_parent_ids(self, revision_ids, allow_leftmost_as_ghost=False):
729
"""Set the parent ids to revision_ids.
731
See also set_parent_trees. This api will try to retrieve the tree data
732
for each element of revision_ids from the trees repository. If you have
733
tree data already available, it is more efficient to use
734
set_parent_trees rather than set_parent_ids. set_parent_ids is however
735
an easier API to use.
737
:param revision_ids: The revision_ids to set as the parent ids of this
738
working tree. Any of these may be ghosts.
740
if len(revision_ids) > 0:
741
leftmost_id = revision_ids[0]
742
if (not allow_leftmost_as_ghost and not
743
self.branch.repository.has_revision(leftmost_id)):
744
raise errors.GhostRevisionUnusableHere(leftmost_id)
745
self.set_last_revision(leftmost_id)
747
self.set_last_revision(None)
748
merges = revision_ids[1:]
749
self._control_files.put_utf8('pending-merges', '\n'.join(merges))
752
def set_parent_trees(self, parents_list, allow_leftmost_as_ghost=False):
753
"""Set the parents of the working tree.
755
:param parents_list: A list of (revision_id, tree) tuples.
756
If tree is None, then that element is treated as an unreachable
757
parent tree - i.e. a ghost.
759
# parent trees are not used in current format trees, delegate to
761
self.set_parent_ids([rev for (rev, tree) in parents_list],
762
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
610
764
@needs_write_lock
611
765
def set_pending_merges(self, rev_list):
612
self._control_files.put_utf8('pending-merges', '\n'.join(rev_list))
766
parents = self.get_parent_ids()
767
leftmost = parents[:1]
768
new_parents = leftmost + rev_list
769
self.set_parent_ids(new_parents)
614
771
@needs_write_lock
615
772
def set_merge_modified(self, modified_hashes):
623
780
my_file = rio_file(stanzas, header)
624
781
self._control_files.put(filename, my_file)
784
def merge_from_branch(self, branch, to_revision=None):
785
"""Merge from a branch into this working tree.
787
:param branch: The branch to merge from.
788
:param to_revision: If non-None, the merge will merge to to_revision, but
789
not beyond it. to_revision does not need to be in the history of
790
the branch when it is supplied. If None, to_revision defaults to
791
branch.last_revision().
793
from bzrlib.merge import Merger, Merge3Merger
794
pb = bzrlib.ui.ui_factory.nested_progress_bar()
796
merger = Merger(self.branch, this_tree=self, pb=pb)
797
merger.pp = ProgressPhase("Merge phase", 5, pb)
798
merger.pp.next_phase()
799
# check that there are no
801
merger.check_basis(check_clean=True, require_commits=False)
802
if to_revision is None:
803
to_revision = branch.last_revision()
804
merger.other_rev_id = to_revision
805
if merger.other_rev_id is None:
806
raise error.NoCommits(branch)
807
self.branch.fetch(branch, last_revision=merger.other_rev_id)
808
merger.other_basis = merger.other_rev_id
809
merger.other_tree = self.branch.repository.revision_tree(
811
merger.pp.next_phase()
813
if merger.base_rev_id == merger.other_rev_id:
814
raise errors.PointlessMerge
815
merger.backup_files = False
816
merger.merge_type = Merge3Merger
817
merger.set_interesting_files(None)
818
merger.show_base = False
819
merger.reprocess = False
820
conflicts = merger.do_merge()
627
827
def merge_modified(self):
666
866
Skips the control directory.
668
868
inv = self._inventory
670
def descend(from_dir_relpath, from_dir_id, dp):
869
# Convert these into local objects to save lookup times
870
pathjoin = osutils.pathjoin
871
file_kind = osutils.file_kind
873
# transport.base ends in a slash, we want the piece
874
# between the last two slashes
875
transport_base_dir = self.bzrdir.transport.base.rsplit('/', 2)[1]
877
fk_entries = {'directory':TreeDirectory, 'file':TreeFile, 'symlink':TreeLink}
879
# directory file_id, relative path, absolute path, reverse sorted children
880
children = os.listdir(self.basedir)
882
# jam 20060527 The kernel sized tree seems equivalent whether we
883
# use a deque and popleft to keep them sorted, or if we use a plain
884
# list and just reverse() them.
885
children = collections.deque(children)
886
stack = [(inv.root.file_id, u'', self.basedir, children)]
888
from_dir_id, from_dir_relpath, from_dir_abspath, children = stack[-1]
891
f = children.popleft()
674
892
## TODO: If we find a subdirectory with its own .bzr
675
893
## directory, then that is a separate tree and we
676
894
## should exclude it.
678
896
# the bzrdir for this tree
679
if self.bzrdir.transport.base.endswith(f + '/'):
897
if transport_base_dir == f:
683
fp = appendpath(from_dir_relpath, f)
900
# we know that from_dir_relpath and from_dir_abspath never end in a slash
901
# and 'f' doesn't begin with one, we can do a string op, rather
902
# than the checks of pathjoin(), all relative paths will have an extra slash
904
fp = from_dir_relpath + '/' + f
686
fap = appendpath(dp, f)
907
fap = from_dir_abspath + '/' + f
688
909
f_ie = inv.get_child(from_dir_id, f)
691
elif self.is_ignored(fp):
912
elif self.is_ignored(fp[1:]):
915
# we may not have found this file, because of a unicode issue
916
f_norm, can_access = osutils.normalized_filename(f)
917
if f == f_norm or not can_access:
918
# No change, so treat this file normally
921
# this file can be accessed by a normalized path
922
# check again if it is versioned
923
# these lines are repeated here for performance
925
fp = from_dir_relpath + '/' + f
926
fap = from_dir_abspath + '/' + f
927
f_ie = inv.get_child(from_dir_id, f)
930
elif self.is_ignored(fp[1:]):
696
935
fk = file_kind(fap)
704
943
# make a last minute entry
945
yield fp[1:], c, fk, f_ie.file_id, f_ie
708
if fk == 'directory':
709
entry = TreeDirectory()
712
elif fk == 'symlink':
948
yield fp[1:], c, fk, None, fk_entries[fk]()
950
yield fp[1:], c, fk, None, TreeEntry()
717
yield fp, c, fk, (f_ie and f_ie.file_id), entry
719
953
if fk != 'directory':
723
# don't descend unversioned directories
726
for ff in descend(fp, f_ie.file_id, fap):
956
# But do this child first
957
new_children = os.listdir(fap)
959
new_children = collections.deque(new_children)
960
stack.append((f_ie.file_id, fp, fap, new_children))
961
# Break out of inner loop, so that we start outer loop with child
964
# if we finished all children, pop it off the stack
729
for f in descend(u'', inv.root.file_id, self.basedir):
732
968
@needs_write_lock
733
969
def move(self, from_paths, to_name):
906
1130
repository = self.branch.repository
907
1131
pb = bzrlib.ui.ui_factory.nested_progress_bar()
1133
new_basis_tree = self.branch.basis_tree()
909
1134
merge_inner(self.branch,
910
self.branch.basis_tree(),
916
self.set_last_revision(self.branch.last_revision())
1141
# TODO - dedup parents list with things merged by pull ?
1142
# reuse the revisiontree we merged against to set the new
1144
parent_trees = [(self.branch.last_revision(), new_basis_tree)]
1145
# we have to pull the merge trees out again, because
1146
# merge_inner has set the ids. - this corner is not yet
1147
# layered well enough to prevent double handling.
1148
merges = self.get_parent_ids()[1:]
1149
parent_trees.extend([
1150
(parent, repository.revision_tree(parent)) for
1152
self.set_parent_trees(parent_trees)
940
1176
for subf in os.listdir(dirabs):
942
and (subf not in dir_entry.children)):
1179
if subf not in dir_entry.children:
1180
subf_norm, can_access = osutils.normalized_filename(subf)
1181
if subf_norm != subf and can_access:
1182
if subf_norm not in dir_entry.children:
1183
fl.append(subf_norm)
947
subp = appendpath(path, subf)
1189
subp = pathjoin(path, subf)
1192
def _translate_ignore_rule(self, rule):
1193
"""Translate a single ignore rule to a regex.
1195
There are two types of ignore rules. Those that do not contain a / are
1196
matched against the tail of the filename (that is, they do not care
1197
what directory the file is in.) Rules which do contain a slash must
1198
match the entire path. As a special case, './' at the start of the
1199
string counts as a slash in the string but is removed before matching
1200
(e.g. ./foo.c, ./src/foo.c)
1202
:return: The translated regex.
1204
if rule[:2] in ('./', '.\\'):
1206
result = fnmatch.translate(rule[2:])
1207
elif '/' in rule or '\\' in rule:
1209
result = fnmatch.translate(rule)
1211
# default rule style.
1212
result = "(?:.*/)?(?!.*/)" + fnmatch.translate(rule)
1213
assert result[-1] == '$', "fnmatch.translate did not add the expected $"
1214
return "(" + result + ")"
1216
def _combine_ignore_rules(self, rules):
1217
"""Combine a list of ignore rules into a single regex object.
1219
Each individual rule is combined with | to form a big regex, which then
1220
has $ added to it to form something like ()|()|()$. The group index for
1221
each subregex's outermost group is placed in a dictionary mapping back
1222
to the rule. This allows quick identification of the matching rule that
1224
:return: a list of the compiled regex and the matching-group index
1225
dictionaries. We return a list because python complains if you try to
1226
combine more than 100 regexes.
1231
translated_rules = []
1233
translated_rule = self._translate_ignore_rule(rule)
1234
compiled_rule = re.compile(translated_rule)
1235
groups[next_group] = rule
1236
next_group += compiled_rule.groups
1237
translated_rules.append(translated_rule)
1238
if next_group == 99:
1239
result.append((re.compile("|".join(translated_rules)), groups))
1242
translated_rules = []
1243
if len(translated_rules):
1244
result.append((re.compile("|".join(translated_rules)), groups))
951
1247
def ignored_files(self):
952
1248
"""Yield list of PATH, IGNORE_PATTERN"""
959
1254
def get_ignore_list(self):
960
1255
"""Return list of ignore patterns.
962
1257
Cached in the Tree object after the first call.
964
if hasattr(self, '_ignorelist'):
965
return self._ignorelist
967
l = bzrlib.DEFAULT_IGNORE[:]
1259
ignoreset = getattr(self, '_ignoreset', None)
1260
if ignoreset is not None:
1263
ignore_globs = set(bzrlib.DEFAULT_IGNORE)
1264
ignore_globs.update(ignores.get_runtime_ignores())
1266
ignore_globs.update(ignores.get_user_ignores())
968
1268
if self.has_filename(bzrlib.IGNORE_FILENAME):
969
1269
f = self.get_file_byname(bzrlib.IGNORE_FILENAME)
970
l.extend([line.rstrip("\n\r") for line in f.readlines()])
1271
ignore_globs.update(ignores.parse_ignore_file(f))
1275
self._ignoreset = ignore_globs
1276
self._ignore_regex = self._combine_ignore_rules(ignore_globs)
1279
def _get_ignore_rules_as_regex(self):
1280
"""Return a regex of the ignore rules and a mapping dict.
1282
:return: (ignore rules compiled regex, dictionary mapping rule group
1283
indices to original rule.)
1285
if getattr(self, '_ignoreset', None) is None:
1286
self.get_ignore_list()
1287
return self._ignore_regex
975
1289
def is_ignored(self, filename):
976
1290
r"""Check whether the filename matches an ignore pattern.
990
1304
# treat dotfiles correctly and allows * to match /.
991
1305
# Eventually it should be replaced with something more
994
for pat in self.get_ignore_list():
995
if '/' in pat or '\\' in pat:
997
# as a special case, you can put ./ at the start of a
998
# pattern; this is good to match in the top-level
1001
if (pat[:2] == './') or (pat[:2] == '.\\'):
1005
if fnmatch.fnmatchcase(filename, newpat):
1008
if fnmatch.fnmatchcase(splitpath(filename)[-1], pat):
1308
rules = self._get_ignore_rules_as_regex()
1309
for regex, mapping in rules:
1310
match = regex.match(filename)
1311
if match is not None:
1312
# one or more of the groups in mapping will have a non-None
1314
groups = match.groups()
1315
rules = [mapping[group] for group in
1316
mapping if groups[group] is not None]
1013
1320
def kind(self, file_id):
1014
1321
return file_kind(self.id2abspath(file_id))
1059
1372
if new_revision is None:
1060
1373
self.branch.set_revision_history([])
1062
# current format is locked in with the branch
1063
revision_history = self.branch.revision_history()
1065
position = revision_history.index(new_revision)
1067
raise errors.NoSuchRevision(self.branch, new_revision)
1068
self.branch.set_revision_history(revision_history[:position + 1])
1376
self.branch.generate_revision_history(new_revision)
1377
except errors.NoSuchRevision:
1378
# not present in the repo - dont try to set it deeper than the tip
1379
self.branch.set_revision_history([new_revision])
1071
1382
def _cache_basis_inventory(self, new_revision):
1072
1383
"""Cache new_revision as the basis inventory."""
1384
# TODO: this should allow the ready-to-use inventory to be passed in,
1385
# as commit already has that ready-to-use [while the format is the
1074
1388
# this double handles the inventory - unpack and repack -
1075
1389
# but is easier to understand. We can/should put a conditional
1076
1390
# in here based on whether the inventory is in the latest format
1077
1391
# - perhaps we should repack all inventories on a repository
1079
inv = self.branch.repository.get_inventory(new_revision)
1080
inv.revision_id = new_revision
1081
xml = bzrlib.xml5.serializer_v5.write_inventory_to_string(inv)
1393
# the fast path is to copy the raw xml from the repository. If the
1394
# xml contains 'revision_id="', then we assume the right
1395
# revision_id is set. We must check for this full string, because a
1396
# root node id can legitimately look like 'revision_id' but cannot
1398
xml = self.branch.repository.get_inventory_xml(new_revision)
1399
if not 'revision_id="' in xml.split('\n', 1)[0]:
1400
inv = self.branch.repository.deserialise_inventory(
1402
inv.revision_id = new_revision
1403
xml = bzrlib.xml5.serializer_v5.write_inventory_to_string(inv)
1404
assert isinstance(xml, str), 'serialised xml must be bytestring.'
1083
1405
path = self._basis_inventory_name()
1084
self._control_files.put_utf8(path, xml)
1085
except WeaveRevisionNotPresent:
1407
self._control_files.put(path, sio)
1408
except (errors.NoSuchRevision, errors.RevisionNotPresent):
1088
1411
def read_basis_inventory(self):
1089
1412
"""Read the cached basis inventory."""
1090
1413
path = self._basis_inventory_name()
1091
return self._control_files.get_utf8(path).read()
1414
return self._control_files.get(path).read()
1093
1416
@needs_read_lock
1094
1417
def read_working_inventory(self):
1129
1452
# TODO: Perhaps make this just a warning, and continue?
1130
1453
# This tends to happen when
1131
1454
raise NotVersionedError(path=f)
1132
mutter("remove inventory entry %s {%s}", quotefn(f), fid)
1134
1456
# having remove it, it must be either ignored or unknown
1135
1457
if self.is_ignored(f):
1136
1458
new_status = 'I'
1138
1460
new_status = '?'
1139
show_status(new_status, inv[fid].kind, quotefn(f))
1461
show_status(new_status, inv[fid].kind, f, to_file=to_file)
1142
1464
self._write_inventory(inv)
1204
1526
between multiple working trees, i.e. via shared storage, then we
1205
1527
would probably want to lock both the local tree, and the branch.
1207
# FIXME: We want to write out the hashcache only when the last lock on
1208
# this working copy is released. Peeking at the lock count is a bit
1209
# of a nasty hack; probably it's better to have a transaction object,
1210
# which can do some finalization when it's either successfully or
1211
# unsuccessfully completed. (Denys's original patch did that.)
1212
# RBC 20060206 hookinhg into transaction will couple lock and transaction
1213
# wrongly. Hookinh into unllock on the control files object is fine though.
1215
# TODO: split this per format so there is no ugly if block
1216
if self._hashcache.needs_write and (
1217
# dedicated lock files
1218
self._control_files._lock_count==1 or
1220
(self._control_files is self.branch.control_files and
1221
self._control_files._lock_count==3)):
1222
self._hashcache.write()
1223
# reverse order of locking.
1224
result = self._control_files.unlock()
1226
self.branch.unlock()
1529
raise NotImplementedError(self.unlock)
1230
1531
@needs_write_lock
1231
1532
def update(self):
1248
1549
Do a 'normal' merge of the old branch basis if it is relevant.
1250
1551
old_tip = self.branch.update()
1251
if old_tip is not None:
1252
self.add_pending_merge(old_tip)
1253
self.branch.lock_read()
1256
if self.last_revision() != self.branch.last_revision():
1257
# merge tree state up to new branch tip.
1258
basis = self.basis_tree()
1259
to_tree = self.branch.basis_tree()
1260
result += merge_inner(self.branch,
1264
self.set_last_revision(self.branch.last_revision())
1265
if old_tip and old_tip != self.last_revision():
1266
# our last revision was not the prior branch last reivison
1267
# and we have converted that last revision to a pending merge.
1268
# base is somewhere between the branch tip now
1269
# and the now pending merge
1270
from bzrlib.revision import common_ancestor
1272
base_rev_id = common_ancestor(self.branch.last_revision(),
1274
self.branch.repository)
1275
except errors.NoCommonAncestor:
1277
base_tree = self.branch.repository.revision_tree(base_rev_id)
1278
other_tree = self.branch.repository.revision_tree(old_tip)
1279
result += merge_inner(self.branch,
1285
self.branch.unlock()
1552
# here if old_tip is not None, it is the old tip of the branch before
1553
# it was updated from the master branch. This should become a pending
1554
# merge in the working tree to preserve the user existing work. we
1555
# cant set that until we update the working trees last revision to be
1556
# one from the new branch, because it will just get absorbed by the
1557
# parent de-duplication logic.
1559
# We MUST save it even if an error occurs, because otherwise the users
1560
# local work is unreferenced and will appear to have been lost.
1563
if self.last_revision() != self.branch.last_revision():
1564
# merge tree state up to new branch tip.
1565
basis = self.basis_tree()
1566
to_tree = self.branch.basis_tree()
1567
result += merge_inner(self.branch,
1571
# TODO - dedup parents list with things merged by pull ?
1572
# reuse the tree we've updated to to set the basis:
1573
parent_trees = [(self.branch.last_revision(), to_tree)]
1574
merges = self.get_parent_ids()[1:]
1575
# Ideally we ask the tree for the trees here, that way the working
1576
# tree can decide whether to give us teh entire tree or give us a
1577
# lazy initialised tree. dirstate for instance will have the trees
1578
# in ram already, whereas a last-revision + basis-inventory tree
1579
# will not, but also does not need them when setting parents.
1580
for parent in merges:
1581
parent_trees.append(
1582
(parent, self.branch.repository.revision_tree(parent)))
1583
if old_tip is not None:
1584
parent_trees.append(
1585
(old_tip, self.branch.repository.revision_tree(old_tip)))
1586
self.set_parent_trees(parent_trees)
1588
# the working tree had the same last-revision as the master
1589
# branch did. We may still have pivot local work from the local
1590
# branch into old_tip:
1591
if old_tip is not None:
1592
self.add_parent_tree_id(old_tip)
1593
if old_tip and old_tip != self.last_revision():
1594
# our last revision was not the prior branch last revision
1595
# and we have converted that last revision to a pending merge.
1596
# base is somewhere between the branch tip now
1597
# and the now pending merge
1598
from bzrlib.revision import common_ancestor
1600
base_rev_id = common_ancestor(self.branch.last_revision(),
1602
self.branch.repository)
1603
except errors.NoCommonAncestor:
1605
base_tree = self.branch.repository.revision_tree(base_rev_id)
1606
other_tree = self.branch.repository.revision_tree(old_tip)
1607
result += merge_inner(self.branch,
1287
1613
@needs_write_lock
1288
1614
def _write_inventory(self, inv):
1306
1635
if file_kind(self.abspath(conflicted)) != "file":
1309
if e.errno == errno.ENOENT:
1637
except errors.NoSuchFile:
1313
1639
if text is True:
1314
1640
for suffix in ('.THIS', '.OTHER'):
1316
1642
kind = file_kind(self.abspath(conflicted+suffix))
1318
if e.errno == errno.ENOENT:
1645
except errors.NoSuchFile:
1326
1649
ctype = {True: 'text conflict', False: 'contents conflict'}[text]
1327
1650
conflicts.append(Conflict.factory(ctype, path=conflicted,
1329
1652
return conflicts
1655
class WorkingTree2(WorkingTree):
1656
"""This is the Format 2 working tree.
1658
This was the first weave based working tree.
1659
- uses os locks for locking.
1660
- uses the branch last-revision.
1664
# we share control files:
1665
if self._hashcache.needs_write and self._control_files._lock_count==3:
1666
self._hashcache.write()
1667
# reverse order of locking.
1669
return self._control_files.unlock()
1671
self.branch.unlock()
1332
1674
class WorkingTree3(WorkingTree):
1333
1675
"""This is the Format 3 working tree.
1486
1840
"""See WorkingTreeFormat.get_format_description()."""
1487
1841
return "Working tree format 2"
1843
def stub_initialize_remote(self, control_files):
1844
"""As a special workaround create critical control files for a remote working tree
1846
This ensures that it can later be updated and dealt with locally,
1847
since BzrDirFormat6 and BzrDirFormat5 cannot represent dirs with
1848
no working tree. (See bug #43064).
1852
bzrlib.xml5.serializer_v5.write_inventory(inv, sio)
1854
control_files.put('inventory', sio)
1856
control_files.put_utf8('pending-merges', '')
1489
1859
def initialize(self, a_bzrdir, revision_id=None):
1490
1860
"""See WorkingTreeFormat.initialize()."""
1491
1861
if not isinstance(a_bzrdir.transport, LocalTransport):
1577
1947
transport = a_bzrdir.get_workingtree_transport(self)
1578
1948
control_files = self._open_control_files(a_bzrdir)
1579
1949
control_files.create_lock()
1950
control_files.lock_write()
1580
1951
control_files.put_utf8('format', self.get_format_string())
1581
1952
branch = a_bzrdir.open_branch()
1582
1953
if revision_id is None:
1583
1954
revision_id = branch.last_revision()
1584
1955
inv = Inventory()
1585
wt = WorkingTree3(a_bzrdir.root_transport.base,
1956
wt = WorkingTree3(a_bzrdir.root_transport.local_abspath('.'),
1588
1959
_internal=True,
1590
1961
_bzrdir=a_bzrdir,
1591
1962
_control_files=control_files)
1592
wt._write_inventory(inv)
1593
wt.set_root_id(inv.root.file_id)
1594
wt.set_last_revision(revision_id)
1595
wt.set_pending_merges([])
1596
build_tree(wt.basis_tree(), wt)
1965
wt._write_inventory(inv)
1966
wt.set_root_id(inv.root.file_id)
1967
basis_tree = branch.repository.revision_tree(revision_id)
1968
if revision_id == bzrlib.revision.NULL_REVISION:
1969
wt.set_parent_trees([])
1971
wt.set_parent_trees([(revision_id, basis_tree)])
1972
build_tree(basis_tree, wt)
1975
control_files.unlock()
1599
1978
def __init__(self):
1611
1990
raise NotImplementedError
1612
1991
if not isinstance(a_bzrdir.transport, LocalTransport):
1613
1992
raise errors.NotLocalUrl(a_bzrdir.transport.base)
1614
control_files = self._open_control_files(a_bzrdir)
1615
return WorkingTree3(a_bzrdir.root_transport.base,
1993
return self._open(a_bzrdir, self._open_control_files(a_bzrdir))
1995
def _open(self, a_bzrdir, control_files):
1996
"""Open the tree itself.
1998
:param a_bzrdir: the dir for the tree.
1999
:param control_files: the control files for the tree.
2001
return WorkingTree3(a_bzrdir.root_transport.local_abspath('.'),
1616
2002
_internal=True,
1618
2004
_bzrdir=a_bzrdir,
1645
2031
self._transport_readonly_server = transport_readonly_server
1646
2032
self._formats = formats
2034
def _clone_test(self, test, bzrdir_format, workingtree_format, variation):
2035
"""Clone test for adaption."""
2036
new_test = deepcopy(test)
2037
new_test.transport_server = self._transport_server
2038
new_test.transport_readonly_server = self._transport_readonly_server
2039
new_test.bzrdir_format = bzrdir_format
2040
new_test.workingtree_format = workingtree_format
2041
def make_new_test_id():
2042
new_id = "%s(%s)" % (test.id(), variation)
2043
return lambda: new_id
2044
new_test.id = make_new_test_id()
1648
2047
def adapt(self, test):
1649
2048
from bzrlib.tests import TestSuite
1650
2049
result = TestSuite()
1651
2050
for workingtree_format, bzrdir_format in self._formats:
1652
new_test = deepcopy(test)
1653
new_test.transport_server = self._transport_server
1654
new_test.transport_readonly_server = self._transport_readonly_server
1655
new_test.bzrdir_format = bzrdir_format
1656
new_test.workingtree_format = workingtree_format
1657
def make_new_test_id():
1658
new_id = "%s(%s)" % (new_test.id(), workingtree_format.__class__.__name__)
1659
return lambda: new_id
1660
new_test.id = make_new_test_id()
2051
new_test = self._clone_test(
2054
workingtree_format, workingtree_format.__class__.__name__)
1661
2055
result.addTest(new_test)