1
# Copyright (C) 2005, 2006 Canonical Ltd
3
# This program is free software; you can redistribute it and/or modify
4
# it under the terms of the GNU General Public License as published by
5
# the Free Software Foundation; either version 2 of the License, or
6
# (at your option) any later version.
8
# This program is distributed in the hope that it will be useful,
9
# but WITHOUT ANY WARRANTY; without even the implied warranty of
10
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11
# GNU General Public License for more details.
13
# You should have received a copy of the GNU General Public License
14
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
17
"""WorkingTree4 format and implementation.
19
WorkingTree4 provides the dirstate based working tree logic.
21
To get a WorkingTree, call bzrdir.open_workingtree() or
22
WorkingTree.open(dir).
25
from cStringIO import StringIO
29
from bzrlib.lazy_import import lazy_import
30
lazy_import(globals(), """
31
from bisect import bisect_left
33
from copy import deepcopy
45
conflicts as _mod_conflicts,
63
from bzrlib.transport import get_transport
67
from bzrlib import symbol_versioning
68
from bzrlib.decorators import needs_read_lock, needs_write_lock
69
from bzrlib.inventory import InventoryEntry, Inventory, ROOT_ID, entry_factory
70
from bzrlib.lockable_files import LockableFiles, TransportLock
71
from bzrlib.lockdir import LockDir
72
import bzrlib.mutabletree
73
from bzrlib.mutabletree import needs_tree_write_lock
74
from bzrlib.osutils import (
86
from bzrlib.trace import mutter, note
87
from bzrlib.transport.local import LocalTransport
88
from bzrlib.tree import InterTree
89
from bzrlib.progress import DummyProgress, ProgressPhase
90
from bzrlib.revision import NULL_REVISION, CURRENT_REVISION
91
from bzrlib.rio import RioReader, rio_file, Stanza
92
from bzrlib.symbol_versioning import (deprecated_passed,
100
from bzrlib.tree import Tree
101
from bzrlib.workingtree import WorkingTree, WorkingTree3, WorkingTreeFormat3
104
class WorkingTree4(WorkingTree3):
105
"""This is the Format 4 working tree.
107
This differs from WorkingTree3 by:
108
- having a consolidated internal dirstate.
109
- not having a regular inventory attribute.
111
This is new in bzr TODO FIXME SETMEBEFORE MERGE.
114
def __init__(self, basedir,
119
"""Construct a WorkingTree for basedir.
121
If the branch is not supplied, it is opened automatically.
122
If the branch is supplied, it must be the branch for this basedir.
123
(branch.base is not cross checked, because for remote branches that
124
would be meaningless).
126
self._format = _format
127
self.bzrdir = _bzrdir
128
from bzrlib.hashcache import HashCache
129
from bzrlib.trace import note, mutter
130
assert isinstance(basedir, basestring), \
131
"base directory %r is not a string" % basedir
132
basedir = safe_unicode(basedir)
133
mutter("opening working tree %r", basedir)
134
self._branch = branch
135
assert isinstance(self.branch, bzrlib.branch.Branch), \
136
"branch %r is not a Branch" % self.branch
137
self.basedir = realpath(basedir)
138
# if branch is at our basedir and is a format 6 or less
139
# assume all other formats have their own control files.
140
assert isinstance(_control_files, LockableFiles), \
141
"_control_files must be a LockableFiles, not %r" % _control_files
142
self._control_files = _control_files
143
# update the whole cache up front and write to disk if anything changed;
144
# in the future we might want to do this more selectively
145
# two possible ways offer themselves : in self._unlock, write the cache
146
# if needed, or, when the cache sees a change, append it to the hash
147
# cache file, and have the parser take the most recent entry for a
149
cache_filename = self.bzrdir.get_workingtree_transport(None).local_abspath('stat-cache')
150
hc = self._hashcache = HashCache(basedir, cache_filename, self._control_files._file_mode)
152
# is this scan needed ? it makes things kinda slow.
161
# during a read or write lock these objects are set, and are
162
# None the rest of the time.
163
self._dirstate = None
164
self._inventory = None
167
@needs_tree_write_lock
168
def _add(self, files, ids, kinds):
169
"""See MutableTree._add."""
170
state = self.current_dirstate()
171
for f, file_id, kind in zip(files, ids, kinds):
176
file_id = generate_ids.gen_file_id(f)
177
# deliberately add the file with no cached stat or sha1
178
# - on the first access it will be gathered, and we can
179
# always change this once tests are all passing.
180
state.add(f, file_id, kind, None, '')
183
def break_lock(self):
184
"""Break a lock if one is present from another instance.
186
Uses the ui factory to ask for confirmation if the lock may be from
189
This will probe the repository for its lock as well.
191
# if the dirstate is locked by an active process, reject the break lock
194
if self._dirstate is None:
198
state = self._current_dirstate()
199
if state._lock_token is not None:
200
# we already have it locked. sheese, cant break our own lock.
201
raise errors.LockActive(self.basedir)
204
# try for a write lock - need permission to get one anyhow
207
except errors.LockContention:
208
# oslocks fail when a process is still live: fail.
209
# TODO: get the locked lockdir info and give to the user to
210
# assist in debugging.
211
raise errors.LockActive(self.basedir)
216
self._dirstate = None
217
self._control_files.break_lock()
218
self.branch.break_lock()
220
def current_dirstate(self):
221
"""Return the current dirstate object.
223
This is not part of the tree interface and only exposed for ease of
226
:raises errors.NotWriteLocked: when not in a lock.
228
if not self._control_files._lock_count:
229
raise errors.ObjectNotLocked(self)
230
return self._current_dirstate()
232
def _current_dirstate(self):
233
"""Internal function that does not check lock status.
235
This is needed for break_lock which also needs the dirstate.
237
if self._dirstate is not None:
238
return self._dirstate
239
local_path = self.bzrdir.get_workingtree_transport(None
240
).local_abspath('dirstate')
241
self._dirstate = dirstate.DirState.on_file(local_path)
242
return self._dirstate
244
def filter_unversioned_files(self, paths):
245
"""Filter out paths that are not versioned.
247
:return: set of paths.
249
# TODO: make a generic multi-bisect routine roughly that should list
250
# the paths, then process one half at a time recursively, and feed the
251
# results of each bisect in further still
252
paths = sorted(paths)
254
state = self.current_dirstate()
255
# TODO we want a paths_to_dirblocks helper I think
257
dirname, basename = os.path.split(path.encode('utf8'))
258
_, _, _, path_is_versioned = state._get_block_entry_index(
259
dirname, basename, 0)
260
if path_is_versioned:
265
"""Write all cached data to disk."""
266
if self._control_files._lock_mode != 'w':
267
raise errors.NotWriteLocked(self)
268
self.current_dirstate().save()
269
self._inventory = None
272
def _generate_inventory(self):
273
"""Create and set self.inventory from the dirstate object.
275
This is relatively expensive: we have to walk the entire dirstate.
276
Ideally we would not, and can deprecate this function.
278
#: uncomment to trap on inventory requests.
279
# import pdb;pdb.set_trace()
280
state = self.current_dirstate()
281
state._read_dirblocks_if_needed()
282
root_key, current_entry = self._get_entry(path='')
283
current_id = root_key[2]
284
assert current_entry[0][0] == 'd' # directory
285
inv = Inventory(root_id=current_id)
286
# Turn some things into local variables
287
minikind_to_kind = dirstate.DirState._minikind_to_kind
288
factory = entry_factory
289
utf8_decode = cache_utf8._utf8_decode
291
# we could do this straight out of the dirstate; it might be fast
292
# and should be profiled - RBC 20070216
293
parent_ies = {'' : inv.root}
294
for block in state._dirblocks[1:]: # skip the root
297
parent_ie = parent_ies[dirname]
299
# all the paths in this block are not versioned in this tree
301
for key, entry in block[1]:
302
minikind, link_or_sha1, size, executable, stat = entry[0]
303
if minikind in ('a', 'r'): # absent, relocated
304
# a parent tree only entry
307
name_unicode = utf8_decode(name)[0]
309
kind = minikind_to_kind[minikind]
310
inv_entry = factory[kind](file_id, name_unicode,
313
# not strictly needed: working tree
314
#entry.executable = executable
315
#entry.text_size = size
316
#entry.text_sha1 = sha1
318
elif kind == 'directory':
319
# add this entry to the parent map.
320
parent_ies[(dirname + '/' + name).strip('/')] = inv_entry
321
# These checks cost us around 40ms on a 55k entry tree
322
assert file_id not in inv_byid, ('file_id %s already in'
323
' inventory as %s' % (file_id, inv_byid[file_id]))
324
assert name_unicode not in parent_ie.children
325
inv_byid[file_id] = inv_entry
326
parent_ie.children[name_unicode] = inv_entry
327
self._inventory = inv
329
def _get_entry(self, file_id=None, path=None):
330
"""Get the dirstate row for file_id or path.
332
If either file_id or path is supplied, it is used as the key to lookup.
333
If both are supplied, the fastest lookup is used, and an error is
334
raised if they do not both point at the same row.
336
:param file_id: An optional unicode file_id to be looked up.
337
:param path: An optional unicode path to be looked up.
338
:return: The dirstate row tuple for path/file_id, or (None, None)
340
if file_id is None and path is None:
341
raise errors.BzrError('must supply file_id or path')
342
state = self.current_dirstate()
344
path = path.encode('utf8')
345
return state._get_entry(0, fileid_utf8=file_id, path_utf8=path)
347
def get_file_sha1(self, file_id, path=None, stat_value=None):
348
# check file id is valid unconditionally.
349
key, details = self._get_entry(file_id=file_id, path=path)
350
assert key is not None, 'what error should this raise'
352
# if row stat is valid, use cached sha1, else, get a new sha1.
354
path = pathjoin(key[0], key[1]).decode('utf8')
355
return self._hashcache.get_sha1(path, stat_value)
357
def _get_inventory(self):
358
"""Get the inventory for the tree. This is only valid within a lock."""
359
if self._inventory is not None:
360
return self._inventory
361
self._generate_inventory()
362
return self._inventory
364
inventory = property(_get_inventory,
365
doc="Inventory of this Tree")
368
def get_parent_ids(self):
369
"""See Tree.get_parent_ids.
371
This implementation requests the ids list from the dirstate file.
373
return self.current_dirstate().get_parent_ids()
376
def get_root_id(self):
377
"""Return the id of this trees root"""
378
return self._get_entry(path='')[0][2]
380
def has_id(self, file_id):
381
state = self.current_dirstate()
382
file_id = osutils.safe_file_id(file_id)
383
row, parents = self._get_entry(file_id=file_id)
386
return osutils.lexists(pathjoin(
387
self.basedir, row[0].decode('utf8'), row[1].decode('utf8')))
390
def id2path(self, file_id):
391
file_id = osutils.safe_file_id(file_id)
392
state = self.current_dirstate()
393
entry = self._get_entry(file_id=file_id)
394
if entry == (None, None):
396
path_utf8 = osutils.pathjoin(entry[0][0], entry[0][1])
397
return path_utf8.decode('utf8')
401
"""Iterate through file_ids for this tree.
403
file_ids are in a WorkingTree if they are in the working inventory
404
and the working file exists.
407
for key, tree_details in self.current_dirstate()._iter_entries():
408
if tree_details[0][0] in ('a', 'r'): # absent, relocated
409
# not relevant to the working tree
411
path = pathjoin(self.basedir, key[0].decode('utf8'), key[1].decode('utf8'))
412
if osutils.lexists(path):
413
result.append(key[2])
417
def _last_revision(self):
418
"""See Mutable.last_revision."""
419
parent_ids = self.current_dirstate().get_parent_ids()
426
super(WorkingTree4, self).lock_read()
427
if self._dirstate is None:
428
self.current_dirstate()
429
self._dirstate.lock_read()
431
def lock_tree_write(self):
432
super(WorkingTree4, self).lock_tree_write()
433
if self._dirstate is None:
434
self.current_dirstate()
435
self._dirstate.lock_write()
437
def lock_write(self):
438
super(WorkingTree4, self).lock_write()
439
if self._dirstate is None:
440
self.current_dirstate()
441
self._dirstate.lock_write()
443
@needs_tree_write_lock
444
def move(self, from_paths, to_dir, after=False):
445
"""See WorkingTree.move()."""
449
state = self.current_dirstate()
451
assert not isinstance(from_paths, basestring)
452
to_dir_utf8 = to_dir.encode('utf8')
453
to_entry_dirname, to_basename = os.path.split(to_dir_utf8)
454
id_index = state._get_id_index()
455
# check destination directory
456
# get the details for it
457
to_entry_block_index, to_entry_entry_index, dir_present, entry_present = \
458
state._get_block_entry_index(to_entry_dirname, to_basename, 0)
459
if not entry_present:
460
raise errors.BzrMoveFailedError('', to_dir,
461
errors.NotInWorkingDirectory(to_dir))
462
to_entry = state._dirblocks[to_entry_block_index][1][to_entry_entry_index]
463
# get a handle on the block itself.
464
to_block_index = state._ensure_block(
465
to_entry_block_index, to_entry_entry_index, to_dir_utf8)
466
to_block = state._dirblocks[to_block_index]
467
to_abs = self.abspath(to_dir)
468
if not isdir(to_abs):
469
raise errors.BzrMoveFailedError('',to_dir,
470
errors.NotADirectory(to_abs))
472
if to_entry[1][0][0] != 'd':
473
raise errors.BzrMoveFailedError('',to_dir,
474
errors.NotADirectory(to_abs))
476
if self._inventory is not None:
477
update_inventory = True
479
to_dir_ie = inv[to_dir_id]
480
to_dir_id = to_entry[0][2]
482
update_inventory = False
485
def move_one(old_entry, from_path_utf8, minikind, executable,
486
fingerprint, packed_stat, size,
487
to_block, to_key, to_path_utf8):
488
state._make_absent(old_entry)
489
from_key = old_entry[0]
491
lambda:state.update_minimal(from_key,
493
executable=executable,
494
fingerprint=fingerprint,
495
packed_stat=packed_stat,
497
path_utf8=from_path_utf8))
498
state.update_minimal(to_key,
500
executable=executable,
501
fingerprint=fingerprint,
502
packed_stat=packed_stat,
504
path_utf8=to_path_utf8)
505
added_entry_index, _ = state._find_entry_index(to_key, to_block[1])
506
new_entry = to_block[1][added_entry_index]
507
rollbacks.append(lambda:state._make_absent(new_entry))
509
# create rename entries and tuples
510
for from_rel in from_paths:
511
# from_rel is 'pathinroot/foo/bar'
512
from_rel_utf8 = from_rel.encode('utf8')
513
from_dirname, from_tail = osutils.split(from_rel)
514
from_dirname, from_tail_utf8 = osutils.split(from_rel_utf8)
515
from_entry = self._get_entry(path=from_rel)
516
if from_entry == (None, None):
517
raise errors.BzrMoveFailedError(from_rel,to_dir,
518
errors.NotVersionedError(path=str(from_rel)))
520
from_id = from_entry[0][2]
521
to_rel = pathjoin(to_dir, from_tail)
522
to_rel_utf8 = pathjoin(to_dir_utf8, from_tail_utf8)
523
item_to_entry = self._get_entry(path=to_rel)
524
if item_to_entry != (None, None):
525
raise errors.BzrMoveFailedError(from_rel, to_rel,
526
"Target is already versioned.")
528
if from_rel == to_rel:
529
raise errors.BzrMoveFailedError(from_rel, to_rel,
530
"Source and target are identical.")
532
from_missing = not self.has_filename(from_rel)
533
to_missing = not self.has_filename(to_rel)
540
raise errors.BzrMoveFailedError(from_rel, to_rel,
541
errors.NoSuchFile(path=to_rel,
542
extra="New file has not been created yet"))
544
# neither path exists
545
raise errors.BzrRenameFailedError(from_rel, to_rel,
546
errors.PathsDoNotExist(paths=(from_rel, to_rel)))
548
if from_missing: # implicitly just update our path mapping
551
raise errors.RenameFailedFilesExist(from_rel, to_rel,
552
extra="(Use --after to update the Bazaar id)")
555
def rollback_rename():
556
"""A single rename has failed, roll it back."""
558
for rollback in reversed(rollbacks):
562
import pdb;pdb.set_trace()
563
exc_info = sys.exc_info()
565
raise exc_info[0], exc_info[1], exc_info[2]
567
# perform the disk move first - its the most likely failure point.
569
from_rel_abs = self.abspath(from_rel)
570
to_rel_abs = self.abspath(to_rel)
572
osutils.rename(from_rel_abs, to_rel_abs)
574
raise errors.BzrMoveFailedError(from_rel, to_rel, e[1])
575
rollbacks.append(lambda: osutils.rename(to_rel_abs, from_rel_abs))
577
# perform the rename in the inventory next if needed: its easy
581
from_entry = inv[from_id]
582
current_parent = from_entry.parent_id
583
inv.rename(from_id, to_dir_id, from_tail)
585
lambda: inv.rename(from_id, current_parent, from_tail))
586
# finally do the rename in the dirstate, which is a little
587
# tricky to rollback, but least likely to need it.
588
old_block_index, old_entry_index, dir_present, file_present = \
589
state._get_block_entry_index(from_dirname, from_tail_utf8, 0)
590
old_block = state._dirblocks[old_block_index][1]
591
old_entry = old_block[old_entry_index]
592
from_key, old_entry_details = old_entry
593
cur_details = old_entry_details[0]
595
to_key = ((to_block[0],) + from_key[1:3])
596
minikind = cur_details[0]
597
move_one(old_entry, from_path_utf8=from_rel_utf8,
599
executable=cur_details[3],
600
fingerprint=cur_details[1],
601
packed_stat=cur_details[4],
605
to_path_utf8=to_rel_utf8)
608
def update_dirblock(from_dir, to_key, to_dir_utf8):
609
"""all entries in this block need updating.
611
TODO: This is pretty ugly, and doesn't support
612
reverting, but it works.
614
assert from_dir != '', "renaming root not supported"
615
from_key = (from_dir, '')
616
from_block_idx, present = \
617
state._find_block_index_from_key(from_key)
619
# This is the old record, if it isn't present, then
620
# there is theoretically nothing to update.
621
# (Unless it isn't present because of lazy loading,
622
# but we don't do that yet)
624
from_block = state._dirblocks[from_block_idx]
625
to_block_index, to_entry_index, _, _ = \
626
state._get_block_entry_index(to_key[0], to_key[1], 0)
627
to_block_index = state._ensure_block(
628
to_block_index, to_entry_index, to_dir_utf8)
629
to_block = state._dirblocks[to_block_index]
630
for entry in from_block[1]:
631
assert entry[0][0] == from_dir
632
cur_details = entry[1][0]
633
to_key = (to_dir_utf8, entry[0][1], entry[0][2])
634
from_path_utf8 = osutils.pathjoin(entry[0][0], entry[0][1])
635
to_path_utf8 = osutils.pathjoin(to_dir_utf8, entry[0][1])
636
minikind = cur_details[0]
637
move_one(entry, from_path_utf8=from_path_utf8,
639
executable=cur_details[3],
640
fingerprint=cur_details[1],
641
packed_stat=cur_details[4],
645
to_path_utf8=to_rel_utf8)
647
# We need to move all the children of this
649
update_dirblock(from_path_utf8, to_key,
651
update_dirblock(from_rel_utf8, to_key, to_rel_utf8)
655
state._dirblock_state = dirstate.DirState.IN_MEMORY_MODIFIED
658
return #rename_tuples
661
"""Initialize the state in this tree to be a new tree."""
665
def path2id(self, path):
666
"""Return the id for path in this tree."""
667
entry = self._get_entry(path=path)
668
if entry == (None, None):
672
def paths2ids(self, paths, trees=[], require_versioned=True):
673
"""See Tree.paths2ids().
675
This specialisation fast-paths the case where all the trees are in the
680
parents = self.get_parent_ids()
682
if not (isinstance(tree, DirStateRevisionTree) and tree._revision_id in
684
return super(WorkingTree4, self).paths2ids(paths, trees, require_versioned)
685
search_indexes = [0] + [1 + parents.index(tree._revision_id) for tree in trees]
686
# -- make all paths utf8 --
689
paths_utf8.add(path.encode('utf8'))
691
# -- paths is now a utf8 path set --
692
# -- get the state object and prepare it.
693
state = self.current_dirstate()
694
if False and (state._dirblock_state == dirstate.DirState.NOT_IN_MEMORY
695
and '' not in paths):
696
paths2ids = self._paths2ids_using_bisect
698
paths2ids = self._paths2ids_in_memory
699
return paths2ids(paths, search_indexes,
700
require_versioned=require_versioned)
702
def _paths2ids_in_memory(self, paths, search_indexes,
703
require_versioned=True):
704
state = self.current_dirstate()
705
state._read_dirblocks_if_needed()
706
def _entries_for_path(path):
707
"""Return a list with all the entries that match path for all ids.
709
dirname, basename = os.path.split(path)
710
key = (dirname, basename, '')
711
block_index, present = state._find_block_index_from_key(key)
713
# the block which should contain path is absent.
716
block = state._dirblocks[block_index][1]
717
entry_index, _ = state._find_entry_index(key, block)
718
# we may need to look at multiple entries at this path: walk while the paths match.
719
while (entry_index < len(block) and
720
block[entry_index][0][0:2] == key[0:2]):
721
result.append(block[entry_index])
724
if require_versioned:
725
# -- check all supplied paths are versioned in a search tree. --
728
path_entries = _entries_for_path(path)
730
# this specified path is not present at all: error
731
all_versioned = False
733
found_versioned = False
734
# for each id at this path
735
for entry in path_entries:
737
for index in search_indexes:
738
if entry[1][index][0] != 'a': # absent
739
found_versioned = True
740
# all good: found a versioned cell
742
if not found_versioned:
743
# none of the indexes was not 'absent' at all ids for this
745
all_versioned = False
747
if not all_versioned:
748
raise errors.PathsNotVersionedError(paths)
749
# -- remove redundancy in supplied paths to prevent over-scanning --
752
other_paths = paths.difference(set([path]))
753
if not osutils.is_inside_any(other_paths, path):
754
# this is a top level path, we must check it.
755
search_paths.add(path)
757
# for all search_indexs in each path at or under each element of
758
# search_paths, if the detail is relocated: add the id, and add the
759
# relocated path as one to search if its not searched already. If the
760
# detail is not relocated, add the id.
761
searched_paths = set()
763
def _process_entry(entry):
764
"""Look at search_indexes within entry.
766
If a specific tree's details are relocated, add the relocation
767
target to search_paths if not searched already. If it is absent, do
768
nothing. Otherwise add the id to found_ids.
770
for index in search_indexes:
771
if entry[1][index][0] == 'r': # relocated
772
if not osutils.is_inside_any(searched_paths, entry[1][index][1]):
773
search_paths.add(entry[1][index][1])
774
elif entry[1][index][0] != 'a': # absent
775
found_ids.add(entry[0][2])
777
current_root = search_paths.pop()
778
searched_paths.add(current_root)
779
# process the entries for this containing directory: the rest will be
780
# found by their parents recursively.
781
root_entries = _entries_for_path(current_root)
783
# this specified path is not present at all, skip it.
785
for entry in root_entries:
786
_process_entry(entry)
787
initial_key = (current_root, '', '')
788
block_index, _ = state._find_block_index_from_key(initial_key)
789
while (block_index < len(state._dirblocks) and
790
osutils.is_inside(current_root, state._dirblocks[block_index][0])):
791
for entry in state._dirblocks[block_index][1]:
792
_process_entry(entry)
796
def _paths2ids_using_bisect(self, paths, search_indexes,
797
require_versioned=True):
798
state = self.current_dirstate()
801
split_paths = sorted(osutils.split(p) for p in paths)
802
found = state._bisect_recursive(split_paths)
804
if require_versioned:
805
found_dir_names = set(dir_name_id[:2] for dir_name_id in found)
806
for dir_name in split_paths:
807
if dir_name not in found_dir_names:
808
raise errors.PathsNotVersionedError(paths)
810
for dir_name_id, trees_info in found.iteritems():
811
for index in search_indexes:
812
if trees_info[index][0] not in ('r', 'a'):
813
found_ids.add(dir_name_id[2])
816
def read_working_inventory(self):
817
"""Read the working inventory.
819
This is a meaningless operation for dirstate, but we obey it anyhow.
821
return self.inventory
824
def revision_tree(self, revision_id):
825
"""See Tree.revision_tree.
827
WorkingTree4 supplies revision_trees for any basis tree.
829
revision_id = osutils.safe_revision_id(revision_id)
830
dirstate = self.current_dirstate()
831
parent_ids = dirstate.get_parent_ids()
832
if revision_id not in parent_ids:
833
raise errors.NoSuchRevisionInTree(self, revision_id)
834
if revision_id in dirstate.get_ghosts():
835
raise errors.NoSuchRevisionInTree(self, revision_id)
836
return DirStateRevisionTree(dirstate, revision_id,
837
self.branch.repository)
839
@needs_tree_write_lock
840
def set_last_revision(self, new_revision):
841
"""Change the last revision in the working tree."""
842
new_revision = osutils.safe_revision_id(new_revision)
843
parents = self.get_parent_ids()
844
if new_revision in (NULL_REVISION, None):
845
assert len(parents) < 2, (
846
"setting the last parent to none with a pending merge is "
848
self.set_parent_ids([])
850
self.set_parent_ids([new_revision] + parents[1:],
851
allow_leftmost_as_ghost=True)
853
@needs_tree_write_lock
854
def set_parent_ids(self, revision_ids, allow_leftmost_as_ghost=False):
855
"""Set the parent ids to revision_ids.
857
See also set_parent_trees. This api will try to retrieve the tree data
858
for each element of revision_ids from the trees repository. If you have
859
tree data already available, it is more efficient to use
860
set_parent_trees rather than set_parent_ids. set_parent_ids is however
861
an easier API to use.
863
:param revision_ids: The revision_ids to set as the parent ids of this
864
working tree. Any of these may be ghosts.
866
revision_ids = [osutils.safe_revision_id(r) for r in revision_ids]
868
for revision_id in revision_ids:
870
revtree = self.branch.repository.revision_tree(revision_id)
871
# TODO: jam 20070213 KnitVersionedFile raises
872
# RevisionNotPresent rather than NoSuchRevision if a
873
# given revision_id is not present. Should Repository be
874
# catching it and re-raising NoSuchRevision?
875
except (errors.NoSuchRevision, errors.RevisionNotPresent):
877
trees.append((revision_id, revtree))
878
self.set_parent_trees(trees,
879
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
881
@needs_tree_write_lock
882
def set_parent_trees(self, parents_list, allow_leftmost_as_ghost=False):
883
"""Set the parents of the working tree.
885
:param parents_list: A list of (revision_id, tree) tuples.
886
If tree is None, then that element is treated as an unreachable
887
parent tree - i.e. a ghost.
889
dirstate = self.current_dirstate()
890
if len(parents_list) > 0:
891
if not allow_leftmost_as_ghost and parents_list[0][1] is None:
892
raise errors.GhostRevisionUnusableHere(parents_list[0][0])
895
# convert absent trees to the null tree, which we convert back to
897
for rev_id, tree in parents_list:
898
rev_id = osutils.safe_revision_id(rev_id)
900
real_trees.append((rev_id, tree))
902
real_trees.append((rev_id,
903
self.branch.repository.revision_tree(None)))
904
ghosts.append(rev_id)
905
dirstate.set_parent_trees(real_trees, ghosts=ghosts)
908
def _set_root_id(self, file_id):
909
"""See WorkingTree.set_root_id."""
910
state = self.current_dirstate()
911
state.set_path_id('', file_id)
912
self._dirty = state._dirblock_state == dirstate.DirState.IN_MEMORY_MODIFIED
915
"""Unlock in format 4 trees needs to write the entire dirstate."""
916
if self._control_files._lock_count == 1:
917
self._write_hashcache_if_dirty()
918
# eventually we should do signature checking during read locks for
920
if self._control_files._lock_mode == 'w':
923
if self._dirstate is not None:
924
self._dirstate.unlock()
925
self._dirstate = None
926
self._inventory = None
927
# reverse order of locking.
929
return self._control_files.unlock()
933
@needs_tree_write_lock
934
def unversion(self, file_ids):
935
"""Remove the file ids in file_ids from the current versioned set.
937
When a file_id is unversioned, all of its children are automatically
940
:param file_ids: The file ids to stop versioning.
941
:raises: NoSuchId if any fileid is not currently versioned.
945
state = self.current_dirstate()
946
state._read_dirblocks_if_needed()
947
ids_to_unversion = set()
948
for file_id in file_ids:
949
ids_to_unversion.add(osutils.safe_file_id(file_id))
950
paths_to_unversion = set()
952
# check if the root is to be unversioned, if so, assert for now.
953
# walk the state marking unversioned things as absent.
954
# if there are any un-unversioned ids at the end, raise
955
for key, details in state._dirblocks[0][1]:
956
if (details[0][0] not in ('a', 'r') and # absent or relocated
957
key[2] in ids_to_unversion):
958
# I haven't written the code to unversion / yet - it should be
960
raise errors.BzrError('Unversioning the / is not currently supported')
962
while block_index < len(state._dirblocks):
963
# process one directory at a time.
964
block = state._dirblocks[block_index]
965
# first check: is the path one to remove - it or its children
967
for path in paths_to_unversion:
968
if (block[0].startswith(path) and
969
(len(block[0]) == len(path) or
970
block[0][len(path)] == '/')):
971
# this entire block should be deleted - its the block for a
972
# path to unversion; or the child of one
975
# TODO: trim paths_to_unversion as we pass by paths
977
# this block is to be deleted: process it.
978
# TODO: we can special case the no-parents case and
979
# just forget the whole block.
981
while entry_index < len(block[1]):
982
# Mark this file id as having been removed
983
ids_to_unversion.discard(block[1][entry_index][0][2])
984
if not state._make_absent(block[1][entry_index]):
986
# go to the next block. (At the moment we dont delete empty
991
while entry_index < len(block[1]):
992
entry = block[1][entry_index]
993
if (entry[1][0][0] in ('a', 'r') or # absent, relocated
995
entry[0][2] not in ids_to_unversion):
996
# ^ not an id to unversion
999
if entry[1][0][0] == 'd':
1000
paths_to_unversion.add(pathjoin(entry[0][0], entry[0][1]))
1001
if not state._make_absent(entry):
1003
# we have unversioned this id
1004
ids_to_unversion.remove(entry[0][2])
1006
if ids_to_unversion:
1007
raise errors.NoSuchId(self, iter(ids_to_unversion).next())
1009
# have to change the legacy inventory too.
1010
if self._inventory is not None:
1011
for file_id in file_ids:
1012
self._inventory.remove_recursive_id(file_id)
1014
@needs_tree_write_lock
1015
def _write_inventory(self, inv):
1016
"""Write inventory as the current inventory."""
1017
assert not self._dirty, "attempting to write an inventory when the dirstate is dirty will cause data loss"
1018
self.current_dirstate().set_state_from_inventory(inv)
1023
class WorkingTreeFormat4(WorkingTreeFormat3):
1024
"""The first consolidated dirstate working tree format.
1027
- exists within a metadir controlling .bzr
1028
- includes an explicit version marker for the workingtree control
1029
files, separate from the BzrDir format
1030
- modifies the hash cache format
1031
- is new in bzr TODO FIXME SETBEFOREMERGE
1032
- uses a LockDir to guard access to it.
1035
def get_format_string(self):
1036
"""See WorkingTreeFormat.get_format_string()."""
1037
return "Bazaar Working Tree format 4\n"
1039
def get_format_description(self):
1040
"""See WorkingTreeFormat.get_format_description()."""
1041
return "Working tree format 4"
1043
def initialize(self, a_bzrdir, revision_id=None):
1044
"""See WorkingTreeFormat.initialize().
1046
revision_id allows creating a working tree at a different
1047
revision than the branch is at.
1049
revision_id = osutils.safe_revision_id(revision_id)
1050
if not isinstance(a_bzrdir.transport, LocalTransport):
1051
raise errors.NotLocalUrl(a_bzrdir.transport.base)
1052
transport = a_bzrdir.get_workingtree_transport(self)
1053
control_files = self._open_control_files(a_bzrdir)
1054
control_files.create_lock()
1055
control_files.lock_write()
1056
control_files.put_utf8('format', self.get_format_string())
1057
branch = a_bzrdir.open_branch()
1058
if revision_id is None:
1059
revision_id = branch.last_revision()
1060
local_path = transport.local_abspath('dirstate')
1061
state = dirstate.DirState.initialize(local_path)
1063
wt = WorkingTree4(a_bzrdir.root_transport.local_abspath('.'),
1067
_control_files=control_files)
1071
#wt.current_dirstate().set_path_id('', NEWROOT)
1072
wt.set_last_revision(revision_id)
1074
basis = wt.basis_tree()
1076
transform.build_tree(basis, wt)
1079
control_files.unlock()
1084
def _open(self, a_bzrdir, control_files):
1085
"""Open the tree itself.
1087
:param a_bzrdir: the dir for the tree.
1088
:param control_files: the control files for the tree.
1090
return WorkingTree4(a_bzrdir.root_transport.local_abspath('.'),
1091
branch=a_bzrdir.open_branch(),
1094
_control_files=control_files)
1097
class DirStateRevisionTree(Tree):
1098
"""A revision tree pulling the inventory from a dirstate."""
1100
def __init__(self, dirstate, revision_id, repository):
1101
self._dirstate = dirstate
1102
self._revision_id = osutils.safe_revision_id(revision_id)
1103
self._repository = repository
1104
self._inventory = None
1106
self._dirstate_locked = False
1108
def annotate_iter(self, file_id):
1109
"""See Tree.annotate_iter"""
1110
w = self._repository.weave_store.get_weave(file_id,
1111
self._repository.get_transaction())
1112
return w.annotate_iter(self.inventory[file_id].revision)
1114
def _comparison_data(self, entry, path):
1115
"""See Tree._comparison_data."""
1117
return None, False, None
1118
# trust the entry as RevisionTree does, but this may not be
1119
# sensible: the entry might not have come from us?
1120
return entry.kind, entry.executable, None
1122
def _file_size(self, entry, stat_value):
1123
return entry.text_size
1125
def filter_unversioned_files(self, paths):
1126
"""Filter out paths that are not versioned.
1128
:return: set of paths.
1130
pred = self.has_filename
1131
return set((p for p in paths if not pred(p)))
1133
def _get_parent_index(self):
1134
"""Return the index in the dirstate referenced by this tree."""
1135
return self._dirstate.get_parent_ids().index(self._revision_id) + 1
1137
def _get_entry(self, file_id=None, path=None):
1138
"""Get the dirstate row for file_id or path.
1140
If either file_id or path is supplied, it is used as the key to lookup.
1141
If both are supplied, the fastest lookup is used, and an error is
1142
raised if they do not both point at the same row.
1144
:param file_id: An optional unicode file_id to be looked up.
1145
:param path: An optional unicode path to be looked up.
1146
:return: The dirstate row tuple for path/file_id, or (None, None)
1148
if file_id is None and path is None:
1149
raise errors.BzrError('must supply file_id or path')
1150
file_id = osutils.safe_file_id(file_id)
1151
if path is not None:
1152
path = path.encode('utf8')
1153
parent_index = self._get_parent_index()
1154
return self._dirstate._get_entry(parent_index, fileid_utf8=file_id, path_utf8=path)
1156
def _generate_inventory(self):
1157
"""Create and set self.inventory from the dirstate object.
1159
This is relatively expensive: we have to walk the entire dirstate.
1160
Ideally we would not, and instead would """
1161
assert self._locked, 'cannot generate inventory of an unlocked '\
1162
'dirstate revision tree'
1163
# separate call for profiling - makes it clear where the costs are.
1164
self._dirstate._read_dirblocks_if_needed()
1165
assert self._revision_id in self._dirstate.get_parent_ids(), \
1166
'parent %s has disappeared from %s' % (
1167
self._revision_id, self._dirstate.get_parent_ids())
1168
parent_index = self._dirstate.get_parent_ids().index(self._revision_id) + 1
1169
# This is identical now to the WorkingTree _generate_inventory except
1170
# for the tree index use.
1171
root_key, current_entry = self._dirstate._get_entry(parent_index, path_utf8='')
1172
current_id = root_key[2]
1173
assert current_entry[parent_index][0] == 'd'
1174
inv = Inventory(root_id=current_id, revision_id=self._revision_id)
1175
inv.root.revision = current_entry[parent_index][4]
1176
# Turn some things into local variables
1177
minikind_to_kind = dirstate.DirState._minikind_to_kind
1178
factory = entry_factory
1179
utf8_decode = cache_utf8._utf8_decode
1180
inv_byid = inv._byid
1181
# we could do this straight out of the dirstate; it might be fast
1182
# and should be profiled - RBC 20070216
1183
parent_ies = {'' : inv.root}
1184
for block in self._dirstate._dirblocks[1:]: #skip root
1187
parent_ie = parent_ies[dirname]
1189
# all the paths in this block are not versioned in this tree
1191
for key, entry in block[1]:
1192
minikind, link_or_sha1, size, executable, revid = entry[parent_index]
1193
if minikind in ('a', 'r'): # absent, relocated
1197
name_unicode = utf8_decode(name)[0]
1199
kind = minikind_to_kind[minikind]
1200
inv_entry = factory[kind](file_id, name_unicode,
1202
inv_entry.revision = revid
1204
inv_entry.executable = executable
1205
inv_entry.text_size = size
1206
inv_entry.text_sha1 = link_or_sha1
1207
elif kind == 'directory':
1208
parent_ies[(dirname + '/' + name).strip('/')] = inv_entry
1209
elif kind == 'symlink':
1210
inv_entry.executable = False
1211
inv_entry.text_size = size
1212
inv_entry.symlink_target = utf8_decode(link_or_sha1)[0]
1214
raise Exception, kind
1215
# These checks cost us around 40ms on a 55k entry tree
1216
assert file_id not in inv_byid
1217
assert name_unicode not in parent_ie.children
1218
inv_byid[file_id] = inv_entry
1219
parent_ie.children[name_unicode] = inv_entry
1220
self._inventory = inv
1222
def get_file_mtime(self, file_id, path=None):
1223
"""Return the modification time for this record.
1225
We return the timestamp of the last-changed revision.
1227
# Make sure the file exists
1228
entry = self._get_entry(file_id, path=path)
1229
if entry == (None, None): # do we raise?
1231
parent_index = self._get_parent_index()
1232
last_changed_revision = entry[1][parent_index][4]
1233
return self._repository.get_revision(last_changed_revision).timestamp
1235
def get_file_sha1(self, file_id, path=None, stat_value=None):
1236
# TODO: if path is present, fast-path on that, as inventory
1237
# might not be present
1238
ie = self.inventory[file_id]
1239
if ie.kind == "file":
1243
def get_file(self, file_id):
1244
return StringIO(self.get_file_text(file_id))
1246
def get_file_lines(self, file_id):
1247
ie = self.inventory[file_id]
1248
return self._repository.weave_store.get_weave(file_id,
1249
self._repository.get_transaction()).get_lines(ie.revision)
1251
def get_file_size(self, file_id):
1252
return self.inventory[file_id].text_size
1254
def get_file_text(self, file_id):
1255
return ''.join(self.get_file_lines(file_id))
1257
def get_symlink_target(self, file_id):
1258
entry = self._get_entry(file_id=file_id)
1259
parent_index = self._get_parent_index()
1260
if entry[1][parent_index][0] != 'l':
1263
# At present, none of the tree implementations supports non-ascii
1264
# symlink targets. So we will just assume that the dirstate path is
1266
return entry[1][parent_index][1]
1268
def get_revision_id(self):
1269
"""Return the revision id for this tree."""
1270
return self._revision_id
1272
def _get_inventory(self):
1273
if self._inventory is not None:
1274
return self._inventory
1275
self._generate_inventory()
1276
return self._inventory
1278
inventory = property(_get_inventory,
1279
doc="Inventory of this Tree")
1281
def get_parent_ids(self):
1282
"""The parents of a tree in the dirstate are not cached."""
1283
return self._repository.get_revision(self._revision_id).parent_ids
1285
def has_filename(self, filename):
1286
return bool(self.path2id(filename))
1288
def kind(self, file_id):
1289
return self.inventory[file_id].kind
1291
def is_executable(self, file_id, path=None):
1292
ie = self.inventory[file_id]
1293
if ie.kind != "file":
1295
return ie.executable
1297
def list_files(self, include_root=False):
1298
# We use a standard implementation, because DirStateRevisionTree is
1299
# dealing with one of the parents of the current state
1300
inv = self._get_inventory()
1301
entries = inv.iter_entries()
1302
if self.inventory.root is not None and not include_root:
1304
for path, entry in entries:
1305
yield path, 'V', entry.kind, entry.file_id, entry
1307
def lock_read(self):
1308
"""Lock the tree for a set of operations."""
1309
if not self._locked:
1310
self._repository.lock_read()
1311
if self._dirstate._lock_token is None:
1312
self._dirstate.lock_read()
1313
self._dirstate_locked = True
1317
def path2id(self, path):
1318
"""Return the id for path in this tree."""
1319
# lookup by path: faster than splitting and walking the ivnentory.
1320
entry = self._get_entry(path=path)
1321
if entry == (None, None):
1326
"""Unlock, freeing any cache memory used during the lock."""
1327
# outside of a lock, the inventory is suspect: release it.
1329
if not self._locked:
1330
self._inventory = None
1332
if self._dirstate_locked:
1333
self._dirstate.unlock()
1334
self._dirstate_locked = False
1335
self._repository.unlock()
1337
def walkdirs(self, prefix=""):
1338
# TODO: jam 20070215 This is the cheap way by cheating and using the
1339
# RevisionTree implementation.
1340
# This should be cleaned up to use the much faster Dirstate code
1341
# This is a little tricky, though, because the dirstate is
1342
# indexed by current path, not by parent path.
1343
# So for now, we just build up the parent inventory, and extract
1344
# it the same way RevisionTree does.
1345
_directory = 'directory'
1346
inv = self._get_inventory()
1347
top_id = inv.path2id(prefix)
1351
pending = [(prefix, top_id)]
1354
relpath, file_id = pending.pop()
1355
# 0 - relpath, 1- file-id
1357
relroot = relpath + '/'
1360
# FIXME: stash the node in pending
1361
entry = inv[file_id]
1362
for name, child in entry.sorted_children():
1363
toppath = relroot + name
1364
dirblock.append((toppath, name, child.kind, None,
1365
child.file_id, child.kind
1367
yield (relpath, entry.file_id), dirblock
1368
# push the user specified dirs from dirblock
1369
for dir in reversed(dirblock):
1370
if dir[2] == _directory:
1371
pending.append((dir[0], dir[4]))
1374
class InterDirStateTree(InterTree):
1375
"""Fast path optimiser for changes_from with dirstate trees."""
1377
def __init__(self, source, target):
1378
super(InterDirStateTree, self).__init__(source, target)
1379
if not InterDirStateTree.is_compatible(source, target):
1380
raise Exception, "invalid source %r and target %r" % (source, target)
1383
def make_source_parent_tree(source, target):
1384
"""Change the source tree into a parent of the target."""
1385
revid = source.commit('record tree')
1386
target.branch.repository.fetch(source.branch.repository, revid)
1387
target.set_parent_ids([revid])
1388
return target.basis_tree(), target
1390
_matching_from_tree_format = WorkingTreeFormat4()
1391
_matching_to_tree_format = WorkingTreeFormat4()
1392
_test_mutable_trees_to_test_trees = make_source_parent_tree
1394
def _iter_changes(self, include_unchanged=False,
1395
specific_files=None, pb=None, extra_trees=[],
1396
require_versioned=True):
1397
"""Return the changes from source to target.
1399
:return: An iterator that yields tuples. See InterTree._iter_changes
1401
:param specific_files: An optional list of file paths to restrict the
1402
comparison to. When mapping filenames to ids, all matches in all
1403
trees (including optional extra_trees) are used, and all children of
1404
matched directories are included.
1405
:param include_unchanged: An optional boolean requesting the inclusion of
1406
unchanged entries in the result.
1407
:param extra_trees: An optional list of additional trees to use when
1408
mapping the contents of specific_files (paths) to file_ids.
1409
:param require_versioned: If True, all files in specific_files must be
1410
versioned in one of source, target, extra_trees or
1411
PathsNotVersionedError is raised.
1413
utf8_decode = cache_utf8._utf8_decode
1414
_minikind_to_kind = dirstate.DirState._minikind_to_kind
1415
# NB: show_status depends on being able to pass in non-versioned files
1416
# and report them as unknown
1417
# TODO: handle extra trees in the dirstate.
1419
for f in super(InterDirStateTree, self)._iter_changes(
1420
include_unchanged, specific_files, pb, extra_trees,
1424
parent_ids = self.target.get_parent_ids()
1426
if self.source._revision_id == NULL_REVISION:
1428
indices = (target_index,)
1430
assert (self.source._revision_id in parent_ids), \
1431
"Failure: source._revision_id: %s not in target.parent_ids(%s)" % (
1432
self.source._revision_id, parent_ids)
1433
source_index = 1 + parent_ids.index(self.source._revision_id)
1434
indices = (source_index,target_index)
1435
# -- make all specific_files utf8 --
1437
specific_files_utf8 = set()
1438
for path in specific_files:
1439
specific_files_utf8.add(path.encode('utf8'))
1440
specific_files = specific_files_utf8
1442
specific_files = set([''])
1443
# -- specific_files is now a utf8 path set --
1444
# -- get the state object and prepare it.
1445
state = self.target.current_dirstate()
1446
state._read_dirblocks_if_needed()
1447
def _entries_for_path(path):
1448
"""Return a list with all the entries that match path for all ids.
1450
dirname, basename = os.path.split(path)
1451
key = (dirname, basename, '')
1452
block_index, present = state._find_block_index_from_key(key)
1454
# the block which should contain path is absent.
1457
block = state._dirblocks[block_index][1]
1458
entry_index, _ = state._find_entry_index(key, block)
1459
# we may need to look at multiple entries at this path: walk while the specific_files match.
1460
while (entry_index < len(block) and
1461
block[entry_index][0][0:2] == key[0:2]):
1462
result.append(block[entry_index])
1465
if require_versioned:
1466
# -- check all supplied paths are versioned in a search tree. --
1467
all_versioned = True
1468
for path in specific_files:
1469
path_entries = _entries_for_path(path)
1470
if not path_entries:
1471
# this specified path is not present at all: error
1472
all_versioned = False
1474
found_versioned = False
1475
# for each id at this path
1476
for entry in path_entries:
1478
for index in indices:
1479
if entry[1][index][0] != 'a': # absent
1480
found_versioned = True
1481
# all good: found a versioned cell
1483
if not found_versioned:
1484
# none of the indexes was not 'absent' at all ids for this
1486
all_versioned = False
1488
if not all_versioned:
1489
raise errors.PathsNotVersionedError(paths)
1490
# -- remove redundancy in supplied specific_files to prevent over-scanning --
1491
search_specific_files = set()
1492
for path in specific_files:
1493
other_specific_files = specific_files.difference(set([path]))
1494
if not osutils.is_inside_any(other_specific_files, path):
1495
# this is a top level path, we must check it.
1496
search_specific_files.add(path)
1498
# compare source_index and target_index at or under each element of search_specific_files.
1499
# follow the following comparison table. Note that we only want to do diff operations when
1500
# the target is fdl because thats when the walkdirs logic will have exposed the pathinfo
1504
# Source | Target | disk | action
1505
# r | fdl | | add source to search, add id path move and perform
1506
# | | | diff check on source-target
1507
# r | fdl | a | dangling file that was present in the basis.
1509
# r | a | | add source to search
1511
# r | r | | this path is present in a non-examined tree, skip.
1512
# r | r | a | this path is present in a non-examined tree, skip.
1513
# a | fdl | | add new id
1514
# a | fdl | a | dangling locally added file, skip
1515
# a | a | | not present in either tree, skip
1516
# a | a | a | not present in any tree, skip
1517
# a | r | | not present in either tree at this path, skip as it
1518
# | | | may not be selected by the users list of paths.
1519
# a | r | a | not present in either tree at this path, skip as it
1520
# | | | may not be selected by the users list of paths.
1521
# fdl | fdl | | content in both: diff them
1522
# fdl | fdl | a | deleted locally, but not unversioned - show as deleted ?
1523
# fdl | a | | unversioned: output deleted id for now
1524
# fdl | a | a | unversioned and deleted: output deleted id
1525
# fdl | r | | relocated in this tree, so add target to search.
1526
# | | | Dont diff, we will see an r,fd; pair when we reach
1527
# | | | this id at the other path.
1528
# fdl | r | a | relocated in this tree, so add target to search.
1529
# | | | Dont diff, we will see an r,fd; pair when we reach
1530
# | | | this id at the other path.
1532
# for all search_indexs in each path at or under each element of
1533
# search_specific_files, if the detail is relocated: add the id, and add the
1534
# relocated path as one to search if its not searched already. If the
1535
# detail is not relocated, add the id.
1536
searched_specific_files = set()
1537
NULL_PARENT_DETAILS = dirstate.DirState.NULL_PARENT_DETAILS
1538
# Using a list so that we can access the values and change them in
1539
# nested scope. Each one is [path, file_id, entry]
1540
last_source_parent = [None, None, None]
1541
last_target_parent = [None, None, None]
1543
def _process_entry(entry, path_info):
1544
"""Compare an entry and real disk to generate delta information.
1546
:param path_info: top_relpath, basename, kind, lstat, abspath for
1547
the path of entry. If None, then the path is considered absent.
1548
(Perhaps we should pass in a concrete entry for this ?)
1550
# TODO: when a parent has been renamed, dont emit path renames for children,
1551
if source_index is None:
1552
source_details = NULL_PARENT_DETAILS
1554
source_details = entry[1][source_index]
1555
target_details = entry[1][target_index]
1556
source_minikind = source_details[0]
1557
target_minikind = target_details[0]
1558
if source_minikind in 'fdlr' and target_minikind in 'fdl':
1559
# claimed content in both: diff
1560
# r | fdl | | add source to search, add id path move and perform
1561
# | | | diff check on source-target
1562
# r | fdl | a | dangling file that was present in the basis.
1564
if source_minikind in 'r':
1565
# add the source to the search path to find any children it
1566
# has. TODO ? : only add if it is a container ?
1567
if not osutils.is_inside_any(searched_specific_files,
1569
search_specific_files.add(source_details[1])
1570
# generate the old path; this is needed for stating later
1572
old_path = source_details[1]
1573
old_dirname, old_basename = os.path.split(old_path)
1574
path = pathjoin(entry[0][0], entry[0][1])
1575
old_entry = state._get_entry(source_index,
1577
# update the source details variable to be the real
1579
source_details = old_entry[1][source_index]
1580
source_minikind = source_details[0]
1582
old_dirname = entry[0][0]
1583
old_basename = entry[0][1]
1584
old_path = path = pathjoin(old_dirname, old_basename)
1585
if path_info is None:
1586
# the file is missing on disk, show as removed.
1587
old_path = pathjoin(entry[0][0], entry[0][1])
1588
content_change = True
1592
# source and target are both versioned and disk file is present.
1593
target_kind = path_info[2]
1594
if target_kind == 'directory':
1595
if source_minikind != 'd':
1596
content_change = True
1598
# directories have no fingerprint
1599
content_change = False
1601
elif target_kind == 'file':
1602
if source_minikind != 'f':
1603
content_change = True
1605
# has it changed? fast path: size, slow path: sha1.
1606
if source_details[2] != path_info[3].st_size:
1607
content_change = True
1609
# maybe the same. Get the hash
1610
new_hash = self.target._hashcache.get_sha1(
1612
content_change = (new_hash != source_details[1])
1614
stat.S_ISREG(path_info[3].st_mode)
1615
and stat.S_IEXEC & path_info[3].st_mode)
1616
elif target_kind == 'symlink':
1617
if source_minikind != 'l':
1618
content_change = True
1620
# TODO: check symlink supported for windows users
1621
# and grab from target state here.
1622
link_target = os.readlink(path_info[4])
1623
content_change = (link_target != source_details[1])
1626
raise Exception, "unknown kind %s" % path_info[2]
1627
# parent id is the entry for the path in the target tree
1628
if old_dirname == last_source_parent[0]:
1629
source_parent_id = last_source_parent[1]
1631
source_parent_entry = state._get_entry(source_index,
1632
path_utf8=old_dirname)
1633
source_parent_id = source_parent_entry[0][2]
1634
if source_parent_id == entry[0][2]:
1635
# This is the root, so the parent is None
1636
source_parent_id = None
1638
last_source_parent[0] = old_dirname
1639
last_source_parent[1] = source_parent_id
1640
last_source_parent[2] = source_parent_entry
1642
new_dirname = entry[0][0]
1643
if new_dirname == last_target_parent[0]:
1644
target_parent_id = last_target_parent[1]
1646
# TODO: We don't always need to do the lookup, because the
1647
# parent entry will be the same as the source entry.
1648
target_parent_entry = state._get_entry(target_index,
1649
path_utf8=new_dirname)
1650
target_parent_id = target_parent_entry[0][2]
1651
if target_parent_id == entry[0][2]:
1652
# This is the root, so the parent is None
1653
target_parent_id = None
1655
last_target_parent[0] = new_dirname
1656
last_target_parent[1] = target_parent_id
1657
last_target_parent[2] = target_parent_entry
1659
source_exec = source_details[3]
1660
path_unicode = utf8_decode(path)[0]
1661
return ((entry[0][2], path_unicode, content_change,
1663
(source_parent_id, target_parent_id),
1664
(old_basename, entry[0][1]),
1665
(_minikind_to_kind[source_minikind], target_kind),
1666
(source_exec, target_exec)),)
1667
elif source_minikind in 'a' and target_minikind in 'fdl':
1668
# looks like a new file
1669
if path_info is not None:
1670
path = pathjoin(entry[0][0], entry[0][1])
1671
# parent id is the entry for the path in the target tree
1672
# TODO: these are the same for an entire directory: cache em.
1673
parent_id = state._get_entry(target_index, path_utf8=entry[0][0])[0][2]
1674
if parent_id == entry[0][2]:
1677
new_executable = bool(
1678
stat.S_ISREG(path_info[3].st_mode)
1679
and stat.S_IEXEC & path_info[3].st_mode)
1680
path_unicode = utf8_decode(path)[0]
1681
return ((entry[0][2], path_unicode, True,
1684
(None, entry[0][1]),
1685
(None, path_info[2]),
1686
(None, new_executable)),)
1688
# but its not on disk: we deliberately treat this as just
1689
# never-present. (Why ?! - RBC 20070224)
1691
elif source_minikind in 'fdl' and target_minikind in 'a':
1692
# unversioned, possibly, or possibly not deleted: we dont care.
1693
# if its still on disk, *and* theres no other entry at this
1694
# path [we dont know this in this routine at the moment -
1695
# perhaps we should change this - then it would be an unknown.
1696
old_path = pathjoin(entry[0][0], entry[0][1])
1697
# parent id is the entry for the path in the target tree
1698
parent_id = state._get_entry(source_index, path_utf8=entry[0][0])[0][2]
1699
if parent_id == entry[0][2]:
1701
old_path_unicode = utf8_decode(old_path)[0]
1702
return ((entry[0][2], old_path_unicode, True,
1705
(entry[0][1], None),
1706
(_minikind_to_kind[source_minikind], None),
1707
(source_details[3], None)),)
1708
elif source_minikind in 'fdl' and target_minikind in 'r':
1709
# a rename; could be a true rename, or a rename inherited from
1710
# a renamed parent. TODO: handle this efficiently. Its not
1711
# common case to rename dirs though, so a correct but slow
1712
# implementation will do.
1713
if not osutils.is_inside_any(searched_specific_files, target_details[1]):
1714
search_specific_files.add(target_details[1])
1716
import pdb;pdb.set_trace()
1718
while search_specific_files:
1719
# TODO: the pending list should be lexically sorted?
1720
current_root = search_specific_files.pop()
1721
searched_specific_files.add(current_root)
1722
# process the entries for this containing directory: the rest will be
1723
# found by their parents recursively.
1724
root_entries = _entries_for_path(current_root)
1725
root_abspath = self.target.abspath(current_root)
1727
root_stat = os.lstat(root_abspath)
1729
if e.errno == errno.ENOENT:
1730
# the path does not exist: let _process_entry know that.
1731
root_dir_info = None
1733
# some other random error: hand it up.
1736
root_dir_info = ('', current_root,
1737
osutils.file_kind_from_stat_mode(root_stat.st_mode), root_stat,
1739
if not root_entries and not root_dir_info:
1740
# this specified path is not present at all, skip it.
1742
for entry in root_entries:
1743
for result in _process_entry(entry, root_dir_info):
1744
# this check should probably be outside the loop: one
1745
# 'iterate two trees' api, and then _iter_changes filters
1746
# unchanged pairs. - RBC 20070226
1747
if (include_unchanged
1748
or result[2] # content change
1749
or result[3][0] != result[3][1] # versioned status
1750
or result[4][0] != result[4][1] # parent id
1751
or result[5][0] != result[5][1] # name
1752
or result[6][0] != result[6][1] # kind
1753
or result[7][0] != result[7][1] # executable
1756
dir_iterator = osutils._walkdirs_utf8(root_abspath, prefix=current_root)
1757
initial_key = (current_root, '', '')
1758
block_index, _ = state._find_block_index_from_key(initial_key)
1759
if block_index == 0:
1760
# we have processed the total root already, but because the
1761
# initial key matched it we should skip it here.
1764
current_dir_info = dir_iterator.next()
1766
if e.errno in (errno.ENOENT, errno.ENOTDIR):
1767
# there may be directories in the inventory even though
1768
# this path is not a file on disk: so mark it as end of
1770
current_dir_info = None
1774
if current_dir_info[0][0] == '':
1775
# remove .bzr from iteration
1776
bzr_index = bisect_left(current_dir_info[1], ('.bzr',))
1777
assert current_dir_info[1][bzr_index][0] == '.bzr'
1778
del current_dir_info[1][bzr_index]
1779
# walk until both the directory listing and the versioned metadata
1780
# are exhausted. TODO: reevaluate this, perhaps we should stop when
1781
# the versioned data runs out.
1782
if (block_index < len(state._dirblocks) and
1783
osutils.is_inside(current_root, state._dirblocks[block_index][0])):
1784
current_block = state._dirblocks[block_index]
1786
current_block = None
1787
while (current_dir_info is not None or
1788
current_block is not None):
1789
if (current_dir_info and current_block
1790
and current_dir_info[0][0] != current_block[0]):
1791
if current_dir_info[0][0] < current_block[0] :
1792
# import pdb; pdb.set_trace()
1793
# print 'unversioned dir'
1794
# filesystem data refers to paths not covered by the dirblock.
1795
# this has two possibilities:
1796
# A) it is versioned but empty, so there is no block for it
1797
# B) it is not versioned.
1798
# in either case it was processed by the containing directories walk:
1799
# if it is root/foo, when we walked root we emitted it,
1800
# or if we ere given root/foo to walk specifically, we
1801
# emitted it when checking the walk-root entries
1802
# advance the iterator and loop - we dont need to emit it.
1804
current_dir_info = dir_iterator.next()
1805
except StopIteration:
1806
current_dir_info = None
1808
# We have a dirblock entry for this location, but there
1809
# is no filesystem path for this. This is most likely
1810
# because a directory was removed from the disk.
1811
# We don't have to report the missing directory,
1812
# because that should have already been handled, but we
1813
# need to handle all of the files that are contained
1815
for current_entry in current_block[1]:
1816
# entry referring to file not present on disk.
1817
# advance the entry only, after processing.
1818
for result in _process_entry(current_entry, None):
1819
# this check should probably be outside the loop: one
1820
# 'iterate two trees' api, and then _iter_changes filters
1821
# unchanged pairs. - RBC 20070226
1822
if (include_unchanged
1823
or result[2] # content change
1824
or result[3][0] != result[3][1] # versioned status
1825
or result[4][0] != result[4][1] # parent id
1826
or result[5][0] != result[5][1] # name
1827
or result[6][0] != result[6][1] # kind
1828
or result[7][0] != result[7][1] # executable
1832
if (block_index < len(state._dirblocks) and
1833
osutils.is_inside(current_root,
1834
state._dirblocks[block_index][0])):
1835
current_block = state._dirblocks[block_index]
1837
current_block = None
1840
if current_block and entry_index < len(current_block[1]):
1841
current_entry = current_block[1][entry_index]
1843
current_entry = None
1844
advance_entry = True
1846
if current_dir_info and path_index < len(current_dir_info[1]):
1847
current_path_info = current_dir_info[1][path_index]
1849
current_path_info = None
1851
while (current_entry is not None or
1852
current_path_info is not None):
1853
if current_entry is None:
1854
# no more entries: yield current_pathinfo as an
1855
# unversioned file: its not the same as a path in any
1856
# tree in the dirstate.
1857
new_executable = bool(
1858
stat.S_ISREG(current_path_info[3].st_mode)
1859
and stat.S_IEXEC & current_path_info[3].st_mode)
1860
yield (None, current_path_info[0], True,
1863
(None, current_path_info[1]),
1864
(None, current_path_info[2]),
1865
(None, new_executable))
1866
elif current_path_info is None:
1867
# no path is fine: the per entry code will handle it.
1868
for result in _process_entry(current_entry, current_path_info):
1869
# this check should probably be outside the loop: one
1870
# 'iterate two trees' api, and then _iter_changes filters
1871
# unchanged pairs. - RBC 20070226
1872
if (include_unchanged
1873
or result[2] # content change
1874
or result[3][0] != result[3][1] # versioned status
1875
or result[4][0] != result[4][1] # parent id
1876
or result[5][0] != result[5][1] # name
1877
or result[6][0] != result[6][1] # kind
1878
or result[7][0] != result[7][1] # executable
1881
elif current_entry[0][1] != current_path_info[1]:
1882
if current_path_info[1] < current_entry[0][1]:
1883
# extra file on disk: pass for now, but only
1884
# increment the path, not the entry
1885
# import pdb; pdb.set_trace()
1886
# print 'unversioned file'
1887
advance_entry = False
1889
# entry referring to file not present on disk.
1890
# advance the entry only, after processing.
1891
for result in _process_entry(current_entry, None):
1892
# this check should probably be outside the loop: one
1893
# 'iterate two trees' api, and then _iter_changes filters
1894
# unchanged pairs. - RBC 20070226
1895
if (include_unchanged
1896
or result[2] # content change
1897
or result[3][0] != result[3][1] # versioned status
1898
or result[4][0] != result[4][1] # parent id
1899
or result[5][0] != result[5][1] # name
1900
or result[6][0] != result[6][1] # kind
1901
or result[7][0] != result[7][1] # executable
1904
advance_path = False
1906
for result in _process_entry(current_entry, current_path_info):
1907
# this check should probably be outside the loop: one
1908
# 'iterate two trees' api, and then _iter_changes filters
1909
# unchanged pairs. - RBC 20070226
1910
if (include_unchanged
1911
or result[2] # content change
1912
or result[3][0] != result[3][1] # versioned status
1913
or result[4][0] != result[4][1] # parent id
1914
or result[5][0] != result[5][1] # name
1915
or result[6][0] != result[6][1] # kind
1916
or result[7][0] != result[7][1] # executable
1919
if advance_entry and current_entry is not None:
1921
if entry_index < len(current_block[1]):
1922
current_entry = current_block[1][entry_index]
1924
current_entry = None
1926
advance_entry = True # reset the advance flaga
1927
if advance_path and current_path_info is not None:
1929
if path_index < len(current_dir_info[1]):
1930
current_path_info = current_dir_info[1][path_index]
1932
current_path_info = None
1934
advance_path = True # reset the advance flagg.
1935
if current_block is not None:
1937
if (block_index < len(state._dirblocks) and
1938
osutils.is_inside(current_root, state._dirblocks[block_index][0])):
1939
current_block = state._dirblocks[block_index]
1941
current_block = None
1942
if current_dir_info is not None:
1944
current_dir_info = dir_iterator.next()
1945
except StopIteration:
1946
current_dir_info = None
1950
def is_compatible(source, target):
1951
# the target must be a dirstate working tree
1952
if not isinstance(target, WorkingTree4):
1954
# the source must be a revtreee or dirstate rev tree.
1955
if not isinstance(source,
1956
(revisiontree.RevisionTree, DirStateRevisionTree)):
1958
# the source revid must be in the target dirstate
1959
if not (source._revision_id == NULL_REVISION or
1960
source._revision_id in target.get_parent_ids()):
1961
# TODO: what about ghosts? it may well need to
1962
# check for them explicitly.
1966
InterTree.register_optimiser(InterDirStateTree)