/brz/remove-bazaar

To get this branch, use:
bzr branch http://gegoxaren.bato24.eu/bzr/brz/remove-bazaar

« back to all changes in this revision

Viewing changes to bzrlib/tests/test_repository.py

  • Committer: Canonical.com Patch Queue Manager
  • Date: 2008-08-01 18:19:44 UTC
  • mfrom: (3514.3.13 multi_walker)
  • Revision ID: pqm@pqm.ubuntu.com-20080801181944-pl10irs3d0p0a5ia
(jam) Add a MultiWalker class to check for changes between >2 trees

Show diffs side-by-side

added added

removed removed

Lines of Context:
1
 
# Copyright (C) 2006-2010 Canonical Ltd
 
1
# Copyright (C) 2006, 2007 Canonical Ltd
2
2
#
3
3
# This program is free software; you can redistribute it and/or modify
4
4
# it under the terms of the GNU General Public License as published by
12
12
#
13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
 
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
16
16
 
17
17
"""Tests for the Repository facility that are not interface tests.
18
18
 
19
 
For interface tests see tests/per_repository/*.py.
 
19
For interface tests see tests/repository_implementations/*.py.
20
20
 
21
21
For concrete class tests see this file, and for storage formats tests
22
22
also see this file.
23
23
"""
24
24
 
 
25
import md5
25
26
from stat import S_ISDIR
26
 
import sys
 
27
from StringIO import StringIO
27
28
 
28
29
import bzrlib
29
 
from bzrlib.errors import (NoSuchFile,
 
30
from bzrlib.errors import (NotBranchError,
 
31
                           NoSuchFile,
30
32
                           UnknownFormatError,
31
33
                           UnsupportedFormatError,
32
34
                           )
33
 
from bzrlib import (
34
 
    graph,
35
 
    tests,
36
 
    )
37
 
from bzrlib.btree_index import BTreeBuilder, BTreeGraphIndex
38
 
from bzrlib.index import GraphIndex
 
35
from bzrlib import graph
 
36
from bzrlib.index import GraphIndex, InMemoryGraphIndex
39
37
from bzrlib.repository import RepositoryFormat
 
38
from bzrlib.smart import server
40
39
from bzrlib.tests import (
41
40
    TestCase,
42
41
    TestCaseWithTransport,
 
42
    TestSkipped,
 
43
    test_knit,
43
44
    )
44
45
from bzrlib.transport import (
 
46
    fakenfs,
45
47
    get_transport,
46
48
    )
 
49
from bzrlib.transport.memory import MemoryServer
 
50
from bzrlib.util import bencode
47
51
from bzrlib import (
48
52
    bzrdir,
49
53
    errors,
50
54
    inventory,
51
 
    osutils,
 
55
    progress,
52
56
    repository,
53
57
    revision as _mod_revision,
 
58
    symbol_versioning,
54
59
    upgrade,
55
 
    versionedfile,
56
60
    workingtree,
57
61
    )
58
 
from bzrlib.repofmt import (
59
 
    groupcompress_repo,
60
 
    knitrepo,
61
 
    pack_repo,
62
 
    weaverepo,
63
 
    )
 
62
from bzrlib.repofmt import knitrepo, weaverepo, pack_repo
64
63
 
65
64
 
66
65
class TestDefaultFormat(TestCase):
95
94
class SampleRepositoryFormat(repository.RepositoryFormat):
96
95
    """A sample format
97
96
 
98
 
    this format is initializable, unsupported to aid in testing the
 
97
    this format is initializable, unsupported to aid in testing the 
99
98
    open and open(unsupported=True) routines.
100
99
    """
101
100
 
122
121
    def test_find_format(self):
123
122
        # is the right format object found for a repository?
124
123
        # create a branch with a few known format objects.
125
 
        # this is not quite the same as
 
124
        # this is not quite the same as 
126
125
        self.build_tree(["foo/", "bar/"])
127
126
        def check_format(format, url):
128
127
            dir = format._matchingbzrdir.initialize(url)
131
130
            found_format = repository.RepositoryFormat.find_format(dir)
132
131
            self.failUnless(isinstance(found_format, format.__class__))
133
132
        check_format(weaverepo.RepositoryFormat7(), "bar")
134
 
 
 
133
        
135
134
    def test_find_format_no_repository(self):
136
135
        dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
137
136
        self.assertRaises(errors.NoRepositoryPresent,
167
166
        """Weaves need topological data insertion."""
168
167
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
169
168
        repo = weaverepo.RepositoryFormat6().initialize(control)
170
 
        self.assertEqual('topological', repo._format._fetch_order)
 
169
        self.assertEqual('topological', repo._fetch_order)
171
170
 
172
171
    def test_attribute__fetch_uses_deltas(self):
173
172
        """Weaves do not reuse deltas."""
174
173
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
175
174
        repo = weaverepo.RepositoryFormat6().initialize(control)
176
 
        self.assertEqual(False, repo._format._fetch_uses_deltas)
 
175
        self.assertEqual(False, repo._fetch_uses_deltas)
177
176
 
178
177
    def test_attribute__fetch_reconcile(self):
179
178
        """Weave repositories need a reconcile after fetch."""
180
179
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
181
180
        repo = weaverepo.RepositoryFormat6().initialize(control)
182
 
        self.assertEqual(True, repo._format._fetch_reconcile)
 
181
        self.assertEqual(True, repo._fetch_reconcile)
183
182
 
184
183
    def test_no_ancestry_weave(self):
185
184
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
202
201
        """Weaves need topological data insertion."""
203
202
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
204
203
        repo = weaverepo.RepositoryFormat7().initialize(control)
205
 
        self.assertEqual('topological', repo._format._fetch_order)
 
204
        self.assertEqual('topological', repo._fetch_order)
206
205
 
207
206
    def test_attribute__fetch_uses_deltas(self):
208
207
        """Weaves do not reuse deltas."""
209
208
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
210
209
        repo = weaverepo.RepositoryFormat7().initialize(control)
211
 
        self.assertEqual(False, repo._format._fetch_uses_deltas)
 
210
        self.assertEqual(False, repo._fetch_uses_deltas)
212
211
 
213
212
    def test_attribute__fetch_reconcile(self):
214
213
        """Weave repositories need a reconcile after fetch."""
215
214
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
216
215
        repo = weaverepo.RepositoryFormat7().initialize(control)
217
 
        self.assertEqual(True, repo._format._fetch_reconcile)
 
216
        self.assertEqual(True, repo._fetch_reconcile)
218
217
 
219
218
    def test_disk_layout(self):
220
219
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
243
242
        tree = control.create_workingtree()
244
243
        tree.add(['foo'], ['Foo:Bar'], ['file'])
245
244
        tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
246
 
        try:
247
 
            tree.commit('first post', rev_id='first')
248
 
        except errors.IllegalPath:
249
 
            if sys.platform != 'win32':
250
 
                raise
251
 
            self.knownFailure('Foo:Bar cannot be used as a file-id on windows'
252
 
                              ' in repo format 7')
253
 
            return
 
245
        tree.commit('first post', rev_id='first')
254
246
        self.assertEqualDiff(
255
247
            '# bzr weave file v5\n'
256
248
            'i\n'
291
283
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
292
284
        repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
293
285
        t = control.get_repository_transport(None)
294
 
        # TODO: Should check there is a 'lock' toplevel directory,
 
286
        # TODO: Should check there is a 'lock' toplevel directory, 
295
287
        # regardless of contents
296
288
        self.assertFalse(t.has('lock/held/info'))
297
289
        repo.lock_write()
350
342
 
351
343
 
352
344
class TestFormatKnit1(TestCaseWithTransport):
353
 
 
 
345
    
354
346
    def test_attribute__fetch_order(self):
355
347
        """Knits need topological data insertion."""
356
348
        repo = self.make_repository('.',
357
349
                format=bzrdir.format_registry.get('knit')())
358
 
        self.assertEqual('topological', repo._format._fetch_order)
 
350
        self.assertEqual('topological', repo._fetch_order)
359
351
 
360
352
    def test_attribute__fetch_uses_deltas(self):
361
353
        """Knits reuse deltas."""
362
354
        repo = self.make_repository('.',
363
355
                format=bzrdir.format_registry.get('knit')())
364
 
        self.assertEqual(True, repo._format._fetch_uses_deltas)
 
356
        self.assertEqual(True, repo._fetch_uses_deltas)
365
357
 
366
358
    def test_disk_layout(self):
367
359
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
454
446
        repo = self.make_repository('.',
455
447
                format=bzrdir.format_registry.get('knit')())
456
448
        inv_xml = '<inventory format="5">\n</inventory>\n'
457
 
        inv = repo._deserialise_inventory('test-rev-id', inv_xml)
 
449
        inv = repo.deserialise_inventory('test-rev-id', inv_xml)
458
450
        self.assertEqual('test-rev-id', inv.root.revision)
459
451
 
460
452
    def test_deserialise_uses_global_revision_id(self):
466
458
        # Arguably, the deserialise_inventory should detect a mismatch, and
467
459
        # raise an error, rather than silently using one revision_id over the
468
460
        # other.
469
 
        self.assertRaises(AssertionError, repo._deserialise_inventory,
 
461
        self.assertRaises(AssertionError, repo.deserialise_inventory,
470
462
            'test-rev-id', inv_xml)
471
 
        inv = repo._deserialise_inventory('other-rev-id', inv_xml)
 
463
        inv = repo.deserialise_inventory('other-rev-id', inv_xml)
472
464
        self.assertEqual('other-rev-id', inv.root.revision)
473
465
 
474
466
    def test_supports_external_lookups(self):
480
472
class DummyRepository(object):
481
473
    """A dummy repository for testing."""
482
474
 
483
 
    _format = None
484
475
    _serializer = None
485
476
 
486
477
    def supports_rich_root(self):
487
 
        if self._format is not None:
488
 
            return self._format.rich_root_data
489
478
        return False
490
479
 
491
 
    def get_graph(self):
492
 
        raise NotImplementedError
493
 
 
494
 
    def get_parent_map(self, revision_ids):
495
 
        raise NotImplementedError
496
 
 
497
480
 
498
481
class InterDummy(repository.InterRepository):
499
482
    """An inter-repository optimised code path for DummyRepository.
506
489
    @staticmethod
507
490
    def is_compatible(repo_source, repo_target):
508
491
        """InterDummy is compatible with DummyRepository."""
509
 
        return (isinstance(repo_source, DummyRepository) and
 
492
        return (isinstance(repo_source, DummyRepository) and 
510
493
            isinstance(repo_target, DummyRepository))
511
494
 
512
495
 
525
508
 
526
509
    def assertGetsDefaultInterRepository(self, repo_a, repo_b):
527
510
        """Asserts that InterRepository.get(repo_a, repo_b) -> the default.
528
 
 
 
511
        
529
512
        The effective default is now InterSameDataRepository because there is
530
513
        no actual sane default in the presence of incompatible data models.
531
514
        """
542
525
        # pair that it returns true on for the is_compatible static method
543
526
        # check
544
527
        dummy_a = DummyRepository()
545
 
        dummy_a._format = RepositoryFormat()
546
528
        dummy_b = DummyRepository()
547
 
        dummy_b._format = RepositoryFormat()
548
529
        repo = self.make_repository('.')
549
530
        # hack dummies to look like repo somewhat.
550
531
        dummy_a._serializer = repo._serializer
551
 
        dummy_a._format.supports_tree_reference = repo._format.supports_tree_reference
552
 
        dummy_a._format.rich_root_data = repo._format.rich_root_data
553
532
        dummy_b._serializer = repo._serializer
554
 
        dummy_b._format.supports_tree_reference = repo._format.supports_tree_reference
555
 
        dummy_b._format.rich_root_data = repo._format.rich_root_data
556
533
        repository.InterRepository.register_optimiser(InterDummy)
557
534
        try:
558
535
            # we should get the default for something InterDummy returns False
621
598
 
622
599
 
623
600
class TestMisc(TestCase):
624
 
 
 
601
    
625
602
    def test_unescape_xml(self):
626
603
        """We get some kind of error when malformed entities are passed"""
627
 
        self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;')
 
604
        self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;') 
628
605
 
629
606
 
630
607
class TestRepositoryFormatKnit3(TestCaseWithTransport):
634
611
        format = bzrdir.BzrDirMetaFormat1()
635
612
        format.repository_format = knitrepo.RepositoryFormatKnit3()
636
613
        repo = self.make_repository('.', format=format)
637
 
        self.assertEqual('topological', repo._format._fetch_order)
 
614
        self.assertEqual('topological', repo._fetch_order)
638
615
 
639
616
    def test_attribute__fetch_uses_deltas(self):
640
617
        """Knits reuse deltas."""
641
618
        format = bzrdir.BzrDirMetaFormat1()
642
619
        format.repository_format = knitrepo.RepositoryFormatKnit3()
643
620
        repo = self.make_repository('.', format=format)
644
 
        self.assertEqual(True, repo._format._fetch_uses_deltas)
 
621
        self.assertEqual(True, repo._fetch_uses_deltas)
645
622
 
646
623
    def test_convert(self):
647
624
        """Ensure the upgrade adds weaves for roots"""
679
656
        self.assertFalse(repo._format.supports_external_lookups)
680
657
 
681
658
 
682
 
class Test2a(tests.TestCaseWithMemoryTransport):
683
 
 
684
 
    def test_fetch_combines_groups(self):
685
 
        builder = self.make_branch_builder('source', format='2a')
686
 
        builder.start_series()
687
 
        builder.build_snapshot('1', None, [
688
 
            ('add', ('', 'root-id', 'directory', '')),
689
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
690
 
        builder.build_snapshot('2', ['1'], [
691
 
            ('modify', ('file-id', 'content-2\n'))])
692
 
        builder.finish_series()
693
 
        source = builder.get_branch()
694
 
        target = self.make_repository('target', format='2a')
695
 
        target.fetch(source.repository)
696
 
        target.lock_read()
697
 
        self.addCleanup(target.unlock)
698
 
        details = target.texts._index.get_build_details(
699
 
            [('file-id', '1',), ('file-id', '2',)])
700
 
        file_1_details = details[('file-id', '1')]
701
 
        file_2_details = details[('file-id', '2')]
702
 
        # The index, and what to read off disk, should be the same for both
703
 
        # versions of the file.
704
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
705
 
 
706
 
    def test_fetch_combines_groups(self):
707
 
        builder = self.make_branch_builder('source', format='2a')
708
 
        builder.start_series()
709
 
        builder.build_snapshot('1', None, [
710
 
            ('add', ('', 'root-id', 'directory', '')),
711
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
712
 
        builder.build_snapshot('2', ['1'], [
713
 
            ('modify', ('file-id', 'content-2\n'))])
714
 
        builder.finish_series()
715
 
        source = builder.get_branch()
716
 
        target = self.make_repository('target', format='2a')
717
 
        target.fetch(source.repository)
718
 
        target.lock_read()
719
 
        self.addCleanup(target.unlock)
720
 
        details = target.texts._index.get_build_details(
721
 
            [('file-id', '1',), ('file-id', '2',)])
722
 
        file_1_details = details[('file-id', '1')]
723
 
        file_2_details = details[('file-id', '2')]
724
 
        # The index, and what to read off disk, should be the same for both
725
 
        # versions of the file.
726
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
727
 
 
728
 
    def test_fetch_combines_groups(self):
729
 
        builder = self.make_branch_builder('source', format='2a')
730
 
        builder.start_series()
731
 
        builder.build_snapshot('1', None, [
732
 
            ('add', ('', 'root-id', 'directory', '')),
733
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
734
 
        builder.build_snapshot('2', ['1'], [
735
 
            ('modify', ('file-id', 'content-2\n'))])
736
 
        builder.finish_series()
737
 
        source = builder.get_branch()
738
 
        target = self.make_repository('target', format='2a')
739
 
        target.fetch(source.repository)
740
 
        target.lock_read()
741
 
        self.addCleanup(target.unlock)
742
 
        details = target.texts._index.get_build_details(
743
 
            [('file-id', '1',), ('file-id', '2',)])
744
 
        file_1_details = details[('file-id', '1')]
745
 
        file_2_details = details[('file-id', '2')]
746
 
        # The index, and what to read off disk, should be the same for both
747
 
        # versions of the file.
748
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
749
 
 
750
 
    def test_format_pack_compresses_True(self):
751
 
        repo = self.make_repository('repo', format='2a')
752
 
        self.assertTrue(repo._format.pack_compresses)
753
 
 
754
 
    def test_inventories_use_chk_map_with_parent_base_dict(self):
755
 
        tree = self.make_branch_and_memory_tree('repo', format="2a")
756
 
        tree.lock_write()
757
 
        tree.add([''], ['TREE_ROOT'])
758
 
        revid = tree.commit("foo")
759
 
        tree.unlock()
760
 
        tree.lock_read()
761
 
        self.addCleanup(tree.unlock)
762
 
        inv = tree.branch.repository.get_inventory(revid)
763
 
        self.assertNotEqual(None, inv.parent_id_basename_to_file_id)
764
 
        inv.parent_id_basename_to_file_id._ensure_root()
765
 
        inv.id_to_entry._ensure_root()
766
 
        self.assertEqual(65536, inv.id_to_entry._root_node.maximum_size)
767
 
        self.assertEqual(65536,
768
 
            inv.parent_id_basename_to_file_id._root_node.maximum_size)
769
 
 
770
 
    def test_autopack_unchanged_chk_nodes(self):
771
 
        # at 20 unchanged commits, chk pages are packed that are split into
772
 
        # two groups such that the new pack being made doesn't have all its
773
 
        # pages in the source packs (though they are in the repository).
774
 
        # Use a memory backed repository, we don't need to hit disk for this
775
 
        tree = self.make_branch_and_memory_tree('tree', format='2a')
776
 
        tree.lock_write()
777
 
        self.addCleanup(tree.unlock)
778
 
        tree.add([''], ['TREE_ROOT'])
779
 
        for pos in range(20):
780
 
            tree.commit(str(pos))
781
 
 
782
 
    def test_pack_with_hint(self):
783
 
        tree = self.make_branch_and_memory_tree('tree', format='2a')
784
 
        tree.lock_write()
785
 
        self.addCleanup(tree.unlock)
786
 
        tree.add([''], ['TREE_ROOT'])
787
 
        # 1 commit to leave untouched
788
 
        tree.commit('1')
789
 
        to_keep = tree.branch.repository._pack_collection.names()
790
 
        # 2 to combine
791
 
        tree.commit('2')
792
 
        tree.commit('3')
793
 
        all = tree.branch.repository._pack_collection.names()
794
 
        combine = list(set(all) - set(to_keep))
795
 
        self.assertLength(3, all)
796
 
        self.assertLength(2, combine)
797
 
        tree.branch.repository.pack(hint=combine)
798
 
        final = tree.branch.repository._pack_collection.names()
799
 
        self.assertLength(2, final)
800
 
        self.assertFalse(combine[0] in final)
801
 
        self.assertFalse(combine[1] in final)
802
 
        self.assertSubset(to_keep, final)
803
 
 
804
 
    def test_stream_source_to_gc(self):
805
 
        source = self.make_repository('source', format='2a')
806
 
        target = self.make_repository('target', format='2a')
807
 
        stream = source._get_source(target._format)
808
 
        self.assertIsInstance(stream, groupcompress_repo.GroupCHKStreamSource)
809
 
 
810
 
    def test_stream_source_to_non_gc(self):
811
 
        source = self.make_repository('source', format='2a')
812
 
        target = self.make_repository('target', format='rich-root-pack')
813
 
        stream = source._get_source(target._format)
814
 
        # We don't want the child GroupCHKStreamSource
815
 
        self.assertIs(type(stream), repository.StreamSource)
816
 
 
817
 
    def test_get_stream_for_missing_keys_includes_all_chk_refs(self):
818
 
        source_builder = self.make_branch_builder('source',
819
 
                            format='2a')
820
 
        # We have to build a fairly large tree, so that we are sure the chk
821
 
        # pages will have split into multiple pages.
822
 
        entries = [('add', ('', 'a-root-id', 'directory', None))]
823
 
        for i in 'abcdefghijklmnopqrstuvwxyz123456789':
824
 
            for j in 'abcdefghijklmnopqrstuvwxyz123456789':
825
 
                fname = i + j
826
 
                fid = fname + '-id'
827
 
                content = 'content for %s\n' % (fname,)
828
 
                entries.append(('add', (fname, fid, 'file', content)))
829
 
        source_builder.start_series()
830
 
        source_builder.build_snapshot('rev-1', None, entries)
831
 
        # Now change a few of them, so we get a few new pages for the second
832
 
        # revision
833
 
        source_builder.build_snapshot('rev-2', ['rev-1'], [
834
 
            ('modify', ('aa-id', 'new content for aa-id\n')),
835
 
            ('modify', ('cc-id', 'new content for cc-id\n')),
836
 
            ('modify', ('zz-id', 'new content for zz-id\n')),
837
 
            ])
838
 
        source_builder.finish_series()
839
 
        source_branch = source_builder.get_branch()
840
 
        source_branch.lock_read()
841
 
        self.addCleanup(source_branch.unlock)
842
 
        target = self.make_repository('target', format='2a')
843
 
        source = source_branch.repository._get_source(target._format)
844
 
        self.assertIsInstance(source, groupcompress_repo.GroupCHKStreamSource)
845
 
 
846
 
        # On a regular pass, getting the inventories and chk pages for rev-2
847
 
        # would only get the newly created chk pages
848
 
        search = graph.SearchResult(set(['rev-2']), set(['rev-1']), 1,
849
 
                                    set(['rev-2']))
850
 
        simple_chk_records = []
851
 
        for vf_name, substream in source.get_stream(search):
852
 
            if vf_name == 'chk_bytes':
853
 
                for record in substream:
854
 
                    simple_chk_records.append(record.key)
855
 
            else:
856
 
                for _ in substream:
857
 
                    continue
858
 
        # 3 pages, the root (InternalNode), + 2 pages which actually changed
859
 
        self.assertEqual([('sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
860
 
                          ('sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
861
 
                          ('sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
862
 
                          ('sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)],
863
 
                         simple_chk_records)
864
 
        # Now, when we do a similar call using 'get_stream_for_missing_keys'
865
 
        # we should get a much larger set of pages.
866
 
        missing = [('inventories', 'rev-2')]
867
 
        full_chk_records = []
868
 
        for vf_name, substream in source.get_stream_for_missing_keys(missing):
869
 
            if vf_name == 'inventories':
870
 
                for record in substream:
871
 
                    self.assertEqual(('rev-2',), record.key)
872
 
            elif vf_name == 'chk_bytes':
873
 
                for record in substream:
874
 
                    full_chk_records.append(record.key)
875
 
            else:
876
 
                self.fail('Should not be getting a stream of %s' % (vf_name,))
877
 
        # We have 257 records now. This is because we have 1 root page, and 256
878
 
        # leaf pages in a complete listing.
879
 
        self.assertEqual(257, len(full_chk_records))
880
 
        self.assertSubset(simple_chk_records, full_chk_records)
881
 
 
882
 
    def test_inconsistency_fatal(self):
883
 
        repo = self.make_repository('repo', format='2a')
884
 
        self.assertTrue(repo.revisions._index._inconsistency_fatal)
885
 
        self.assertFalse(repo.texts._index._inconsistency_fatal)
886
 
        self.assertFalse(repo.inventories._index._inconsistency_fatal)
887
 
        self.assertFalse(repo.signatures._index._inconsistency_fatal)
888
 
        self.assertFalse(repo.chk_bytes._index._inconsistency_fatal)
889
 
 
890
 
 
891
 
class TestKnitPackStreamSource(tests.TestCaseWithMemoryTransport):
892
 
 
893
 
    def test_source_to_exact_pack_092(self):
894
 
        source = self.make_repository('source', format='pack-0.92')
895
 
        target = self.make_repository('target', format='pack-0.92')
896
 
        stream_source = source._get_source(target._format)
897
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
898
 
 
899
 
    def test_source_to_exact_pack_rich_root_pack(self):
900
 
        source = self.make_repository('source', format='rich-root-pack')
901
 
        target = self.make_repository('target', format='rich-root-pack')
902
 
        stream_source = source._get_source(target._format)
903
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
904
 
 
905
 
    def test_source_to_exact_pack_19(self):
906
 
        source = self.make_repository('source', format='1.9')
907
 
        target = self.make_repository('target', format='1.9')
908
 
        stream_source = source._get_source(target._format)
909
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
910
 
 
911
 
    def test_source_to_exact_pack_19_rich_root(self):
912
 
        source = self.make_repository('source', format='1.9-rich-root')
913
 
        target = self.make_repository('target', format='1.9-rich-root')
914
 
        stream_source = source._get_source(target._format)
915
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
916
 
 
917
 
    def test_source_to_remote_exact_pack_19(self):
918
 
        trans = self.make_smart_server('target')
919
 
        trans.ensure_base()
920
 
        source = self.make_repository('source', format='1.9')
921
 
        target = self.make_repository('target', format='1.9')
922
 
        target = repository.Repository.open(trans.base)
923
 
        stream_source = source._get_source(target._format)
924
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
925
 
 
926
 
    def test_stream_source_to_non_exact(self):
927
 
        source = self.make_repository('source', format='pack-0.92')
928
 
        target = self.make_repository('target', format='1.9')
929
 
        stream = source._get_source(target._format)
930
 
        self.assertIs(type(stream), repository.StreamSource)
931
 
 
932
 
    def test_stream_source_to_non_exact_rich_root(self):
933
 
        source = self.make_repository('source', format='1.9')
934
 
        target = self.make_repository('target', format='1.9-rich-root')
935
 
        stream = source._get_source(target._format)
936
 
        self.assertIs(type(stream), repository.StreamSource)
937
 
 
938
 
    def test_source_to_remote_non_exact_pack_19(self):
939
 
        trans = self.make_smart_server('target')
940
 
        trans.ensure_base()
941
 
        source = self.make_repository('source', format='1.9')
942
 
        target = self.make_repository('target', format='1.6')
943
 
        target = repository.Repository.open(trans.base)
944
 
        stream_source = source._get_source(target._format)
945
 
        self.assertIs(type(stream_source), repository.StreamSource)
946
 
 
947
 
    def test_stream_source_to_knit(self):
948
 
        source = self.make_repository('source', format='pack-0.92')
949
 
        target = self.make_repository('target', format='dirstate')
950
 
        stream = source._get_source(target._format)
951
 
        self.assertIs(type(stream), repository.StreamSource)
952
 
 
953
 
 
954
 
class TestDevelopment6FindParentIdsOfRevisions(TestCaseWithTransport):
955
 
    """Tests for _find_parent_ids_of_revisions."""
956
 
 
957
 
    def setUp(self):
958
 
        super(TestDevelopment6FindParentIdsOfRevisions, self).setUp()
959
 
        self.builder = self.make_branch_builder('source',
960
 
            format='development6-rich-root')
961
 
        self.builder.start_series()
962
 
        self.builder.build_snapshot('initial', None,
963
 
            [('add', ('', 'tree-root', 'directory', None))])
964
 
        self.repo = self.builder.get_branch().repository
965
 
        self.addCleanup(self.builder.finish_series)
966
 
 
967
 
    def assertParentIds(self, expected_result, rev_set):
968
 
        self.assertEqual(sorted(expected_result),
969
 
            sorted(self.repo._find_parent_ids_of_revisions(rev_set)))
970
 
 
971
 
    def test_simple(self):
972
 
        self.builder.build_snapshot('revid1', None, [])
973
 
        self.builder.build_snapshot('revid2', ['revid1'], [])
974
 
        rev_set = ['revid2']
975
 
        self.assertParentIds(['revid1'], rev_set)
976
 
 
977
 
    def test_not_first_parent(self):
978
 
        self.builder.build_snapshot('revid1', None, [])
979
 
        self.builder.build_snapshot('revid2', ['revid1'], [])
980
 
        self.builder.build_snapshot('revid3', ['revid2'], [])
981
 
        rev_set = ['revid3', 'revid2']
982
 
        self.assertParentIds(['revid1'], rev_set)
983
 
 
984
 
    def test_not_null(self):
985
 
        rev_set = ['initial']
986
 
        self.assertParentIds([], rev_set)
987
 
 
988
 
    def test_not_null_set(self):
989
 
        self.builder.build_snapshot('revid1', None, [])
990
 
        rev_set = [_mod_revision.NULL_REVISION]
991
 
        self.assertParentIds([], rev_set)
992
 
 
993
 
    def test_ghost(self):
994
 
        self.builder.build_snapshot('revid1', None, [])
995
 
        rev_set = ['ghost', 'revid1']
996
 
        self.assertParentIds(['initial'], rev_set)
997
 
 
998
 
    def test_ghost_parent(self):
999
 
        self.builder.build_snapshot('revid1', None, [])
1000
 
        self.builder.build_snapshot('revid2', ['revid1', 'ghost'], [])
1001
 
        rev_set = ['revid2', 'revid1']
1002
 
        self.assertParentIds(['ghost', 'initial'], rev_set)
1003
 
 
1004
 
    def test_righthand_parent(self):
1005
 
        self.builder.build_snapshot('revid1', None, [])
1006
 
        self.builder.build_snapshot('revid2a', ['revid1'], [])
1007
 
        self.builder.build_snapshot('revid2b', ['revid1'], [])
1008
 
        self.builder.build_snapshot('revid3', ['revid2a', 'revid2b'], [])
1009
 
        rev_set = ['revid3', 'revid2a']
1010
 
        self.assertParentIds(['revid1', 'revid2b'], rev_set)
1011
 
 
1012
 
 
1013
659
class TestWithBrokenRepo(TestCaseWithTransport):
1014
660
    """These tests seem to be more appropriate as interface tests?"""
1015
661
 
1028
674
            inv = inventory.Inventory(revision_id='rev1a')
1029
675
            inv.root.revision = 'rev1a'
1030
676
            self.add_file(repo, inv, 'file1', 'rev1a', [])
1031
 
            repo.texts.add_lines((inv.root.file_id, 'rev1a'), [], [])
1032
677
            repo.add_inventory('rev1a', inv, [])
1033
678
            revision = _mod_revision.Revision('rev1a',
1034
679
                committer='jrandom@example.com', timestamp=0,
1069
714
    def add_revision(self, repo, revision_id, inv, parent_ids):
1070
715
        inv.revision_id = revision_id
1071
716
        inv.root.revision = revision_id
1072
 
        repo.texts.add_lines((inv.root.file_id, revision_id), [], [])
1073
717
        repo.add_inventory(revision_id, inv, parent_ids)
1074
718
        revision = _mod_revision.Revision(revision_id,
1075
719
            committer='jrandom@example.com', timestamp=0, inventory_sha1='',
1092
736
        """
1093
737
        broken_repo = self.make_broken_repository()
1094
738
        empty_repo = self.make_repository('empty-repo')
1095
 
        try:
1096
 
            empty_repo.fetch(broken_repo)
1097
 
        except (errors.RevisionNotPresent, errors.BzrCheckError):
1098
 
            # Test successful: compression parent not being copied leads to
1099
 
            # error.
1100
 
            return
1101
 
        empty_repo.lock_read()
1102
 
        self.addCleanup(empty_repo.unlock)
1103
 
        text = empty_repo.texts.get_record_stream(
1104
 
            [('file2-id', 'rev3')], 'topological', True).next()
1105
 
        self.assertEqual('line\n', text.get_bytes_as('fulltext'))
 
739
        self.assertRaises(errors.RevisionNotPresent, empty_repo.fetch, broken_repo)
 
740
 
 
741
 
 
742
class TestKnitPackNoSubtrees(TestCaseWithTransport):
 
743
 
 
744
    def get_format(self):
 
745
        return bzrdir.format_registry.make_bzrdir('pack-0.92')
 
746
 
 
747
    def test_attribute__fetch_order(self):
 
748
        """Packs do not need ordered data retrieval."""
 
749
        format = self.get_format()
 
750
        repo = self.make_repository('.', format=format)
 
751
        self.assertEqual('unsorted', repo._fetch_order)
 
752
 
 
753
    def test_attribute__fetch_uses_deltas(self):
 
754
        """Packs reuse deltas."""
 
755
        format = self.get_format()
 
756
        repo = self.make_repository('.', format=format)
 
757
        self.assertEqual(True, repo._fetch_uses_deltas)
 
758
 
 
759
    def test_disk_layout(self):
 
760
        format = self.get_format()
 
761
        repo = self.make_repository('.', format=format)
 
762
        # in case of side effects of locking.
 
763
        repo.lock_write()
 
764
        repo.unlock()
 
765
        t = repo.bzrdir.get_repository_transport(None)
 
766
        self.check_format(t)
 
767
        # XXX: no locks left when unlocked at the moment
 
768
        # self.assertEqualDiff('', t.get('lock').read())
 
769
        self.check_databases(t)
 
770
 
 
771
    def check_format(self, t):
 
772
        self.assertEqualDiff(
 
773
            "Bazaar pack repository format 1 (needs bzr 0.92)\n",
 
774
                             t.get('format').read())
 
775
 
 
776
    def assertHasNoKndx(self, t, knit_name):
 
777
        """Assert that knit_name has no index on t."""
 
778
        self.assertFalse(t.has(knit_name + '.kndx'))
 
779
 
 
780
    def assertHasNoKnit(self, t, knit_name):
 
781
        """Assert that knit_name exists on t."""
 
782
        # no default content
 
783
        self.assertFalse(t.has(knit_name + '.knit'))
 
784
 
 
785
    def check_databases(self, t):
 
786
        """check knit content for a repository."""
 
787
        # check conversion worked
 
788
        self.assertHasNoKndx(t, 'inventory')
 
789
        self.assertHasNoKnit(t, 'inventory')
 
790
        self.assertHasNoKndx(t, 'revisions')
 
791
        self.assertHasNoKnit(t, 'revisions')
 
792
        self.assertHasNoKndx(t, 'signatures')
 
793
        self.assertHasNoKnit(t, 'signatures')
 
794
        self.assertFalse(t.has('knits'))
 
795
        # revision-indexes file-container directory
 
796
        self.assertEqual([],
 
797
            list(GraphIndex(t, 'pack-names', None).iter_all_entries()))
 
798
        self.assertTrue(S_ISDIR(t.stat('packs').st_mode))
 
799
        self.assertTrue(S_ISDIR(t.stat('upload').st_mode))
 
800
        self.assertTrue(S_ISDIR(t.stat('indices').st_mode))
 
801
        self.assertTrue(S_ISDIR(t.stat('obsolete_packs').st_mode))
 
802
 
 
803
    def test_shared_disk_layout(self):
 
804
        format = self.get_format()
 
805
        repo = self.make_repository('.', shared=True, format=format)
 
806
        # we want:
 
807
        t = repo.bzrdir.get_repository_transport(None)
 
808
        self.check_format(t)
 
809
        # XXX: no locks left when unlocked at the moment
 
810
        # self.assertEqualDiff('', t.get('lock').read())
 
811
        # We should have a 'shared-storage' marker file.
 
812
        self.assertEqualDiff('', t.get('shared-storage').read())
 
813
        self.check_databases(t)
 
814
 
 
815
    def test_shared_no_tree_disk_layout(self):
 
816
        format = self.get_format()
 
817
        repo = self.make_repository('.', shared=True, format=format)
 
818
        repo.set_make_working_trees(False)
 
819
        # we want:
 
820
        t = repo.bzrdir.get_repository_transport(None)
 
821
        self.check_format(t)
 
822
        # XXX: no locks left when unlocked at the moment
 
823
        # self.assertEqualDiff('', t.get('lock').read())
 
824
        # We should have a 'shared-storage' marker file.
 
825
        self.assertEqualDiff('', t.get('shared-storage').read())
 
826
        # We should have a marker for the no-working-trees flag.
 
827
        self.assertEqualDiff('', t.get('no-working-trees').read())
 
828
        # The marker should go when we toggle the setting.
 
829
        repo.set_make_working_trees(True)
 
830
        self.assertFalse(t.has('no-working-trees'))
 
831
        self.check_databases(t)
 
832
 
 
833
    def test_adding_revision_creates_pack_indices(self):
 
834
        format = self.get_format()
 
835
        tree = self.make_branch_and_tree('.', format=format)
 
836
        trans = tree.branch.repository.bzrdir.get_repository_transport(None)
 
837
        self.assertEqual([],
 
838
            list(GraphIndex(trans, 'pack-names', None).iter_all_entries()))
 
839
        tree.commit('foobarbaz')
 
840
        index = GraphIndex(trans, 'pack-names', None)
 
841
        index_nodes = list(index.iter_all_entries())
 
842
        self.assertEqual(1, len(index_nodes))
 
843
        node = index_nodes[0]
 
844
        name = node[1][0]
 
845
        # the pack sizes should be listed in the index
 
846
        pack_value = node[2]
 
847
        sizes = [int(digits) for digits in pack_value.split(' ')]
 
848
        for size, suffix in zip(sizes, ['.rix', '.iix', '.tix', '.six']):
 
849
            stat = trans.stat('indices/%s%s' % (name, suffix))
 
850
            self.assertEqual(size, stat.st_size)
 
851
 
 
852
    def test_pulling_nothing_leads_to_no_new_names(self):
 
853
        format = self.get_format()
 
854
        tree1 = self.make_branch_and_tree('1', format=format)
 
855
        tree2 = self.make_branch_and_tree('2', format=format)
 
856
        tree1.branch.repository.fetch(tree2.branch.repository)
 
857
        trans = tree1.branch.repository.bzrdir.get_repository_transport(None)
 
858
        self.assertEqual([],
 
859
            list(GraphIndex(trans, 'pack-names', None).iter_all_entries()))
 
860
 
 
861
    def test_commit_across_pack_shape_boundary_autopacks(self):
 
862
        format = self.get_format()
 
863
        tree = self.make_branch_and_tree('.', format=format)
 
864
        trans = tree.branch.repository.bzrdir.get_repository_transport(None)
 
865
        # This test could be a little cheaper by replacing the packs
 
866
        # attribute on the repository to allow a different pack distribution
 
867
        # and max packs policy - so we are checking the policy is honoured
 
868
        # in the test. But for now 11 commits is not a big deal in a single
 
869
        # test.
 
870
        for x in range(9):
 
871
            tree.commit('commit %s' % x)
 
872
        # there should be 9 packs:
 
873
        index = GraphIndex(trans, 'pack-names', None)
 
874
        self.assertEqual(9, len(list(index.iter_all_entries())))
 
875
        # insert some files in obsolete_packs which should be removed by pack.
 
876
        trans.put_bytes('obsolete_packs/foo', '123')
 
877
        trans.put_bytes('obsolete_packs/bar', '321')
 
878
        # committing one more should coalesce to 1 of 10.
 
879
        tree.commit('commit triggering pack')
 
880
        index = GraphIndex(trans, 'pack-names', None)
 
881
        self.assertEqual(1, len(list(index.iter_all_entries())))
 
882
        # packing should not damage data
 
883
        tree = tree.bzrdir.open_workingtree()
 
884
        check_result = tree.branch.repository.check(
 
885
            [tree.branch.last_revision()])
 
886
        # We should have 50 (10x5) files in the obsolete_packs directory.
 
887
        obsolete_files = list(trans.list_dir('obsolete_packs'))
 
888
        self.assertFalse('foo' in obsolete_files)
 
889
        self.assertFalse('bar' in obsolete_files)
 
890
        self.assertEqual(50, len(obsolete_files))
 
891
        # XXX: Todo check packs obsoleted correctly - old packs and indices
 
892
        # in the obsolete_packs directory.
 
893
        large_pack_name = list(index.iter_all_entries())[0][1][0]
 
894
        # finally, committing again should not touch the large pack.
 
895
        tree.commit('commit not triggering pack')
 
896
        index = GraphIndex(trans, 'pack-names', None)
 
897
        self.assertEqual(2, len(list(index.iter_all_entries())))
 
898
        pack_names = [node[1][0] for node in index.iter_all_entries()]
 
899
        self.assertTrue(large_pack_name in pack_names)
 
900
 
 
901
    def test_fail_obsolete_deletion(self):
 
902
        # failing to delete obsolete packs is not fatal
 
903
        format = self.get_format()
 
904
        server = fakenfs.FakeNFSServer()
 
905
        server.setUp()
 
906
        self.addCleanup(server.tearDown)
 
907
        transport = get_transport(server.get_url())
 
908
        bzrdir = self.get_format().initialize_on_transport(transport)
 
909
        repo = bzrdir.create_repository()
 
910
        repo_transport = bzrdir.get_repository_transport(None)
 
911
        self.assertTrue(repo_transport.has('obsolete_packs'))
 
912
        # these files are in use by another client and typically can't be deleted
 
913
        repo_transport.put_bytes('obsolete_packs/.nfsblahblah', 'contents')
 
914
        repo._pack_collection._clear_obsolete_packs()
 
915
        self.assertTrue(repo_transport.has('obsolete_packs/.nfsblahblah'))
 
916
 
 
917
    def test_pack_after_two_commits_packs_everything(self):
 
918
        format = self.get_format()
 
919
        tree = self.make_branch_and_tree('.', format=format)
 
920
        trans = tree.branch.repository.bzrdir.get_repository_transport(None)
 
921
        tree.commit('start')
 
922
        tree.commit('more work')
 
923
        tree.branch.repository.pack()
 
924
        # there should be 1 pack:
 
925
        index = GraphIndex(trans, 'pack-names', None)
 
926
        self.assertEqual(1, len(list(index.iter_all_entries())))
 
927
        self.assertEqual(2, len(tree.branch.repository.all_revision_ids()))
 
928
 
 
929
    def test_pack_layout(self):
 
930
        format = self.get_format()
 
931
        tree = self.make_branch_and_tree('.', format=format)
 
932
        trans = tree.branch.repository.bzrdir.get_repository_transport(None)
 
933
        tree.commit('start', rev_id='1')
 
934
        tree.commit('more work', rev_id='2')
 
935
        tree.branch.repository.pack()
 
936
        tree.lock_read()
 
937
        self.addCleanup(tree.unlock)
 
938
        pack = tree.branch.repository._pack_collection.get_pack_by_name(
 
939
            tree.branch.repository._pack_collection.names()[0])
 
940
        # revision access tends to be tip->ancestor, so ordering that way on 
 
941
        # disk is a good idea.
 
942
        for _1, key, val, refs in pack.revision_index.iter_all_entries():
 
943
            if key == ('1',):
 
944
                pos_1 = int(val[1:].split()[0])
 
945
            else:
 
946
                pos_2 = int(val[1:].split()[0])
 
947
        self.assertTrue(pos_2 < pos_1)
 
948
 
 
949
    def test_pack_repositories_support_multiple_write_locks(self):
 
950
        format = self.get_format()
 
951
        self.make_repository('.', shared=True, format=format)
 
952
        r1 = repository.Repository.open('.')
 
953
        r2 = repository.Repository.open('.')
 
954
        r1.lock_write()
 
955
        self.addCleanup(r1.unlock)
 
956
        r2.lock_write()
 
957
        r2.unlock()
 
958
 
 
959
    def _add_text(self, repo, fileid):
 
960
        """Add a text to the repository within a write group."""
 
961
        repo.texts.add_lines((fileid, 'samplerev+'+fileid), [], [])
 
962
 
 
963
    def test_concurrent_writers_merge_new_packs(self):
 
964
        format = self.get_format()
 
965
        self.make_repository('.', shared=True, format=format)
 
966
        r1 = repository.Repository.open('.')
 
967
        r2 = repository.Repository.open('.')
 
968
        r1.lock_write()
 
969
        try:
 
970
            # access enough data to load the names list
 
971
            list(r1.all_revision_ids())
 
972
            r2.lock_write()
 
973
            try:
 
974
                # access enough data to load the names list
 
975
                list(r2.all_revision_ids())
 
976
                r1.start_write_group()
 
977
                try:
 
978
                    r2.start_write_group()
 
979
                    try:
 
980
                        self._add_text(r1, 'fileidr1')
 
981
                        self._add_text(r2, 'fileidr2')
 
982
                    except:
 
983
                        r2.abort_write_group()
 
984
                        raise
 
985
                except:
 
986
                    r1.abort_write_group()
 
987
                    raise
 
988
                # both r1 and r2 have open write groups with data in them
 
989
                # created while the other's write group was open.
 
990
                # Commit both which requires a merge to the pack-names.
 
991
                try:
 
992
                    r1.commit_write_group()
 
993
                except:
 
994
                    r1.abort_write_group()
 
995
                    r2.abort_write_group()
 
996
                    raise
 
997
                r2.commit_write_group()
 
998
                # tell r1 to reload from disk
 
999
                r1._pack_collection.reset()
 
1000
                # Now both repositories should know about both names
 
1001
                r1._pack_collection.ensure_loaded()
 
1002
                r2._pack_collection.ensure_loaded()
 
1003
                self.assertEqual(r1._pack_collection.names(), r2._pack_collection.names())
 
1004
                self.assertEqual(2, len(r1._pack_collection.names()))
 
1005
            finally:
 
1006
                r2.unlock()
 
1007
        finally:
 
1008
            r1.unlock()
 
1009
 
 
1010
    def test_concurrent_writer_second_preserves_dropping_a_pack(self):
 
1011
        format = self.get_format()
 
1012
        self.make_repository('.', shared=True, format=format)
 
1013
        r1 = repository.Repository.open('.')
 
1014
        r2 = repository.Repository.open('.')
 
1015
        # add a pack to drop
 
1016
        r1.lock_write()
 
1017
        try:
 
1018
            r1.start_write_group()
 
1019
            try:
 
1020
                self._add_text(r1, 'fileidr1')
 
1021
            except:
 
1022
                r1.abort_write_group()
 
1023
                raise
 
1024
            else:
 
1025
                r1.commit_write_group()
 
1026
            r1._pack_collection.ensure_loaded()
 
1027
            name_to_drop = r1._pack_collection.all_packs()[0].name
 
1028
        finally:
 
1029
            r1.unlock()
 
1030
        r1.lock_write()
 
1031
        try:
 
1032
            # access enough data to load the names list
 
1033
            list(r1.all_revision_ids())
 
1034
            r2.lock_write()
 
1035
            try:
 
1036
                # access enough data to load the names list
 
1037
                list(r2.all_revision_ids())
 
1038
                r1._pack_collection.ensure_loaded()
 
1039
                try:
 
1040
                    r2.start_write_group()
 
1041
                    try:
 
1042
                        # in r1, drop the pack
 
1043
                        r1._pack_collection._remove_pack_from_memory(
 
1044
                            r1._pack_collection.get_pack_by_name(name_to_drop))
 
1045
                        # in r2, add a pack
 
1046
                        self._add_text(r2, 'fileidr2')
 
1047
                    except:
 
1048
                        r2.abort_write_group()
 
1049
                        raise
 
1050
                except:
 
1051
                    r1._pack_collection.reset()
 
1052
                    raise
 
1053
                # r1 has a changed names list, and r2 an open write groups with
 
1054
                # changes.
 
1055
                # save r1, and then commit the r2 write group, which requires a
 
1056
                # merge to the pack-names, which should not reinstate
 
1057
                # name_to_drop
 
1058
                try:
 
1059
                    r1._pack_collection._save_pack_names()
 
1060
                    r1._pack_collection.reset()
 
1061
                except:
 
1062
                    r2.abort_write_group()
 
1063
                    raise
 
1064
                try:
 
1065
                    r2.commit_write_group()
 
1066
                except:
 
1067
                    r2.abort_write_group()
 
1068
                    raise
 
1069
                # Now both repositories should now about just one name.
 
1070
                r1._pack_collection.ensure_loaded()
 
1071
                r2._pack_collection.ensure_loaded()
 
1072
                self.assertEqual(r1._pack_collection.names(), r2._pack_collection.names())
 
1073
                self.assertEqual(1, len(r1._pack_collection.names()))
 
1074
                self.assertFalse(name_to_drop in r1._pack_collection.names())
 
1075
            finally:
 
1076
                r2.unlock()
 
1077
        finally:
 
1078
            r1.unlock()
 
1079
 
 
1080
    def test_lock_write_does_not_physically_lock(self):
 
1081
        repo = self.make_repository('.', format=self.get_format())
 
1082
        repo.lock_write()
 
1083
        self.addCleanup(repo.unlock)
 
1084
        self.assertFalse(repo.get_physical_lock_status())
 
1085
 
 
1086
    def prepare_for_break_lock(self):
 
1087
        # Setup the global ui factory state so that a break-lock method call
 
1088
        # will find usable input in the input stream.
 
1089
        old_factory = bzrlib.ui.ui_factory
 
1090
        def restoreFactory():
 
1091
            bzrlib.ui.ui_factory = old_factory
 
1092
        self.addCleanup(restoreFactory)
 
1093
        bzrlib.ui.ui_factory = bzrlib.ui.SilentUIFactory()
 
1094
        bzrlib.ui.ui_factory.stdin = StringIO("y\n")
 
1095
 
 
1096
    def test_break_lock_breaks_physical_lock(self):
 
1097
        repo = self.make_repository('.', format=self.get_format())
 
1098
        repo._pack_collection.lock_names()
 
1099
        repo2 = repository.Repository.open('.')
 
1100
        self.assertTrue(repo.get_physical_lock_status())
 
1101
        self.prepare_for_break_lock()
 
1102
        repo2.break_lock()
 
1103
        self.assertFalse(repo.get_physical_lock_status())
 
1104
 
 
1105
    def test_broken_physical_locks_error_on__unlock_names_lock(self):
 
1106
        repo = self.make_repository('.', format=self.get_format())
 
1107
        repo._pack_collection.lock_names()
 
1108
        self.assertTrue(repo.get_physical_lock_status())
 
1109
        repo2 = repository.Repository.open('.')
 
1110
        self.prepare_for_break_lock()
 
1111
        repo2.break_lock()
 
1112
        self.assertRaises(errors.LockBroken, repo._pack_collection._unlock_names)
 
1113
 
 
1114
    def test_fetch_without_find_ghosts_ignores_ghosts(self):
 
1115
        # we want two repositories at this point:
 
1116
        # one with a revision that is a ghost in the other
 
1117
        # repository.
 
1118
        # 'ghost' is present in has_ghost, 'ghost' is absent in 'missing_ghost'.
 
1119
        # 'references' is present in both repositories, and 'tip' is present
 
1120
        # just in has_ghost.
 
1121
        # has_ghost       missing_ghost
 
1122
        #------------------------------
 
1123
        # 'ghost'             -
 
1124
        # 'references'    'references'
 
1125
        # 'tip'               -
 
1126
        # In this test we fetch 'tip' which should not fetch 'ghost'
 
1127
        has_ghost = self.make_repository('has_ghost', format=self.get_format())
 
1128
        missing_ghost = self.make_repository('missing_ghost',
 
1129
            format=self.get_format())
 
1130
 
 
1131
        def add_commit(repo, revision_id, parent_ids):
 
1132
            repo.lock_write()
 
1133
            repo.start_write_group()
 
1134
            inv = inventory.Inventory(revision_id=revision_id)
 
1135
            inv.root.revision = revision_id
 
1136
            root_id = inv.root.file_id
 
1137
            sha1 = repo.add_inventory(revision_id, inv, [])
 
1138
            repo.texts.add_lines((root_id, revision_id), [], [])
 
1139
            rev = bzrlib.revision.Revision(timestamp=0,
 
1140
                                           timezone=None,
 
1141
                                           committer="Foo Bar <foo@example.com>",
 
1142
                                           message="Message",
 
1143
                                           inventory_sha1=sha1,
 
1144
                                           revision_id=revision_id)
 
1145
            rev.parent_ids = parent_ids
 
1146
            repo.add_revision(revision_id, rev)
 
1147
            repo.commit_write_group()
 
1148
            repo.unlock()
 
1149
        add_commit(has_ghost, 'ghost', [])
 
1150
        add_commit(has_ghost, 'references', ['ghost'])
 
1151
        add_commit(missing_ghost, 'references', ['ghost'])
 
1152
        add_commit(has_ghost, 'tip', ['references'])
 
1153
        missing_ghost.fetch(has_ghost, 'tip')
 
1154
        # missing ghost now has tip and not ghost.
 
1155
        rev = missing_ghost.get_revision('tip')
 
1156
        inv = missing_ghost.get_inventory('tip')
 
1157
        self.assertRaises(errors.NoSuchRevision,
 
1158
            missing_ghost.get_revision, 'ghost')
 
1159
        self.assertRaises(errors.NoSuchRevision,
 
1160
            missing_ghost.get_inventory, 'ghost')
 
1161
 
 
1162
    def test_supports_external_lookups(self):
 
1163
        repo = self.make_repository('.', format=self.get_format())
 
1164
        self.assertFalse(repo._format.supports_external_lookups)
 
1165
 
 
1166
 
 
1167
class TestKnitPackSubtrees(TestKnitPackNoSubtrees):
 
1168
 
 
1169
    def get_format(self):
 
1170
        return bzrdir.format_registry.make_bzrdir(
 
1171
            'pack-0.92-subtree')
 
1172
 
 
1173
    def check_format(self, t):
 
1174
        self.assertEqualDiff(
 
1175
            "Bazaar pack repository format 1 with subtree support (needs bzr 0.92)\n",
 
1176
            t.get('format').read())
 
1177
 
 
1178
 
 
1179
class TestDevelopment0(TestKnitPackNoSubtrees):
 
1180
 
 
1181
    def get_format(self):
 
1182
        return bzrdir.format_registry.make_bzrdir(
 
1183
            'development0')
 
1184
 
 
1185
    def check_format(self, t):
 
1186
        self.assertEqualDiff(
 
1187
            "Bazaar development format 0 (needs bzr.dev from before 1.3)\n",
 
1188
            t.get('format').read())
 
1189
 
 
1190
 
 
1191
class TestDevelopment0Subtree(TestKnitPackNoSubtrees):
 
1192
 
 
1193
    def get_format(self):
 
1194
        return bzrdir.format_registry.make_bzrdir(
 
1195
            'development0-subtree')
 
1196
 
 
1197
    def check_format(self, t):
 
1198
        self.assertEqualDiff(
 
1199
            "Bazaar development format 0 with subtree support "
 
1200
            "(needs bzr.dev from before 1.3)\n",
 
1201
            t.get('format').read())
 
1202
 
 
1203
 
 
1204
class TestExternalDevelopment1(object):
 
1205
 
 
1206
    # mixin class for testing stack-supporting development formats
 
1207
 
 
1208
    def test_stack_checks_compatibility(self):
 
1209
        # early versions of the packing code relied on pack internals to
 
1210
        # stack, but the current version should be able to stack on any
 
1211
        # format.
 
1212
        #
 
1213
        # TODO: Possibly this should be run per-repository-format and raise
 
1214
        # TestNotApplicable on formats that don't support stacking. -- mbp
 
1215
        # 20080729
 
1216
        repo = self.make_repository('repo', format=self.get_format())
 
1217
        if repo.supports_rich_root():
 
1218
            # can only stack on repositories that have compatible internal
 
1219
            # metadata
 
1220
            matching_format_name = 'pack-0.92-subtree'
 
1221
            mismatching_format_name = 'pack-0.92'
 
1222
        else:
 
1223
            matching_format_name = 'pack-0.92'
 
1224
            mismatching_format_name = 'pack-0.92-subtree'
 
1225
        base = self.make_repository('base', format=matching_format_name)
 
1226
        repo.add_fallback_repository(base)
 
1227
        # you can't stack on something with incompatible data
 
1228
        bad_repo = self.make_repository('mismatch',
 
1229
            format=mismatching_format_name)
 
1230
        e = self.assertRaises(errors.IncompatibleRepositories,
 
1231
            repo.add_fallback_repository, bad_repo)
 
1232
        self.assertContainsRe(str(e),
 
1233
            r'(?m)KnitPackRepository.*/mismatch/.*\nis not compatible with\n'
 
1234
            r'KnitPackRepository.*/repo/.*\n'
 
1235
            r'different rich-root support')
 
1236
 
 
1237
    def test_adding_pack_does_not_record_pack_names_from_other_repositories(self):
 
1238
        base = self.make_branch_and_tree('base', format=self.get_format())
 
1239
        base.commit('foo')
 
1240
        referencing = self.make_branch_and_tree('repo', format=self.get_format())
 
1241
        referencing.branch.repository.add_fallback_repository(base.branch.repository)
 
1242
        referencing.commit('bar')
 
1243
        new_instance = referencing.bzrdir.open_repository()
 
1244
        new_instance.lock_read()
 
1245
        self.addCleanup(new_instance.unlock)
 
1246
        new_instance._pack_collection.ensure_loaded()
 
1247
        self.assertEqual(1, len(new_instance._pack_collection.all_packs()))
 
1248
 
 
1249
    def test_autopack_only_considers_main_repo_packs(self):
 
1250
        base = self.make_branch_and_tree('base', format=self.get_format())
 
1251
        base.commit('foo')
 
1252
        tree = self.make_branch_and_tree('repo', format=self.get_format())
 
1253
        tree.branch.repository.add_fallback_repository(base.branch.repository)
 
1254
        trans = tree.branch.repository.bzrdir.get_repository_transport(None)
 
1255
        # This test could be a little cheaper by replacing the packs
 
1256
        # attribute on the repository to allow a different pack distribution
 
1257
        # and max packs policy - so we are checking the policy is honoured
 
1258
        # in the test. But for now 11 commits is not a big deal in a single
 
1259
        # test.
 
1260
        for x in range(9):
 
1261
            tree.commit('commit %s' % x)
 
1262
        # there should be 9 packs:
 
1263
        index = GraphIndex(trans, 'pack-names', None)
 
1264
        self.assertEqual(9, len(list(index.iter_all_entries())))
 
1265
        # committing one more should coalesce to 1 of 10.
 
1266
        tree.commit('commit triggering pack')
 
1267
        index = GraphIndex(trans, 'pack-names', None)
 
1268
        self.assertEqual(1, len(list(index.iter_all_entries())))
 
1269
        # packing should not damage data
 
1270
        tree = tree.bzrdir.open_workingtree()
 
1271
        check_result = tree.branch.repository.check(
 
1272
            [tree.branch.last_revision()])
 
1273
        # We should have 50 (10x5) files in the obsolete_packs directory.
 
1274
        obsolete_files = list(trans.list_dir('obsolete_packs'))
 
1275
        self.assertFalse('foo' in obsolete_files)
 
1276
        self.assertFalse('bar' in obsolete_files)
 
1277
        self.assertEqual(50, len(obsolete_files))
 
1278
        # XXX: Todo check packs obsoleted correctly - old packs and indices
 
1279
        # in the obsolete_packs directory.
 
1280
        large_pack_name = list(index.iter_all_entries())[0][1][0]
 
1281
        # finally, committing again should not touch the large pack.
 
1282
        tree.commit('commit not triggering pack')
 
1283
        index = GraphIndex(trans, 'pack-names', None)
 
1284
        self.assertEqual(2, len(list(index.iter_all_entries())))
 
1285
        pack_names = [node[1][0] for node in index.iter_all_entries()]
 
1286
        self.assertTrue(large_pack_name in pack_names)
 
1287
 
 
1288
 
 
1289
class TestDevelopment1(TestKnitPackNoSubtrees, TestExternalDevelopment1):
 
1290
 
 
1291
    def get_format(self):
 
1292
        return bzrdir.format_registry.make_bzrdir(
 
1293
            'development')
 
1294
 
 
1295
    def check_format(self, t):
 
1296
        self.assertEqualDiff(
 
1297
            "Bazaar development format 1 (needs bzr.dev from before 1.6)\n",
 
1298
            t.get('format').read())
 
1299
 
 
1300
    def test_supports_external_lookups(self):
 
1301
        repo = self.make_repository('.', format=self.get_format())
 
1302
        self.assertTrue(repo._format.supports_external_lookups)
 
1303
 
 
1304
 
 
1305
class TestDevelopment1Subtree(TestKnitPackNoSubtrees, TestExternalDevelopment1):
 
1306
 
 
1307
    def get_format(self):
 
1308
        return bzrdir.format_registry.make_bzrdir(
 
1309
            'development-subtree')
 
1310
 
 
1311
    def check_format(self, t):
 
1312
        self.assertEqualDiff(
 
1313
            "Bazaar development format 1 with subtree support "
 
1314
            "(needs bzr.dev from before 1.6)\n",
 
1315
            t.get('format').read())
 
1316
 
 
1317
    def test_supports_external_lookups(self):
 
1318
        repo = self.make_repository('.', format=self.get_format())
 
1319
        self.assertTrue(repo._format.supports_external_lookups)
1106
1320
 
1107
1321
 
1108
1322
class TestRepositoryPackCollection(TestCaseWithTransport):
1110
1324
    def get_format(self):
1111
1325
        return bzrdir.format_registry.make_bzrdir('pack-0.92')
1112
1326
 
1113
 
    def get_packs(self):
1114
 
        format = self.get_format()
1115
 
        repo = self.make_repository('.', format=format)
1116
 
        return repo._pack_collection
1117
 
 
1118
 
    def make_packs_and_alt_repo(self, write_lock=False):
1119
 
        """Create a pack repo with 3 packs, and access it via a second repo."""
1120
 
        tree = self.make_branch_and_tree('.', format=self.get_format())
1121
 
        tree.lock_write()
1122
 
        self.addCleanup(tree.unlock)
1123
 
        rev1 = tree.commit('one')
1124
 
        rev2 = tree.commit('two')
1125
 
        rev3 = tree.commit('three')
1126
 
        r = repository.Repository.open('.')
1127
 
        if write_lock:
1128
 
            r.lock_write()
1129
 
        else:
1130
 
            r.lock_read()
1131
 
        self.addCleanup(r.unlock)
1132
 
        packs = r._pack_collection
1133
 
        packs.ensure_loaded()
1134
 
        return tree, r, packs, [rev1, rev2, rev3]
1135
 
 
1136
 
    def test__clear_obsolete_packs(self):
1137
 
        packs = self.get_packs()
1138
 
        obsolete_pack_trans = packs.transport.clone('obsolete_packs')
1139
 
        obsolete_pack_trans.put_bytes('a-pack.pack', 'content\n')
1140
 
        obsolete_pack_trans.put_bytes('a-pack.rix', 'content\n')
1141
 
        obsolete_pack_trans.put_bytes('a-pack.iix', 'content\n')
1142
 
        obsolete_pack_trans.put_bytes('another-pack.pack', 'foo\n')
1143
 
        obsolete_pack_trans.put_bytes('not-a-pack.rix', 'foo\n')
1144
 
        res = packs._clear_obsolete_packs()
1145
 
        self.assertEqual(['a-pack', 'another-pack'], sorted(res))
1146
 
        self.assertEqual([], obsolete_pack_trans.list_dir('.'))
1147
 
 
1148
 
    def test__clear_obsolete_packs_preserve(self):
1149
 
        packs = self.get_packs()
1150
 
        obsolete_pack_trans = packs.transport.clone('obsolete_packs')
1151
 
        obsolete_pack_trans.put_bytes('a-pack.pack', 'content\n')
1152
 
        obsolete_pack_trans.put_bytes('a-pack.rix', 'content\n')
1153
 
        obsolete_pack_trans.put_bytes('a-pack.iix', 'content\n')
1154
 
        obsolete_pack_trans.put_bytes('another-pack.pack', 'foo\n')
1155
 
        obsolete_pack_trans.put_bytes('not-a-pack.rix', 'foo\n')
1156
 
        res = packs._clear_obsolete_packs(preserve=set(['a-pack']))
1157
 
        self.assertEqual(['a-pack', 'another-pack'], sorted(res))
1158
 
        self.assertEqual(['a-pack.iix', 'a-pack.pack', 'a-pack.rix'],
1159
 
                         sorted(obsolete_pack_trans.list_dir('.')))
1160
 
 
1161
1327
    def test__max_pack_count(self):
1162
1328
        """The maximum pack count is a function of the number of revisions."""
 
1329
        format = self.get_format()
 
1330
        repo = self.make_repository('.', format=format)
 
1331
        packs = repo._pack_collection
1163
1332
        # no revisions - one pack, so that we can have a revision free repo
1164
1333
        # without it blowing up
1165
 
        packs = self.get_packs()
1166
1334
        self.assertEqual(1, packs._max_pack_count(0))
1167
1335
        # after that the sum of the digits, - check the first 1-9
1168
1336
        self.assertEqual(1, packs._max_pack_count(1))
1183
1351
        # check some arbitrary big numbers
1184
1352
        self.assertEqual(25, packs._max_pack_count(112894))
1185
1353
 
1186
 
    def test_repr(self):
1187
 
        packs = self.get_packs()
1188
 
        self.assertContainsRe(repr(packs),
1189
 
            'RepositoryPackCollection(.*Repository(.*))')
1190
 
 
1191
 
    def test__obsolete_packs(self):
1192
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1193
 
        names = packs.names()
1194
 
        pack = packs.get_pack_by_name(names[0])
1195
 
        # Schedule this one for removal
1196
 
        packs._remove_pack_from_memory(pack)
1197
 
        # Simulate a concurrent update by renaming the .pack file and one of
1198
 
        # the indices
1199
 
        packs.transport.rename('packs/%s.pack' % (names[0],),
1200
 
                               'obsolete_packs/%s.pack' % (names[0],))
1201
 
        packs.transport.rename('indices/%s.iix' % (names[0],),
1202
 
                               'obsolete_packs/%s.iix' % (names[0],))
1203
 
        # Now trigger the obsoletion, and ensure that all the remaining files
1204
 
        # are still renamed
1205
 
        packs._obsolete_packs([pack])
1206
 
        self.assertEqual([n + '.pack' for n in names[1:]],
1207
 
                         sorted(packs._pack_transport.list_dir('.')))
1208
 
        # names[0] should not be present in the index anymore
1209
 
        self.assertEqual(names[1:],
1210
 
            sorted(set([osutils.splitext(n)[0] for n in
1211
 
                        packs._index_transport.list_dir('.')])))
1212
 
 
1213
1354
    def test_pack_distribution_zero(self):
1214
 
        packs = self.get_packs()
 
1355
        format = self.get_format()
 
1356
        repo = self.make_repository('.', format=format)
 
1357
        packs = repo._pack_collection
1215
1358
        self.assertEqual([0], packs.pack_distribution(0))
1216
1359
 
1217
1360
    def test_ensure_loaded_unlocked(self):
1218
 
        packs = self.get_packs()
 
1361
        format = self.get_format()
 
1362
        repo = self.make_repository('.', format=format)
1219
1363
        self.assertRaises(errors.ObjectNotLocked,
1220
 
                          packs.ensure_loaded)
 
1364
                          repo._pack_collection.ensure_loaded)
1221
1365
 
1222
1366
    def test_pack_distribution_one_to_nine(self):
1223
 
        packs = self.get_packs()
 
1367
        format = self.get_format()
 
1368
        repo = self.make_repository('.', format=format)
 
1369
        packs = repo._pack_collection
1224
1370
        self.assertEqual([1],
1225
1371
            packs.pack_distribution(1))
1226
1372
        self.assertEqual([1, 1],
1242
1388
 
1243
1389
    def test_pack_distribution_stable_at_boundaries(self):
1244
1390
        """When there are multi-rev packs the counts are stable."""
1245
 
        packs = self.get_packs()
 
1391
        format = self.get_format()
 
1392
        repo = self.make_repository('.', format=format)
 
1393
        packs = repo._pack_collection
1246
1394
        # in 10s:
1247
1395
        self.assertEqual([10], packs.pack_distribution(10))
1248
1396
        self.assertEqual([10, 1], packs.pack_distribution(11))
1257
1405
        self.assertEqual([100, 100, 10, 1], packs.pack_distribution(211))
1258
1406
 
1259
1407
    def test_plan_pack_operations_2009_revisions_skip_all_packs(self):
1260
 
        packs = self.get_packs()
 
1408
        format = self.get_format()
 
1409
        repo = self.make_repository('.', format=format)
 
1410
        packs = repo._pack_collection
1261
1411
        existing_packs = [(2000, "big"), (9, "medium")]
1262
1412
        # rev count - 2009 -> 2x1000 + 9x1
1263
1413
        pack_operations = packs.plan_autopack_combinations(
1265
1415
        self.assertEqual([], pack_operations)
1266
1416
 
1267
1417
    def test_plan_pack_operations_2010_revisions_skip_all_packs(self):
1268
 
        packs = self.get_packs()
 
1418
        format = self.get_format()
 
1419
        repo = self.make_repository('.', format=format)
 
1420
        packs = repo._pack_collection
1269
1421
        existing_packs = [(2000, "big"), (9, "medium"), (1, "single")]
1270
1422
        # rev count - 2010 -> 2x1000 + 1x10
1271
1423
        pack_operations = packs.plan_autopack_combinations(
1273
1425
        self.assertEqual([], pack_operations)
1274
1426
 
1275
1427
    def test_plan_pack_operations_2010_combines_smallest_two(self):
1276
 
        packs = self.get_packs()
 
1428
        format = self.get_format()
 
1429
        repo = self.make_repository('.', format=format)
 
1430
        packs = repo._pack_collection
1277
1431
        existing_packs = [(1999, "big"), (9, "medium"), (1, "single2"),
1278
1432
            (1, "single1")]
1279
1433
        # rev count - 2010 -> 2x1000 + 1x10 (3)
1280
1434
        pack_operations = packs.plan_autopack_combinations(
1281
1435
            existing_packs, [1000, 1000, 10])
1282
 
        self.assertEqual([[2, ["single2", "single1"]]], pack_operations)
1283
 
 
1284
 
    def test_plan_pack_operations_creates_a_single_op(self):
1285
 
        packs = self.get_packs()
1286
 
        existing_packs = [(50, 'a'), (40, 'b'), (30, 'c'), (10, 'd'),
1287
 
                          (10, 'e'), (6, 'f'), (4, 'g')]
1288
 
        # rev count 150 -> 1x100 and 5x10
1289
 
        # The two size 10 packs do not need to be touched. The 50, 40, 30 would
1290
 
        # be combined into a single 120 size pack, and the 6 & 4 would
1291
 
        # becombined into a size 10 pack. However, if we have to rewrite them,
1292
 
        # we save a pack file with no increased I/O by putting them into the
1293
 
        # same file.
1294
 
        distribution = packs.pack_distribution(150)
1295
 
        pack_operations = packs.plan_autopack_combinations(existing_packs,
1296
 
                                                           distribution)
1297
 
        self.assertEqual([[130, ['a', 'b', 'c', 'f', 'g']]], pack_operations)
 
1436
        self.assertEqual([[2, ["single2", "single1"]], [0, []]], pack_operations)
1298
1437
 
1299
1438
    def test_all_packs_none(self):
1300
1439
        format = self.get_format()
1338
1477
        tree.lock_read()
1339
1478
        self.addCleanup(tree.unlock)
1340
1479
        packs = tree.branch.repository._pack_collection
1341
 
        packs.reset()
1342
1480
        packs.ensure_loaded()
1343
1481
        name = packs.names()[0]
1344
1482
        pack_1 = packs.get_pack_by_name(name)
1353
1491
        # and the same instance should be returned on successive calls.
1354
1492
        self.assertTrue(pack_1 is packs.get_pack_by_name(name))
1355
1493
 
1356
 
    def test_reload_pack_names_new_entry(self):
1357
 
        tree, r, packs, revs = self.make_packs_and_alt_repo()
1358
 
        names = packs.names()
1359
 
        # Add a new pack file into the repository
1360
 
        rev4 = tree.commit('four')
1361
 
        new_names = tree.branch.repository._pack_collection.names()
1362
 
        new_name = set(new_names).difference(names)
1363
 
        self.assertEqual(1, len(new_name))
1364
 
        new_name = new_name.pop()
1365
 
        # The old collection hasn't noticed yet
1366
 
        self.assertEqual(names, packs.names())
1367
 
        self.assertTrue(packs.reload_pack_names())
1368
 
        self.assertEqual(new_names, packs.names())
1369
 
        # And the repository can access the new revision
1370
 
        self.assertEqual({rev4:(revs[-1],)}, r.get_parent_map([rev4]))
1371
 
        self.assertFalse(packs.reload_pack_names())
1372
 
 
1373
 
    def test_reload_pack_names_added_and_removed(self):
1374
 
        tree, r, packs, revs = self.make_packs_and_alt_repo()
1375
 
        names = packs.names()
1376
 
        # Now repack the whole thing
1377
 
        tree.branch.repository.pack()
1378
 
        new_names = tree.branch.repository._pack_collection.names()
1379
 
        # The other collection hasn't noticed yet
1380
 
        self.assertEqual(names, packs.names())
1381
 
        self.assertTrue(packs.reload_pack_names())
1382
 
        self.assertEqual(new_names, packs.names())
1383
 
        self.assertEqual({revs[-1]:(revs[-2],)}, r.get_parent_map([revs[-1]]))
1384
 
        self.assertFalse(packs.reload_pack_names())
1385
 
 
1386
 
    def test_reload_pack_names_preserves_pending(self):
1387
 
        # TODO: Update this to also test for pending-deleted names
1388
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1389
 
        # We will add one pack (via start_write_group + insert_record_stream),
1390
 
        # and remove another pack (via _remove_pack_from_memory)
1391
 
        orig_names = packs.names()
1392
 
        orig_at_load = packs._packs_at_load
1393
 
        to_remove_name = iter(orig_names).next()
1394
 
        r.start_write_group()
1395
 
        self.addCleanup(r.abort_write_group)
1396
 
        r.texts.insert_record_stream([versionedfile.FulltextContentFactory(
1397
 
            ('text', 'rev'), (), None, 'content\n')])
1398
 
        new_pack = packs._new_pack
1399
 
        self.assertTrue(new_pack.data_inserted())
1400
 
        new_pack.finish()
1401
 
        packs.allocate(new_pack)
1402
 
        packs._new_pack = None
1403
 
        removed_pack = packs.get_pack_by_name(to_remove_name)
1404
 
        packs._remove_pack_from_memory(removed_pack)
1405
 
        names = packs.names()
1406
 
        all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1407
 
        new_names = set([x[0][0] for x in new_nodes])
1408
 
        self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1409
 
        self.assertEqual(set(names) - set(orig_names), new_names)
1410
 
        self.assertEqual(set([new_pack.name]), new_names)
1411
 
        self.assertEqual([to_remove_name],
1412
 
                         sorted([x[0][0] for x in deleted_nodes]))
1413
 
        packs.reload_pack_names()
1414
 
        reloaded_names = packs.names()
1415
 
        self.assertEqual(orig_at_load, packs._packs_at_load)
1416
 
        self.assertEqual(names, reloaded_names)
1417
 
        all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1418
 
        new_names = set([x[0][0] for x in new_nodes])
1419
 
        self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1420
 
        self.assertEqual(set(names) - set(orig_names), new_names)
1421
 
        self.assertEqual(set([new_pack.name]), new_names)
1422
 
        self.assertEqual([to_remove_name],
1423
 
                         sorted([x[0][0] for x in deleted_nodes]))
1424
 
 
1425
 
    def test_autopack_obsoletes_new_pack(self):
1426
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1427
 
        packs._max_pack_count = lambda x: 1
1428
 
        packs.pack_distribution = lambda x: [10]
1429
 
        r.start_write_group()
1430
 
        r.revisions.insert_record_stream([versionedfile.FulltextContentFactory(
1431
 
            ('bogus-rev',), (), None, 'bogus-content\n')])
1432
 
        # This should trigger an autopack, which will combine everything into a
1433
 
        # single pack file.
1434
 
        new_names = r.commit_write_group()
1435
 
        names = packs.names()
1436
 
        self.assertEqual(1, len(names))
1437
 
        self.assertEqual([names[0] + '.pack'],
1438
 
                         packs._pack_transport.list_dir('.'))
1439
 
 
1440
 
    def test_autopack_reloads_and_stops(self):
1441
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1442
 
        # After we have determined what needs to be autopacked, trigger a
1443
 
        # full-pack via the other repo which will cause us to re-evaluate and
1444
 
        # decide we don't need to do anything
1445
 
        orig_execute = packs._execute_pack_operations
1446
 
        def _munged_execute_pack_ops(*args, **kwargs):
1447
 
            tree.branch.repository.pack()
1448
 
            return orig_execute(*args, **kwargs)
1449
 
        packs._execute_pack_operations = _munged_execute_pack_ops
1450
 
        packs._max_pack_count = lambda x: 1
1451
 
        packs.pack_distribution = lambda x: [10]
1452
 
        self.assertFalse(packs.autopack())
1453
 
        self.assertEqual(1, len(packs.names()))
1454
 
        self.assertEqual(tree.branch.repository._pack_collection.names(),
1455
 
                         packs.names())
1456
 
 
1457
 
    def test__save_pack_names(self):
1458
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1459
 
        names = packs.names()
1460
 
        pack = packs.get_pack_by_name(names[0])
1461
 
        packs._remove_pack_from_memory(pack)
1462
 
        packs._save_pack_names(obsolete_packs=[pack])
1463
 
        cur_packs = packs._pack_transport.list_dir('.')
1464
 
        self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1465
 
        # obsolete_packs will also have stuff like .rix and .iix present.
1466
 
        obsolete_packs = packs.transport.list_dir('obsolete_packs')
1467
 
        obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1468
 
        self.assertEqual([pack.name], sorted(obsolete_names))
1469
 
 
1470
 
    def test__save_pack_names_already_obsoleted(self):
1471
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1472
 
        names = packs.names()
1473
 
        pack = packs.get_pack_by_name(names[0])
1474
 
        packs._remove_pack_from_memory(pack)
1475
 
        # We are going to simulate a concurrent autopack by manually obsoleting
1476
 
        # the pack directly.
1477
 
        packs._obsolete_packs([pack])
1478
 
        packs._save_pack_names(clear_obsolete_packs=True,
1479
 
                               obsolete_packs=[pack])
1480
 
        cur_packs = packs._pack_transport.list_dir('.')
1481
 
        self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1482
 
        # Note that while we set clear_obsolete_packs=True, it should not
1483
 
        # delete a pack file that we have also scheduled for obsoletion.
1484
 
        obsolete_packs = packs.transport.list_dir('obsolete_packs')
1485
 
        obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1486
 
        self.assertEqual([pack.name], sorted(obsolete_names))
1487
 
 
1488
 
 
1489
1494
 
1490
1495
class TestPack(TestCaseWithTransport):
1491
1496
    """Tests for the Pack object."""
1545
1550
        pack_transport = self.get_transport('pack')
1546
1551
        index_transport = self.get_transport('index')
1547
1552
        upload_transport.mkdir('.')
1548
 
        collection = pack_repo.RepositoryPackCollection(
1549
 
            repo=None,
1550
 
            transport=self.get_transport('.'),
1551
 
            index_transport=index_transport,
1552
 
            upload_transport=upload_transport,
1553
 
            pack_transport=pack_transport,
1554
 
            index_builder_class=BTreeBuilder,
1555
 
            index_class=BTreeGraphIndex,
1556
 
            use_chk_index=False)
1557
 
        pack = pack_repo.NewPack(collection)
1558
 
        self.addCleanup(pack.abort) # Make sure the write stream gets closed
1559
 
        self.assertIsInstance(pack.revision_index, BTreeBuilder)
1560
 
        self.assertIsInstance(pack.inventory_index, BTreeBuilder)
1561
 
        self.assertIsInstance(pack._hash, type(osutils.md5()))
 
1553
        pack = pack_repo.NewPack(upload_transport, index_transport,
 
1554
            pack_transport)
 
1555
        self.assertIsInstance(pack.revision_index, InMemoryGraphIndex)
 
1556
        self.assertIsInstance(pack.inventory_index, InMemoryGraphIndex)
 
1557
        self.assertIsInstance(pack._hash, type(md5.new()))
1562
1558
        self.assertTrue(pack.upload_transport is upload_transport)
1563
1559
        self.assertTrue(pack.index_transport is index_transport)
1564
1560
        self.assertTrue(pack.pack_transport is pack_transport)
1571
1567
class TestPacker(TestCaseWithTransport):
1572
1568
    """Tests for the packs repository Packer class."""
1573
1569
 
1574
 
    def test_pack_optimizes_pack_order(self):
1575
 
        builder = self.make_branch_builder('.', format="1.9")
1576
 
        builder.start_series()
1577
 
        builder.build_snapshot('A', None, [
1578
 
            ('add', ('', 'root-id', 'directory', None)),
1579
 
            ('add', ('f', 'f-id', 'file', 'content\n'))])
1580
 
        builder.build_snapshot('B', ['A'],
1581
 
            [('modify', ('f-id', 'new-content\n'))])
1582
 
        builder.build_snapshot('C', ['B'],
1583
 
            [('modify', ('f-id', 'third-content\n'))])
1584
 
        builder.build_snapshot('D', ['C'],
1585
 
            [('modify', ('f-id', 'fourth-content\n'))])
1586
 
        b = builder.get_branch()
1587
 
        b.lock_read()
1588
 
        builder.finish_series()
1589
 
        self.addCleanup(b.unlock)
1590
 
        # At this point, we should have 4 pack files available
1591
 
        # Because of how they were built, they correspond to
1592
 
        # ['D', 'C', 'B', 'A']
1593
 
        packs = b.repository._pack_collection.packs
1594
 
        packer = pack_repo.Packer(b.repository._pack_collection,
1595
 
                                  packs, 'testing',
1596
 
                                  revision_ids=['B', 'C'])
1597
 
        # Now, when we are copying the B & C revisions, their pack files should
1598
 
        # be moved to the front of the stack
1599
 
        # The new ordering moves B & C to the front of the .packs attribute,
1600
 
        # and leaves the others in the original order.
1601
 
        new_packs = [packs[1], packs[2], packs[0], packs[3]]
1602
 
        new_pack = packer.pack()
1603
 
        self.assertEqual(new_packs, packer.packs)
1604
 
 
1605
 
 
1606
 
class TestOptimisingPacker(TestCaseWithTransport):
1607
 
    """Tests for the OptimisingPacker class."""
1608
 
 
1609
 
    def get_pack_collection(self):
1610
 
        repo = self.make_repository('.')
1611
 
        return repo._pack_collection
1612
 
 
1613
 
    def test_open_pack_will_optimise(self):
1614
 
        packer = pack_repo.OptimisingPacker(self.get_pack_collection(),
1615
 
                                            [], '.test')
1616
 
        new_pack = packer.open_pack()
1617
 
        self.addCleanup(new_pack.abort) # ensure cleanup
1618
 
        self.assertIsInstance(new_pack, pack_repo.NewPack)
1619
 
        self.assertTrue(new_pack.revision_index._optimize_for_size)
1620
 
        self.assertTrue(new_pack.inventory_index._optimize_for_size)
1621
 
        self.assertTrue(new_pack.text_index._optimize_for_size)
1622
 
        self.assertTrue(new_pack.signature_index._optimize_for_size)
1623
 
 
1624
 
 
1625
 
class TestCrossFormatPacks(TestCaseWithTransport):
1626
 
 
1627
 
    def log_pack(self, hint=None):
1628
 
        self.calls.append(('pack', hint))
1629
 
        self.orig_pack(hint=hint)
1630
 
        if self.expect_hint:
1631
 
            self.assertTrue(hint)
1632
 
 
1633
 
    def run_stream(self, src_fmt, target_fmt, expect_pack_called):
1634
 
        self.expect_hint = expect_pack_called
1635
 
        self.calls = []
1636
 
        source_tree = self.make_branch_and_tree('src', format=src_fmt)
1637
 
        source_tree.lock_write()
1638
 
        self.addCleanup(source_tree.unlock)
1639
 
        tip = source_tree.commit('foo')
1640
 
        target = self.make_repository('target', format=target_fmt)
1641
 
        target.lock_write()
1642
 
        self.addCleanup(target.unlock)
1643
 
        source = source_tree.branch.repository._get_source(target._format)
1644
 
        self.orig_pack = target.pack
1645
 
        target.pack = self.log_pack
1646
 
        search = target.search_missing_revision_ids(
1647
 
            source_tree.branch.repository, tip)
1648
 
        stream = source.get_stream(search)
1649
 
        from_format = source_tree.branch.repository._format
1650
 
        sink = target._get_sink()
1651
 
        sink.insert_stream(stream, from_format, [])
1652
 
        if expect_pack_called:
1653
 
            self.assertLength(1, self.calls)
1654
 
        else:
1655
 
            self.assertLength(0, self.calls)
1656
 
 
1657
 
    def run_fetch(self, src_fmt, target_fmt, expect_pack_called):
1658
 
        self.expect_hint = expect_pack_called
1659
 
        self.calls = []
1660
 
        source_tree = self.make_branch_and_tree('src', format=src_fmt)
1661
 
        source_tree.lock_write()
1662
 
        self.addCleanup(source_tree.unlock)
1663
 
        tip = source_tree.commit('foo')
1664
 
        target = self.make_repository('target', format=target_fmt)
1665
 
        target.lock_write()
1666
 
        self.addCleanup(target.unlock)
1667
 
        source = source_tree.branch.repository
1668
 
        self.orig_pack = target.pack
1669
 
        target.pack = self.log_pack
1670
 
        target.fetch(source)
1671
 
        if expect_pack_called:
1672
 
            self.assertLength(1, self.calls)
1673
 
        else:
1674
 
            self.assertLength(0, self.calls)
1675
 
 
1676
 
    def test_sink_format_hint_no(self):
1677
 
        # When the target format says packing makes no difference, pack is not
1678
 
        # called.
1679
 
        self.run_stream('1.9', 'rich-root-pack', False)
1680
 
 
1681
 
    def test_sink_format_hint_yes(self):
1682
 
        # When the target format says packing makes a difference, pack is
1683
 
        # called.
1684
 
        self.run_stream('1.9', '2a', True)
1685
 
 
1686
 
    def test_sink_format_same_no(self):
1687
 
        # When the formats are the same, pack is not called.
1688
 
        self.run_stream('2a', '2a', False)
1689
 
 
1690
 
    def test_IDS_format_hint_no(self):
1691
 
        # When the target format says packing makes no difference, pack is not
1692
 
        # called.
1693
 
        self.run_fetch('1.9', 'rich-root-pack', False)
1694
 
 
1695
 
    def test_IDS_format_hint_yes(self):
1696
 
        # When the target format says packing makes a difference, pack is
1697
 
        # called.
1698
 
        self.run_fetch('1.9', '2a', True)
1699
 
 
1700
 
    def test_IDS_format_same_no(self):
1701
 
        # When the formats are the same, pack is not called.
1702
 
        self.run_fetch('2a', '2a', False)
 
1570
    # To date, this class has been factored out and nothing new added to it;
 
1571
    # thus there are not yet any tests.
 
1572
 
 
1573
 
 
1574
class TestInterDifferingSerializer(TestCaseWithTransport):
 
1575
 
 
1576
    def test_progress_bar(self):
 
1577
        tree = self.make_branch_and_tree('tree')
 
1578
        tree.commit('rev1', rev_id='rev-1')
 
1579
        tree.commit('rev2', rev_id='rev-2')
 
1580
        tree.commit('rev3', rev_id='rev-3')
 
1581
        repo = self.make_repository('repo')
 
1582
        inter_repo = repository.InterDifferingSerializer(
 
1583
            tree.branch.repository, repo)
 
1584
        pb = progress.InstrumentedProgress(to_file=StringIO())
 
1585
        pb.never_throttle = True
 
1586
        inter_repo.fetch('rev-1', pb)
 
1587
        self.assertEqual('Transferring revisions', pb.last_msg)
 
1588
        self.assertEqual(1, pb.last_cnt)
 
1589
        self.assertEqual(1, pb.last_total)
 
1590
        inter_repo.fetch('rev-3', pb)
 
1591
        self.assertEqual(2, pb.last_cnt)
 
1592
        self.assertEqual(2, pb.last_total)