124
137
# create a branch with a few known format objects.
125
138
# this is not quite the same as
126
139
self.build_tree(["foo/", "bar/"])
127
141
def check_format(format, url):
128
dir = format._matchingbzrdir.initialize(url)
142
dir = format._matchingcontroldir.initialize(url)
129
143
format.initialize(dir)
130
t = get_transport(url)
131
found_format = repository.RepositoryFormat.find_format(dir)
132
self.failUnless(isinstance(found_format, format.__class__))
133
check_format(weaverepo.RepositoryFormat7(), "bar")
144
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
146
self.assertIsInstance(found_format, format.__class__)
147
check_format(repository.format_registry.get_default(), "bar")
135
149
def test_find_format_no_repository(self):
136
150
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
137
151
self.assertRaises(errors.NoRepositoryPresent,
138
repository.RepositoryFormat.find_format,
152
bzrrepository.RepositoryFormatMetaDir.find_format,
155
def test_from_string(self):
156
self.assertIsInstance(
157
SampleRepositoryFormat.from_string(
158
b"Sample .bzr repository format."),
159
SampleRepositoryFormat)
160
self.assertRaises(AssertionError,
161
SampleRepositoryFormat.from_string,
162
b"Different .bzr repository format.")
141
164
def test_find_format_unknown_format(self):
142
165
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
143
166
SampleRepositoryFormat().initialize(dir)
144
167
self.assertRaises(UnknownFormatError,
145
repository.RepositoryFormat.find_format,
168
bzrrepository.RepositoryFormatMetaDir.find_format,
171
def test_find_format_with_features(self):
172
tree = self.make_branch_and_tree('.', format='2a')
173
tree.branch.repository.update_feature_flags({b"name": b"necessity"})
174
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
176
self.assertIsInstance(
177
found_format, bzrrepository.RepositoryFormatMetaDir)
178
self.assertEqual(found_format.features.get(b"name"), b"necessity")
180
bzrdir.MissingFeature, found_format.check_support_status, True)
182
bzrrepository.RepositoryFormatMetaDir.unregister_feature, b"name")
183
bzrrepository.RepositoryFormatMetaDir.register_feature(b"name")
184
found_format.check_support_status(True)
187
class TestRepositoryFormatRegistry(TestCase):
190
super(TestRepositoryFormatRegistry, self).setUp()
191
self.registry = repository.RepositoryFormatRegistry()
148
193
def test_register_unregister_format(self):
149
194
format = SampleRepositoryFormat()
151
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
153
format.initialize(dir)
154
# register a format for it.
155
repository.RepositoryFormat.register_format(format)
156
# which repository.Open will refuse (not supported)
157
self.assertRaises(UnsupportedFormatError, repository.Repository.open, self.get_url())
158
# but open(unsupported) will work
159
self.assertEqual(format.open(dir), "opened repository.")
160
# unregister the format
161
repository.RepositoryFormat.unregister_format(format)
164
class TestFormat6(TestCaseWithTransport):
166
def test_attribute__fetch_order(self):
167
"""Weaves need topological data insertion."""
168
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
169
repo = weaverepo.RepositoryFormat6().initialize(control)
170
self.assertEqual('topological', repo._format._fetch_order)
172
def test_attribute__fetch_uses_deltas(self):
173
"""Weaves do not reuse deltas."""
174
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
175
repo = weaverepo.RepositoryFormat6().initialize(control)
176
self.assertEqual(False, repo._format._fetch_uses_deltas)
178
def test_attribute__fetch_reconcile(self):
179
"""Weave repositories need a reconcile after fetch."""
180
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
181
repo = weaverepo.RepositoryFormat6().initialize(control)
182
self.assertEqual(True, repo._format._fetch_reconcile)
184
def test_no_ancestry_weave(self):
185
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
186
repo = weaverepo.RepositoryFormat6().initialize(control)
187
# We no longer need to create the ancestry.weave file
188
# since it is *never* used.
189
self.assertRaises(NoSuchFile,
190
control.transport.get,
193
def test_supports_external_lookups(self):
194
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
195
repo = weaverepo.RepositoryFormat6().initialize(control)
196
self.assertFalse(repo._format.supports_external_lookups)
199
class TestFormat7(TestCaseWithTransport):
201
def test_attribute__fetch_order(self):
202
"""Weaves need topological data insertion."""
203
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
204
repo = weaverepo.RepositoryFormat7().initialize(control)
205
self.assertEqual('topological', repo._format._fetch_order)
207
def test_attribute__fetch_uses_deltas(self):
208
"""Weaves do not reuse deltas."""
209
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
210
repo = weaverepo.RepositoryFormat7().initialize(control)
211
self.assertEqual(False, repo._format._fetch_uses_deltas)
213
def test_attribute__fetch_reconcile(self):
214
"""Weave repositories need a reconcile after fetch."""
215
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
216
repo = weaverepo.RepositoryFormat7().initialize(control)
217
self.assertEqual(True, repo._format._fetch_reconcile)
219
def test_disk_layout(self):
220
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
221
repo = weaverepo.RepositoryFormat7().initialize(control)
222
# in case of side effects of locking.
226
# format 'Bazaar-NG Repository format 7'
228
# inventory.weave == empty_weave
229
# empty revision-store directory
230
# empty weaves directory
231
t = control.get_repository_transport(None)
232
self.assertEqualDiff('Bazaar-NG Repository format 7',
233
t.get('format').read())
234
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
235
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
236
self.assertEqualDiff('# bzr weave file v5\n'
239
t.get('inventory.weave').read())
240
# Creating a file with id Foo:Bar results in a non-escaped file name on
242
control.create_branch()
243
tree = control.create_workingtree()
244
tree.add(['foo'], ['Foo:Bar'], ['file'])
245
tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
247
tree.commit('first post', rev_id='first')
248
except errors.IllegalPath:
249
if sys.platform != 'win32':
251
self.knownFailure('Foo:Bar cannot be used as a file-id on windows'
254
self.assertEqualDiff(
255
'# bzr weave file v5\n'
257
'1 7fe70820e08a1aac0ef224d9c66ab66831cc4ab1\n'
265
t.get('weaves/74/Foo%3ABar.weave').read())
267
def test_shared_disk_layout(self):
268
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
269
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
271
# format 'Bazaar-NG Repository format 7'
272
# inventory.weave == empty_weave
273
# empty revision-store directory
274
# empty weaves directory
275
# a 'shared-storage' marker file.
276
# lock is not present when unlocked
277
t = control.get_repository_transport(None)
278
self.assertEqualDiff('Bazaar-NG Repository format 7',
279
t.get('format').read())
280
self.assertEqualDiff('', t.get('shared-storage').read())
281
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
282
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
283
self.assertEqualDiff('# bzr weave file v5\n'
286
t.get('inventory.weave').read())
287
self.assertFalse(t.has('branch-lock'))
289
def test_creates_lockdir(self):
290
"""Make sure it appears to be controlled by a LockDir existence"""
291
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
292
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
293
t = control.get_repository_transport(None)
294
# TODO: Should check there is a 'lock' toplevel directory,
295
# regardless of contents
296
self.assertFalse(t.has('lock/held/info'))
299
self.assertTrue(t.has('lock/held/info'))
301
# unlock so we don't get a warning about failing to do so
304
def test_uses_lockdir(self):
305
"""repo format 7 actually locks on lockdir"""
306
base_url = self.get_url()
307
control = bzrdir.BzrDirMetaFormat1().initialize(base_url)
308
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
309
t = control.get_repository_transport(None)
313
# make sure the same lock is created by opening it
314
repo = repository.Repository.open(base_url)
316
self.assertTrue(t.has('lock/held/info'))
318
self.assertFalse(t.has('lock/held/info'))
320
def test_shared_no_tree_disk_layout(self):
321
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
322
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
323
repo.set_make_working_trees(False)
325
# format 'Bazaar-NG Repository format 7'
327
# inventory.weave == empty_weave
328
# empty revision-store directory
329
# empty weaves directory
330
# a 'shared-storage' marker file.
331
t = control.get_repository_transport(None)
332
self.assertEqualDiff('Bazaar-NG Repository format 7',
333
t.get('format').read())
334
## self.assertEqualDiff('', t.get('lock').read())
335
self.assertEqualDiff('', t.get('shared-storage').read())
336
self.assertEqualDiff('', t.get('no-working-trees').read())
337
repo.set_make_working_trees(True)
338
self.assertFalse(t.has('no-working-trees'))
339
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
340
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
341
self.assertEqualDiff('# bzr weave file v5\n'
344
t.get('inventory.weave').read())
346
def test_supports_external_lookups(self):
347
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
348
repo = weaverepo.RepositoryFormat7().initialize(control)
349
self.assertFalse(repo._format.supports_external_lookups)
195
self.registry.register(format)
196
self.assertEqual(format, self.registry.get(
197
b"Sample .bzr repository format."))
198
self.registry.remove(format)
199
self.assertRaises(KeyError, self.registry.get,
200
b"Sample .bzr repository format.")
202
def test_get_all(self):
203
format = SampleRepositoryFormat()
204
self.assertEqual([], self.registry._get_all())
205
self.registry.register(format)
206
self.assertEqual([format], self.registry._get_all())
208
def test_register_extra(self):
209
format = SampleExtraRepositoryFormat()
210
self.assertEqual([], self.registry._get_all())
211
self.registry.register_extra(format)
212
self.assertEqual([format], self.registry._get_all())
214
def test_register_extra_lazy(self):
215
self.assertEqual([], self.registry._get_all())
216
self.registry.register_extra_lazy("breezy.tests.test_repository",
217
"SampleExtraRepositoryFormat")
218
formats = self.registry._get_all()
219
self.assertEqual(1, len(formats))
220
self.assertIsInstance(formats[0], SampleExtraRepositoryFormat)
352
223
class TestFormatKnit1(TestCaseWithTransport):
354
225
def test_attribute__fetch_order(self):
355
226
"""Knits need topological data insertion."""
356
repo = self.make_repository('.',
357
format=bzrdir.format_registry.get('knit')())
227
repo = self.make_repository(
228
'.', format=controldir.format_registry.get('knit')())
358
229
self.assertEqual('topological', repo._format._fetch_order)
360
231
def test_attribute__fetch_uses_deltas(self):
361
232
"""Knits reuse deltas."""
362
repo = self.make_repository('.',
363
format=bzrdir.format_registry.get('knit')())
233
repo = self.make_repository(
234
'.', format=controldir.format_registry.get('knit')())
364
235
self.assertEqual(True, repo._format._fetch_uses_deltas)
366
237
def test_disk_layout(self):
571
456
self.assertGetsDefaultInterRepository(dummy_a, dummy_b)
574
class TestInterWeaveRepo(TestCaseWithTransport):
576
def test_is_compatible_and_registered(self):
577
# InterWeaveRepo is compatible when either side
578
# is a format 5/6/7 branch
579
from bzrlib.repofmt import knitrepo, weaverepo
580
formats = [weaverepo.RepositoryFormat5(),
581
weaverepo.RepositoryFormat6(),
582
weaverepo.RepositoryFormat7()]
583
incompatible_formats = [weaverepo.RepositoryFormat4(),
584
knitrepo.RepositoryFormatKnit1(),
586
repo_a = self.make_repository('a')
587
repo_b = self.make_repository('b')
588
is_compatible = repository.InterWeaveRepo.is_compatible
589
for source in incompatible_formats:
590
# force incompatible left then right
591
repo_a._format = source
592
repo_b._format = formats[0]
593
self.assertFalse(is_compatible(repo_a, repo_b))
594
self.assertFalse(is_compatible(repo_b, repo_a))
595
for source in formats:
596
repo_a._format = source
597
for target in formats:
598
repo_b._format = target
599
self.assertTrue(is_compatible(repo_a, repo_b))
600
self.assertEqual(repository.InterWeaveRepo,
601
repository.InterRepository.get(repo_a,
459
class TestRepositoryFormat1(knitrepo.RepositoryFormatKnit1):
462
def get_format_string(cls):
463
return b"Test Format 1"
466
class TestRepositoryFormat2(knitrepo.RepositoryFormatKnit1):
469
def get_format_string(cls):
470
return b"Test Format 2"
605
473
class TestRepositoryConverter(TestCaseWithTransport):
607
475
def test_convert_empty(self):
608
t = get_transport(self.get_url('.'))
476
source_format = TestRepositoryFormat1()
477
target_format = TestRepositoryFormat2()
478
repository.format_registry.register(source_format)
479
self.addCleanup(repository.format_registry.remove,
481
repository.format_registry.register(target_format)
482
self.addCleanup(repository.format_registry.remove,
484
t = self.get_transport()
609
485
t.mkdir('repository')
610
486
repo_dir = bzrdir.BzrDirMetaFormat1().initialize('repository')
611
repo = weaverepo.RepositoryFormat7().initialize(repo_dir)
612
target_format = knitrepo.RepositoryFormatKnit1()
487
repo = TestRepositoryFormat1().initialize(repo_dir)
613
488
converter = repository.CopyConverter(target_format)
614
pb = bzrlib.ui.ui_factory.nested_progress_bar()
489
with breezy.ui.ui_factory.nested_progress_bar() as pb:
616
490
converter.convert(repo, pb)
619
491
repo = repo_dir.open_repository()
620
492
self.assertTrue(isinstance(target_format, repo._format.__class__))
623
class TestMisc(TestCase):
625
def test_unescape_xml(self):
626
"""We get some kind of error when malformed entities are passed"""
627
self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;')
630
495
class TestRepositoryFormatKnit3(TestCaseWithTransport):
632
497
def test_attribute__fetch_order(self):
682
549
class Test2a(tests.TestCaseWithMemoryTransport):
684
def test_fetch_combines_groups(self):
685
builder = self.make_branch_builder('source', format='2a')
686
builder.start_series()
687
builder.build_snapshot('1', None, [
688
('add', ('', 'root-id', 'directory', '')),
689
('add', ('file', 'file-id', 'file', 'content\n'))])
690
builder.build_snapshot('2', ['1'], [
691
('modify', ('file-id', 'content-2\n'))])
692
builder.finish_series()
693
source = builder.get_branch()
694
target = self.make_repository('target', format='2a')
695
target.fetch(source.repository)
697
self.addCleanup(target.unlock)
698
details = target.texts._index.get_build_details(
699
[('file-id', '1',), ('file-id', '2',)])
700
file_1_details = details[('file-id', '1')]
701
file_2_details = details[('file-id', '2')]
702
# The index, and what to read off disk, should be the same for both
703
# versions of the file.
704
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
706
def test_fetch_combines_groups(self):
707
builder = self.make_branch_builder('source', format='2a')
708
builder.start_series()
709
builder.build_snapshot('1', None, [
710
('add', ('', 'root-id', 'directory', '')),
711
('add', ('file', 'file-id', 'file', 'content\n'))])
712
builder.build_snapshot('2', ['1'], [
713
('modify', ('file-id', 'content-2\n'))])
714
builder.finish_series()
715
source = builder.get_branch()
716
target = self.make_repository('target', format='2a')
717
target.fetch(source.repository)
719
self.addCleanup(target.unlock)
720
details = target.texts._index.get_build_details(
721
[('file-id', '1',), ('file-id', '2',)])
722
file_1_details = details[('file-id', '1')]
723
file_2_details = details[('file-id', '2')]
724
# The index, and what to read off disk, should be the same for both
725
# versions of the file.
726
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
728
def test_fetch_combines_groups(self):
729
builder = self.make_branch_builder('source', format='2a')
730
builder.start_series()
731
builder.build_snapshot('1', None, [
732
('add', ('', 'root-id', 'directory', '')),
733
('add', ('file', 'file-id', 'file', 'content\n'))])
734
builder.build_snapshot('2', ['1'], [
735
('modify', ('file-id', 'content-2\n'))])
736
builder.finish_series()
737
source = builder.get_branch()
738
target = self.make_repository('target', format='2a')
739
target.fetch(source.repository)
741
self.addCleanup(target.unlock)
742
details = target.texts._index.get_build_details(
743
[('file-id', '1',), ('file-id', '2',)])
744
file_1_details = details[('file-id', '1')]
745
file_2_details = details[('file-id', '2')]
551
def test_chk_bytes_uses_custom_btree_parser(self):
552
mt = self.make_branch_and_memory_tree('test', format='2a')
554
self.addCleanup(mt.unlock)
555
mt.add([''], [b'root-id'])
557
index = mt.branch.repository.chk_bytes._index._graph_index._indices[0]
558
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
559
# It should also work if we re-open the repo
560
repo = mt.branch.repository.controldir.open_repository()
562
self.addCleanup(repo.unlock)
563
index = repo.chk_bytes._index._graph_index._indices[0]
564
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
566
def test_fetch_combines_groups(self):
567
builder = self.make_branch_builder('source', format='2a')
568
builder.start_series()
569
builder.build_snapshot(None, [
570
('add', ('', b'root-id', 'directory', '')),
571
('add', ('file', b'file-id', 'file', b'content\n'))],
573
builder.build_snapshot([b'1'], [
574
('modify', ('file', b'content-2\n'))],
576
builder.finish_series()
577
source = builder.get_branch()
578
target = self.make_repository('target', format='2a')
579
target.fetch(source.repository)
581
self.addCleanup(target.unlock)
582
details = target.texts._index.get_build_details(
583
[(b'file-id', b'1',), (b'file-id', b'2',)])
584
file_1_details = details[(b'file-id', b'1')]
585
file_2_details = details[(b'file-id', b'2')]
586
# The index, and what to read off disk, should be the same for both
587
# versions of the file.
588
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
590
def test_fetch_combines_groups(self):
591
builder = self.make_branch_builder('source', format='2a')
592
builder.start_series()
593
builder.build_snapshot(None, [
594
('add', ('', b'root-id', 'directory', '')),
595
('add', ('file', b'file-id', 'file', b'content\n'))],
597
builder.build_snapshot([b'1'], [
598
('modify', ('file', b'content-2\n'))],
600
builder.finish_series()
601
source = builder.get_branch()
602
target = self.make_repository('target', format='2a')
603
target.fetch(source.repository)
605
self.addCleanup(target.unlock)
606
details = target.texts._index.get_build_details(
607
[(b'file-id', b'1',), (b'file-id', b'2',)])
608
file_1_details = details[(b'file-id', b'1')]
609
file_2_details = details[(b'file-id', b'2')]
610
# The index, and what to read off disk, should be the same for both
611
# versions of the file.
612
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
614
def test_fetch_combines_groups(self):
615
builder = self.make_branch_builder('source', format='2a')
616
builder.start_series()
617
builder.build_snapshot(None, [
618
('add', ('', b'root-id', 'directory', '')),
619
('add', ('file', b'file-id', 'file', b'content\n'))],
621
builder.build_snapshot([b'1'], [
622
('modify', ('file', b'content-2\n'))],
624
builder.finish_series()
625
source = builder.get_branch()
626
target = self.make_repository('target', format='2a')
627
target.fetch(source.repository)
629
self.addCleanup(target.unlock)
630
details = target.texts._index.get_build_details(
631
[(b'file-id', b'1',), (b'file-id', b'2',)])
632
file_1_details = details[(b'file-id', b'1')]
633
file_2_details = details[(b'file-id', b'2')]
746
634
# The index, and what to read off disk, should be the same for both
747
635
# versions of the file.
748
636
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
812
700
target = self.make_repository('target', format='rich-root-pack')
813
701
stream = source._get_source(target._format)
814
702
# We don't want the child GroupCHKStreamSource
815
self.assertIs(type(stream), repository.StreamSource)
703
self.assertIs(type(stream), vf_repository.StreamSource)
817
705
def test_get_stream_for_missing_keys_includes_all_chk_refs(self):
818
706
source_builder = self.make_branch_builder('source',
820
708
# We have to build a fairly large tree, so that we are sure the chk
821
709
# pages will have split into multiple pages.
822
entries = [('add', ('', 'a-root-id', 'directory', None))]
710
entries = [('add', ('', b'a-root-id', 'directory', None))]
823
711
for i in 'abcdefghijklmnopqrstuvwxyz123456789':
824
712
for j in 'abcdefghijklmnopqrstuvwxyz123456789':
827
content = 'content for %s\n' % (fname,)
714
fid = fname.encode('utf-8') + b'-id'
715
content = b'content for %s\n' % (fname.encode('utf-8'),)
828
716
entries.append(('add', (fname, fid, 'file', content)))
829
717
source_builder.start_series()
830
source_builder.build_snapshot('rev-1', None, entries)
718
source_builder.build_snapshot(None, entries, revision_id=b'rev-1')
831
719
# Now change a few of them, so we get a few new pages for the second
833
source_builder.build_snapshot('rev-2', ['rev-1'], [
834
('modify', ('aa-id', 'new content for aa-id\n')),
835
('modify', ('cc-id', 'new content for cc-id\n')),
836
('modify', ('zz-id', 'new content for zz-id\n')),
721
source_builder.build_snapshot([b'rev-1'], [
722
('modify', ('aa', b'new content for aa-id\n')),
723
('modify', ('cc', b'new content for cc-id\n')),
724
('modify', ('zz', b'new content for zz-id\n')),
725
], revision_id=b'rev-2')
838
726
source_builder.finish_series()
839
727
source_branch = source_builder.get_branch()
840
728
source_branch.lock_read()
846
734
# On a regular pass, getting the inventories and chk pages for rev-2
847
735
# would only get the newly created chk pages
848
search = graph.SearchResult(set(['rev-2']), set(['rev-1']), 1,
850
simple_chk_records = []
736
search = vf_search.SearchResult({b'rev-2'}, {b'rev-1'}, 1,
738
simple_chk_records = set()
851
739
for vf_name, substream in source.get_stream(search):
852
740
if vf_name == 'chk_bytes':
853
741
for record in substream:
854
simple_chk_records.append(record.key)
742
simple_chk_records.add(record.key)
856
744
for _ in substream:
858
746
# 3 pages, the root (InternalNode), + 2 pages which actually changed
859
self.assertEqual([('sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
860
('sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
861
('sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
862
('sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)],
747
self.assertEqual({(b'sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
748
(b'sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
749
(b'sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
750
(b'sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)},
751
set(simple_chk_records))
864
752
# Now, when we do a similar call using 'get_stream_for_missing_keys'
865
753
# we should get a much larger set of pages.
866
missing = [('inventories', 'rev-2')]
867
full_chk_records = []
754
missing = [('inventories', b'rev-2')]
755
full_chk_records = set()
868
756
for vf_name, substream in source.get_stream_for_missing_keys(missing):
869
757
if vf_name == 'inventories':
870
758
for record in substream:
871
self.assertEqual(('rev-2',), record.key)
759
self.assertEqual((b'rev-2',), record.key)
872
760
elif vf_name == 'chk_bytes':
873
761
for record in substream:
874
full_chk_records.append(record.key)
762
full_chk_records.add(record.key)
876
764
self.fail('Should not be getting a stream of %s' % (vf_name,))
877
765
# We have 257 records now. This is because we have 1 root page, and 256
958
851
super(TestDevelopment6FindParentIdsOfRevisions, self).setUp()
959
self.builder = self.make_branch_builder('source',
960
format='development6-rich-root')
852
self.builder = self.make_branch_builder('source')
961
853
self.builder.start_series()
962
self.builder.build_snapshot('initial', None,
963
[('add', ('', 'tree-root', 'directory', None))])
854
self.builder.build_snapshot(
856
[('add', ('', b'tree-root', 'directory', None))],
857
revision_id=b'initial')
964
858
self.repo = self.builder.get_branch().repository
965
859
self.addCleanup(self.builder.finish_series)
967
861
def assertParentIds(self, expected_result, rev_set):
968
self.assertEqual(sorted(expected_result),
863
sorted(expected_result),
969
864
sorted(self.repo._find_parent_ids_of_revisions(rev_set)))
971
866
def test_simple(self):
972
self.builder.build_snapshot('revid1', None, [])
973
self.builder.build_snapshot('revid2', ['revid1'], [])
975
self.assertParentIds(['revid1'], rev_set)
867
self.builder.build_snapshot(None, [], revision_id=b'revid1')
868
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
869
rev_set = [b'revid2']
870
self.assertParentIds([b'revid1'], rev_set)
977
872
def test_not_first_parent(self):
978
self.builder.build_snapshot('revid1', None, [])
979
self.builder.build_snapshot('revid2', ['revid1'], [])
980
self.builder.build_snapshot('revid3', ['revid2'], [])
981
rev_set = ['revid3', 'revid2']
982
self.assertParentIds(['revid1'], rev_set)
873
self.builder.build_snapshot(None, [], revision_id=b'revid1')
874
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
875
self.builder.build_snapshot([b'revid2'], [], revision_id=b'revid3')
876
rev_set = [b'revid3', b'revid2']
877
self.assertParentIds([b'revid1'], rev_set)
984
879
def test_not_null(self):
985
rev_set = ['initial']
880
rev_set = [b'initial']
986
881
self.assertParentIds([], rev_set)
988
883
def test_not_null_set(self):
989
self.builder.build_snapshot('revid1', None, [])
884
self.builder.build_snapshot(None, [], revision_id=b'revid1')
990
885
rev_set = [_mod_revision.NULL_REVISION]
991
886
self.assertParentIds([], rev_set)
993
888
def test_ghost(self):
994
self.builder.build_snapshot('revid1', None, [])
995
rev_set = ['ghost', 'revid1']
996
self.assertParentIds(['initial'], rev_set)
889
self.builder.build_snapshot(None, [], revision_id=b'revid1')
890
rev_set = [b'ghost', b'revid1']
891
self.assertParentIds([b'initial'], rev_set)
998
893
def test_ghost_parent(self):
999
self.builder.build_snapshot('revid1', None, [])
1000
self.builder.build_snapshot('revid2', ['revid1', 'ghost'], [])
1001
rev_set = ['revid2', 'revid1']
1002
self.assertParentIds(['ghost', 'initial'], rev_set)
894
self.builder.build_snapshot(None, [], revision_id=b'revid1')
895
self.builder.build_snapshot(
896
[b'revid1', b'ghost'], [], revision_id=b'revid2')
897
rev_set = [b'revid2', b'revid1']
898
self.assertParentIds([b'ghost', b'initial'], rev_set)
1004
900
def test_righthand_parent(self):
1005
self.builder.build_snapshot('revid1', None, [])
1006
self.builder.build_snapshot('revid2a', ['revid1'], [])
1007
self.builder.build_snapshot('revid2b', ['revid1'], [])
1008
self.builder.build_snapshot('revid3', ['revid2a', 'revid2b'], [])
1009
rev_set = ['revid3', 'revid2a']
1010
self.assertParentIds(['revid1', 'revid2b'], rev_set)
901
self.builder.build_snapshot(None, [], revision_id=b'revid1')
902
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2a')
903
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2b')
904
self.builder.build_snapshot([b'revid2a', b'revid2b'], [],
905
revision_id=b'revid3')
906
rev_set = [b'revid3', b'revid2a']
907
self.assertParentIds([b'revid1', b'revid2b'], rev_set)
1013
910
class TestWithBrokenRepo(TestCaseWithTransport):
1025
922
repo.start_write_group()
1026
923
cleanups.append(repo.commit_write_group)
1027
924
# make rev1a: A well-formed revision, containing 'file1'
1028
inv = inventory.Inventory(revision_id='rev1a')
1029
inv.root.revision = 'rev1a'
1030
self.add_file(repo, inv, 'file1', 'rev1a', [])
1031
repo.texts.add_lines((inv.root.file_id, 'rev1a'), [], [])
1032
repo.add_inventory('rev1a', inv, [])
1033
revision = _mod_revision.Revision('rev1a',
925
inv = inventory.Inventory(revision_id=b'rev1a')
926
inv.root.revision = b'rev1a'
927
self.add_file(repo, inv, 'file1', b'rev1a', [])
928
repo.texts.add_lines((inv.root.file_id, b'rev1a'), [], [])
929
repo.add_inventory(b'rev1a', inv, [])
930
revision = _mod_revision.Revision(
1034
932
committer='jrandom@example.com', timestamp=0,
1035
933
inventory_sha1='', timezone=0, message='foo', parent_ids=[])
1036
repo.add_revision('rev1a',revision, inv)
934
repo.add_revision(b'rev1a', revision, inv)
1038
936
# make rev1b, which has no Revision, but has an Inventory, and
1040
inv = inventory.Inventory(revision_id='rev1b')
1041
inv.root.revision = 'rev1b'
1042
self.add_file(repo, inv, 'file1', 'rev1b', [])
1043
repo.add_inventory('rev1b', inv, [])
938
inv = inventory.Inventory(revision_id=b'rev1b')
939
inv.root.revision = b'rev1b'
940
self.add_file(repo, inv, 'file1', b'rev1b', [])
941
repo.add_inventory(b'rev1b', inv, [])
1045
943
# make rev2, with file1 and file2
1047
945
# file1 has 'rev1b' as an ancestor, even though this is not
1048
946
# mentioned by 'rev1a', making it an unreferenced ancestor
1049
947
inv = inventory.Inventory()
1050
self.add_file(repo, inv, 'file1', 'rev2', ['rev1a', 'rev1b'])
1051
self.add_file(repo, inv, 'file2', 'rev2', [])
1052
self.add_revision(repo, 'rev2', inv, ['rev1a'])
948
self.add_file(repo, inv, 'file1', b'rev2', [b'rev1a', b'rev1b'])
949
self.add_file(repo, inv, 'file2', b'rev2', [])
950
self.add_revision(repo, b'rev2', inv, [b'rev1a'])
1054
952
# make ghost revision rev1c
1055
953
inv = inventory.Inventory()
1056
self.add_file(repo, inv, 'file2', 'rev1c', [])
954
self.add_file(repo, inv, 'file2', b'rev1c', [])
1058
956
# make rev3 with file2
1059
957
# file2 refers to 'rev1c', which is a ghost in this repository, so
1060
958
# file2 cannot have rev1c as its ancestor.
1061
959
inv = inventory.Inventory()
1062
self.add_file(repo, inv, 'file2', 'rev3', ['rev1c'])
1063
self.add_revision(repo, 'rev3', inv, ['rev1c'])
960
self.add_file(repo, inv, 'file2', b'rev3', [b'rev1c'])
961
self.add_revision(repo, b'rev3', inv, [b'rev1c'])
1066
964
for cleanup in reversed(cleanups):
1622
1553
self.assertTrue(new_pack.signature_index._optimize_for_size)
1556
class TestGCCHKPacker(TestCaseWithTransport):
1558
def make_abc_branch(self):
1559
builder = self.make_branch_builder('source')
1560
builder.start_series()
1561
builder.build_snapshot(None, [
1562
('add', ('', b'root-id', 'directory', None)),
1563
('add', ('file', b'file-id', 'file', b'content\n')),
1564
], revision_id=b'A')
1565
builder.build_snapshot([b'A'], [
1566
('add', ('dir', b'dir-id', 'directory', None))],
1568
builder.build_snapshot([b'B'], [
1569
('modify', ('file', b'new content\n'))],
1571
builder.finish_series()
1572
return builder.get_branch()
1574
def make_branch_with_disjoint_inventory_and_revision(self):
1575
"""a repo with separate packs for a revisions Revision and Inventory.
1577
There will be one pack file that holds the Revision content, and one
1578
for the Inventory content.
1580
:return: (repository,
1581
pack_name_with_rev_A_Revision,
1582
pack_name_with_rev_A_Inventory,
1583
pack_name_with_rev_C_content)
1585
b_source = self.make_abc_branch()
1586
b_base = b_source.controldir.sprout(
1587
'base', revision_id=b'A').open_branch()
1588
b_stacked = b_base.controldir.sprout(
1589
'stacked', stacked=True).open_branch()
1590
b_stacked.lock_write()
1591
self.addCleanup(b_stacked.unlock)
1592
b_stacked.fetch(b_source, b'B')
1593
# Now re-open the stacked repo directly (no fallbacks) so that we can
1594
# fill in the A rev.
1595
repo_not_stacked = b_stacked.controldir.open_repository()
1596
repo_not_stacked.lock_write()
1597
self.addCleanup(repo_not_stacked.unlock)
1598
# Now we should have a pack file with A's inventory, but not its
1600
self.assertEqual([(b'A',), (b'B',)],
1601
sorted(repo_not_stacked.inventories.keys()))
1602
self.assertEqual([(b'B',)],
1603
sorted(repo_not_stacked.revisions.keys()))
1604
stacked_pack_names = repo_not_stacked._pack_collection.names()
1605
# We have a couple names here, figure out which has A's inventory
1606
for name in stacked_pack_names:
1607
pack = repo_not_stacked._pack_collection.get_pack_by_name(name)
1608
keys = [n[1] for n in pack.inventory_index.iter_all_entries()]
1610
inv_a_pack_name = name
1613
self.fail('Could not find pack containing A\'s inventory')
1614
repo_not_stacked.fetch(b_source.repository, b'A')
1615
self.assertEqual([(b'A',), (b'B',)],
1616
sorted(repo_not_stacked.revisions.keys()))
1617
new_pack_names = set(repo_not_stacked._pack_collection.names())
1618
rev_a_pack_names = new_pack_names.difference(stacked_pack_names)
1619
self.assertEqual(1, len(rev_a_pack_names))
1620
rev_a_pack_name = list(rev_a_pack_names)[0]
1621
# Now fetch 'C', so we have a couple pack files to join
1622
repo_not_stacked.fetch(b_source.repository, b'C')
1623
rev_c_pack_names = set(repo_not_stacked._pack_collection.names())
1624
rev_c_pack_names = rev_c_pack_names.difference(new_pack_names)
1625
self.assertEqual(1, len(rev_c_pack_names))
1626
rev_c_pack_name = list(rev_c_pack_names)[0]
1627
return (repo_not_stacked, rev_a_pack_name, inv_a_pack_name,
1630
def test_pack_with_distant_inventories(self):
1631
# See https://bugs.launchpad.net/bzr/+bug/437003
1632
# When repacking, it is possible to have an inventory in a different
1633
# pack file than the associated revision. An autopack can then come
1634
# along, and miss that inventory, and complain.
1635
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1636
) = self.make_branch_with_disjoint_inventory_and_revision()
1637
a_pack = repo._pack_collection.get_pack_by_name(rev_a_pack_name)
1638
c_pack = repo._pack_collection.get_pack_by_name(rev_c_pack_name)
1639
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1640
[a_pack, c_pack], '.test-pack')
1641
# This would raise ValueError in bug #437003, but should not raise an
1645
def test_pack_with_missing_inventory(self):
1646
# Similar to test_pack_with_missing_inventory, but this time, we force
1647
# the A inventory to actually be gone from the repository.
1648
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1649
) = self.make_branch_with_disjoint_inventory_and_revision()
1650
inv_a_pack = repo._pack_collection.get_pack_by_name(inv_a_pack_name)
1651
repo._pack_collection._remove_pack_from_memory(inv_a_pack)
1652
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1653
repo._pack_collection.all_packs(), '.test-pack')
1654
e = self.assertRaises(ValueError, packer.pack)
1655
packer.new_pack.abort()
1656
self.assertContainsRe(str(e),
1657
r"We are missing inventories for revisions: .*'A'")
1625
1660
class TestCrossFormatPacks(TestCaseWithTransport):
1627
1662
def log_pack(self, hint=None):