137
124
# create a branch with a few known format objects.
138
125
# this is not quite the same as
139
126
self.build_tree(["foo/", "bar/"])
141
127
def check_format(format, url):
142
dir = format._matchingcontroldir.initialize(url)
128
dir = format._matchingbzrdir.initialize(url)
143
129
format.initialize(dir)
144
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
146
self.assertIsInstance(found_format, format.__class__)
147
check_format(repository.format_registry.get_default(), "bar")
130
t = get_transport(url)
131
found_format = repository.RepositoryFormat.find_format(dir)
132
self.failUnless(isinstance(found_format, format.__class__))
133
check_format(weaverepo.RepositoryFormat7(), "bar")
149
135
def test_find_format_no_repository(self):
150
136
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
151
137
self.assertRaises(errors.NoRepositoryPresent,
152
bzrrepository.RepositoryFormatMetaDir.find_format,
138
repository.RepositoryFormat.find_format,
155
def test_from_string(self):
156
self.assertIsInstance(
157
SampleRepositoryFormat.from_string(
158
b"Sample .bzr repository format."),
159
SampleRepositoryFormat)
160
self.assertRaises(AssertionError,
161
SampleRepositoryFormat.from_string,
162
b"Different .bzr repository format.")
164
141
def test_find_format_unknown_format(self):
165
142
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
166
143
SampleRepositoryFormat().initialize(dir)
167
144
self.assertRaises(UnknownFormatError,
168
bzrrepository.RepositoryFormatMetaDir.find_format,
145
repository.RepositoryFormat.find_format,
171
def test_find_format_with_features(self):
172
tree = self.make_branch_and_tree('.', format='2a')
173
tree.branch.repository.update_feature_flags({b"name": b"necessity"})
174
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
176
self.assertIsInstance(
177
found_format, bzrrepository.RepositoryFormatMetaDir)
178
self.assertEqual(found_format.features.get(b"name"), b"necessity")
180
bzrdir.MissingFeature, found_format.check_support_status, True)
182
bzrrepository.RepositoryFormatMetaDir.unregister_feature, b"name")
183
bzrrepository.RepositoryFormatMetaDir.register_feature(b"name")
184
found_format.check_support_status(True)
187
class TestRepositoryFormatRegistry(TestCase):
190
super(TestRepositoryFormatRegistry, self).setUp()
191
self.registry = repository.RepositoryFormatRegistry()
193
148
def test_register_unregister_format(self):
194
149
format = SampleRepositoryFormat()
195
self.registry.register(format)
196
self.assertEqual(format, self.registry.get(
197
b"Sample .bzr repository format."))
198
self.registry.remove(format)
199
self.assertRaises(KeyError, self.registry.get,
200
b"Sample .bzr repository format.")
202
def test_get_all(self):
203
format = SampleRepositoryFormat()
204
self.assertEqual([], self.registry._get_all())
205
self.registry.register(format)
206
self.assertEqual([format], self.registry._get_all())
208
def test_register_extra(self):
209
format = SampleExtraRepositoryFormat()
210
self.assertEqual([], self.registry._get_all())
211
self.registry.register_extra(format)
212
self.assertEqual([format], self.registry._get_all())
214
def test_register_extra_lazy(self):
215
self.assertEqual([], self.registry._get_all())
216
self.registry.register_extra_lazy("breezy.tests.test_repository",
217
"SampleExtraRepositoryFormat")
218
formats = self.registry._get_all()
219
self.assertEqual(1, len(formats))
220
self.assertIsInstance(formats[0], SampleExtraRepositoryFormat)
151
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
153
format.initialize(dir)
154
# register a format for it.
155
repository.RepositoryFormat.register_format(format)
156
# which repository.Open will refuse (not supported)
157
self.assertRaises(UnsupportedFormatError, repository.Repository.open, self.get_url())
158
# but open(unsupported) will work
159
self.assertEqual(format.open(dir), "opened repository.")
160
# unregister the format
161
repository.RepositoryFormat.unregister_format(format)
164
class TestFormat6(TestCaseWithTransport):
166
def test_attribute__fetch_order(self):
167
"""Weaves need topological data insertion."""
168
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
169
repo = weaverepo.RepositoryFormat6().initialize(control)
170
self.assertEqual('topological', repo._format._fetch_order)
172
def test_attribute__fetch_uses_deltas(self):
173
"""Weaves do not reuse deltas."""
174
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
175
repo = weaverepo.RepositoryFormat6().initialize(control)
176
self.assertEqual(False, repo._format._fetch_uses_deltas)
178
def test_attribute__fetch_reconcile(self):
179
"""Weave repositories need a reconcile after fetch."""
180
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
181
repo = weaverepo.RepositoryFormat6().initialize(control)
182
self.assertEqual(True, repo._format._fetch_reconcile)
184
def test_no_ancestry_weave(self):
185
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
186
repo = weaverepo.RepositoryFormat6().initialize(control)
187
# We no longer need to create the ancestry.weave file
188
# since it is *never* used.
189
self.assertRaises(NoSuchFile,
190
control.transport.get,
193
def test_supports_external_lookups(self):
194
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
195
repo = weaverepo.RepositoryFormat6().initialize(control)
196
self.assertFalse(repo._format.supports_external_lookups)
199
class TestFormat7(TestCaseWithTransport):
201
def test_attribute__fetch_order(self):
202
"""Weaves need topological data insertion."""
203
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
204
repo = weaverepo.RepositoryFormat7().initialize(control)
205
self.assertEqual('topological', repo._format._fetch_order)
207
def test_attribute__fetch_uses_deltas(self):
208
"""Weaves do not reuse deltas."""
209
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
210
repo = weaverepo.RepositoryFormat7().initialize(control)
211
self.assertEqual(False, repo._format._fetch_uses_deltas)
213
def test_attribute__fetch_reconcile(self):
214
"""Weave repositories need a reconcile after fetch."""
215
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
216
repo = weaverepo.RepositoryFormat7().initialize(control)
217
self.assertEqual(True, repo._format._fetch_reconcile)
219
def test_disk_layout(self):
220
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
221
repo = weaverepo.RepositoryFormat7().initialize(control)
222
# in case of side effects of locking.
226
# format 'Bazaar-NG Repository format 7'
228
# inventory.weave == empty_weave
229
# empty revision-store directory
230
# empty weaves directory
231
t = control.get_repository_transport(None)
232
self.assertEqualDiff('Bazaar-NG Repository format 7',
233
t.get('format').read())
234
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
235
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
236
self.assertEqualDiff('# bzr weave file v5\n'
239
t.get('inventory.weave').read())
240
# Creating a file with id Foo:Bar results in a non-escaped file name on
242
control.create_branch()
243
tree = control.create_workingtree()
244
tree.add(['foo'], ['Foo:Bar'], ['file'])
245
tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
247
tree.commit('first post', rev_id='first')
248
except errors.IllegalPath:
249
if sys.platform != 'win32':
251
self.knownFailure('Foo:Bar cannot be used as a file-id on windows'
254
self.assertEqualDiff(
255
'# bzr weave file v5\n'
257
'1 7fe70820e08a1aac0ef224d9c66ab66831cc4ab1\n'
265
t.get('weaves/74/Foo%3ABar.weave').read())
267
def test_shared_disk_layout(self):
268
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
269
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
271
# format 'Bazaar-NG Repository format 7'
272
# inventory.weave == empty_weave
273
# empty revision-store directory
274
# empty weaves directory
275
# a 'shared-storage' marker file.
276
# lock is not present when unlocked
277
t = control.get_repository_transport(None)
278
self.assertEqualDiff('Bazaar-NG Repository format 7',
279
t.get('format').read())
280
self.assertEqualDiff('', t.get('shared-storage').read())
281
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
282
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
283
self.assertEqualDiff('# bzr weave file v5\n'
286
t.get('inventory.weave').read())
287
self.assertFalse(t.has('branch-lock'))
289
def test_creates_lockdir(self):
290
"""Make sure it appears to be controlled by a LockDir existence"""
291
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
292
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
293
t = control.get_repository_transport(None)
294
# TODO: Should check there is a 'lock' toplevel directory,
295
# regardless of contents
296
self.assertFalse(t.has('lock/held/info'))
299
self.assertTrue(t.has('lock/held/info'))
301
# unlock so we don't get a warning about failing to do so
304
def test_uses_lockdir(self):
305
"""repo format 7 actually locks on lockdir"""
306
base_url = self.get_url()
307
control = bzrdir.BzrDirMetaFormat1().initialize(base_url)
308
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
309
t = control.get_repository_transport(None)
313
# make sure the same lock is created by opening it
314
repo = repository.Repository.open(base_url)
316
self.assertTrue(t.has('lock/held/info'))
318
self.assertFalse(t.has('lock/held/info'))
320
def test_shared_no_tree_disk_layout(self):
321
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
322
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
323
repo.set_make_working_trees(False)
325
# format 'Bazaar-NG Repository format 7'
327
# inventory.weave == empty_weave
328
# empty revision-store directory
329
# empty weaves directory
330
# a 'shared-storage' marker file.
331
t = control.get_repository_transport(None)
332
self.assertEqualDiff('Bazaar-NG Repository format 7',
333
t.get('format').read())
334
## self.assertEqualDiff('', t.get('lock').read())
335
self.assertEqualDiff('', t.get('shared-storage').read())
336
self.assertEqualDiff('', t.get('no-working-trees').read())
337
repo.set_make_working_trees(True)
338
self.assertFalse(t.has('no-working-trees'))
339
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
340
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
341
self.assertEqualDiff('# bzr weave file v5\n'
344
t.get('inventory.weave').read())
346
def test_supports_external_lookups(self):
347
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
348
repo = weaverepo.RepositoryFormat7().initialize(control)
349
self.assertFalse(repo._format.supports_external_lookups)
223
352
class TestFormatKnit1(TestCaseWithTransport):
225
354
def test_attribute__fetch_order(self):
226
355
"""Knits need topological data insertion."""
227
repo = self.make_repository(
228
'.', format=controldir.format_registry.get('knit')())
356
repo = self.make_repository('.',
357
format=bzrdir.format_registry.get('knit')())
229
358
self.assertEqual('topological', repo._format._fetch_order)
231
360
def test_attribute__fetch_uses_deltas(self):
232
361
"""Knits reuse deltas."""
233
repo = self.make_repository(
234
'.', format=controldir.format_registry.get('knit')())
362
repo = self.make_repository('.',
363
format=bzrdir.format_registry.get('knit')())
235
364
self.assertEqual(True, repo._format._fetch_uses_deltas)
237
366
def test_disk_layout(self):
456
571
self.assertGetsDefaultInterRepository(dummy_a, dummy_b)
459
class TestRepositoryFormat1(knitrepo.RepositoryFormatKnit1):
462
def get_format_string(cls):
463
return b"Test Format 1"
466
class TestRepositoryFormat2(knitrepo.RepositoryFormatKnit1):
469
def get_format_string(cls):
470
return b"Test Format 2"
574
class TestInterWeaveRepo(TestCaseWithTransport):
576
def test_is_compatible_and_registered(self):
577
# InterWeaveRepo is compatible when either side
578
# is a format 5/6/7 branch
579
from bzrlib.repofmt import knitrepo, weaverepo
580
formats = [weaverepo.RepositoryFormat5(),
581
weaverepo.RepositoryFormat6(),
582
weaverepo.RepositoryFormat7()]
583
incompatible_formats = [weaverepo.RepositoryFormat4(),
584
knitrepo.RepositoryFormatKnit1(),
586
repo_a = self.make_repository('a')
587
repo_b = self.make_repository('b')
588
is_compatible = repository.InterWeaveRepo.is_compatible
589
for source in incompatible_formats:
590
# force incompatible left then right
591
repo_a._format = source
592
repo_b._format = formats[0]
593
self.assertFalse(is_compatible(repo_a, repo_b))
594
self.assertFalse(is_compatible(repo_b, repo_a))
595
for source in formats:
596
repo_a._format = source
597
for target in formats:
598
repo_b._format = target
599
self.assertTrue(is_compatible(repo_a, repo_b))
600
self.assertEqual(repository.InterWeaveRepo,
601
repository.InterRepository.get(repo_a,
473
605
class TestRepositoryConverter(TestCaseWithTransport):
475
607
def test_convert_empty(self):
476
source_format = TestRepositoryFormat1()
477
target_format = TestRepositoryFormat2()
478
repository.format_registry.register(source_format)
479
self.addCleanup(repository.format_registry.remove,
481
repository.format_registry.register(target_format)
482
self.addCleanup(repository.format_registry.remove,
484
t = self.get_transport()
608
t = get_transport(self.get_url('.'))
485
609
t.mkdir('repository')
486
610
repo_dir = bzrdir.BzrDirMetaFormat1().initialize('repository')
487
repo = TestRepositoryFormat1().initialize(repo_dir)
611
repo = weaverepo.RepositoryFormat7().initialize(repo_dir)
612
target_format = knitrepo.RepositoryFormatKnit1()
488
613
converter = repository.CopyConverter(target_format)
489
with breezy.ui.ui_factory.nested_progress_bar() as pb:
614
pb = bzrlib.ui.ui_factory.nested_progress_bar()
490
616
converter.convert(repo, pb)
491
619
repo = repo_dir.open_repository()
492
620
self.assertTrue(isinstance(target_format, repo._format.__class__))
623
class TestMisc(TestCase):
625
def test_unescape_xml(self):
626
"""We get some kind of error when malformed entities are passed"""
627
self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;')
495
630
class TestRepositoryFormatKnit3(TestCaseWithTransport):
497
632
def test_attribute__fetch_order(self):
541
682
class Test2a(tests.TestCaseWithMemoryTransport):
543
def test_chk_bytes_uses_custom_btree_parser(self):
544
mt = self.make_branch_and_memory_tree('test', format='2a')
546
self.addCleanup(mt.unlock)
547
mt.add([''], [b'root-id'])
549
index = mt.branch.repository.chk_bytes._index._graph_index._indices[0]
550
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
551
# It should also work if we re-open the repo
552
repo = mt.branch.repository.controldir.open_repository()
554
self.addCleanup(repo.unlock)
555
index = repo.chk_bytes._index._graph_index._indices[0]
556
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
558
def test_fetch_combines_groups(self):
559
builder = self.make_branch_builder('source', format='2a')
560
builder.start_series()
561
builder.build_snapshot(None, [
562
('add', ('', b'root-id', 'directory', '')),
563
('add', ('file', b'file-id', 'file', b'content\n'))],
565
builder.build_snapshot([b'1'], [
566
('modify', ('file', b'content-2\n'))],
568
builder.finish_series()
569
source = builder.get_branch()
570
target = self.make_repository('target', format='2a')
571
target.fetch(source.repository)
573
self.addCleanup(target.unlock)
574
details = target.texts._index.get_build_details(
575
[(b'file-id', b'1',), (b'file-id', b'2',)])
576
file_1_details = details[(b'file-id', b'1')]
577
file_2_details = details[(b'file-id', b'2')]
578
# The index, and what to read off disk, should be the same for both
579
# versions of the file.
580
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
582
def test_fetch_combines_groups(self):
583
builder = self.make_branch_builder('source', format='2a')
584
builder.start_series()
585
builder.build_snapshot(None, [
586
('add', ('', b'root-id', 'directory', '')),
587
('add', ('file', b'file-id', 'file', b'content\n'))],
589
builder.build_snapshot([b'1'], [
590
('modify', ('file', b'content-2\n'))],
592
builder.finish_series()
593
source = builder.get_branch()
594
target = self.make_repository('target', format='2a')
595
target.fetch(source.repository)
597
self.addCleanup(target.unlock)
598
details = target.texts._index.get_build_details(
599
[(b'file-id', b'1',), (b'file-id', b'2',)])
600
file_1_details = details[(b'file-id', b'1')]
601
file_2_details = details[(b'file-id', b'2')]
602
# The index, and what to read off disk, should be the same for both
603
# versions of the file.
604
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
606
def test_fetch_combines_groups(self):
607
builder = self.make_branch_builder('source', format='2a')
608
builder.start_series()
609
builder.build_snapshot(None, [
610
('add', ('', b'root-id', 'directory', '')),
611
('add', ('file', b'file-id', 'file', b'content\n'))],
613
builder.build_snapshot([b'1'], [
614
('modify', ('file', b'content-2\n'))],
616
builder.finish_series()
617
source = builder.get_branch()
618
target = self.make_repository('target', format='2a')
619
target.fetch(source.repository)
621
self.addCleanup(target.unlock)
622
details = target.texts._index.get_build_details(
623
[(b'file-id', b'1',), (b'file-id', b'2',)])
624
file_1_details = details[(b'file-id', b'1')]
625
file_2_details = details[(b'file-id', b'2')]
684
def test_fetch_combines_groups(self):
685
builder = self.make_branch_builder('source', format='2a')
686
builder.start_series()
687
builder.build_snapshot('1', None, [
688
('add', ('', 'root-id', 'directory', '')),
689
('add', ('file', 'file-id', 'file', 'content\n'))])
690
builder.build_snapshot('2', ['1'], [
691
('modify', ('file-id', 'content-2\n'))])
692
builder.finish_series()
693
source = builder.get_branch()
694
target = self.make_repository('target', format='2a')
695
target.fetch(source.repository)
697
self.addCleanup(target.unlock)
698
details = target.texts._index.get_build_details(
699
[('file-id', '1',), ('file-id', '2',)])
700
file_1_details = details[('file-id', '1')]
701
file_2_details = details[('file-id', '2')]
702
# The index, and what to read off disk, should be the same for both
703
# versions of the file.
704
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
706
def test_fetch_combines_groups(self):
707
builder = self.make_branch_builder('source', format='2a')
708
builder.start_series()
709
builder.build_snapshot('1', None, [
710
('add', ('', 'root-id', 'directory', '')),
711
('add', ('file', 'file-id', 'file', 'content\n'))])
712
builder.build_snapshot('2', ['1'], [
713
('modify', ('file-id', 'content-2\n'))])
714
builder.finish_series()
715
source = builder.get_branch()
716
target = self.make_repository('target', format='2a')
717
target.fetch(source.repository)
719
self.addCleanup(target.unlock)
720
details = target.texts._index.get_build_details(
721
[('file-id', '1',), ('file-id', '2',)])
722
file_1_details = details[('file-id', '1')]
723
file_2_details = details[('file-id', '2')]
724
# The index, and what to read off disk, should be the same for both
725
# versions of the file.
726
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
728
def test_fetch_combines_groups(self):
729
builder = self.make_branch_builder('source', format='2a')
730
builder.start_series()
731
builder.build_snapshot('1', None, [
732
('add', ('', 'root-id', 'directory', '')),
733
('add', ('file', 'file-id', 'file', 'content\n'))])
734
builder.build_snapshot('2', ['1'], [
735
('modify', ('file-id', 'content-2\n'))])
736
builder.finish_series()
737
source = builder.get_branch()
738
target = self.make_repository('target', format='2a')
739
target.fetch(source.repository)
741
self.addCleanup(target.unlock)
742
details = target.texts._index.get_build_details(
743
[('file-id', '1',), ('file-id', '2',)])
744
file_1_details = details[('file-id', '1')]
745
file_2_details = details[('file-id', '2')]
626
746
# The index, and what to read off disk, should be the same for both
627
747
# versions of the file.
628
748
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
692
812
target = self.make_repository('target', format='rich-root-pack')
693
813
stream = source._get_source(target._format)
694
814
# We don't want the child GroupCHKStreamSource
695
self.assertIs(type(stream), vf_repository.StreamSource)
815
self.assertIs(type(stream), repository.StreamSource)
697
817
def test_get_stream_for_missing_keys_includes_all_chk_refs(self):
698
818
source_builder = self.make_branch_builder('source',
700
820
# We have to build a fairly large tree, so that we are sure the chk
701
821
# pages will have split into multiple pages.
702
entries = [('add', ('', b'a-root-id', 'directory', None))]
822
entries = [('add', ('', 'a-root-id', 'directory', None))]
703
823
for i in 'abcdefghijklmnopqrstuvwxyz123456789':
704
824
for j in 'abcdefghijklmnopqrstuvwxyz123456789':
706
fid = fname.encode('utf-8') + b'-id'
707
content = b'content for %s\n' % (fname.encode('utf-8'),)
827
content = 'content for %s\n' % (fname,)
708
828
entries.append(('add', (fname, fid, 'file', content)))
709
829
source_builder.start_series()
710
source_builder.build_snapshot(None, entries, revision_id=b'rev-1')
830
source_builder.build_snapshot('rev-1', None, entries)
711
831
# Now change a few of them, so we get a few new pages for the second
713
source_builder.build_snapshot([b'rev-1'], [
714
('modify', ('aa', b'new content for aa-id\n')),
715
('modify', ('cc', b'new content for cc-id\n')),
716
('modify', ('zz', b'new content for zz-id\n')),
717
], revision_id=b'rev-2')
833
source_builder.build_snapshot('rev-2', ['rev-1'], [
834
('modify', ('aa-id', 'new content for aa-id\n')),
835
('modify', ('cc-id', 'new content for cc-id\n')),
836
('modify', ('zz-id', 'new content for zz-id\n')),
718
838
source_builder.finish_series()
719
839
source_branch = source_builder.get_branch()
720
840
source_branch.lock_read()
726
846
# On a regular pass, getting the inventories and chk pages for rev-2
727
847
# would only get the newly created chk pages
728
search = vf_search.SearchResult({b'rev-2'}, {b'rev-1'}, 1,
730
simple_chk_records = set()
848
search = graph.SearchResult(set(['rev-2']), set(['rev-1']), 1,
850
simple_chk_records = []
731
851
for vf_name, substream in source.get_stream(search):
732
852
if vf_name == 'chk_bytes':
733
853
for record in substream:
734
simple_chk_records.add(record.key)
854
simple_chk_records.append(record.key)
736
856
for _ in substream:
738
858
# 3 pages, the root (InternalNode), + 2 pages which actually changed
739
self.assertEqual({(b'sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
740
(b'sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
741
(b'sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
742
(b'sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)},
743
set(simple_chk_records))
859
self.assertEqual([('sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
860
('sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
861
('sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
862
('sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)],
744
864
# Now, when we do a similar call using 'get_stream_for_missing_keys'
745
865
# we should get a much larger set of pages.
746
missing = [('inventories', b'rev-2')]
747
full_chk_records = set()
866
missing = [('inventories', 'rev-2')]
867
full_chk_records = []
748
868
for vf_name, substream in source.get_stream_for_missing_keys(missing):
749
869
if vf_name == 'inventories':
750
870
for record in substream:
751
self.assertEqual((b'rev-2',), record.key)
871
self.assertEqual(('rev-2',), record.key)
752
872
elif vf_name == 'chk_bytes':
753
873
for record in substream:
754
full_chk_records.add(record.key)
874
full_chk_records.append(record.key)
756
876
self.fail('Should not be getting a stream of %s' % (vf_name,))
757
877
# We have 257 records now. This is because we have 1 root page, and 256
843
958
super(TestDevelopment6FindParentIdsOfRevisions, self).setUp()
844
self.builder = self.make_branch_builder('source')
959
self.builder = self.make_branch_builder('source',
960
format='development6-rich-root')
845
961
self.builder.start_series()
846
self.builder.build_snapshot(
848
[('add', ('', b'tree-root', 'directory', None))],
849
revision_id=b'initial')
962
self.builder.build_snapshot('initial', None,
963
[('add', ('', 'tree-root', 'directory', None))])
850
964
self.repo = self.builder.get_branch().repository
851
965
self.addCleanup(self.builder.finish_series)
853
967
def assertParentIds(self, expected_result, rev_set):
855
sorted(expected_result),
968
self.assertEqual(sorted(expected_result),
856
969
sorted(self.repo._find_parent_ids_of_revisions(rev_set)))
858
971
def test_simple(self):
859
self.builder.build_snapshot(None, [], revision_id=b'revid1')
860
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
861
rev_set = [b'revid2']
862
self.assertParentIds([b'revid1'], rev_set)
972
self.builder.build_snapshot('revid1', None, [])
973
self.builder.build_snapshot('revid2', ['revid1'], [])
975
self.assertParentIds(['revid1'], rev_set)
864
977
def test_not_first_parent(self):
865
self.builder.build_snapshot(None, [], revision_id=b'revid1')
866
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
867
self.builder.build_snapshot([b'revid2'], [], revision_id=b'revid3')
868
rev_set = [b'revid3', b'revid2']
869
self.assertParentIds([b'revid1'], rev_set)
978
self.builder.build_snapshot('revid1', None, [])
979
self.builder.build_snapshot('revid2', ['revid1'], [])
980
self.builder.build_snapshot('revid3', ['revid2'], [])
981
rev_set = ['revid3', 'revid2']
982
self.assertParentIds(['revid1'], rev_set)
871
984
def test_not_null(self):
872
rev_set = [b'initial']
985
rev_set = ['initial']
873
986
self.assertParentIds([], rev_set)
875
988
def test_not_null_set(self):
876
self.builder.build_snapshot(None, [], revision_id=b'revid1')
989
self.builder.build_snapshot('revid1', None, [])
877
990
rev_set = [_mod_revision.NULL_REVISION]
878
991
self.assertParentIds([], rev_set)
880
993
def test_ghost(self):
881
self.builder.build_snapshot(None, [], revision_id=b'revid1')
882
rev_set = [b'ghost', b'revid1']
883
self.assertParentIds([b'initial'], rev_set)
994
self.builder.build_snapshot('revid1', None, [])
995
rev_set = ['ghost', 'revid1']
996
self.assertParentIds(['initial'], rev_set)
885
998
def test_ghost_parent(self):
886
self.builder.build_snapshot(None, [], revision_id=b'revid1')
887
self.builder.build_snapshot(
888
[b'revid1', b'ghost'], [], revision_id=b'revid2')
889
rev_set = [b'revid2', b'revid1']
890
self.assertParentIds([b'ghost', b'initial'], rev_set)
999
self.builder.build_snapshot('revid1', None, [])
1000
self.builder.build_snapshot('revid2', ['revid1', 'ghost'], [])
1001
rev_set = ['revid2', 'revid1']
1002
self.assertParentIds(['ghost', 'initial'], rev_set)
892
1004
def test_righthand_parent(self):
893
self.builder.build_snapshot(None, [], revision_id=b'revid1')
894
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2a')
895
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2b')
896
self.builder.build_snapshot([b'revid2a', b'revid2b'], [],
897
revision_id=b'revid3')
898
rev_set = [b'revid3', b'revid2a']
899
self.assertParentIds([b'revid1', b'revid2b'], rev_set)
1005
self.builder.build_snapshot('revid1', None, [])
1006
self.builder.build_snapshot('revid2a', ['revid1'], [])
1007
self.builder.build_snapshot('revid2b', ['revid1'], [])
1008
self.builder.build_snapshot('revid3', ['revid2a', 'revid2b'], [])
1009
rev_set = ['revid3', 'revid2a']
1010
self.assertParentIds(['revid1', 'revid2b'], rev_set)
902
1013
class TestWithBrokenRepo(TestCaseWithTransport):
914
1025
repo.start_write_group()
915
1026
cleanups.append(repo.commit_write_group)
916
1027
# make rev1a: A well-formed revision, containing 'file1'
917
inv = inventory.Inventory(revision_id=b'rev1a')
918
inv.root.revision = b'rev1a'
919
self.add_file(repo, inv, 'file1', b'rev1a', [])
920
repo.texts.add_lines((inv.root.file_id, b'rev1a'), [], [])
921
repo.add_inventory(b'rev1a', inv, [])
922
revision = _mod_revision.Revision(
1028
inv = inventory.Inventory(revision_id='rev1a')
1029
inv.root.revision = 'rev1a'
1030
self.add_file(repo, inv, 'file1', 'rev1a', [])
1031
repo.texts.add_lines((inv.root.file_id, 'rev1a'), [], [])
1032
repo.add_inventory('rev1a', inv, [])
1033
revision = _mod_revision.Revision('rev1a',
924
1034
committer='jrandom@example.com', timestamp=0,
925
1035
inventory_sha1='', timezone=0, message='foo', parent_ids=[])
926
repo.add_revision(b'rev1a', revision, inv)
1036
repo.add_revision('rev1a',revision, inv)
928
1038
# make rev1b, which has no Revision, but has an Inventory, and
930
inv = inventory.Inventory(revision_id=b'rev1b')
931
inv.root.revision = b'rev1b'
932
self.add_file(repo, inv, 'file1', b'rev1b', [])
933
repo.add_inventory(b'rev1b', inv, [])
1040
inv = inventory.Inventory(revision_id='rev1b')
1041
inv.root.revision = 'rev1b'
1042
self.add_file(repo, inv, 'file1', 'rev1b', [])
1043
repo.add_inventory('rev1b', inv, [])
935
1045
# make rev2, with file1 and file2
937
1047
# file1 has 'rev1b' as an ancestor, even though this is not
938
1048
# mentioned by 'rev1a', making it an unreferenced ancestor
939
1049
inv = inventory.Inventory()
940
self.add_file(repo, inv, 'file1', b'rev2', [b'rev1a', b'rev1b'])
941
self.add_file(repo, inv, 'file2', b'rev2', [])
942
self.add_revision(repo, b'rev2', inv, [b'rev1a'])
1050
self.add_file(repo, inv, 'file1', 'rev2', ['rev1a', 'rev1b'])
1051
self.add_file(repo, inv, 'file2', 'rev2', [])
1052
self.add_revision(repo, 'rev2', inv, ['rev1a'])
944
1054
# make ghost revision rev1c
945
1055
inv = inventory.Inventory()
946
self.add_file(repo, inv, 'file2', b'rev1c', [])
1056
self.add_file(repo, inv, 'file2', 'rev1c', [])
948
1058
# make rev3 with file2
949
1059
# file2 refers to 'rev1c', which is a ghost in this repository, so
950
1060
# file2 cannot have rev1c as its ancestor.
951
1061
inv = inventory.Inventory()
952
self.add_file(repo, inv, 'file2', b'rev3', [b'rev1c'])
953
self.add_revision(repo, b'rev3', inv, [b'rev1c'])
1062
self.add_file(repo, inv, 'file2', 'rev3', ['rev1c'])
1063
self.add_revision(repo, 'rev3', inv, ['rev1c'])
956
1066
for cleanup in reversed(cleanups):
1545
1622
self.assertTrue(new_pack.signature_index._optimize_for_size)
1548
class TestGCCHKPacker(TestCaseWithTransport):
1550
def make_abc_branch(self):
1551
builder = self.make_branch_builder('source')
1552
builder.start_series()
1553
builder.build_snapshot(None, [
1554
('add', ('', b'root-id', 'directory', None)),
1555
('add', ('file', b'file-id', 'file', b'content\n')),
1556
], revision_id=b'A')
1557
builder.build_snapshot([b'A'], [
1558
('add', ('dir', b'dir-id', 'directory', None))],
1560
builder.build_snapshot([b'B'], [
1561
('modify', ('file', b'new content\n'))],
1563
builder.finish_series()
1564
return builder.get_branch()
1566
def make_branch_with_disjoint_inventory_and_revision(self):
1567
"""a repo with separate packs for a revisions Revision and Inventory.
1569
There will be one pack file that holds the Revision content, and one
1570
for the Inventory content.
1572
:return: (repository,
1573
pack_name_with_rev_A_Revision,
1574
pack_name_with_rev_A_Inventory,
1575
pack_name_with_rev_C_content)
1577
b_source = self.make_abc_branch()
1578
b_base = b_source.controldir.sprout(
1579
'base', revision_id=b'A').open_branch()
1580
b_stacked = b_base.controldir.sprout(
1581
'stacked', stacked=True).open_branch()
1582
b_stacked.lock_write()
1583
self.addCleanup(b_stacked.unlock)
1584
b_stacked.fetch(b_source, b'B')
1585
# Now re-open the stacked repo directly (no fallbacks) so that we can
1586
# fill in the A rev.
1587
repo_not_stacked = b_stacked.controldir.open_repository()
1588
repo_not_stacked.lock_write()
1589
self.addCleanup(repo_not_stacked.unlock)
1590
# Now we should have a pack file with A's inventory, but not its
1592
self.assertEqual([(b'A',), (b'B',)],
1593
sorted(repo_not_stacked.inventories.keys()))
1594
self.assertEqual([(b'B',)],
1595
sorted(repo_not_stacked.revisions.keys()))
1596
stacked_pack_names = repo_not_stacked._pack_collection.names()
1597
# We have a couple names here, figure out which has A's inventory
1598
for name in stacked_pack_names:
1599
pack = repo_not_stacked._pack_collection.get_pack_by_name(name)
1600
keys = [n[1] for n in pack.inventory_index.iter_all_entries()]
1602
inv_a_pack_name = name
1605
self.fail('Could not find pack containing A\'s inventory')
1606
repo_not_stacked.fetch(b_source.repository, b'A')
1607
self.assertEqual([(b'A',), (b'B',)],
1608
sorted(repo_not_stacked.revisions.keys()))
1609
new_pack_names = set(repo_not_stacked._pack_collection.names())
1610
rev_a_pack_names = new_pack_names.difference(stacked_pack_names)
1611
self.assertEqual(1, len(rev_a_pack_names))
1612
rev_a_pack_name = list(rev_a_pack_names)[0]
1613
# Now fetch 'C', so we have a couple pack files to join
1614
repo_not_stacked.fetch(b_source.repository, b'C')
1615
rev_c_pack_names = set(repo_not_stacked._pack_collection.names())
1616
rev_c_pack_names = rev_c_pack_names.difference(new_pack_names)
1617
self.assertEqual(1, len(rev_c_pack_names))
1618
rev_c_pack_name = list(rev_c_pack_names)[0]
1619
return (repo_not_stacked, rev_a_pack_name, inv_a_pack_name,
1622
def test_pack_with_distant_inventories(self):
1623
# See https://bugs.launchpad.net/bzr/+bug/437003
1624
# When repacking, it is possible to have an inventory in a different
1625
# pack file than the associated revision. An autopack can then come
1626
# along, and miss that inventory, and complain.
1627
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1628
) = self.make_branch_with_disjoint_inventory_and_revision()
1629
a_pack = repo._pack_collection.get_pack_by_name(rev_a_pack_name)
1630
c_pack = repo._pack_collection.get_pack_by_name(rev_c_pack_name)
1631
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1632
[a_pack, c_pack], '.test-pack')
1633
# This would raise ValueError in bug #437003, but should not raise an
1637
def test_pack_with_missing_inventory(self):
1638
# Similar to test_pack_with_missing_inventory, but this time, we force
1639
# the A inventory to actually be gone from the repository.
1640
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1641
) = self.make_branch_with_disjoint_inventory_and_revision()
1642
inv_a_pack = repo._pack_collection.get_pack_by_name(inv_a_pack_name)
1643
repo._pack_collection._remove_pack_from_memory(inv_a_pack)
1644
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1645
repo._pack_collection.all_packs(), '.test-pack')
1646
e = self.assertRaises(ValueError, packer.pack)
1647
packer.new_pack.abort()
1648
self.assertContainsRe(str(e),
1649
r"We are missing inventories for revisions: .*'A'")
1652
1625
class TestCrossFormatPacks(TestCaseWithTransport):
1654
1627
def log_pack(self, hint=None):