139
124
# create a branch with a few known format objects.
140
125
# this is not quite the same as
141
126
self.build_tree(["foo/", "bar/"])
143
127
def check_format(format, url):
144
dir = format._matchingcontroldir.initialize(url)
128
dir = format._matchingbzrdir.initialize(url)
145
129
format.initialize(dir)
146
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
148
self.assertIsInstance(found_format, format.__class__)
149
check_format(repository.format_registry.get_default(), "bar")
130
t = get_transport(url)
131
found_format = repository.RepositoryFormat.find_format(dir)
132
self.failUnless(isinstance(found_format, format.__class__))
133
check_format(weaverepo.RepositoryFormat7(), "bar")
151
135
def test_find_format_no_repository(self):
152
136
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
153
137
self.assertRaises(errors.NoRepositoryPresent,
154
bzrrepository.RepositoryFormatMetaDir.find_format,
138
repository.RepositoryFormat.find_format,
157
def test_from_string(self):
158
self.assertIsInstance(
159
SampleRepositoryFormat.from_string(
160
b"Sample .bzr repository format."),
161
SampleRepositoryFormat)
162
self.assertRaises(AssertionError,
163
SampleRepositoryFormat.from_string,
164
b"Different .bzr repository format.")
166
141
def test_find_format_unknown_format(self):
167
142
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
168
143
SampleRepositoryFormat().initialize(dir)
169
144
self.assertRaises(UnknownFormatError,
170
bzrrepository.RepositoryFormatMetaDir.find_format,
145
repository.RepositoryFormat.find_format,
173
def test_find_format_with_features(self):
174
tree = self.make_branch_and_tree('.', format='2a')
175
tree.branch.repository.update_feature_flags({b"name": b"necessity"})
176
found_format = bzrrepository.RepositoryFormatMetaDir.find_format(
178
self.assertIsInstance(
179
found_format, bzrrepository.RepositoryFormatMetaDir)
180
self.assertEqual(found_format.features.get(b"name"), b"necessity")
182
bzrdir.MissingFeature, found_format.check_support_status, True)
184
bzrrepository.RepositoryFormatMetaDir.unregister_feature, b"name")
185
bzrrepository.RepositoryFormatMetaDir.register_feature(b"name")
186
found_format.check_support_status(True)
189
class TestRepositoryFormatRegistry(TestCase):
192
super(TestRepositoryFormatRegistry, self).setUp()
193
self.registry = repository.RepositoryFormatRegistry()
195
148
def test_register_unregister_format(self):
196
149
format = SampleRepositoryFormat()
197
self.registry.register(format)
198
self.assertEqual(format, self.registry.get(
199
b"Sample .bzr repository format."))
200
self.registry.remove(format)
201
self.assertRaises(KeyError, self.registry.get,
202
b"Sample .bzr repository format.")
204
def test_get_all(self):
205
format = SampleRepositoryFormat()
206
self.assertEqual([], self.registry._get_all())
207
self.registry.register(format)
208
self.assertEqual([format], self.registry._get_all())
210
def test_register_extra(self):
211
format = SampleExtraRepositoryFormat()
212
self.assertEqual([], self.registry._get_all())
213
self.registry.register_extra(format)
214
self.assertEqual([format], self.registry._get_all())
216
def test_register_extra_lazy(self):
217
self.assertEqual([], self.registry._get_all())
218
self.registry.register_extra_lazy(__name__,
219
"SampleExtraRepositoryFormat")
220
formats = self.registry._get_all()
221
self.assertEqual(1, len(formats))
222
self.assertIsInstance(formats[0], SampleExtraRepositoryFormat)
151
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
153
format.initialize(dir)
154
# register a format for it.
155
repository.RepositoryFormat.register_format(format)
156
# which repository.Open will refuse (not supported)
157
self.assertRaises(UnsupportedFormatError, repository.Repository.open, self.get_url())
158
# but open(unsupported) will work
159
self.assertEqual(format.open(dir), "opened repository.")
160
# unregister the format
161
repository.RepositoryFormat.unregister_format(format)
164
class TestFormat6(TestCaseWithTransport):
166
def test_attribute__fetch_order(self):
167
"""Weaves need topological data insertion."""
168
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
169
repo = weaverepo.RepositoryFormat6().initialize(control)
170
self.assertEqual('topological', repo._format._fetch_order)
172
def test_attribute__fetch_uses_deltas(self):
173
"""Weaves do not reuse deltas."""
174
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
175
repo = weaverepo.RepositoryFormat6().initialize(control)
176
self.assertEqual(False, repo._format._fetch_uses_deltas)
178
def test_attribute__fetch_reconcile(self):
179
"""Weave repositories need a reconcile after fetch."""
180
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
181
repo = weaverepo.RepositoryFormat6().initialize(control)
182
self.assertEqual(True, repo._format._fetch_reconcile)
184
def test_no_ancestry_weave(self):
185
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
186
repo = weaverepo.RepositoryFormat6().initialize(control)
187
# We no longer need to create the ancestry.weave file
188
# since it is *never* used.
189
self.assertRaises(NoSuchFile,
190
control.transport.get,
193
def test_supports_external_lookups(self):
194
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
195
repo = weaverepo.RepositoryFormat6().initialize(control)
196
self.assertFalse(repo._format.supports_external_lookups)
199
class TestFormat7(TestCaseWithTransport):
201
def test_attribute__fetch_order(self):
202
"""Weaves need topological data insertion."""
203
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
204
repo = weaverepo.RepositoryFormat7().initialize(control)
205
self.assertEqual('topological', repo._format._fetch_order)
207
def test_attribute__fetch_uses_deltas(self):
208
"""Weaves do not reuse deltas."""
209
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
210
repo = weaverepo.RepositoryFormat7().initialize(control)
211
self.assertEqual(False, repo._format._fetch_uses_deltas)
213
def test_attribute__fetch_reconcile(self):
214
"""Weave repositories need a reconcile after fetch."""
215
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
216
repo = weaverepo.RepositoryFormat7().initialize(control)
217
self.assertEqual(True, repo._format._fetch_reconcile)
219
def test_disk_layout(self):
220
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
221
repo = weaverepo.RepositoryFormat7().initialize(control)
222
# in case of side effects of locking.
226
# format 'Bazaar-NG Repository format 7'
228
# inventory.weave == empty_weave
229
# empty revision-store directory
230
# empty weaves directory
231
t = control.get_repository_transport(None)
232
self.assertEqualDiff('Bazaar-NG Repository format 7',
233
t.get('format').read())
234
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
235
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
236
self.assertEqualDiff('# bzr weave file v5\n'
239
t.get('inventory.weave').read())
240
# Creating a file with id Foo:Bar results in a non-escaped file name on
242
control.create_branch()
243
tree = control.create_workingtree()
244
tree.add(['foo'], ['Foo:Bar'], ['file'])
245
tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
247
tree.commit('first post', rev_id='first')
248
except errors.IllegalPath:
249
if sys.platform != 'win32':
251
self.knownFailure('Foo:Bar cannot be used as a file-id on windows'
254
self.assertEqualDiff(
255
'# bzr weave file v5\n'
257
'1 7fe70820e08a1aac0ef224d9c66ab66831cc4ab1\n'
265
t.get('weaves/74/Foo%3ABar.weave').read())
267
def test_shared_disk_layout(self):
268
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
269
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
271
# format 'Bazaar-NG Repository format 7'
272
# inventory.weave == empty_weave
273
# empty revision-store directory
274
# empty weaves directory
275
# a 'shared-storage' marker file.
276
# lock is not present when unlocked
277
t = control.get_repository_transport(None)
278
self.assertEqualDiff('Bazaar-NG Repository format 7',
279
t.get('format').read())
280
self.assertEqualDiff('', t.get('shared-storage').read())
281
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
282
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
283
self.assertEqualDiff('# bzr weave file v5\n'
286
t.get('inventory.weave').read())
287
self.assertFalse(t.has('branch-lock'))
289
def test_creates_lockdir(self):
290
"""Make sure it appears to be controlled by a LockDir existence"""
291
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
292
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
293
t = control.get_repository_transport(None)
294
# TODO: Should check there is a 'lock' toplevel directory,
295
# regardless of contents
296
self.assertFalse(t.has('lock/held/info'))
299
self.assertTrue(t.has('lock/held/info'))
301
# unlock so we don't get a warning about failing to do so
304
def test_uses_lockdir(self):
305
"""repo format 7 actually locks on lockdir"""
306
base_url = self.get_url()
307
control = bzrdir.BzrDirMetaFormat1().initialize(base_url)
308
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
309
t = control.get_repository_transport(None)
313
# make sure the same lock is created by opening it
314
repo = repository.Repository.open(base_url)
316
self.assertTrue(t.has('lock/held/info'))
318
self.assertFalse(t.has('lock/held/info'))
320
def test_shared_no_tree_disk_layout(self):
321
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
322
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
323
repo.set_make_working_trees(False)
325
# format 'Bazaar-NG Repository format 7'
327
# inventory.weave == empty_weave
328
# empty revision-store directory
329
# empty weaves directory
330
# a 'shared-storage' marker file.
331
t = control.get_repository_transport(None)
332
self.assertEqualDiff('Bazaar-NG Repository format 7',
333
t.get('format').read())
334
## self.assertEqualDiff('', t.get('lock').read())
335
self.assertEqualDiff('', t.get('shared-storage').read())
336
self.assertEqualDiff('', t.get('no-working-trees').read())
337
repo.set_make_working_trees(True)
338
self.assertFalse(t.has('no-working-trees'))
339
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
340
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
341
self.assertEqualDiff('# bzr weave file v5\n'
344
t.get('inventory.weave').read())
346
def test_supports_external_lookups(self):
347
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
348
repo = weaverepo.RepositoryFormat7().initialize(control)
349
self.assertFalse(repo._format.supports_external_lookups)
225
352
class TestFormatKnit1(TestCaseWithTransport):
227
354
def test_attribute__fetch_order(self):
228
355
"""Knits need topological data insertion."""
229
repo = self.make_repository(
230
'.', format=controldir.format_registry.get('knit')())
356
repo = self.make_repository('.',
357
format=bzrdir.format_registry.get('knit')())
231
358
self.assertEqual('topological', repo._format._fetch_order)
233
360
def test_attribute__fetch_uses_deltas(self):
234
361
"""Knits reuse deltas."""
235
repo = self.make_repository(
236
'.', format=controldir.format_registry.get('knit')())
362
repo = self.make_repository('.',
363
format=bzrdir.format_registry.get('knit')())
237
364
self.assertEqual(True, repo._format._fetch_uses_deltas)
239
366
def test_disk_layout(self):
458
571
self.assertGetsDefaultInterRepository(dummy_a, dummy_b)
461
class TestRepositoryFormat1(knitrepo.RepositoryFormatKnit1):
464
def get_format_string(cls):
465
return b"Test Format 1"
468
class TestRepositoryFormat2(knitrepo.RepositoryFormatKnit1):
471
def get_format_string(cls):
472
return b"Test Format 2"
574
class TestInterWeaveRepo(TestCaseWithTransport):
576
def test_is_compatible_and_registered(self):
577
# InterWeaveRepo is compatible when either side
578
# is a format 5/6/7 branch
579
from bzrlib.repofmt import knitrepo, weaverepo
580
formats = [weaverepo.RepositoryFormat5(),
581
weaverepo.RepositoryFormat6(),
582
weaverepo.RepositoryFormat7()]
583
incompatible_formats = [weaverepo.RepositoryFormat4(),
584
knitrepo.RepositoryFormatKnit1(),
586
repo_a = self.make_repository('a')
587
repo_b = self.make_repository('b')
588
is_compatible = repository.InterWeaveRepo.is_compatible
589
for source in incompatible_formats:
590
# force incompatible left then right
591
repo_a._format = source
592
repo_b._format = formats[0]
593
self.assertFalse(is_compatible(repo_a, repo_b))
594
self.assertFalse(is_compatible(repo_b, repo_a))
595
for source in formats:
596
repo_a._format = source
597
for target in formats:
598
repo_b._format = target
599
self.assertTrue(is_compatible(repo_a, repo_b))
600
self.assertEqual(repository.InterWeaveRepo,
601
repository.InterRepository.get(repo_a,
475
605
class TestRepositoryConverter(TestCaseWithTransport):
477
607
def test_convert_empty(self):
478
source_format = TestRepositoryFormat1()
479
target_format = TestRepositoryFormat2()
480
repository.format_registry.register(source_format)
481
self.addCleanup(repository.format_registry.remove,
483
repository.format_registry.register(target_format)
484
self.addCleanup(repository.format_registry.remove,
486
t = self.get_transport()
608
t = get_transport(self.get_url('.'))
487
609
t.mkdir('repository')
488
610
repo_dir = bzrdir.BzrDirMetaFormat1().initialize('repository')
489
repo = TestRepositoryFormat1().initialize(repo_dir)
611
repo = weaverepo.RepositoryFormat7().initialize(repo_dir)
612
target_format = knitrepo.RepositoryFormatKnit1()
490
613
converter = repository.CopyConverter(target_format)
491
with breezy.ui.ui_factory.nested_progress_bar() as pb:
614
pb = bzrlib.ui.ui_factory.nested_progress_bar()
492
616
converter.convert(repo, pb)
493
619
repo = repo_dir.open_repository()
494
620
self.assertTrue(isinstance(target_format, repo._format.__class__))
623
class TestMisc(TestCase):
625
def test_unescape_xml(self):
626
"""We get some kind of error when malformed entities are passed"""
627
self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;')
497
630
class TestRepositoryFormatKnit3(TestCaseWithTransport):
499
632
def test_attribute__fetch_order(self):
543
682
class Test2a(tests.TestCaseWithMemoryTransport):
545
def test_chk_bytes_uses_custom_btree_parser(self):
546
mt = self.make_branch_and_memory_tree('test', format='2a')
548
self.addCleanup(mt.unlock)
549
mt.add([''], [b'root-id'])
551
index = mt.branch.repository.chk_bytes._index._graph_index._indices[0]
552
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
553
# It should also work if we re-open the repo
554
repo = mt.branch.repository.controldir.open_repository()
556
self.addCleanup(repo.unlock)
557
index = repo.chk_bytes._index._graph_index._indices[0]
558
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
560
def test_fetch_combines_groups(self):
561
builder = self.make_branch_builder('source', format='2a')
562
builder.start_series()
563
builder.build_snapshot(None, [
564
('add', ('', b'root-id', 'directory', '')),
565
('add', ('file', b'file-id', 'file', b'content\n'))],
567
builder.build_snapshot([b'1'], [
568
('modify', ('file', b'content-2\n'))],
570
builder.finish_series()
571
source = builder.get_branch()
572
target = self.make_repository('target', format='2a')
573
target.fetch(source.repository)
575
self.addCleanup(target.unlock)
576
details = target.texts._index.get_build_details(
577
[(b'file-id', b'1',), (b'file-id', b'2',)])
578
file_1_details = details[(b'file-id', b'1')]
579
file_2_details = details[(b'file-id', b'2')]
580
# The index, and what to read off disk, should be the same for both
581
# versions of the file.
582
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
584
def test_fetch_combines_groups(self):
585
builder = self.make_branch_builder('source', format='2a')
586
builder.start_series()
587
builder.build_snapshot(None, [
588
('add', ('', b'root-id', 'directory', '')),
589
('add', ('file', b'file-id', 'file', b'content\n'))],
591
builder.build_snapshot([b'1'], [
592
('modify', ('file', b'content-2\n'))],
594
builder.finish_series()
595
source = builder.get_branch()
596
target = self.make_repository('target', format='2a')
597
target.fetch(source.repository)
599
self.addCleanup(target.unlock)
600
details = target.texts._index.get_build_details(
601
[(b'file-id', b'1',), (b'file-id', b'2',)])
602
file_1_details = details[(b'file-id', b'1')]
603
file_2_details = details[(b'file-id', b'2')]
604
# The index, and what to read off disk, should be the same for both
605
# versions of the file.
606
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
608
def test_fetch_combines_groups(self):
609
builder = self.make_branch_builder('source', format='2a')
610
builder.start_series()
611
builder.build_snapshot(None, [
612
('add', ('', b'root-id', 'directory', '')),
613
('add', ('file', b'file-id', 'file', b'content\n'))],
615
builder.build_snapshot([b'1'], [
616
('modify', ('file', b'content-2\n'))],
618
builder.finish_series()
619
source = builder.get_branch()
620
target = self.make_repository('target', format='2a')
621
target.fetch(source.repository)
623
self.addCleanup(target.unlock)
624
details = target.texts._index.get_build_details(
625
[(b'file-id', b'1',), (b'file-id', b'2',)])
626
file_1_details = details[(b'file-id', b'1')]
627
file_2_details = details[(b'file-id', b'2')]
684
def test_fetch_combines_groups(self):
685
builder = self.make_branch_builder('source', format='2a')
686
builder.start_series()
687
builder.build_snapshot('1', None, [
688
('add', ('', 'root-id', 'directory', '')),
689
('add', ('file', 'file-id', 'file', 'content\n'))])
690
builder.build_snapshot('2', ['1'], [
691
('modify', ('file-id', 'content-2\n'))])
692
builder.finish_series()
693
source = builder.get_branch()
694
target = self.make_repository('target', format='2a')
695
target.fetch(source.repository)
697
self.addCleanup(target.unlock)
698
details = target.texts._index.get_build_details(
699
[('file-id', '1',), ('file-id', '2',)])
700
file_1_details = details[('file-id', '1')]
701
file_2_details = details[('file-id', '2')]
702
# The index, and what to read off disk, should be the same for both
703
# versions of the file.
704
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
706
def test_fetch_combines_groups(self):
707
builder = self.make_branch_builder('source', format='2a')
708
builder.start_series()
709
builder.build_snapshot('1', None, [
710
('add', ('', 'root-id', 'directory', '')),
711
('add', ('file', 'file-id', 'file', 'content\n'))])
712
builder.build_snapshot('2', ['1'], [
713
('modify', ('file-id', 'content-2\n'))])
714
builder.finish_series()
715
source = builder.get_branch()
716
target = self.make_repository('target', format='2a')
717
target.fetch(source.repository)
719
self.addCleanup(target.unlock)
720
details = target.texts._index.get_build_details(
721
[('file-id', '1',), ('file-id', '2',)])
722
file_1_details = details[('file-id', '1')]
723
file_2_details = details[('file-id', '2')]
724
# The index, and what to read off disk, should be the same for both
725
# versions of the file.
726
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
728
def test_fetch_combines_groups(self):
729
builder = self.make_branch_builder('source', format='2a')
730
builder.start_series()
731
builder.build_snapshot('1', None, [
732
('add', ('', 'root-id', 'directory', '')),
733
('add', ('file', 'file-id', 'file', 'content\n'))])
734
builder.build_snapshot('2', ['1'], [
735
('modify', ('file-id', 'content-2\n'))])
736
builder.finish_series()
737
source = builder.get_branch()
738
target = self.make_repository('target', format='2a')
739
target.fetch(source.repository)
741
self.addCleanup(target.unlock)
742
details = target.texts._index.get_build_details(
743
[('file-id', '1',), ('file-id', '2',)])
744
file_1_details = details[('file-id', '1')]
745
file_2_details = details[('file-id', '2')]
628
746
# The index, and what to read off disk, should be the same for both
629
747
# versions of the file.
630
748
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
694
812
target = self.make_repository('target', format='rich-root-pack')
695
813
stream = source._get_source(target._format)
696
814
# We don't want the child GroupCHKStreamSource
697
self.assertIs(type(stream), vf_repository.StreamSource)
815
self.assertIs(type(stream), repository.StreamSource)
699
817
def test_get_stream_for_missing_keys_includes_all_chk_refs(self):
700
818
source_builder = self.make_branch_builder('source',
702
820
# We have to build a fairly large tree, so that we are sure the chk
703
821
# pages will have split into multiple pages.
704
entries = [('add', ('', b'a-root-id', 'directory', None))]
822
entries = [('add', ('', 'a-root-id', 'directory', None))]
705
823
for i in 'abcdefghijklmnopqrstuvwxyz123456789':
706
824
for j in 'abcdefghijklmnopqrstuvwxyz123456789':
708
fid = fname.encode('utf-8') + b'-id'
709
content = b'content for %s\n' % (fname.encode('utf-8'),)
827
content = 'content for %s\n' % (fname,)
710
828
entries.append(('add', (fname, fid, 'file', content)))
711
829
source_builder.start_series()
712
source_builder.build_snapshot(None, entries, revision_id=b'rev-1')
830
source_builder.build_snapshot('rev-1', None, entries)
713
831
# Now change a few of them, so we get a few new pages for the second
715
source_builder.build_snapshot([b'rev-1'], [
716
('modify', ('aa', b'new content for aa-id\n')),
717
('modify', ('cc', b'new content for cc-id\n')),
718
('modify', ('zz', b'new content for zz-id\n')),
719
], revision_id=b'rev-2')
833
source_builder.build_snapshot('rev-2', ['rev-1'], [
834
('modify', ('aa-id', 'new content for aa-id\n')),
835
('modify', ('cc-id', 'new content for cc-id\n')),
836
('modify', ('zz-id', 'new content for zz-id\n')),
720
838
source_builder.finish_series()
721
839
source_branch = source_builder.get_branch()
722
840
source_branch.lock_read()
728
846
# On a regular pass, getting the inventories and chk pages for rev-2
729
847
# would only get the newly created chk pages
730
search = vf_search.SearchResult({b'rev-2'}, {b'rev-1'}, 1,
732
simple_chk_records = set()
848
search = graph.SearchResult(set(['rev-2']), set(['rev-1']), 1,
850
simple_chk_records = []
733
851
for vf_name, substream in source.get_stream(search):
734
852
if vf_name == 'chk_bytes':
735
853
for record in substream:
736
simple_chk_records.add(record.key)
854
simple_chk_records.append(record.key)
738
856
for _ in substream:
740
858
# 3 pages, the root (InternalNode), + 2 pages which actually changed
741
self.assertEqual({(b'sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
742
(b'sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
743
(b'sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
744
(b'sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)},
745
set(simple_chk_records))
859
self.assertEqual([('sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
860
('sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
861
('sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
862
('sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)],
746
864
# Now, when we do a similar call using 'get_stream_for_missing_keys'
747
865
# we should get a much larger set of pages.
748
missing = [('inventories', b'rev-2')]
749
full_chk_records = set()
866
missing = [('inventories', 'rev-2')]
867
full_chk_records = []
750
868
for vf_name, substream in source.get_stream_for_missing_keys(missing):
751
869
if vf_name == 'inventories':
752
870
for record in substream:
753
self.assertEqual((b'rev-2',), record.key)
871
self.assertEqual(('rev-2',), record.key)
754
872
elif vf_name == 'chk_bytes':
755
873
for record in substream:
756
full_chk_records.add(record.key)
874
full_chk_records.append(record.key)
758
876
self.fail('Should not be getting a stream of %s' % (vf_name,))
759
877
# We have 257 records now. This is because we have 1 root page, and 256
845
958
super(TestDevelopment6FindParentIdsOfRevisions, self).setUp()
846
self.builder = self.make_branch_builder('source')
959
self.builder = self.make_branch_builder('source',
960
format='development6-rich-root')
847
961
self.builder.start_series()
848
self.builder.build_snapshot(
850
[('add', ('', b'tree-root', 'directory', None))],
851
revision_id=b'initial')
962
self.builder.build_snapshot('initial', None,
963
[('add', ('', 'tree-root', 'directory', None))])
852
964
self.repo = self.builder.get_branch().repository
853
965
self.addCleanup(self.builder.finish_series)
855
967
def assertParentIds(self, expected_result, rev_set):
857
sorted(expected_result),
968
self.assertEqual(sorted(expected_result),
858
969
sorted(self.repo._find_parent_ids_of_revisions(rev_set)))
860
971
def test_simple(self):
861
self.builder.build_snapshot(None, [], revision_id=b'revid1')
862
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
863
rev_set = [b'revid2']
864
self.assertParentIds([b'revid1'], rev_set)
972
self.builder.build_snapshot('revid1', None, [])
973
self.builder.build_snapshot('revid2', ['revid1'], [])
975
self.assertParentIds(['revid1'], rev_set)
866
977
def test_not_first_parent(self):
867
self.builder.build_snapshot(None, [], revision_id=b'revid1')
868
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2')
869
self.builder.build_snapshot([b'revid2'], [], revision_id=b'revid3')
870
rev_set = [b'revid3', b'revid2']
871
self.assertParentIds([b'revid1'], rev_set)
978
self.builder.build_snapshot('revid1', None, [])
979
self.builder.build_snapshot('revid2', ['revid1'], [])
980
self.builder.build_snapshot('revid3', ['revid2'], [])
981
rev_set = ['revid3', 'revid2']
982
self.assertParentIds(['revid1'], rev_set)
873
984
def test_not_null(self):
874
rev_set = [b'initial']
985
rev_set = ['initial']
875
986
self.assertParentIds([], rev_set)
877
988
def test_not_null_set(self):
878
self.builder.build_snapshot(None, [], revision_id=b'revid1')
989
self.builder.build_snapshot('revid1', None, [])
879
990
rev_set = [_mod_revision.NULL_REVISION]
880
991
self.assertParentIds([], rev_set)
882
993
def test_ghost(self):
883
self.builder.build_snapshot(None, [], revision_id=b'revid1')
884
rev_set = [b'ghost', b'revid1']
885
self.assertParentIds([b'initial'], rev_set)
994
self.builder.build_snapshot('revid1', None, [])
995
rev_set = ['ghost', 'revid1']
996
self.assertParentIds(['initial'], rev_set)
887
998
def test_ghost_parent(self):
888
self.builder.build_snapshot(None, [], revision_id=b'revid1')
889
self.builder.build_snapshot(
890
[b'revid1', b'ghost'], [], revision_id=b'revid2')
891
rev_set = [b'revid2', b'revid1']
892
self.assertParentIds([b'ghost', b'initial'], rev_set)
999
self.builder.build_snapshot('revid1', None, [])
1000
self.builder.build_snapshot('revid2', ['revid1', 'ghost'], [])
1001
rev_set = ['revid2', 'revid1']
1002
self.assertParentIds(['ghost', 'initial'], rev_set)
894
1004
def test_righthand_parent(self):
895
self.builder.build_snapshot(None, [], revision_id=b'revid1')
896
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2a')
897
self.builder.build_snapshot([b'revid1'], [], revision_id=b'revid2b')
898
self.builder.build_snapshot([b'revid2a', b'revid2b'], [],
899
revision_id=b'revid3')
900
rev_set = [b'revid3', b'revid2a']
901
self.assertParentIds([b'revid1', b'revid2b'], rev_set)
1005
self.builder.build_snapshot('revid1', None, [])
1006
self.builder.build_snapshot('revid2a', ['revid1'], [])
1007
self.builder.build_snapshot('revid2b', ['revid1'], [])
1008
self.builder.build_snapshot('revid3', ['revid2a', 'revid2b'], [])
1009
rev_set = ['revid3', 'revid2a']
1010
self.assertParentIds(['revid1', 'revid2b'], rev_set)
904
1013
class TestWithBrokenRepo(TestCaseWithTransport):
916
1025
repo.start_write_group()
917
1026
cleanups.append(repo.commit_write_group)
918
1027
# make rev1a: A well-formed revision, containing 'file1'
919
inv = inventory.Inventory(revision_id=b'rev1a')
920
inv.root.revision = b'rev1a'
921
self.add_file(repo, inv, 'file1', b'rev1a', [])
922
repo.texts.add_lines((inv.root.file_id, b'rev1a'), [], [])
923
repo.add_inventory(b'rev1a', inv, [])
924
revision = _mod_revision.Revision(
1028
inv = inventory.Inventory(revision_id='rev1a')
1029
inv.root.revision = 'rev1a'
1030
self.add_file(repo, inv, 'file1', 'rev1a', [])
1031
repo.texts.add_lines((inv.root.file_id, 'rev1a'), [], [])
1032
repo.add_inventory('rev1a', inv, [])
1033
revision = _mod_revision.Revision('rev1a',
926
1034
committer='jrandom@example.com', timestamp=0,
927
1035
inventory_sha1='', timezone=0, message='foo', parent_ids=[])
928
repo.add_revision(b'rev1a', revision, inv)
1036
repo.add_revision('rev1a',revision, inv)
930
1038
# make rev1b, which has no Revision, but has an Inventory, and
932
inv = inventory.Inventory(revision_id=b'rev1b')
933
inv.root.revision = b'rev1b'
934
self.add_file(repo, inv, 'file1', b'rev1b', [])
935
repo.add_inventory(b'rev1b', inv, [])
1040
inv = inventory.Inventory(revision_id='rev1b')
1041
inv.root.revision = 'rev1b'
1042
self.add_file(repo, inv, 'file1', 'rev1b', [])
1043
repo.add_inventory('rev1b', inv, [])
937
1045
# make rev2, with file1 and file2
939
1047
# file1 has 'rev1b' as an ancestor, even though this is not
940
1048
# mentioned by 'rev1a', making it an unreferenced ancestor
941
1049
inv = inventory.Inventory()
942
self.add_file(repo, inv, 'file1', b'rev2', [b'rev1a', b'rev1b'])
943
self.add_file(repo, inv, 'file2', b'rev2', [])
944
self.add_revision(repo, b'rev2', inv, [b'rev1a'])
1050
self.add_file(repo, inv, 'file1', 'rev2', ['rev1a', 'rev1b'])
1051
self.add_file(repo, inv, 'file2', 'rev2', [])
1052
self.add_revision(repo, 'rev2', inv, ['rev1a'])
946
1054
# make ghost revision rev1c
947
1055
inv = inventory.Inventory()
948
self.add_file(repo, inv, 'file2', b'rev1c', [])
1056
self.add_file(repo, inv, 'file2', 'rev1c', [])
950
1058
# make rev3 with file2
951
1059
# file2 refers to 'rev1c', which is a ghost in this repository, so
952
1060
# file2 cannot have rev1c as its ancestor.
953
1061
inv = inventory.Inventory()
954
self.add_file(repo, inv, 'file2', b'rev3', [b'rev1c'])
955
self.add_revision(repo, b'rev3', inv, [b'rev1c'])
1062
self.add_file(repo, inv, 'file2', 'rev3', ['rev1c'])
1063
self.add_revision(repo, 'rev3', inv, ['rev1c'])
958
1066
for cleanup in reversed(cleanups):
1547
1622
self.assertTrue(new_pack.signature_index._optimize_for_size)
1550
class TestGCCHKPacker(TestCaseWithTransport):
1552
def make_abc_branch(self):
1553
builder = self.make_branch_builder('source')
1554
builder.start_series()
1555
builder.build_snapshot(None, [
1556
('add', ('', b'root-id', 'directory', None)),
1557
('add', ('file', b'file-id', 'file', b'content\n')),
1558
], revision_id=b'A')
1559
builder.build_snapshot([b'A'], [
1560
('add', ('dir', b'dir-id', 'directory', None))],
1562
builder.build_snapshot([b'B'], [
1563
('modify', ('file', b'new content\n'))],
1565
builder.finish_series()
1566
return builder.get_branch()
1568
def make_branch_with_disjoint_inventory_and_revision(self):
1569
"""a repo with separate packs for a revisions Revision and Inventory.
1571
There will be one pack file that holds the Revision content, and one
1572
for the Inventory content.
1574
:return: (repository,
1575
pack_name_with_rev_A_Revision,
1576
pack_name_with_rev_A_Inventory,
1577
pack_name_with_rev_C_content)
1579
b_source = self.make_abc_branch()
1580
b_base = b_source.controldir.sprout(
1581
'base', revision_id=b'A').open_branch()
1582
b_stacked = b_base.controldir.sprout(
1583
'stacked', stacked=True).open_branch()
1584
b_stacked.lock_write()
1585
self.addCleanup(b_stacked.unlock)
1586
b_stacked.fetch(b_source, b'B')
1587
# Now re-open the stacked repo directly (no fallbacks) so that we can
1588
# fill in the A rev.
1589
repo_not_stacked = b_stacked.controldir.open_repository()
1590
repo_not_stacked.lock_write()
1591
self.addCleanup(repo_not_stacked.unlock)
1592
# Now we should have a pack file with A's inventory, but not its
1594
self.assertEqual([(b'A',), (b'B',)],
1595
sorted(repo_not_stacked.inventories.keys()))
1596
self.assertEqual([(b'B',)],
1597
sorted(repo_not_stacked.revisions.keys()))
1598
stacked_pack_names = repo_not_stacked._pack_collection.names()
1599
# We have a couple names here, figure out which has A's inventory
1600
for name in stacked_pack_names:
1601
pack = repo_not_stacked._pack_collection.get_pack_by_name(name)
1602
keys = [n[1] for n in pack.inventory_index.iter_all_entries()]
1604
inv_a_pack_name = name
1607
self.fail('Could not find pack containing A\'s inventory')
1608
repo_not_stacked.fetch(b_source.repository, b'A')
1609
self.assertEqual([(b'A',), (b'B',)],
1610
sorted(repo_not_stacked.revisions.keys()))
1611
new_pack_names = set(repo_not_stacked._pack_collection.names())
1612
rev_a_pack_names = new_pack_names.difference(stacked_pack_names)
1613
self.assertEqual(1, len(rev_a_pack_names))
1614
rev_a_pack_name = list(rev_a_pack_names)[0]
1615
# Now fetch 'C', so we have a couple pack files to join
1616
repo_not_stacked.fetch(b_source.repository, b'C')
1617
rev_c_pack_names = set(repo_not_stacked._pack_collection.names())
1618
rev_c_pack_names = rev_c_pack_names.difference(new_pack_names)
1619
self.assertEqual(1, len(rev_c_pack_names))
1620
rev_c_pack_name = list(rev_c_pack_names)[0]
1621
return (repo_not_stacked, rev_a_pack_name, inv_a_pack_name,
1624
def test_pack_with_distant_inventories(self):
1625
# See https://bugs.launchpad.net/bzr/+bug/437003
1626
# When repacking, it is possible to have an inventory in a different
1627
# pack file than the associated revision. An autopack can then come
1628
# along, and miss that inventory, and complain.
1629
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1630
) = self.make_branch_with_disjoint_inventory_and_revision()
1631
a_pack = repo._pack_collection.get_pack_by_name(rev_a_pack_name)
1632
c_pack = repo._pack_collection.get_pack_by_name(rev_c_pack_name)
1633
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1634
[a_pack, c_pack], '.test-pack')
1635
# This would raise ValueError in bug #437003, but should not raise an
1639
def test_pack_with_missing_inventory(self):
1640
# Similar to test_pack_with_missing_inventory, but this time, we force
1641
# the A inventory to actually be gone from the repository.
1642
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1643
) = self.make_branch_with_disjoint_inventory_and_revision()
1644
inv_a_pack = repo._pack_collection.get_pack_by_name(inv_a_pack_name)
1645
repo._pack_collection._remove_pack_from_memory(inv_a_pack)
1646
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1647
repo._pack_collection.all_packs(), '.test-pack')
1648
e = self.assertRaises(ValueError, packer.pack)
1649
packer.new_pack.abort()
1650
self.assertContainsRe(str(e),
1651
r"We are missing inventories for revisions: .*'A'")
1654
1625
class TestCrossFormatPacks(TestCaseWithTransport):
1656
1627
def log_pack(self, hint=None):