43
41
condition_isinstance(TestBTreeNodes))
44
42
import bzrlib._btree_serializer_py as py_module
45
43
scenarios = [('python', {'parse_btree': py_module})]
46
if compiled_btreeparser_feature.available():
47
scenarios.append(('C', {'parse_btree':
48
compiled_btreeparser_feature.module}))
44
if CompiledBtreeParserFeature.available():
45
# Is there a way to do this that gets missing feature failures rather
46
# than no indication to the user?
47
import bzrlib._btree_serializer_pyx as c_module
48
scenarios.append(('C', {'parse_btree': c_module}))
49
49
return multiply_tests(node_tests, scenarios, others)
52
compiled_btreeparser_feature = tests.ModuleAvailableFeature(
53
'bzrlib._btree_serializer_pyx')
52
class _CompiledBtreeParserFeature(tests.Feature):
55
import bzrlib._btree_serializer_pyx
60
def feature_name(self):
61
return 'bzrlib._btree_serializer_pyx'
63
CompiledBtreeParserFeature = _CompiledBtreeParserFeature()
56
66
class BTreeTestCase(TestCaseWithTransport):
61
71
TestCaseWithTransport.setUp(self)
62
self.overrideAttr(btree_index, '_RESERVED_HEADER_BYTES', 100)
72
self._original_header = btree_index._RESERVED_HEADER_BYTES
74
btree_index._RESERVED_HEADER_BYTES = self._original_header
75
self.addCleanup(restore)
76
btree_index._RESERVED_HEADER_BYTES = 100
64
78
def make_nodes(self, count, key_elements, reference_lists):
65
79
"""Generate count*key_elements sample nodes."""
100
114
def shrink_page_size(self):
101
115
"""Shrink the default page size so that less fits in a page."""
102
self.overrideAttr(btree_index, '_PAGE_SIZE')
116
old_page_size = btree_index._PAGE_SIZE
118
btree_index._PAGE_SIZE = old_page_size
119
self.addCleanup(cleanup)
103
120
btree_index._PAGE_SIZE = 2048
106
123
class TestBTreeBuilder(BTreeTestCase):
108
def test_clear_cache(self):
109
builder = btree_index.BTreeBuilder(reference_lists=0, key_elements=1)
110
# This is a no-op, but we need the api to be consistent with other
111
# BTreeGraphIndex apis.
112
builder.clear_cache()
114
125
def test_empty_1_0(self):
115
126
builder = btree_index.BTreeBuilder(key_elements=1, reference_lists=0)
116
127
# NamedTemporaryFile dies on builder.finish().read(). weird.
340
351
# Test the parts of the index that take up memory are doing so
342
353
self.assertEqual(1, len(builder._nodes))
354
self.assertEqual(1, len(builder._keys))
343
355
self.assertIs(None, builder._nodes_by_key)
344
356
builder.add_node(*nodes[1])
345
357
self.assertEqual(0, len(builder._nodes))
358
self.assertEqual(0, len(builder._keys))
346
359
self.assertIs(None, builder._nodes_by_key)
347
360
self.assertEqual(1, len(builder._backing_indices))
348
361
self.assertEqual(2, builder._backing_indices[0].key_count())
349
362
# now back to memory
350
363
builder.add_node(*nodes[2])
351
364
self.assertEqual(1, len(builder._nodes))
365
self.assertEqual(1, len(builder._keys))
352
366
self.assertIs(None, builder._nodes_by_key)
353
367
# And spills to a second backing index combing all
354
368
builder.add_node(*nodes[3])
355
369
self.assertEqual(0, len(builder._nodes))
370
self.assertEqual(0, len(builder._keys))
356
371
self.assertIs(None, builder._nodes_by_key)
357
372
self.assertEqual(2, len(builder._backing_indices))
358
373
self.assertEqual(None, builder._backing_indices[0])
361
376
builder.add_node(*nodes[4])
362
377
builder.add_node(*nodes[5])
363
378
self.assertEqual(0, len(builder._nodes))
379
self.assertEqual(0, len(builder._keys))
364
380
self.assertIs(None, builder._nodes_by_key)
365
381
self.assertEqual(2, len(builder._backing_indices))
366
382
self.assertEqual(2, builder._backing_indices[0].key_count())
424
440
# Test the parts of the index that take up memory are doing so
426
442
self.assertEqual(1, len(builder._nodes))
443
self.assertEqual(1, len(builder._keys))
427
444
self.assertIs(None, builder._nodes_by_key)
428
445
builder.add_node(*nodes[1])
429
446
self.assertEqual(0, len(builder._nodes))
447
self.assertEqual(0, len(builder._keys))
430
448
self.assertIs(None, builder._nodes_by_key)
431
449
self.assertEqual(1, len(builder._backing_indices))
432
450
self.assertEqual(2, builder._backing_indices[0].key_count())
433
451
# now back to memory
434
452
builder.add_node(*nodes[2])
435
453
self.assertEqual(1, len(builder._nodes))
454
self.assertEqual(1, len(builder._keys))
436
455
self.assertIs(None, builder._nodes_by_key)
437
456
# And spills to a second backing index but doesn't combine
438
457
builder.add_node(*nodes[3])
439
458
self.assertEqual(0, len(builder._nodes))
459
self.assertEqual(0, len(builder._keys))
440
460
self.assertIs(None, builder._nodes_by_key)
441
461
self.assertEqual(2, len(builder._backing_indices))
442
462
for backing_index in builder._backing_indices:
445
465
builder.add_node(*nodes[4])
446
466
builder.add_node(*nodes[5])
447
467
self.assertEqual(0, len(builder._nodes))
468
self.assertEqual(0, len(builder._keys))
448
469
self.assertIs(None, builder._nodes_by_key)
449
470
self.assertEqual(3, len(builder._backing_indices))
450
471
for backing_index in builder._backing_indices:
509
530
builder.add_node(*nodes[0])
510
531
# Test the parts of the index that take up memory are doing so
533
self.assertEqual(1, len(builder._keys))
512
534
self.assertEqual(1, len(builder._nodes))
513
535
self.assertIs(None, builder._nodes_by_key)
514
536
builder.add_node(*nodes[1])
537
self.assertEqual(0, len(builder._keys))
515
538
self.assertEqual(0, len(builder._nodes))
516
539
self.assertIs(None, builder._nodes_by_key)
517
540
self.assertEqual(1, len(builder._backing_indices))
520
543
old = dict(builder._get_nodes_by_key()) #Build up the nodes by key dict
521
544
builder.add_node(*nodes[2])
522
545
self.assertEqual(1, len(builder._nodes))
546
self.assertEqual(1, len(builder._keys))
523
547
self.assertIsNot(None, builder._nodes_by_key)
524
548
self.assertNotEqual({}, builder._nodes_by_key)
525
549
# We should have a new entry
527
551
# And spills to a second backing index combing all
528
552
builder.add_node(*nodes[3])
529
553
self.assertEqual(0, len(builder._nodes))
554
self.assertEqual(0, len(builder._keys))
530
555
self.assertIs(None, builder._nodes_by_key)
531
556
self.assertEqual(2, len(builder._backing_indices))
532
557
self.assertEqual(None, builder._backing_indices[0])
535
560
builder.add_node(*nodes[4])
536
561
builder.add_node(*nodes[5])
537
562
self.assertEqual(0, len(builder._nodes))
563
self.assertEqual(0, len(builder._keys))
538
564
self.assertIs(None, builder._nodes_by_key)
539
565
self.assertEqual(2, len(builder._backing_indices))
540
566
self.assertEqual(2, builder._backing_indices[0].key_count())
611
637
size = trans.put_file('index', stream)
612
638
return btree_index.BTreeGraphIndex(trans, 'index', size)
614
def make_index_with_offset(self, ref_lists=1, key_elements=1, nodes=[],
616
builder = btree_index.BTreeBuilder(key_elements=key_elements,
617
reference_lists=ref_lists)
618
builder.add_nodes(nodes)
619
transport = self.get_transport('')
620
# NamedTemporaryFile dies on builder.finish().read(). weird.
621
temp_file = builder.finish()
622
content = temp_file.read()
625
transport.put_bytes('index', (' '*offset)+content)
626
return btree_index.BTreeGraphIndex(transport, 'index', size=size,
629
def test_clear_cache(self):
630
nodes = self.make_nodes(160, 2, 2)
631
index = self.make_index(ref_lists=2, key_elements=2, nodes=nodes)
632
self.assertEqual(1, len(list(index.iter_entries([nodes[30][0]]))))
633
self.assertEqual([1, 4], index._row_lengths)
634
self.assertIsNot(None, index._root_node)
635
internal_node_pre_clear = index._internal_node_cache.keys()
636
self.assertTrue(len(index._leaf_node_cache) > 0)
638
# We don't touch _root_node or _internal_node_cache, both should be
639
# small, and can save a round trip or two
640
self.assertIsNot(None, index._root_node)
641
# NOTE: We don't want to affect the _internal_node_cache, as we expect
642
# it will be small, and if we ever do touch this index again, it
643
# will save round-trips. This assertion isn't very strong,
644
# becuase without a 3-level index, we don't have any internal
646
self.assertEqual(internal_node_pre_clear,
647
index._internal_node_cache.keys())
648
self.assertEqual(0, len(index._leaf_node_cache))
650
640
def test_trivial_constructor(self):
651
641
transport = get_transport('trace+' + self.get_url(''))
652
642
index = btree_index.BTreeGraphIndex(transport, 'index', None)
699
689
# The entire index should have been read, as it is one page long.
700
690
self.assertEqual([('readv', 'index', [(0, size)], False, None)],
701
691
transport._activity)
702
self.assertEqual(1173, size)
704
def test_with_offset_no_size(self):
705
index = self.make_index_with_offset(key_elements=1, ref_lists=1,
707
nodes=self.make_nodes(200, 1, 1))
708
index._size = None # throw away the size info
709
self.assertEqual(200, index.key_count())
711
def test_with_small_offset(self):
712
index = self.make_index_with_offset(key_elements=1, ref_lists=1,
714
nodes=self.make_nodes(200, 1, 1))
715
self.assertEqual(200, index.key_count())
717
def test_with_large_offset(self):
718
index = self.make_index_with_offset(key_elements=1, ref_lists=1,
720
nodes=self.make_nodes(200, 1, 1))
721
self.assertEqual(200, index.key_count())
692
self.assertEqual(1199, size)
723
694
def test__read_nodes_no_size_one_page_reads_once(self):
724
695
self.make_index(nodes=[(('key',), 'value', ())])
772
743
# The entire index should have been read linearly.
773
744
self.assertEqual([('readv', 'index', [(0, size)], False, None)],
774
745
transport._activity)
775
self.assertEqual(1488, size)
746
self.assertEqual(1514, size)
777
748
def test_validate_two_pages(self):
778
749
builder = btree_index.BTreeBuilder(key_elements=2, reference_lists=2)
1144
1115
self.assertEqual({}, parent_map)
1145
1116
self.assertEqual(set([('one',), ('two',)]), missing_keys)
1147
def test_supports_unlimited_cache(self):
1148
builder = btree_index.BTreeBuilder(reference_lists=0, key_elements=1)
1149
# We need enough nodes to cause a page split (so we have both an
1150
# internal node and a couple leaf nodes. 500 seems to be enough.)
1151
nodes = self.make_nodes(500, 1, 0)
1153
builder.add_node(*node)
1154
stream = builder.finish()
1155
trans = get_transport(self.get_url())
1156
size = trans.put_file('index', stream)
1157
index = btree_index.BTreeGraphIndex(trans, 'index', size)
1158
self.assertEqual(500, index.key_count())
1159
# We have an internal node
1160
self.assertEqual(2, len(index._row_lengths))
1161
# We have at least 2 leaf nodes
1162
self.assertTrue(index._row_lengths[-1] >= 2)
1163
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1164
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1165
index._leaf_node_cache._max_cache)
1166
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1167
self.assertEqual(100, index._internal_node_cache._max_cache)
1168
# No change if unlimited_cache=False is passed
1169
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1170
unlimited_cache=False)
1171
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1172
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1173
index._leaf_node_cache._max_cache)
1174
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1175
self.assertEqual(100, index._internal_node_cache._max_cache)
1176
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1177
unlimited_cache=True)
1178
self.assertIsInstance(index._leaf_node_cache, dict)
1179
self.assertIs(type(index._internal_node_cache), dict)
1180
# Exercise the lookup code
1181
entries = set(index.iter_entries([n[0] for n in nodes]))
1182
self.assertEqual(500, len(entries))
1185
1119
class TestBTreeNodes(BTreeTestCase):
1121
def restore_parser(self):
1122
btree_index._btree_serializer = self.saved_parser
1187
1124
def setUp(self):
1188
1125
BTreeTestCase.setUp(self)
1189
self.overrideAttr(btree_index, '_btree_serializer', self.parse_btree)
1126
self.saved_parser = btree_index._btree_serializer
1127
self.addCleanup(self.restore_parser)
1128
btree_index._btree_serializer = self.parse_btree
1191
1130
def test_LeafNode_1_0(self):
1192
1131
node_bytes = ("type=leaf\n"