104
116
f = self.reopen_file()
119
def test_add_unicode_content(self):
120
# unicode content is not permitted in versioned files.
121
# versioned files version sequences of bytes only.
123
self.assertRaises(errors.BzrBadParameterUnicode,
124
vf.add_lines, 'a', [], ['a\n', u'b\n', 'c\n'])
126
(errors.BzrBadParameterUnicode, NotImplementedError),
127
vf.add_lines_with_ghosts, 'a', [], ['a\n', u'b\n', 'c\n'])
129
def test_inline_newline_throws(self):
130
# \r characters are not permitted in lines being added
132
self.assertRaises(errors.BzrBadParameterContainsNewline,
133
vf.add_lines, 'a', [], ['a\n\n'])
135
(errors.BzrBadParameterContainsNewline, NotImplementedError),
136
vf.add_lines_with_ghosts, 'a', [], ['a\n\n'])
137
# but inline CR's are allowed
138
vf.add_lines('a', [], ['a\r\n'])
140
vf.add_lines_with_ghosts('b', [], ['a\r\n'])
141
except NotImplementedError:
144
def test_add_reserved(self):
146
self.assertRaises(errors.ReservedId,
147
vf.add_lines, 'a:', [], ['a\n', 'b\n', 'c\n'])
149
self.assertRaises(errors.ReservedId,
150
vf.add_delta, 'a:', [], None, 'sha1', False, ((0, 0, 0, []),))
152
def test_get_reserved(self):
154
self.assertRaises(errors.ReservedId, vf.get_delta, 'b:')
155
self.assertRaises(errors.ReservedId, vf.get_texts, ['b:'])
156
self.assertRaises(errors.ReservedId, vf.get_lines, 'b:')
157
self.assertRaises(errors.ReservedId, vf.get_text, 'b:')
107
159
def test_get_delta(self):
108
160
f = self.get_file()
109
161
sha1s = self._setup_for_deltas(f)
162
214
self.assertEqual(expected_delta, deltas['noeol'])
163
215
# smoke tests for eol support - two noeol in a row same content
164
216
expected_deltas = (('noeol', '3ad7ee82dbd8f29ecba073f96e43e414b3f70a4d', True,
165
[(0, 1, 2, [(u'noeolsecond', 'line\n'), (u'noeolsecond', 'line\n')])]),
217
[(0, 1, 2, [('noeolsecond', 'line\n'), ('noeolsecond', 'line\n')])]),
166
218
('noeol', '3ad7ee82dbd8f29ecba073f96e43e414b3f70a4d', True,
167
219
[(0, 0, 1, [('noeolsecond', 'line\n')]), (1, 1, 0, [])]))
168
220
self.assertEqual(['line\n', 'line'], f.get_lines('noeolsecond'))
169
221
self.assertTrue(deltas['noeolsecond'] in expected_deltas)
170
222
# two no-eol in a row, different content
171
223
expected_delta = ('noeolsecond', '8bb553a84e019ef1149db082d65f3133b195223b', True,
172
[(1, 2, 1, [(u'noeolnotshared', 'phone\n')])])
224
[(1, 2, 1, [('noeolnotshared', 'phone\n')])])
173
225
self.assertEqual(['line\n', 'phone'], f.get_lines('noeolnotshared'))
174
226
self.assertEqual(expected_delta, deltas['noeolnotshared'])
175
227
# eol folling a no-eol with content change
176
228
expected_delta = ('noeol', 'a61f6fb6cfc4596e8d88c34a308d1e724caf8977', False,
177
[(0, 1, 1, [(u'eol', 'phone\n')])])
229
[(0, 1, 1, [('eol', 'phone\n')])])
178
230
self.assertEqual(['phone\n'], f.get_lines('eol'))
179
231
self.assertEqual(expected_delta, deltas['eol'])
180
232
# eol folling a no-eol with content change
181
233
expected_delta = ('noeol', '6bfa09d82ce3e898ad4641ae13dd4fdb9cf0d76b', False,
182
[(0, 1, 1, [(u'eolline', 'line\n')])])
234
[(0, 1, 1, [('eolline', 'line\n')])])
183
235
self.assertEqual(['line\n'], f.get_lines('eolline'))
184
236
self.assertEqual(expected_delta, deltas['eolline'])
185
237
# eol with no parents
186
238
expected_delta = (None, '264f39cab871e4cfd65b3a002f7255888bb5ed97', True,
187
[(0, 0, 1, [(u'noeolbase', 'line\n')])])
239
[(0, 0, 1, [('noeolbase', 'line\n')])])
188
240
self.assertEqual(['line'], f.get_lines('noeolbase'))
189
241
self.assertEqual(expected_delta, deltas['noeolbase'])
190
242
# eol with two parents, in inverse insertion order
191
243
expected_deltas = (('noeolbase', '264f39cab871e4cfd65b3a002f7255888bb5ed97', True,
192
[(0, 1, 1, [(u'eolbeforefirstparent', 'line\n')])]),
244
[(0, 1, 1, [('eolbeforefirstparent', 'line\n')])]),
193
245
('noeolbase', '264f39cab871e4cfd65b3a002f7255888bb5ed97', True,
194
[(0, 1, 1, [(u'eolbeforefirstparent', 'line\n')])]))
246
[(0, 1, 1, [('eolbeforefirstparent', 'line\n')])]))
195
247
self.assertEqual(['line'], f.get_lines('eolbeforefirstparent'))
196
248
#self.assertTrue(deltas['eolbeforefirstparent'] in expected_deltas)
391
443
# and should be a list
392
444
self.assertTrue(isinstance(f.__class__.get_suffixes(), list))
446
def build_graph(self, file, graph):
447
for node in topo_sort(graph.items()):
448
file.add_lines(node, graph[node], [])
394
450
def test_get_graph(self):
395
451
f = self.get_file()
396
f.add_lines('v1', [], ['hello\n'])
397
f.add_lines('v2', ['v1'], ['hello\n', 'world\n'])
398
f.add_lines('v3', ['v2'], ['hello\n', 'cruel\n', 'world\n'])
399
self.assertEqual({'v1': [],
456
self.build_graph(f, graph)
457
self.assertEqual(graph, f.get_graph())
459
def test_get_graph_partial(self):
467
complex_graph.update(simple_a)
472
complex_graph.update(simple_b)
479
complex_graph.update(simple_gam)
481
simple_b_gam.update(simple_gam)
482
simple_b_gam.update(simple_b)
483
self.build_graph(f, complex_graph)
484
self.assertEqual(simple_a, f.get_graph(['a']))
485
self.assertEqual(simple_b, f.get_graph(['b']))
486
self.assertEqual(simple_gam, f.get_graph(['gam']))
487
self.assertEqual(simple_b_gam, f.get_graph(['b', 'gam']))
404
489
def test_get_parents(self):
405
490
f = self.get_file()
499
595
'otherchild\n':0,
597
progress = InstrumentedProgress()
501
598
# iterate over the lines
502
for line in vf.iter_lines_added_or_present_in_versions(versions):
599
for line in vf.iter_lines_added_or_present_in_versions(versions,
602
if []!= progress.updates:
603
self.assertEqual(expected, progress.updates)
505
lines = iter_with_versions(['child', 'otherchild'])
605
lines = iter_with_versions(['child', 'otherchild'],
606
[('Walking content.', 0, 2),
607
('Walking content.', 1, 2),
608
('Walking content.', 2, 2)])
506
609
# we must see child and otherchild
507
610
self.assertTrue(lines['child\n'] > 0)
508
611
self.assertTrue(lines['otherchild\n'] > 0)
509
612
# we dont care if we got more than that.
512
lines = iter_with_versions(None)
615
lines = iter_with_versions(None, [('Walking content.', 0, 5),
616
('Walking content.', 1, 5),
617
('Walking content.', 2, 5),
618
('Walking content.', 3, 5),
619
('Walking content.', 4, 5),
620
('Walking content.', 5, 5)])
513
621
# all lines must be seen at least once
514
622
self.assertTrue(lines['base\n'] > 0)
515
623
self.assertTrue(lines['lancestor\n'] > 0)
570
681
self.assertRaises(NotImplementedError, vf.get_parents_with_ghosts, 'foo')
571
682
self.assertRaises(NotImplementedError, vf.get_graph_with_ghosts)
684
vf = self.reopen_file()
573
685
# test key graph related apis: getncestry, _graph, get_parents
575
687
# - these are ghost unaware and must not be reflect ghosts
576
self.assertEqual([u'notbxbfse'], vf.get_ancestry(u'notbxbfse'))
577
self.assertEqual([], vf.get_parents(u'notbxbfse'))
578
self.assertEqual({u'notbxbfse':[]}, vf.get_graph())
579
self.assertFalse(vf.has_version(u'b\xbfse'))
688
self.assertEqual(['notbxbfse'], vf.get_ancestry('notbxbfse'))
689
self.assertEqual([], vf.get_parents('notbxbfse'))
690
self.assertEqual({'notbxbfse':[]}, vf.get_graph())
691
self.assertFalse(self.callDeprecated([osutils._revision_id_warning],
692
vf.has_version, parent_id_unicode))
693
self.assertFalse(vf.has_version(parent_id_utf8))
580
694
# we have _with_ghost apis to give us ghost information.
581
self.assertEqual([u'b\xbfse', u'notbxbfse'], vf.get_ancestry_with_ghosts([u'notbxbfse']))
582
self.assertEqual([u'b\xbfse'], vf.get_parents_with_ghosts(u'notbxbfse'))
583
self.assertEqual({u'notbxbfse':[u'b\xbfse']}, vf.get_graph_with_ghosts())
584
self.assertTrue(vf.has_ghost(u'b\xbfse'))
695
self.assertEqual([parent_id_utf8, 'notbxbfse'], vf.get_ancestry_with_ghosts(['notbxbfse']))
696
self.assertEqual([parent_id_utf8], vf.get_parents_with_ghosts('notbxbfse'))
697
self.assertEqual({'notbxbfse':[parent_id_utf8]}, vf.get_graph_with_ghosts())
698
self.assertTrue(self.callDeprecated([osutils._revision_id_warning],
699
vf.has_ghost, parent_id_unicode))
700
self.assertTrue(vf.has_ghost(parent_id_utf8))
585
701
# if we add something that is a ghost of another, it should correct the
586
702
# results of the prior apis
587
vf.add_lines(u'b\xbfse', [], [])
588
self.assertEqual([u'b\xbfse', u'notbxbfse'], vf.get_ancestry([u'notbxbfse']))
589
self.assertEqual([u'b\xbfse'], vf.get_parents(u'notbxbfse'))
590
self.assertEqual({u'b\xbfse':[],
591
u'notbxbfse':[u'b\xbfse'],
703
self.callDeprecated([osutils._revision_id_warning],
704
vf.add_lines, parent_id_unicode, [], [])
705
self.assertEqual([parent_id_utf8, 'notbxbfse'], vf.get_ancestry(['notbxbfse']))
706
self.assertEqual([parent_id_utf8], vf.get_parents('notbxbfse'))
707
self.assertEqual({parent_id_utf8:[],
708
'notbxbfse':[parent_id_utf8],
594
self.assertTrue(vf.has_version(u'b\xbfse'))
711
self.assertTrue(self.callDeprecated([osutils._revision_id_warning],
712
vf.has_version, parent_id_unicode))
713
self.assertTrue(vf.has_version(parent_id_utf8))
595
714
# we have _with_ghost apis to give us ghost information.
596
self.assertEqual([u'b\xbfse', u'notbxbfse'], vf.get_ancestry_with_ghosts([u'notbxbfse']))
597
self.assertEqual([u'b\xbfse'], vf.get_parents_with_ghosts(u'notbxbfse'))
598
self.assertEqual({u'b\xbfse':[],
599
u'notbxbfse':[u'b\xbfse'],
715
self.assertEqual([parent_id_utf8, 'notbxbfse'], vf.get_ancestry_with_ghosts(['notbxbfse']))
716
self.assertEqual([parent_id_utf8], vf.get_parents_with_ghosts('notbxbfse'))
717
self.assertEqual({parent_id_utf8:[],
718
'notbxbfse':[parent_id_utf8],
601
720
vf.get_graph_with_ghosts())
602
self.assertFalse(vf.has_ghost(u'b\xbfse'))
721
self.assertFalse(self.callDeprecated([osutils._revision_id_warning],
722
vf.has_ghost, parent_id_unicode))
723
self.assertFalse(vf.has_ghost(parent_id_utf8))
604
725
def test_add_lines_with_ghosts_after_normal_revs(self):
605
726
# some versioned file formats allow lines to be added with parent
783
921
versionedfile.InterVersionedFile.unregister_optimiser(InterString)
784
922
# now we should get the default InterVersionedFile object again.
785
923
self.assertGetsDefaultInterVersionedFile(dummy_a, dummy_b)
926
class TestReadonlyHttpMixin(object):
928
def test_readonly_http_works(self):
929
# we should be able to read from http with a versioned file.
931
# try an empty file access
932
readonly_vf = self.get_factory()('foo', get_transport(self.get_readonly_url('.')))
933
self.assertEqual([], readonly_vf.versions())
935
vf.add_lines('1', [], ['a\n'])
936
vf.add_lines('2', ['1'], ['b\n', 'a\n'])
937
readonly_vf = self.get_factory()('foo', get_transport(self.get_readonly_url('.')))
938
self.assertEqual(['1', '2'], vf.versions())
939
for version in readonly_vf.versions():
940
readonly_vf.get_lines(version)
943
class TestWeaveHTTP(TestCaseWithWebserver, TestReadonlyHttpMixin):
946
return WeaveFile('foo', get_transport(self.get_url('.')), create=True)
948
def get_factory(self):
952
class TestKnitHTTP(TestCaseWithWebserver, TestReadonlyHttpMixin):
955
return KnitVersionedFile('foo', get_transport(self.get_url('.')),
956
delta=True, create=True)
958
def get_factory(self):
959
return KnitVersionedFile
962
class MergeCasesMixin(object):
964
def doMerge(self, base, a, b, mp):
965
from cStringIO import StringIO
966
from textwrap import dedent
972
w.add_lines('text0', [], map(addcrlf, base))
973
w.add_lines('text1', ['text0'], map(addcrlf, a))
974
w.add_lines('text2', ['text0'], map(addcrlf, b))
978
self.log('merge plan:')
979
p = list(w.plan_merge('text1', 'text2'))
980
for state, line in p:
982
self.log('%12s | %s' % (state, line[:-1]))
986
mt.writelines(w.weave_merge(p))
988
self.log(mt.getvalue())
990
mp = map(addcrlf, mp)
991
self.assertEqual(mt.readlines(), mp)
994
def testOneInsert(self):
1000
def testSeparateInserts(self):
1001
self.doMerge(['aaa', 'bbb', 'ccc'],
1002
['aaa', 'xxx', 'bbb', 'ccc'],
1003
['aaa', 'bbb', 'yyy', 'ccc'],
1004
['aaa', 'xxx', 'bbb', 'yyy', 'ccc'])
1006
def testSameInsert(self):
1007
self.doMerge(['aaa', 'bbb', 'ccc'],
1008
['aaa', 'xxx', 'bbb', 'ccc'],
1009
['aaa', 'xxx', 'bbb', 'yyy', 'ccc'],
1010
['aaa', 'xxx', 'bbb', 'yyy', 'ccc'])
1011
overlappedInsertExpected = ['aaa', 'xxx', 'yyy', 'bbb']
1012
def testOverlappedInsert(self):
1013
self.doMerge(['aaa', 'bbb'],
1014
['aaa', 'xxx', 'yyy', 'bbb'],
1015
['aaa', 'xxx', 'bbb'], self.overlappedInsertExpected)
1017
# really it ought to reduce this to
1018
# ['aaa', 'xxx', 'yyy', 'bbb']
1021
def testClashReplace(self):
1022
self.doMerge(['aaa'],
1025
['<<<<<<< ', 'xxx', '=======', 'yyy', 'zzz',
1028
def testNonClashInsert1(self):
1029
self.doMerge(['aaa'],
1032
['<<<<<<< ', 'xxx', 'aaa', '=======', 'yyy', 'zzz',
1035
def testNonClashInsert2(self):
1036
self.doMerge(['aaa'],
1042
def testDeleteAndModify(self):
1043
"""Clashing delete and modification.
1045
If one side modifies a region and the other deletes it then
1046
there should be a conflict with one side blank.
1049
#######################################
1050
# skippd, not working yet
1053
self.doMerge(['aaa', 'bbb', 'ccc'],
1054
['aaa', 'ddd', 'ccc'],
1056
['<<<<<<<< ', 'aaa', '=======', '>>>>>>> ', 'ccc'])
1058
def _test_merge_from_strings(self, base, a, b, expected):
1060
w.add_lines('text0', [], base.splitlines(True))
1061
w.add_lines('text1', ['text0'], a.splitlines(True))
1062
w.add_lines('text2', ['text0'], b.splitlines(True))
1063
self.log('merge plan:')
1064
p = list(w.plan_merge('text1', 'text2'))
1065
for state, line in p:
1067
self.log('%12s | %s' % (state, line[:-1]))
1068
self.log('merge result:')
1069
result_text = ''.join(w.weave_merge(p))
1070
self.log(result_text)
1071
self.assertEqualDiff(result_text, expected)
1073
def test_weave_merge_conflicts(self):
1074
# does weave merge properly handle plans that end with unchanged?
1075
result = ''.join(self.get_file().weave_merge([('new-a', 'hello\n')]))
1076
self.assertEqual(result, 'hello\n')
1078
def test_deletion_extended(self):
1079
"""One side deletes, the other deletes more.
1096
self._test_merge_from_strings(base, a, b, result)
1098
def test_deletion_overlap(self):
1099
"""Delete overlapping regions with no other conflict.
1101
Arguably it'd be better to treat these as agreement, rather than
1102
conflict, but for now conflict is safer.
1130
self._test_merge_from_strings(base, a, b, result)
1132
def test_agreement_deletion(self):
1133
"""Agree to delete some lines, without conflicts."""
1155
self._test_merge_from_strings(base, a, b, result)
1157
def test_sync_on_deletion(self):
1158
"""Specific case of merge where we can synchronize incorrectly.
1160
A previous version of the weave merge concluded that the two versions
1161
agreed on deleting line 2, and this could be a synchronization point.
1162
Line 1 was then considered in isolation, and thought to be deleted on
1165
It's better to consider the whole thing as a disagreement region.
1176
a's replacement line 2
1189
a's replacement line 2
1196
self._test_merge_from_strings(base, a, b, result)
1199
class TestKnitMerge(TestCaseWithTransport, MergeCasesMixin):
1201
def get_file(self, name='foo'):
1202
return KnitVersionedFile(name, get_transport(self.get_url('.')),
1203
delta=True, create=True)
1205
def log_contents(self, w):
1209
class TestWeaveMerge(TestCaseWithTransport, MergeCasesMixin):
1211
def get_file(self, name='foo'):
1212
return WeaveFile(name, get_transport(self.get_url('.')), create=True)
1214
def log_contents(self, w):
1215
self.log('weave is:')
1217
write_weave(w, tmpf)
1218
self.log(tmpf.getvalue())
1220
overlappedInsertExpected = ['aaa', '<<<<<<< ', 'xxx', 'yyy', '=======',
1221
'xxx', '>>>>>>> ', 'bbb']