14
14
# along with this program; if not, write to the Free Software
15
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
17
from bzrlib import osutils, ui, urlutils
18
from bzrlib.errors import InvalidRevisionId
19
from bzrlib.inventory import Inventory
20
from bzrlib.repository import InterRepository
21
from bzrlib.trace import info
22
from bzrlib.tsort import topo_sort
17
from dulwich.objects import (
22
from dulwich.object_store import (
36
from bzrlib.errors import (
40
from bzrlib.inventory import (
47
from bzrlib.repository import (
50
from bzrlib.revision import (
53
from bzrlib.tsort import (
56
from bzrlib.versionedfile import (
57
FulltextContentFactory,
60
from bzrlib.plugins.git.mapping import (
62
inventory_to_tree_and_blobs,
67
from bzrlib.plugins.git.object_store import (
70
from bzrlib.plugins.git.remote import (
24
73
from bzrlib.plugins.git.repository import (
29
from bzrlib.plugins.git.remote import RemoteGitRepository
32
from dulwich.client import SimpleFetchGraphWalker
33
from dulwich.objects import Commit
35
from cStringIO import StringIO
38
class BzrFetchGraphWalker(object):
39
"""GraphWalker implementation that uses a Bazaar repository."""
41
def __init__(self, repository, mapping):
42
self.repository = repository
43
self.mapping = mapping
45
self.heads = set(repository.all_revision_ids())
49
return iter(self.next, None)
52
revid = self.mapping.revision_id_foreign_to_bzr(sha)
55
def remove(self, revid):
57
if revid in self.heads:
58
self.heads.remove(revid)
59
if revid in self.parents:
60
for p in self.parents[revid]:
65
ret = self.heads.pop()
66
ps = self.repository.get_parent_map([ret])[ret]
67
self.parents[ret] = ps
68
self.heads.update([p for p in ps if not p in self.done])
71
return self.mapping.revision_id_bzr_to_foreign(ret)[0]
72
except InvalidRevisionId:
77
def import_git_blob(repo, mapping, path, blob, inv, parent_invs, executable):
80
MAX_INV_CACHE_SIZE = 50 * 1024 * 1024
83
def import_git_blob(texts, mapping, path, hexsha, base_inv, base_ie, parent_id,
84
revision_id, parent_invs, shagitmap, lookup_object, executable, symlink):
78
85
"""Import a git blob object into a bzr repository.
80
:param repo: bzr repository
87
:param texts: VersionedFiles to add to
81
88
:param path: Path in the tree
82
89
:param blob: A git blob
90
:return: Inventory delta for this file
84
92
file_id = mapping.generate_file_id(path)
85
text_revision = inv.revision_id
86
repo.texts.add_lines((file_id, text_revision),
87
[(file_id, p[file_id].revision) for p in parent_invs if file_id in p],
88
osutils.split_lines(blob.data))
89
ie = inv.add_path(path, "file", file_id)
90
ie.revision = text_revision
91
ie.text_size = len(blob.data)
92
ie.text_sha1 = osutils.sha_string(blob.data)
97
# We just have to hope this is indeed utf-8:
98
ie = cls(file_id, urlutils.basename(path).decode("utf-8"), parent_id)
93
99
ie.executable = executable
96
def import_git_tree(repo, mapping, path, tree, inv, parent_invs, lookup_object):
100
# See if this has changed at all
105
base_sha = shagitmap.lookup_blob(file_id, base_ie.revision)
109
if (base_sha == hexsha and base_ie.executable == ie.executable
110
and base_ie.kind == ie.kind):
111
# If nothing has changed since the base revision, we're done
113
if base_sha == hexsha and base_ie.kind == ie.kind:
114
ie.text_size = base_ie.text_size
115
ie.text_sha1 = base_ie.text_sha1
116
ie.symlink_target = base_ie.symlink_target
117
if ie.executable == base_ie.executable:
118
ie.revision = base_ie.revision
120
blob = lookup_object(hexsha)
122
blob = lookup_object(hexsha)
123
if ie.kind == "symlink":
125
ie.symlink_target = blob.data
129
ie.text_size = len(blob.data)
130
ie.text_sha1 = osutils.sha_string(blob.data)
131
# Check what revision we should store
133
for pinv in parent_invs:
134
if pinv.revision_id == base_inv.revision_id:
143
if pie.text_sha1 == ie.text_sha1 and pie.executable == ie.executable and pie.symlink_target == ie.symlink_target:
144
# found a revision in one of the parents to use
145
ie.revision = pie.revision
147
parent_keys.append((file_id, pie.revision))
148
if ie.revision is None:
149
# Need to store a new revision
150
ie.revision = revision_id
151
assert file_id is not None
152
assert ie.revision is not None
153
texts.insert_record_stream([FulltextContentFactory((file_id, ie.revision), tuple(parent_keys), ie.text_sha1, blob.data)])
154
shamap = [(hexsha, "blob", (ie.file_id, ie.revision))]
158
if base_ie is not None:
159
old_path = base_inv.id2path(file_id)
160
if base_ie.kind == "directory":
161
invdelta.extend(remove_disappeared_children(old_path, base_ie.children, []))
164
invdelta.append((old_path, path, file_id, ie))
165
return (invdelta, shamap)
168
class SubmodulesRequireSubtrees(BzrError):
169
_fmt = """The repository you are fetching from contains submodules. To continue, upgrade your Bazaar repository to a format that supports nested trees, such as 'development-subtree'."""
173
def import_git_submodule(texts, mapping, path, hexsha, base_inv, base_ie,
174
parent_id, revision_id, parent_invs, shagitmap, lookup_object):
175
file_id = mapping.generate_file_id(path)
176
ie = TreeReference(file_id, urlutils.basename(path.decode("utf-8")),
178
ie.revision = revision_id
183
if base_ie.kind == ie.kind and base_ie.reference_revision == ie.reference_revision:
184
ie.revision = base_ie.revision
185
ie.reference_revision = mapping.revision_id_foreign_to_bzr(hexsha)
186
texts.insert_record_stream([FulltextContentFactory((file_id, ie.revision), (), None, "")])
187
invdelta = [(oldpath, path, file_id, ie)]
188
return invdelta, {}, {}
191
def remove_disappeared_children(path, base_children, existing_children):
193
deletable = [(osutils.pathjoin(path, k), v) for k,v in base_children.iteritems() if k not in existing_children]
195
(path, ie) = deletable.pop()
196
ret.append((path, None, ie.file_id, None))
197
if ie.kind == "directory":
198
for name, child_ie in ie.children.iteritems():
199
deletable.append((osutils.pathjoin(path, name), child_ie))
203
def import_git_tree(texts, mapping, path, hexsha, base_inv, base_ie, parent_id,
204
revision_id, parent_invs, shagitmap, lookup_object, allow_submodules=False):
97
205
"""Import a git tree object into a bzr repository.
99
:param repo: A Bzr repository object
207
:param texts: VersionedFiles object to add to
100
208
:param path: Path in the tree
101
209
:param tree: A git tree object
102
:param inv: Inventory object
210
:param base_inv: Base inventory against which to return inventory delta
211
:return: Inventory delta for this subtree
104
214
file_id = mapping.generate_file_id(path)
105
text_revision = inv.revision_id
106
repo.texts.add_lines((file_id, text_revision),
107
[(file_id, p[file_id].revision) for p in parent_invs if file_id in p],
109
ie = inv.add_path(path, "directory", file_id)
110
ie.revision = text_revision
111
for mode, name, hexsha in tree.entries():
112
entry_kind = (mode & 0700000) / 0100000
215
# We just have to hope this is indeed utf-8:
216
ie = InventoryDirectory(file_id, urlutils.basename(path.decode("utf-8")),
219
# Newly appeared here
220
ie.revision = revision_id
221
texts.insert_record_stream([FulltextContentFactory((file_id, ie.revision), (), None, "")])
222
invdelta.append((None, path, file_id, ie))
224
# See if this has changed at all
226
base_sha = shagitmap.lookup_tree(file_id, base_inv.revision_id)
230
if base_sha == hexsha:
231
# If nothing has changed since the base revision, we're done
233
if base_ie.kind != "directory":
234
ie.revision = revision_id
235
texts.insert_record_stream([FulltextContentFactory((ie.file_id, ie.revision), (), None, "")])
236
invdelta.append((base_inv.id2path(ie.file_id), path, ie.file_id, ie))
237
if base_ie is not None and base_ie.kind == "directory":
238
base_children = base_ie.children
241
# Remember for next time
242
existing_children = set()
245
tree = lookup_object(hexsha)
246
for mode, name, child_hexsha in tree.entries():
113
247
basename = name.decode("utf-8")
117
child_path = urlutils.join(path, name)
119
tree = lookup_object(hexsha)
120
import_git_tree(repo, mapping, child_path, tree, inv, parent_invs, lookup_object)
121
elif entry_kind == 1:
122
blob = lookup_object(hexsha)
123
fs_mode = mode & 0777
124
import_git_blob(repo, mapping, child_path, blob, inv, parent_invs, bool(fs_mode & 0111))
126
raise AssertionError("Unknown blob kind, perms=%r." % (mode,))
129
def import_git_objects(repo, mapping, object_iter, pb=None):
248
existing_children.add(basename)
249
child_path = osutils.pathjoin(path, name)
250
if stat.S_ISDIR(mode):
251
subinvdelta, grandchildmodes, subshamap = import_git_tree(
252
texts, mapping, child_path, child_hexsha, base_inv,
253
base_children.get(basename), file_id, revision_id,
254
parent_invs, shagitmap, lookup_object,
255
allow_submodules=allow_submodules)
256
invdelta.extend(subinvdelta)
257
child_modes.update(grandchildmodes)
258
shamap.extend(subshamap)
259
elif S_ISGITLINK(mode): # submodule
260
if not allow_submodules:
261
raise SubmodulesRequireSubtrees()
262
subinvdelta, grandchildmodes, subshamap = import_git_submodule(
263
texts, mapping, child_path, child_hexsha, base_inv, base_children.get(basename),
264
file_id, revision_id, parent_invs, shagitmap, lookup_object)
265
invdelta.extend(subinvdelta)
266
child_modes.update(grandchildmodes)
267
shamap.extend(subshamap)
269
subinvdelta, subshamap = import_git_blob(texts, mapping,
270
child_path, child_hexsha, base_inv, base_children.get(basename), file_id,
271
revision_id, parent_invs, shagitmap, lookup_object,
272
mode_is_executable(mode), stat.S_ISLNK(mode))
273
invdelta.extend(subinvdelta)
274
shamap.extend(subshamap)
275
if mode not in (stat.S_IFDIR, DEFAULT_FILE_MODE,
276
stat.S_IFLNK, DEFAULT_FILE_MODE|0111):
277
child_modes[child_path] = mode
278
# Remove any children that have disappeared
279
if base_ie is not None and base_ie.kind == "directory":
280
invdelta.extend(remove_disappeared_children(base_inv.id2path(file_id),
281
base_children, existing_children))
282
shamap.append((hexsha, "tree", (file_id, revision_id)))
283
return invdelta, child_modes, shamap
286
def approx_inv_size(inv):
287
# Very rough estimate, 1k per inventory entry
288
return len(inv) * 1024
291
def import_git_commit(repo, mapping, head, lookup_object,
292
target_git_object_retriever, parent_invs_cache):
293
o = lookup_object(head)
294
rev = mapping.import_commit(o)
295
# We have to do this here, since we have to walk the tree and
296
# we need to make sure to import the blobs / trees with the right
297
# path; this may involve adding them more than once.
299
for parent_id in rev.parent_ids:
301
parent_invs.append(parent_invs_cache[parent_id])
303
parent_inv = repo.get_inventory(parent_id)
304
parent_invs.append(parent_inv)
305
parent_invs_cache[parent_id] = parent_inv
306
if parent_invs == []:
307
base_inv = Inventory(root_id=None)
310
base_inv = parent_invs[0]
311
base_ie = base_inv.root
312
inv_delta, unusual_modes, shamap = import_git_tree(repo.texts,
313
mapping, "", o.tree, base_inv, base_ie, None, rev.revision_id,
314
parent_invs, target_git_object_retriever._idmap, lookup_object,
315
allow_submodules=getattr(repo._format, "supports_tree_reference", False))
316
target_git_object_retriever._idmap.add_entries(shamap)
317
if unusual_modes != {}:
318
for path, mode in unusual_modes.iteritems():
319
warn_unusual_mode(rev.foreign_revid, path, mode)
320
mapping.import_unusual_file_modes(rev, unusual_modes)
322
basis_id = rev.parent_ids[0]
324
basis_id = NULL_REVISION
326
rev.inventory_sha1, inv = repo.add_inventory_by_delta(basis_id,
327
inv_delta, rev.revision_id, rev.parent_ids,
329
parent_invs_cache[rev.revision_id] = inv
330
repo.add_revision(rev.revision_id, rev)
331
if "verify" in debug.debug_flags:
332
new_unusual_modes = mapping.export_unusual_file_modes(rev)
333
if new_unusual_modes != unusual_modes:
334
raise AssertionError("unusual modes don't match: %r != %r" % (unusual_modes, new_unusual_modes))
335
objs = inventory_to_tree_and_blobs(inv, repo.texts, mapping, unusual_modes)
336
for sha1, newobj, path in objs:
337
assert path is not None
338
oldobj = tree_lookup_path(lookup_object, o.tree, path)
340
raise AssertionError("%r != %r in %s" % (oldobj, newobj, path))
343
def import_git_objects(repo, mapping, object_iter, target_git_object_retriever,
130
345
"""Import a set of git objects into a bzr repository.
132
:param repo: Bazaar repository
347
:param repo: Target Bazaar repository
133
348
:param mapping: Mapping to use
134
349
:param object_iter: Iterator over Git objects.
351
target_git_object_retriever._idmap.start_write_group() # FIXME: try/finally
352
def lookup_object(sha):
354
return object_iter[sha]
356
return target_git_object_retriever[sha]
136
357
# TODO: a more (memory-)efficient implementation of this
361
parent_invs_cache = lru_cache.LRUSizeCache(compute_size=approx_inv_size,
362
max_size=MAX_INV_CACHE_SIZE)
140
363
# Find and convert commit objects
141
for o in object_iter.iterobjects():
366
pb.update("finding revisions to fetch", len(graph), None)
368
assert isinstance(head, str)
370
o = lookup_object(head)
372
trace.mutter('missing head %s', head)
142
374
if isinstance(o, Commit):
143
375
rev = mapping.import_commit(o)
144
root_trees[rev.revision_id] = object_iter[o.tree]
145
revisions[rev.revision_id] = rev
146
graph.append((rev.revision_id, rev.parent_ids))
376
if repo.has_revision(rev.revision_id):
378
squash_revision(repo, rev)
379
graph.append((o.id, o.parents))
380
target_git_object_retriever._idmap.add_entry(o.id, "commit",
381
(rev.revision_id, o.tree))
382
heads.extend([p for p in o.parents if p not in checked])
383
elif isinstance(o, Tag):
384
heads.append(o.object[1])
386
trace.warning("Unable to import head object %r" % o)
147
389
# Order the revisions
148
390
# Create the inventory objects
149
for i, revid in enumerate(topo_sort(graph)):
151
pb.update("fetching revisions", i, len(graph))
152
root_tree = root_trees[revid]
153
rev = revisions[revid]
154
# We have to do this here, since we have to walk the tree and
155
# we need to make sure to import the blobs / trees with the riht
156
# path; this may involve adding them more than once.
158
inv.revision_id = rev.revision_id
159
def lookup_object(sha):
160
if sha in object_iter:
161
return object_iter[sha]
162
return reconstruct_git_object(repo, mapping, sha)
163
parent_invs = [repo.get_inventory(r) for r in rev.parent_ids]
164
import_git_tree(repo, mapping, "", root_tree, inv, parent_invs,
166
repo.add_revision(rev.revision_id, rev, inv)
169
def reconstruct_git_commit(repo, rev):
170
raise NotImplementedError(self.reconstruct_git_commit)
173
def reconstruct_git_object(repo, mapping, sha):
175
revid = mapping.revision_id_foreign_to_bzr(sha)
177
rev = repo.get_revision(revid)
178
except NoSuchRevision:
181
return reconstruct_git_commit(rev)
185
raise KeyError("No such object %s" % sha)
188
class InterGitNonGitRepository(InterRepository):
190
_matching_repo_format = GitFormat()
392
revision_ids = topo_sort(graph)
394
for offset in range(0, len(revision_ids), batch_size):
395
repo.start_write_group()
397
for i, head in enumerate(revision_ids[offset:offset+batch_size]):
399
pb.update("fetching revisions", offset+i, len(revision_ids))
400
import_git_commit(repo, mapping, head, lookup_object,
401
target_git_object_retriever,
404
repo.abort_write_group()
407
hint = repo.commit_write_group()
409
pack_hints.extend(hint)
410
target_git_object_retriever._idmap.commit_write_group()
414
class InterGitRepository(InterRepository):
416
_matching_repo_format = GitRepositoryFormat()
193
419
def _get_repo_format_to_test():
197
423
"""See InterRepository.copy_content."""
198
424
self.fetch(revision_id, pb, find_ghosts=False)
200
def fetch(self, revision_id=None, pb=None, find_ghosts=False,
426
def fetch(self, revision_id=None, pb=None, find_ghosts=False, mapping=None,
428
self.fetch_refs(revision_id=revision_id, pb=pb, find_ghosts=find_ghosts,
429
mapping=mapping, fetch_spec=fetch_spec)
432
class InterGitNonGitRepository(InterGitRepository):
433
"""Base InterRepository that copies revisions from a Git into a non-Git
436
def fetch_refs(self, revision_id=None, pb=None, find_ghosts=False,
437
mapping=None, fetch_spec=None):
202
438
if mapping is None:
203
439
mapping = self.source.get_mapping()
205
pb.update("git: %s" % text.rstrip("\r\n"), 0, 0)
206
def determine_wants(heads):
207
if revision_id is None:
440
if revision_id is not None:
441
interesting_heads = [revision_id]
442
elif fetch_spec is not None:
443
interesting_heads = fetch_spec.heads
445
interesting_heads = None
447
def determine_wants(refs):
449
if interesting_heads is None:
450
ret = [sha for (ref, sha) in refs.iteritems() if not ref.endswith("^{}")]
210
ret = [mapping.revision_id_bzr_to_foreign(revision_id)[0]]
452
ret = [mapping.revision_id_bzr_to_foreign(revid)[0] for revid in interesting_heads if revid not in (None, NULL_REVISION)]
211
453
return [rev for rev in ret if not self.target.has_revision(mapping.revision_id_foreign_to_bzr(rev))]
212
graph_walker = BzrFetchGraphWalker(self.target, mapping)
454
pack_hint = self.fetch_objects(determine_wants, mapping, pb)
455
if pack_hint is not None and self.target._format.pack_compresses:
456
self.target.pack(hint=pack_hint)
457
if interesting_heads is not None:
458
present_interesting_heads = self.target.has_revisions(interesting_heads)
459
missing_interesting_heads = set(interesting_heads) - present_interesting_heads
460
if missing_interesting_heads:
461
raise AssertionError("Missing interesting heads: %r" % missing_interesting_heads)
465
_GIT_PROGRESS_RE = re.compile(r"(.*?): +(\d+)% \((\d+)/(\d+)\)")
466
def report_git_progress(pb, text):
467
text = text.rstrip("\r\n")
468
g = _GIT_PROGRESS_RE.match(text)
470
(text, pct, current, total) = g.groups()
471
pb.update(text, int(current), int(total))
473
pb.update(text, 0, 0)
476
class InterRemoteGitNonGitRepository(InterGitNonGitRepository):
477
"""InterRepository that copies revisions from a remote Git into a non-Git
480
def get_target_heads(self):
481
# FIXME: This should be more efficient
482
all_revs = self.target.all_revision_ids()
483
parent_map = self.target.get_parent_map(all_revs)
485
map(all_parents.update, parent_map.itervalues())
486
return set(all_revs) - all_parents
488
def fetch_objects(self, determine_wants, mapping, pb=None):
490
report_git_progress(pb, text)
491
store = BazaarObjectStore(self.target, mapping)
492
self.target.lock_write()
494
heads = self.get_target_heads()
495
graph_walker = store.get_graph_walker(
496
[store._lookup_revision_sha1(head) for head in heads])
499
def record_determine_wants(heads):
500
wants = determine_wants(heads)
501
recorded_wants.extend(wants)
506
create_pb = pb = ui.ui_factory.nested_progress_bar()
508
objects_iter = self.source.fetch_objects(
509
record_determine_wants, graph_walker,
510
store.get_raw, progress)
511
return import_git_objects(self.target, mapping,
512
objects_iter, store, recorded_wants, pb)
520
def is_compatible(source, target):
521
"""Be compatible with GitRepository."""
522
# FIXME: Also check target uses VersionedFile
523
return (isinstance(source, RemoteGitRepository) and
524
target.supports_rich_root() and
525
not isinstance(target, GitRepository))
528
class InterLocalGitNonGitRepository(InterGitNonGitRepository):
529
"""InterRepository that copies revisions from a local Git into a non-Git
532
def fetch_objects(self, determine_wants, mapping, pb=None):
533
wants = determine_wants(self.source._git.get_refs())
215
536
create_pb = pb = ui.ui_factory.nested_progress_bar()
537
target_git_object_retriever = BazaarObjectStore(self.target, mapping)
217
539
self.target.lock_write()
219
self.target.start_write_group()
221
objects_iter = self.source.fetch_objects(determine_wants,
222
graph_walker, progress)
223
import_git_objects(self.target, mapping, objects_iter, pb)
225
self.target.commit_write_group()
541
return import_git_objects(self.target, mapping,
542
self.source._git.object_store, target_git_object_retriever,
227
545
self.target.unlock()