/brz/remove-bazaar

To get this branch, use:
bzr branch http://gegoxaren.bato24.eu/bzr/brz/remove-bazaar
4988.10.3 by John Arbash Meinel
Merge bzr.dev 5007, resolve conflict, update NEWS
1
# Copyright (C) 2005-2010 Canonical Ltd
1887.1.1 by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines,
2
#
974.1.27 by aaron.bentley at utoronto
Initial greedy fetch work
3
# This program is free software; you can redistribute it and/or modify
4
# it under the terms of the GNU General Public License as published by
5
# the Free Software Foundation; either version 2 of the License, or
6
# (at your option) any later version.
1887.1.1 by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines,
7
#
974.1.27 by aaron.bentley at utoronto
Initial greedy fetch work
8
# This program is distributed in the hope that it will be useful,
9
# but WITHOUT ANY WARRANTY; without even the implied warranty of
10
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
11
# GNU General Public License for more details.
1887.1.1 by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines,
12
#
974.1.27 by aaron.bentley at utoronto
Initial greedy fetch work
13
# You should have received a copy of the GNU General Public License
14
# along with this program; if not, write to the Free Software
4183.7.1 by Sabin Iacob
update FSF mailing address
15
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
1218 by Martin Pool
- fix up import
16
1231 by Martin Pool
- more progress on fetch on top of weaves
17
18
"""Copying of history from one branch to another.
19
20
The basic plan is that every branch knows the history of everything
21
that has merged into it.  As the first step of a merge, pull, or
22
branch operation we copy history from the source into the destination
23
branch.
24
"""
25
3350.6.4 by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores.
26
import operator
27
4476.3.9 by Andrew Bennetts
Further reduce duplication.
28
from bzrlib.lazy_import import lazy_import
29
lazy_import(globals(), """
30
from bzrlib import (
5539.2.1 by Andrew Bennetts
Start defining an 'everything' fetch spec.
31
    graph,
4476.3.9 by Andrew Bennetts
Further reduce duplication.
32
    tsort,
33
    versionedfile,
34
    )
35
""")
1534.1.31 by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo.
36
import bzrlib
4110.2.4 by Martin Pool
Deprecate passing a pb in to RepoFetcher
37
from bzrlib import (
38
    errors,
4819.2.4 by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits.
39
    ui,
4110.2.4 by Martin Pool
Deprecate passing a pb in to RepoFetcher
40
    )
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
41
from bzrlib.revision import NULL_REVISION
2094.3.5 by John Arbash Meinel
Fix imports to ensure modules are loaded before they are used
42
from bzrlib.trace import mutter
1534.1.31 by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo.
43
1238 by Martin Pool
- remove a lot of dead code from fetch
44
1534.4.41 by Robert Collins
Branch now uses BzrDir reasonably sanely.
45
class RepoFetcher(object):
46
    """Pull revisions and texts from one repository to another.
47
2592.4.5 by Martin Pool
Add Repository.base on all repositories.
48
    This should not be used directly, it's essential a object to encapsulate
1534.1.33 by Robert Collins
Move copy_content_into into InterRepository and InterWeaveRepo, and disable the default codepath test as we have optimised paths for all current combinations.
49
    the logic in InterRepository.fetch().
1260 by Martin Pool
- some updates for fetch/update function
50
    """
3172.4.1 by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is
51
4070.9.2 by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations.
52
    def __init__(self, to_repository, from_repository, last_revision=None,
4961.2.3 by Martin Pool
Delete deprecated pb parameter to RepoFetcher
53
        find_ghosts=True, fetch_spec=None):
3172.4.1 by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is
54
        """Create a repo fetcher.
55
4110.2.2 by Martin Pool
Remove obsolete comments
56
        :param last_revision: If set, try to limit to the data this revision
57
            references.
3172.4.1 by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is
58
        :param find_ghosts: If True search the entire history for ghosts.
59
        """
4509.3.18 by Martin Pool
RepoFetcher relies on Repository.fetch to shortcircuit no-op fetches
60
        # repository.fetch has the responsibility for short-circuiting
61
        # attempts to copy between a repository and itself.
1534.4.41 by Robert Collins
Branch now uses BzrDir reasonably sanely.
62
        self.to_repository = to_repository
63
        self.from_repository = from_repository
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
64
        self.sink = to_repository._get_sink()
1534.4.41 by Robert Collins
Branch now uses BzrDir reasonably sanely.
65
        # must not mutate self._last_revision as its potentially a shared instance
1185.65.27 by Robert Collins
Tweak storage towards mergability.
66
        self._last_revision = last_revision
4070.9.2 by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations.
67
        self._fetch_spec = fetch_spec
3172.4.1 by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is
68
        self.find_ghosts = find_ghosts
1534.4.41 by Robert Collins
Branch now uses BzrDir reasonably sanely.
69
        self.from_repository.lock_read()
4110.2.22 by Martin Pool
Re-add mutter calls during fetch
70
        mutter("Using fetch logic to copy between %s(%s) and %s(%s)",
71
               self.from_repository, self.from_repository._format,
72
               self.to_repository, self.to_repository._format)
3842.3.5 by Andrew Bennetts
Remove some debugging cruft, make more tests pass.
73
        try:
4110.2.3 by Martin Pool
Remove redundant variable from fetch.
74
            self.__fetch()
3842.3.5 by Andrew Bennetts
Remove some debugging cruft, make more tests pass.
75
        finally:
76
            self.from_repository.unlock()
1185.65.27 by Robert Collins
Tweak storage towards mergability.
77
78
    def __fetch(self):
79
        """Primary worker function.
80
3943.8.1 by Marius Kruger
remove all trailing whitespace from bzr source
81
        This initialises all the needed variables, and then fetches the
1185.65.27 by Robert Collins
Tweak storage towards mergability.
82
        requested revisions, finally clearing the progress bar.
83
        """
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
84
        # Roughly this is what we're aiming for fetch to become:
85
        #
86
        # missing = self.sink.insert_stream(self.source.get_stream(search))
87
        # if missing:
88
        #     missing = self.sink.insert_stream(self.source.get_items(missing))
89
        # assert not missing
1240 by Martin Pool
- clean up fetch code and add progress bar
90
        self.count_total = 0
1185.33.55 by Martin Pool
[patch] weave fetch optimizations (Goffredo Baroncelli)
91
        self.file_ids_names = {}
4819.2.4 by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits.
92
        pb = ui.ui_factory.nested_progress_bar()
4110.2.14 by Martin Pool
Small fetch progress tweaks
93
        pb.show_pct = pb.show_count = False
4110.2.9 by Martin Pool
Re-add very basic top-level pb for fetch
94
        try:
4110.2.14 by Martin Pool
Small fetch progress tweaks
95
            pb.update("Finding revisions", 0, 2)
4110.2.9 by Martin Pool
Re-add very basic top-level pb for fetch
96
            search = self._revids_to_fetch()
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
97
            mutter('fetching: %s', search)
98
            if search.is_empty():
4110.2.9 by Martin Pool
Re-add very basic top-level pb for fetch
99
                return
4110.2.14 by Martin Pool
Small fetch progress tweaks
100
            pb.update("Fetching revisions", 1, 2)
4110.2.9 by Martin Pool
Re-add very basic top-level pb for fetch
101
            self._fetch_everything_for_search(search)
102
        finally:
103
            pb.finished()
2535.3.6 by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository.
104
4110.2.6 by Martin Pool
Remove more progressbar cruft from fetch
105
    def _fetch_everything_for_search(self, search):
2535.3.6 by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository.
106
        """Fetch all data for the given set of revisions."""
2535.3.9 by Andrew Bennetts
More comments.
107
        # The first phase is "file".  We pass the progress bar for it directly
2668.2.8 by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by.
108
        # into item_keys_introduced_by, which has more information about how
2535.3.9 by Andrew Bennetts
More comments.
109
        # that phase is progressing than we do.  Progress updates for the other
110
        # phases are taken care of in this function.
111
        # XXX: there should be a clear owner of the progress reporting.  Perhaps
2668.2.8 by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by.
112
        # item_keys_introduced_by should have a richer API than it does at the
113
        # moment, so that it can feed the progress information back to this
2535.3.9 by Andrew Bennetts
More comments.
114
        # function?
4060.1.3 by Robert Collins
Implement the separate source component for fetch - repository.StreamSource.
115
        if (self.from_repository._format.rich_root_data and
116
            not self.to_repository._format.rich_root_data):
117
            raise errors.IncompatibleRepositories(
118
                self.from_repository, self.to_repository,
119
                "different rich-root support")
4819.2.4 by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits.
120
        pb = ui.ui_factory.nested_progress_bar()
2535.3.7 by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch.
121
        try:
4110.2.12 by Martin Pool
Add more fetch progress
122
            pb.update("Get stream source")
4060.1.3 by Robert Collins
Implement the separate source component for fetch - repository.StreamSource.
123
            source = self.from_repository._get_source(
124
                self.to_repository._format)
125
            stream = source.get_stream(search)
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
126
            from_format = self.from_repository._format
4110.2.12 by Martin Pool
Add more fetch progress
127
            pb.update("Inserting stream")
4032.3.7 by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink.
128
            resume_tokens, missing_keys = self.sink.insert_stream(
5195.3.14 by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records
129
                stream, from_format, [])
4257.3.2 by Andrew Bennetts
Check during fetch if we are going to be missing data necessary to calculate altered fileids for stacked revisions.
130
            if self.to_repository._fallback_repositories:
5539.2.1 by Andrew Bennetts
Start defining an 'everything' fetch spec.
131
                if not isinstance(search, graph.EverythingResult):
5539.2.7 by Andrew Bennetts
Add a test, revise a comment.
132
                    # If search is EverythingResult this is be unnecessary,
133
                    # so we can skip this step.  The source will send us
134
                    # every revision it has, and their parent inventories.
135
                    # (Unless the source is damaged!  but not really worth
136
                    # optimising for that case.  The pack code will reject bad
137
                    # streams anyway.)
5539.2.1 by Andrew Bennetts
Start defining an 'everything' fetch spec.
138
                    missing_keys.update(
139
                        self._parent_inventories(search.get_keys()))
4029.2.1 by Robert Collins
Support streaming push to stacked branches.
140
            if missing_keys:
4110.2.12 by Martin Pool
Add more fetch progress
141
                pb.update("Missing keys")
4060.1.3 by Robert Collins
Implement the separate source component for fetch - repository.StreamSource.
142
                stream = source.get_stream_for_missing_keys(missing_keys)
4110.2.12 by Martin Pool
Add more fetch progress
143
                pb.update("Inserting missing keys")
4032.3.7 by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink.
144
                resume_tokens, missing_keys = self.sink.insert_stream(
5195.3.14 by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records
145
                    stream, from_format, resume_tokens)
4029.2.1 by Robert Collins
Support streaming push to stacked branches.
146
            if missing_keys:
147
                raise AssertionError(
148
                    "second push failed to complete a fetch %r." % (
149
                        missing_keys,))
4032.3.7 by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink.
150
            if resume_tokens:
151
                raise AssertionError(
152
                    "second push failed to commit the fetch %r." % (
153
                        resume_tokens,))
4110.2.12 by Martin Pool
Add more fetch progress
154
            pb.update("Finishing stream")
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
155
            self.sink.finished()
2535.3.7 by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch.
156
        finally:
4110.2.6 by Martin Pool
Remove more progressbar cruft from fetch
157
            pb.finished()
4029.2.1 by Robert Collins
Support streaming push to stacked branches.
158
1185.65.30 by Robert Collins
Merge integration.
159
    def _revids_to_fetch(self):
2535.3.7 by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch.
160
        """Determines the exact revisions needed from self.from_repository to
161
        install self._last_revision in self.to_repository.
162
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
163
        :returns: A SearchResult of some sort.  (Possibly a
164
        PendingAncestryResult, EmptySearchResult, etc.)
2535.3.7 by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch.
165
        """
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
166
        mutter("self._fetch_spec, self._last_revision: %r, %r",
167
                self._fetch_spec, self._last_revision)
5539.2.19 by Andrew Bennetts
Define SearchResult/Search interfaces with explicit abstract base classes, add some docstrings and change a method name.
168
        get_search_result = getattr(self._fetch_spec, 'get_search_result', None)
169
        if get_search_result is not None:
170
            mutter(
171
                'resolving fetch_spec into search result: %s', self._fetch_spec)
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
172
            # This is EverythingNotInOther or a similar kind of fetch_spec.
173
            # Turn it into a search result.
5539.2.19 by Andrew Bennetts
Define SearchResult/Search interfaces with explicit abstract base classes, add some docstrings and change a method name.
174
            return get_search_result()
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
175
        elif self._fetch_spec is not None:
176
            # The fetch spec is already a concrete search result.
4070.9.2 by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations.
177
            return self._fetch_spec
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
178
        elif self._last_revision == NULL_REVISION:
179
            # fetch_spec is None + last_revision is null => empty fetch.
1534.4.50 by Robert Collins
Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running.
180
            # explicit limit of no revisions needed
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
181
            return graph.EmptySearchResult()
182
        elif self._last_revision is not None:
5539.2.10 by Andrew Bennetts
s/NotInOtherForRev/NotInOtherForRevs/, and allow passing multiple revision_ids to search_missing_revision_ids.
183
            return graph.NotInOtherForRevs(self.to_repository,
184
                self.from_repository, [self._last_revision],
5539.2.19 by Andrew Bennetts
Define SearchResult/Search interfaces with explicit abstract base classes, add some docstrings and change a method name.
185
                find_ghosts=self.find_ghosts).get_search_result()
5539.2.8 by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'.
186
        else: # self._last_revision is None:
187
            return graph.EverythingNotInOther(self.to_repository,
5539.2.19 by Andrew Bennetts
Define SearchResult/Search interfaces with explicit abstract base classes, add some docstrings and change a method name.
188
                self.from_repository,
189
                find_ghosts=self.find_ghosts).get_search_result()
1185.64.3 by Goffredo Baroncelli
This patch changes the fetch code. Before, the original code expanded every inventory and
190
4257.4.12 by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher.
191
    def _parent_inventories(self, revision_ids):
192
        # Find all the parent revisions referenced by the stream, but
4257.4.13 by Andrew Bennetts
Tweak comment.
193
        # not present in the stream, and make sure we send their
4257.4.12 by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher.
194
        # inventories.
195
        parent_maps = self.to_repository.get_parent_map(revision_ids)
196
        parents = set()
197
        map(parents.update, parent_maps.itervalues())
198
        parents.discard(NULL_REVISION)
199
        parents.difference_update(revision_ids)
200
        missing_keys = set(('inventories', rev_id) for rev_id in parents)
201
        return missing_keys
202
3565.3.3 by Robert Collins
* Fetching data between repositories that have the same model but no
203
1910.2.24 by Aaron Bentley
Got intra-repository fetch working between model1 and 2 for all types
204
class Inter1and2Helper(object):
1910.2.48 by Aaron Bentley
Update from review comments
205
    """Helper for operations that convert data from model 1 and 2
3943.8.1 by Marius Kruger
remove all trailing whitespace from bzr source
206
1910.2.48 by Aaron Bentley
Update from review comments
207
    This is for use by fetchers and converters.
208
    """
209
5050.32.1 by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories.
210
    # This is a class variable so that the test suite can override it.
211
    known_graph_threshold = 100
212
4022.1.1 by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts)
213
    def __init__(self, source):
1910.2.48 by Aaron Bentley
Update from review comments
214
        """Constructor.
215
216
        :param source: The repository data comes from
217
        """
218
        self.source = source
219
220
    def iter_rev_trees(self, revs):
221
        """Iterate through RevisionTrees efficiently.
222
223
        Additionally, the inventory's revision_id is set if unset.
224
225
        Trees are retrieved in batches of 100, and then yielded in the order
226
        they were requested.
227
228
        :param revs: A list of revision ids
229
        """
3172.4.4 by Robert Collins
Review feedback.
230
        # In case that revs is not a list.
231
        revs = list(revs)
1910.2.48 by Aaron Bentley
Update from review comments
232
        while revs:
233
            for tree in self.source.revision_trees(revs[:100]):
1910.2.44 by Aaron Bentley
Retrieve only 500 revision trees at once
234
                if tree.inventory.revision_id is None:
235
                    tree.inventory.revision_id = tree.get_revision_id()
236
                yield tree
1910.2.48 by Aaron Bentley
Update from review comments
237
            revs = revs[100:]
1910.2.44 by Aaron Bentley
Retrieve only 500 revision trees at once
238
3380.2.4 by Aaron Bentley
Updates from review
239
    def _find_root_ids(self, revs, parent_map, graph):
240
        revision_root = {}
1910.2.48 by Aaron Bentley
Update from review comments
241
        for tree in self.iter_rev_trees(revs):
1910.2.18 by Aaron Bentley
Implement creation of knits for tree roots
242
            revision_id = tree.inventory.root.revision
2946.3.3 by John Arbash Meinel
Prefer tree.get_root_id() as more explicit than tree.path2id('')
243
            root_id = tree.get_root_id()
3380.1.3 by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed
244
            revision_root[revision_id] = root_id
245
        # Find out which parents we don't already know root ids for
246
        parents = set()
247
        for revision_parents in parent_map.itervalues():
248
            parents.update(revision_parents)
249
        parents.difference_update(revision_root.keys() + [NULL_REVISION])
3380.2.7 by Aaron Bentley
Update docs
250
        # Limit to revisions present in the versionedfile
3380.1.3 by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed
251
        parents = graph.get_parent_map(parents).keys()
252
        for tree in self.iter_rev_trees(parents):
253
            root_id = tree.get_root_id()
254
            revision_root[tree.get_revision_id()] = root_id
4476.3.11 by Andrew Bennetts
All fetch and interrepo tests passing.
255
        return revision_root
3380.2.4 by Aaron Bentley
Updates from review
256
257
    def generate_root_texts(self, revs):
258
        """Generate VersionedFiles for all root ids.
259
260
        :param revs: the revisions to include
261
        """
262
        graph = self.source.get_graph()
263
        parent_map = graph.get_parent_map(revs)
4476.3.9 by Andrew Bennetts
Further reduce duplication.
264
        rev_order = tsort.topo_sort(parent_map)
4476.3.11 by Andrew Bennetts
All fetch and interrepo tests passing.
265
        rev_id_to_root_id = self._find_root_ids(revs, parent_map, graph)
3350.6.4 by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores.
266
        root_id_order = [(rev_id_to_root_id[rev_id], rev_id) for rev_id in
267
            rev_order]
268
        # Guaranteed stable, this groups all the file id operations together
269
        # retaining topological order within the revisions of a file id.
270
        # File id splits and joins would invalidate this, but they don't exist
271
        # yet, and are unlikely to in non-rich-root environments anyway.
272
        root_id_order.sort(key=operator.itemgetter(0))
273
        # Create a record stream containing the roots to create.
5050.32.1 by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories.
274
        if len(revs) > self.known_graph_threshold:
275
            graph = self.source.get_known_graph_ancestry(revs)
4476.3.9 by Andrew Bennetts
Further reduce duplication.
276
        new_roots_stream = _new_root_data_stream(
4476.3.41 by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation.
277
            root_id_order, rev_id_to_root_id, parent_map, self.source, graph)
4476.3.9 by Andrew Bennetts
Further reduce duplication.
278
        return [('texts', new_roots_stream)]
279
280
281
def _new_root_data_stream(
4476.3.41 by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation.
282
    root_keys_to_create, rev_id_to_root_id_map, parent_map, repo, graph=None):
4476.3.69 by Andrew Bennetts
Elaborate some docstrings.
283
    """Generate a texts substream of synthesised root entries.
284
285
    Used in fetches that do rich-root upgrades.
286
    
287
    :param root_keys_to_create: iterable of (root_id, rev_id) pairs describing
288
        the root entries to create.
289
    :param rev_id_to_root_id_map: dict of known rev_id -> root_id mappings for
290
        calculating the parents.  If a parent rev_id is not found here then it
291
        will be recalculated.
292
    :param parent_map: a parent map for all the revisions in
293
        root_keys_to_create.
294
    :param graph: a graph to use instead of repo.get_graph().
295
    """
4476.3.9 by Andrew Bennetts
Further reduce duplication.
296
    for root_key in root_keys_to_create:
297
        root_id, rev_id = root_key
298
        parent_keys = _parent_keys_for_root_version(
4476.3.41 by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation.
299
            root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph)
4476.3.9 by Andrew Bennetts
Further reduce duplication.
300
        yield versionedfile.FulltextContentFactory(
301
            root_key, parent_keys, None, '')
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
302
303
304
def _parent_keys_for_root_version(
4476.3.41 by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation.
305
    root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph=None):
4476.3.69 by Andrew Bennetts
Elaborate some docstrings.
306
    """Get the parent keys for a given root id.
307
    
308
    A helper function for _new_root_data_stream.
309
    """
4476.3.9 by Andrew Bennetts
Further reduce duplication.
310
    # Include direct parents of the revision, but only if they used the same
311
    # root_id and are heads.
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
312
    rev_parents = parent_map[rev_id]
313
    parent_ids = []
314
    for parent_id in rev_parents:
315
        if parent_id == NULL_REVISION:
316
            continue
317
        if parent_id not in rev_id_to_root_id_map:
4476.3.9 by Andrew Bennetts
Further reduce duplication.
318
            # We probably didn't read this revision, go spend the extra effort
319
            # to actually check
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
320
            try:
321
                tree = repo.revision_tree(parent_id)
322
            except errors.NoSuchRevision:
4476.3.9 by Andrew Bennetts
Further reduce duplication.
323
                # Ghost, fill out rev_id_to_root_id in case we encounter this
324
                # again.
325
                # But set parent_root_id to None since we don't really know
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
326
                parent_root_id = None
327
            else:
328
                parent_root_id = tree.get_root_id()
329
            rev_id_to_root_id_map[parent_id] = None
4476.3.21 by Andrew Bennetts
Clarify some code and comments, and s/1.17/1.18/ in a few places.
330
            # XXX: why not:
331
            #   rev_id_to_root_id_map[parent_id] = parent_root_id
332
            # memory consumption maybe?
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
333
        else:
334
            parent_root_id = rev_id_to_root_id_map[parent_id]
335
        if root_id == parent_root_id:
4476.3.9 by Andrew Bennetts
Further reduce duplication.
336
            # With stacking we _might_ want to refer to a non-local revision,
337
            # but this code path only applies when we have the full content
338
            # available, so ghosts really are ghosts, not just the edge of
339
            # local data.
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
340
            parent_ids.append(parent_id)
341
        else:
342
            # root_id may be in the parent anyway.
343
            try:
344
                tree = repo.revision_tree(parent_id)
345
            except errors.NoSuchRevision:
346
                # ghost, can't refer to it.
347
                pass
348
            else:
349
                try:
4476.3.9 by Andrew Bennetts
Further reduce duplication.
350
                    parent_ids.append(tree.inventory[root_id].revision)
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
351
                except errors.NoSuchId:
352
                    # not in the tree
353
                    pass
354
    # Drop non-head parents
4476.3.41 by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation.
355
    if graph is None:
356
        graph = repo.get_graph()
357
    heads = graph.heads(parent_ids)
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
358
    selected_ids = []
359
    for parent_id in parent_ids:
360
        if parent_id in heads and parent_id not in selected_ids:
361
            selected_ids.append(parent_id)
4476.3.9 by Andrew Bennetts
Further reduce duplication.
362
    parent_keys = [(root_id, parent_id) for parent_id in selected_ids]
4476.3.6 by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer.
363
    return parent_keys