bzr branch
http://gegoxaren.bato24.eu/bzr/brz/remove-bazaar
| 
1616.1.2
by Martin Pool
 Emit trace message about hash cache performance.  | 
1  | 
# Copyright (C) 2005, 2006 by Canonical Ltd
 | 
| 
1887.1.1
by Adeodato Simó
 Do not separate paragraphs in the copyright statement with blank lines,  | 
2  | 
#
 | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
3  | 
# This program is free software; you can redistribute it and/or modify
 | 
4  | 
# it under the terms of the GNU General Public License as published by
 | 
|
5  | 
# the Free Software Foundation; either version 2 of the License, or
 | 
|
6  | 
# (at your option) any later version.
 | 
|
| 
1887.1.1
by Adeodato Simó
 Do not separate paragraphs in the copyright statement with blank lines,  | 
7  | 
#
 | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
8  | 
# This program is distributed in the hope that it will be useful,
 | 
9  | 
# but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
|
10  | 
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
|
11  | 
# GNU General Public License for more details.
 | 
|
| 
1887.1.1
by Adeodato Simó
 Do not separate paragraphs in the copyright statement with blank lines,  | 
12  | 
#
 | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
13  | 
# You should have received a copy of the GNU General Public License
 | 
14  | 
# along with this program; if not, write to the Free Software
 | 
|
15  | 
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 | 
|
16  | 
||
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
17  | 
# TODO: Up-front, stat all files in order and remove those which are deleted or 
 | 
18  | 
# out-of-date.  Don't actually re-read them until they're needed.  That ought 
 | 
|
19  | 
# to bring all the inodes into core so that future stats to them are fast, and 
 | 
|
20  | 
# it preserves the nice property that any caller will always get up-to-date
 | 
|
21  | 
# data except in unavoidable cases.
 | 
|
| 
864
by Martin Pool
 doc  | 
22  | 
|
23  | 
# TODO: Perhaps return more details on the file to avoid statting it
 | 
|
24  | 
# again: nonexistent, file type, size, etc
 | 
|
25  | 
||
| 
1213
by Martin Pool
 - move import in hashcache  | 
26  | 
# TODO: Perhaps use a Python pickle instead of a text file; might be faster.
 | 
27  | 
||
| 
864
by Martin Pool
 doc  | 
28  | 
|
29  | 
||
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
30  | 
CACHE_HEADER = "### bzr hashcache v5\n"  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
31  | 
|
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
32  | 
import os, stat, time  | 
| 
1092.2.6
by Robert Collins
 symlink support updated to work  | 
33  | 
import sha  | 
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
34  | 
|
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
35  | 
from bzrlib.osutils import sha_file, pathjoin, safe_unicode  | 
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
36  | 
from bzrlib.trace import mutter, warning  | 
| 
1213
by Martin Pool
 - move import in hashcache  | 
37  | 
from bzrlib.atomicfile import AtomicFile  | 
| 
1185.59.8
by Denys Duchier
 hashcache: missing import for BzrError  | 
38  | 
from bzrlib.errors import BzrError  | 
| 
1540.1.1
by Martin Pool
 [patch] stat-cache fixes from Denys  | 
39  | 
|
40  | 
||
| 
1185.59.10
by Denys Duchier
 hashcache: new constants and improved comment  | 
41  | 
FP_MTIME_COLUMN = 1  | 
42  | 
FP_CTIME_COLUMN = 2  | 
|
| 
1092.2.6
by Robert Collins
 symlink support updated to work  | 
43  | 
FP_MODE_COLUMN = 5  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
44  | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
45  | 
|
46  | 
||
47  | 
class HashCache(object):  | 
|
48  | 
"""Cache for looking up file SHA-1.  | 
|
49  | 
||
50  | 
    Files are considered to match the cached value if the fingerprint
 | 
|
51  | 
    of the file has not changed.  This includes its mtime, ctime,
 | 
|
52  | 
    device number, inode number, and size.  This should catch
 | 
|
53  | 
    modifications or replacement of the file by a new one.
 | 
|
54  | 
||
55  | 
    This may not catch modifications that do not change the file's
 | 
|
56  | 
    size and that occur within the resolution window of the
 | 
|
57  | 
    timestamps.  To handle this we specifically do not cache files
 | 
|
58  | 
    which have changed since the start of the present second, since
 | 
|
59  | 
    they could undetectably change again.
 | 
|
60  | 
||
61  | 
    This scheme may fail if the machine's clock steps backwards.
 | 
|
62  | 
    Don't do that.
 | 
|
63  | 
||
64  | 
    This does not canonicalize the paths passed in; that should be
 | 
|
65  | 
    done by the caller.
 | 
|
66  | 
||
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
67  | 
    _cache
 | 
68  | 
        Indexed by path, points to a two-tuple of the SHA-1 of the file.
 | 
|
69  | 
        and its fingerprint.
 | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
70  | 
|
71  | 
    stat_count
 | 
|
72  | 
        number of times files have been statted
 | 
|
73  | 
||
74  | 
    hit_count
 | 
|
75  | 
        number of times files have been retrieved from the cache, avoiding a
 | 
|
76  | 
        re-read
 | 
|
77  | 
        
 | 
|
78  | 
    miss_count
 | 
|
79  | 
        number of misses (times files have been completely re-read)
 | 
|
80  | 
    """
 | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
81  | 
needs_write = False  | 
82  | 
||
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
83  | 
def __init__(self, root, cache_file_name, mode=None):  | 
| 
1534.4.50
by Robert Collins
 Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running.  | 
84  | 
"""Create a hash cache in base dir, and set the file mode to mode."""  | 
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
85  | 
self.root = safe_unicode(root)  | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
86  | 
self.hit_count = 0  | 
87  | 
self.miss_count = 0  | 
|
88  | 
self.stat_count = 0  | 
|
89  | 
self.danger_count = 0  | 
|
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
90  | 
self.removed_count = 0  | 
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
91  | 
self.update_count = 0  | 
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
92  | 
self._cache = {}  | 
| 
1534.4.50
by Robert Collins
 Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running.  | 
93  | 
self._mode = mode  | 
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
94  | 
self._cache_file_name = safe_unicode(cache_file_name)  | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
95  | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
96  | 
def cache_file_name(self):  | 
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
97  | 
return self._cache_file_name  | 
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
98  | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
99  | 
def clear(self):  | 
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
100  | 
"""Discard all cached information.  | 
101  | 
||
102  | 
        This does not reset the counters."""
 | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
103  | 
if self._cache:  | 
104  | 
self.needs_write = True  | 
|
105  | 
self._cache = {}  | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
106  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
107  | 
def scan(self):  | 
108  | 
"""Scan all files and remove entries where the cache entry is obsolete.  | 
|
109  | 
        
 | 
|
110  | 
        Obsolete entries are those where the file has been modified or deleted
 | 
|
111  | 
        since the entry was inserted.        
 | 
|
112  | 
        """
 | 
|
| 
1534.4.50
by Robert Collins
 Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running.  | 
113  | 
        # FIXME optimisation opportunity, on linux [and check other oses]:
 | 
114  | 
        # rather than iteritems order, stat in inode order.
 | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
115  | 
prep = [(ce[1][3], path, ce) for (path, ce) in self._cache.iteritems()]  | 
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
116  | 
prep.sort()  | 
117  | 
||
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
118  | 
for inum, path, cache_entry in prep:  | 
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
119  | 
abspath = pathjoin(self.root, path)  | 
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
120  | 
fp = self._fingerprint(abspath)  | 
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
121  | 
self.stat_count += 1  | 
122  | 
||
123  | 
cache_fp = cache_entry[1]  | 
|
124  | 
||
125  | 
if (not fp) or (cache_fp != fp):  | 
|
126  | 
                # not here or not a regular file anymore
 | 
|
127  | 
self.removed_count += 1  | 
|
128  | 
self.needs_write = True  | 
|
129  | 
del self._cache[path]  | 
|
130  | 
||
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
131  | 
def get_sha1(self, path):  | 
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
132  | 
"""Return the sha1 of a file.  | 
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
133  | 
        """
 | 
| 
1534.4.51
by Robert Collins
 Test the disk layout of format3 working trees.  | 
134  | 
abspath = pathjoin(self.root, path)  | 
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
135  | 
self.stat_count += 1  | 
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
136  | 
file_fp = self._fingerprint(abspath)  | 
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
137  | 
|
138  | 
if not file_fp:  | 
|
139  | 
            # not a regular file or not existing
 | 
|
140  | 
if path in self._cache:  | 
|
141  | 
self.removed_count += 1  | 
|
142  | 
self.needs_write = True  | 
|
143  | 
del self._cache[path]  | 
|
144  | 
return None  | 
|
| 
953
by Martin Pool
 - refactor imports and stats for hashcache  | 
145  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
146  | 
if path in self._cache:  | 
147  | 
cache_sha1, cache_fp = self._cache[path]  | 
|
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
148  | 
else:  | 
149  | 
cache_sha1, cache_fp = None, None  | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
150  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
151  | 
if cache_fp == file_fp:  | 
| 
1845.1.2
by mbp at sourcefrog
 Use larger time window on hashcache to be safe with fractional times  | 
152  | 
            ## mutter("hashcache hit for %s %r -> %s", path, file_fp, cache_sha1)
 | 
153  | 
            ## mutter("now = %s", time.time())
 | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
154  | 
self.hit_count += 1  | 
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
155  | 
return cache_sha1  | 
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
156  | 
|
157  | 
self.miss_count += 1  | 
|
| 
1092.2.6
by Robert Collins
 symlink support updated to work  | 
158  | 
|
159  | 
mode = file_fp[FP_MODE_COLUMN]  | 
|
160  | 
if stat.S_ISREG(mode):  | 
|
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
161  | 
digest = self._really_sha1_file(abspath)  | 
| 
1092.2.6
by Robert Collins
 symlink support updated to work  | 
162  | 
elif stat.S_ISLNK(mode):  | 
163  | 
digest = sha.new(os.readlink(abspath)).hexdigest()  | 
|
164  | 
else:  | 
|
165  | 
raise BzrError("file %r: unknown file stat mode: %o"%(abspath,mode))  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
166  | 
|
| 
1845.1.2
by mbp at sourcefrog
 Use larger time window on hashcache to be safe with fractional times  | 
167  | 
        # window of 3 seconds to allow for 2s resolution on windows,
 | 
168  | 
        # unsynchronized file servers, etc.
 | 
|
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
169  | 
cutoff = self._cutoff_time()  | 
| 
1845.1.2
by mbp at sourcefrog
 Use larger time window on hashcache to be safe with fractional times  | 
170  | 
if file_fp[FP_MTIME_COLUMN] >= cutoff \  | 
171  | 
or file_fp[FP_CTIME_COLUMN] >= cutoff:  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
172  | 
            # changed too recently; can't be cached.  we can
 | 
173  | 
            # return the result and it could possibly be cached
 | 
|
174  | 
            # next time.
 | 
|
| 
1185.59.10
by Denys Duchier
 hashcache: new constants and improved comment  | 
175  | 
            #
 | 
176  | 
            # the point is that we only want to cache when we are sure that any
 | 
|
177  | 
            # subsequent modifications of the file can be detected.  If a
 | 
|
178  | 
            # modification neither changes the inode, the device, the size, nor
 | 
|
179  | 
            # the mode, then we can only distinguish it by time; therefore we
 | 
|
180  | 
            # need to let sufficient time elapse before we may cache this entry
 | 
|
181  | 
            # again.  If we didn't do this, then, for example, a very quick 1
 | 
|
182  | 
            # byte replacement in the file might go undetected.
 | 
|
| 
1845.1.2
by mbp at sourcefrog
 Use larger time window on hashcache to be safe with fractional times  | 
183  | 
            ## mutter('%r modified too recently; not caching', path)
 | 
184  | 
self.danger_count += 1  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
185  | 
if cache_fp:  | 
186  | 
self.removed_count += 1  | 
|
187  | 
self.needs_write = True  | 
|
188  | 
del self._cache[path]  | 
|
| 
846
by Martin Pool
 - start adding refactored/simplified hash cache  | 
189  | 
else:  | 
| 
1845.1.2
by mbp at sourcefrog
 Use larger time window on hashcache to be safe with fractional times  | 
190  | 
            ## mutter('%r added to cache: now=%f, mtime=%d, ctime=%d',
 | 
191  | 
            ##        path, time.time(), file_fp[FP_MTIME_COLUMN],
 | 
|
192  | 
            ##        file_fp[FP_CTIME_COLUMN])
 | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
193  | 
self.update_count += 1  | 
194  | 
self.needs_write = True  | 
|
195  | 
self._cache[path] = (digest, file_fp)  | 
|
196  | 
return digest  | 
|
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
197  | 
|
198  | 
def _really_sha1_file(self, abspath):  | 
|
199  | 
"""Calculate the SHA1 of a file by reading the full text"""  | 
|
200  | 
return sha_file(file(abspath, 'rb', buffering=65000))  | 
|
| 
954
by Martin Pool
 - separate out code that just scans the hash cache to find files that are possibly  | 
201  | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
202  | 
def write(self):  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
203  | 
"""Write contents of cache to file."""  | 
| 
1534.4.50
by Robert Collins
 Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running.  | 
204  | 
outf = AtomicFile(self.cache_file_name(), 'wb', new_mode=self._mode)  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
205  | 
try:  | 
| 
1908.4.8
by John Arbash Meinel
 Small tweak to hashcache to make it write out faster  | 
206  | 
outf.write(CACHE_HEADER)  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
207  | 
|
| 
860
by Martin Pool
 - refactor hashcache to use just one dictionary  | 
208  | 
for path, c in self._cache.iteritems():  | 
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
209  | 
assert '//' not in path, path  | 
| 
1908.4.8
by John Arbash Meinel
 Small tweak to hashcache to make it write out faster  | 
210  | 
line_info = [path.encode('utf-8'), '// ', c[0], ' ']  | 
211  | 
line_info.append(' '.join([str(fld) for fld in c[1]]))  | 
|
212  | 
line_info.append('\n')  | 
|
213  | 
outf.write(''.join(line_info))  | 
|
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
214  | 
outf.commit()  | 
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
215  | 
self.needs_write = False  | 
| 
1845.1.1
by mbp at sourcefrog
 Refactor and improve hashcache tests  | 
216  | 
            ## mutter("write hash cache: %s hits=%d misses=%d stat=%d recent=%d updates=%d",
 | 
217  | 
            ##        self.cache_file_name(), self.hit_count, self.miss_count,
 | 
|
218  | 
            ##        self.stat_count,
 | 
|
219  | 
            ##        self.danger_count, self.update_count)
 | 
|
| 
859
by Martin Pool
 - add HashCache.write and a simple test for it  | 
220  | 
finally:  | 
| 
1755.3.1
by Robert Collins
 Tune the time to build our kernel_like tree : make LocalTransport.put faster, AtomicFile faster, LocalTransport.append faster.  | 
221  | 
outf.close()  | 
| 
862
by Martin Pool
 - code to re-read hashcache from file  | 
222  | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
223  | 
def read(self):  | 
| 
862
by Martin Pool
 - code to re-read hashcache from file  | 
224  | 
"""Reinstate cache from file.  | 
225  | 
||
226  | 
        Overwrites existing cache.
 | 
|
227  | 
||
228  | 
        If the cache file has the wrong version marker, this just clears 
 | 
|
229  | 
        the cache."""
 | 
|
230  | 
self._cache = {}  | 
|
231  | 
||
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
232  | 
fn = self.cache_file_name()  | 
233  | 
try:  | 
|
| 
948
by Martin Pool
 - more buffering when reading/writing hashcache  | 
234  | 
inf = file(fn, 'rb', buffering=65000)  | 
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
235  | 
except IOError, e:  | 
| 
1185.31.4
by John Arbash Meinel
 Fixing mutter() calls to not have to do string processing.  | 
236  | 
mutter("failed to open %s: %s", fn, e)  | 
| 
1214
by Martin Pool
 - hashcache should be written out if it can't be read  | 
237  | 
            # better write it now so it is valid
 | 
238  | 
self.needs_write = True  | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
239  | 
            return
 | 
240  | 
||
| 
862
by Martin Pool
 - code to re-read hashcache from file  | 
241  | 
hdr = inf.readline()  | 
242  | 
if hdr != CACHE_HEADER:  | 
|
| 
1185.31.4
by John Arbash Meinel
 Fixing mutter() calls to not have to do string processing.  | 
243  | 
mutter('cache header marker not found at top of %s;'  | 
244  | 
' discarding cache', fn)  | 
|
| 
1214
by Martin Pool
 - hashcache should be written out if it can't be read  | 
245  | 
self.needs_write = True  | 
| 
862
by Martin Pool
 - code to re-read hashcache from file  | 
246  | 
            return
 | 
247  | 
||
248  | 
for l in inf:  | 
|
249  | 
pos = l.index('// ')  | 
|
250  | 
path = l[:pos].decode('utf-8')  | 
|
251  | 
if path in self._cache:  | 
|
252  | 
warning('duplicated path %r in cache' % path)  | 
|
253  | 
                continue
 | 
|
254  | 
||
255  | 
pos += 3  | 
|
256  | 
fields = l[pos:].split(' ')  | 
|
| 
1092.2.6
by Robert Collins
 symlink support updated to work  | 
257  | 
if len(fields) != 7:  | 
| 
862
by Martin Pool
 - code to re-read hashcache from file  | 
258  | 
warning("bad line in hashcache: %r" % l)  | 
259  | 
                continue
 | 
|
260  | 
||
261  | 
sha1 = fields[0]  | 
|
262  | 
if len(sha1) != 40:  | 
|
263  | 
warning("bad sha1 in hashcache: %r" % sha1)  | 
|
264  | 
                continue
 | 
|
265  | 
||
266  | 
fp = tuple(map(long, fields[1:]))  | 
|
267  | 
||
268  | 
self._cache[path] = (sha1, fp)  | 
|
269  | 
||
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
270  | 
self.needs_write = False  | 
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
271  | 
|
272  | 
def _cutoff_time(self):  | 
|
273  | 
"""Return cutoff time.  | 
|
274  | 
||
275  | 
        Files modified more recently than this time are at risk of being
 | 
|
276  | 
        undetectably modified and so can't be cached.
 | 
|
277  | 
        """
 | 
|
278  | 
return int(time.time()) - 3  | 
|
| 
866
by Martin Pool
 - use new path-based hashcache for WorkingTree- squash mtime/ctime to whole seconds- update and if necessary write out hashcache when WorkingTree object is created.  | 
279  | 
|
| 
1845.1.3
by Martin Pool
 Improvements to hashcache testing:  | 
280  | 
def _fingerprint(self, abspath):  | 
281  | 
try:  | 
|
282  | 
fs = os.lstat(abspath)  | 
|
283  | 
except OSError:  | 
|
284  | 
            # might be missing, etc
 | 
|
285  | 
return None  | 
|
286  | 
if stat.S_ISDIR(fs.st_mode):  | 
|
287  | 
return None  | 
|
288  | 
        # we discard any high precision because it's not reliable; perhaps we
 | 
|
289  | 
        # could do better on some systems?
 | 
|
290  | 
return (fs.st_size, long(fs.st_mtime),  | 
|
291  | 
long(fs.st_ctime), fs.st_ino, fs.st_dev, fs.st_mode)  |