Mercurial > hg > mercurial-crew-with-dirclash
view mercurial/streamclone.py @ 5377:756a43a30e34
convert: readd --filemap
To handle merges correctly, this revision adds a filemap_source class
that wraps a converter_source and does the work necessary to calculate
the subgraph we're interested in.
The wrapped converter_source must provide a new getchangedfiles method
that, given a revision rev, and an index N, returns the list of files
that are different in rev and its Nth parent.
The implementation depends on the ability to skip some revisions and to
change the parents field of the commit objects that we returned earlier.
To make the conversion restartable, we assume the revisons in the
revmapfile are topologically sorted.
author | Alexis S. L. Carvalho <alexis@cecm.usp.br> |
---|---|
date | Thu, 04 Oct 2007 23:21:37 -0300 |
parents | 97b734fb9c6f |
children | 5105b119edd2 |
line wrap: on
line source
# streamclone.py - streaming clone server support for mercurial # # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com> # # This software may be used and distributed according to the terms # of the GNU General Public License, incorporated herein by reference. from i18n import _ import os, stat, util, lock # if server supports streaming clone, it advertises "stream" # capability with value that is version+flags of repo it is serving. # client only streams if it can read that repo format. def walkrepo(root): '''iterate over metadata files in repository. walk in natural (sorted) order. yields 2-tuples: name of .d or .i file, size of file.''' strip_count = len(root) + len(os.sep) def walk(path, recurse): ents = os.listdir(path) ents.sort() for e in ents: pe = os.path.join(path, e) st = os.lstat(pe) if stat.S_ISDIR(st.st_mode): if recurse: for x in walk(pe, True): yield x else: if not stat.S_ISREG(st.st_mode) or len(e) < 2: continue sfx = e[-2:] if sfx in ('.d', '.i'): yield pe[strip_count:], st.st_size # write file data first for x in walk(os.path.join(root, 'data'), True): yield x # write manifest before changelog meta = list(walk(root, False)) meta.sort() meta.reverse() for x in meta: yield x # stream file format is simple. # # server writes out line that says how many files, how many total # bytes. separator is ascii space, byte counts are strings. # # then for each file: # # server writes out line that says file name, how many bytes in # file. separator is ascii nul, byte count is string. # # server writes out raw file data. def stream_out(repo, fileobj, untrusted=False): '''stream out all metadata files in repository. writes to file-like object, must support write() and optional flush().''' if not repo.ui.configbool('server', 'uncompressed', untrusted=untrusted): fileobj.write('1\n') return # get consistent snapshot of repo. lock during scan so lock not # needed while we stream, and commits can happen. lock = None try: try: repolock = repo.lock() except (lock.LockHeld, lock.LockUnavailable), inst: repo.ui.warn('locking the repository failed: %s\n' % (inst,)) fileobj.write('2\n') return fileobj.write('0\n') repo.ui.debug('scanning\n') entries = [] total_bytes = 0 for name, size in walkrepo(repo.spath): name = repo.decodefn(util.pconvert(name)) entries.append((name, size)) total_bytes += size finally: del repolock repo.ui.debug('%d files, %d bytes to transfer\n' % (len(entries), total_bytes)) fileobj.write('%d %d\n' % (len(entries), total_bytes)) for name, size in entries: repo.ui.debug('sending %s (%d bytes)\n' % (name, size)) fileobj.write('%s\0%d\n' % (name, size)) for chunk in util.filechunkiter(repo.sopener(name), limit=size): fileobj.write(chunk) flush = getattr(fileobj, 'flush', None) if flush: flush()