X-Git-Url: https://git.mxchange.org/?a=blobdiff_plain;ds=sidebyside;f=apt_dht%2FMirrorManager.py;h=b4995e9445a1a9ca175399430b4196b5582a1e1e;hb=85d160514bda0e66058f0fb1ac70c8cf52b4435d;hp=67bb183acadd0b130c9890eb45f4a90c1bb02614;hpb=20478a588c72621406547d1f1dbc24c231e02d5a;p=quix0rs-apt-p2p.git diff --git a/apt_dht/MirrorManager.py b/apt_dht/MirrorManager.py index 67bb183..b4995e9 100644 --- a/apt_dht/MirrorManager.py +++ b/apt_dht/MirrorManager.py @@ -1,30 +1,88 @@ +from urlparse import urlparse import os -from twisted.python import log +from twisted.python import log, filepath from twisted.internet import defer from twisted.trial import unittest +from twisted.web2 import stream +from twisted.web2.http import splitHostPort from AptPackages import AptPackages aptpkg_dir='.apt-dht' +class MirrorError(Exception): + """Exception raised when there's a problem with the mirror.""" + +class ProxyFileStream(stream.SimpleStream): + """Saves a stream to a file while providing a new stream.""" + + def __init__(self, stream, outFile, modtime = None): + """Initializes the proxy. + + @type stream: C{twisted.web2.stream.IByteStream} + @param stream: the input stream to read from + @type outFile: C{twisted.python.filepath.FilePath} + @param outFile: the file to write to + @type modtime: C{int} + @param modtime: the modification time to set for the file + """ + self.stream = stream + self.outFile = outFile + self.openFile = outFile.open('w') + self.modtime = modtime + self.length = self.stream.length + self.start = 0 + + def _done(self): + """Close the output file.""" + if not self.openFile.closed: + self.openFile.close() + if self.modtime: + os.utime(self.outFile.path, (self.modtime, self.modtime)) + + def read(self): + """Read some data from the stream.""" + if self.openFile.closed: + return None + + data = self.stream.read() + if isinstance(data, defer.Deferred): + data.addCallbacks(self._write, self._done) + return data + + self._write(data) + return data + + def _write(self, data): + """Write the stream data to the file and return it for others to use.""" + if data is None: + self._done() + return data + + self.openFile.write(data) + return data + + def close(self): + """Clean everything up and return None to future reads.""" + self.length = 0 + self._done() + self.stream.close() + class MirrorManager: """Manages all requests for mirror objects.""" def __init__(self, cache_dir): self.cache_dir = cache_dir + self.cache = filepath.FilePath(self.cache_dir) self.apt_caches = {} - def extractPath(self, path): - site, path = path.split('/',1) - if not site: - site, path = path.split('/',1) - path = '/'+path - - # Make sure a port is included for consistency - if site.find(':') < 0: - site = site + ":80" + def extractPath(self, url): + parsed = urlparse(url) + host, port = splitHostPort(parsed[0], parsed[1]) + site = host + ":" + str(port) + path = parsed[2] i = max(path.rfind('/dists/'), path.rfind('/pool/')) if i >= 0: @@ -58,22 +116,47 @@ class MirrorManager: site_cache = os.path.join(self.cache_dir, aptpkg_dir, 'mirrors', site + baseDir.replace('/', '_')) self.apt_caches[site][baseDir] = AptPackages(site_cache) - def updatedFile(self, path, file_path): - site, baseDir, path = self.extractPath(path) + def updatedFile(self, url, file_path): + site, baseDir, path = self.extractPath(url) self.init(site, baseDir) self.apt_caches[site][baseDir].file_updated(path, file_path) - def findHash(self, path): - site, baseDir, path = self.extractPath(path) + def findHash(self, url): + site, baseDir, path = self.extractPath(url) if site in self.apt_caches and baseDir in self.apt_caches[site]: return self.apt_caches[site][baseDir].findHash(path) d = defer.Deferred() - d.errback("Not Found") + d.errback(MirrorError("Site Not Found")) return d + def save_file(self, response, hash, size, url): + """Save a downloaded file to the cache and stream it.""" + log.msg('Returning file: %s' % url) + + parsed = urlparse(url) + destFile = self.cache.preauthChild(parsed[1] + parsed[2]) + log.msg('Saving returned %r byte file to cache: %s' % (response.stream.length, destFile.path)) + + if destFile.exists(): + log.msg('File already exists, removing: %s' % destFile.path) + destFile.remove() + else: + destFile.parent().makedirs() + + orig_stream = response.stream + response.stream = ProxyFileStream(orig_stream, destFile, response.headers.getHeader('Last-Modified')) + return response + + def save_error(self, failure, url): + """An error has occurred in downloadign or saving the file.""" + log.msg('Error occurred downloading %s' % url) + log.err(failure) + return failure + class TestMirrorManager(unittest.TestCase): """Unit tests for the mirror manager.""" + timeout = 20 pending_calls = [] client = None @@ -81,16 +164,21 @@ class TestMirrorManager(unittest.TestCase): self.client = MirrorManager('/tmp') def test_extractPath(self): - site, baseDir, path = self.client.extractPath('/ftp.us.debian.org/debian/dists/unstable/Release') + site, baseDir, path = self.client.extractPath('http://ftp.us.debian.org/debian/dists/unstable/Release') self.failUnless(site == "ftp.us.debian.org:80", "no match: %s" % site) self.failUnless(baseDir == "/debian", "no match: %s" % baseDir) self.failUnless(path == "/dists/unstable/Release", "no match: %s" % path) - site, baseDir, path = self.client.extractPath('/ftp.us.debian.org:16999/debian/pool/d/dpkg/dpkg_1.2.1-1.tar.gz') + site, baseDir, path = self.client.extractPath('http://ftp.us.debian.org:16999/debian/pool/d/dpkg/dpkg_1.2.1-1.tar.gz') self.failUnless(site == "ftp.us.debian.org:16999", "no match: %s" % site) self.failUnless(baseDir == "/debian", "no match: %s" % baseDir) self.failUnless(path == "/pool/d/dpkg/dpkg_1.2.1-1.tar.gz", "no match: %s" % path) + site, baseDir, path = self.client.extractPath('http://debian.camrdale.org/dists/unstable/Release') + self.failUnless(site == "debian.camrdale.org:80", "no match: %s" % site) + self.failUnless(baseDir == "", "no match: %s" % baseDir) + self.failUnless(path == "/dists/unstable/Release", "no match: %s" % path) + def verifyHash(self, found_hash, path, true_hash): self.failUnless(found_hash[0] == true_hash, "%s hashes don't match: %s != %s" % (path, found_hash[0], true_hash)) @@ -103,12 +191,12 @@ class TestMirrorManager(unittest.TestCase): self.releaseFile = f break - self.client.updatedFile('/' + self.releaseFile.replace('_','/'), + self.client.updatedFile('http://' + self.releaseFile.replace('_','/'), '/var/lib/apt/lists/' + self.releaseFile) - self.client.updatedFile('/' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + + self.client.updatedFile('http://' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + self.packagesFile[self.packagesFile.find('_dists_')+1:].replace('_','/'), '/var/lib/apt/lists/' + self.packagesFile) - self.client.updatedFile('/' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + + self.client.updatedFile('http://' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + self.sourcesFile[self.sourcesFile.find('_dists_')+1:].replace('_','/'), '/var/lib/apt/lists/' + self.sourcesFile) @@ -118,7 +206,7 @@ class TestMirrorManager(unittest.TestCase): '/var/lib/apt/lists/' + self.releaseFile + ' | grep -E " main/binary-i386/Packages.bz2$"' ' | head -n 1 | cut -d\ -f 2').read().rstrip('\n') - idx_path = '/' + self.releaseFile.replace('_','/')[:-7] + 'main/binary-i386/Packages.bz2' + idx_path = 'http://' + self.releaseFile.replace('_','/')[:-7] + 'main/binary-i386/Packages.bz2' d = self.client.findHash(idx_path) d.addCallback(self.verifyHash, idx_path, idx_hash) @@ -127,7 +215,7 @@ class TestMirrorManager(unittest.TestCase): '/var/lib/apt/lists/' + self.packagesFile + ' | grep -E "^SHA1:" | head -n 1' + ' | cut -d\ -f 2').read().rstrip('\n') - pkg_path = '/' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + \ + pkg_path = 'http://' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + \ os.popen('grep -A 30 -E "^Package: dpkg$" ' + '/var/lib/apt/lists/' + self.packagesFile + ' | grep -E "^Filename:" | head -n 1' + @@ -150,7 +238,7 @@ class TestMirrorManager(unittest.TestCase): ' | cut -d\ -f 4').read().split('\n')[:-1] for i in range(len(src_hashes)): - src_path = '/' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + src_dir + '/' + src_paths[i] + src_path = 'http://' + self.releaseFile[:self.releaseFile.find('_dists_')+1].replace('_','/') + src_dir + '/' + src_paths[i] d = self.client.findHash(src_path) d.addCallback(self.verifyHash, src_path, src_hashes[i]) @@ -158,12 +246,12 @@ class TestMirrorManager(unittest.TestCase): '/var/lib/apt/lists/' + self.releaseFile + ' | grep -E " main/source/Sources.bz2$"' ' | head -n 1 | cut -d\ -f 2').read().rstrip('\n') - idx_path = '/' + self.releaseFile.replace('_','/')[:-7] + 'main/source/Sources.bz2' + idx_path = 'http://' + self.releaseFile.replace('_','/')[:-7] + 'main/source/Sources.bz2' d = self.client.findHash(idx_path) d.addCallback(self.verifyHash, idx_path, idx_hash) - d.addCallback(lastDefer.callback) + d.addBoth(lastDefer.callback) return lastDefer def tearDown(self):