X-Git-Url: https://git.mxchange.org/?a=blobdiff_plain;f=apt_dht%2Fapt_dht.py;h=94260bd4c396a4c216e2952c9094ae2a1cdaf0cb;hb=d900237088b7832d2554c31b7436977bc5669348;hp=3679685e5f70ad873b2076bad04f21265b6786e8;hpb=8297bc9a2aa8132ea1a7363761d9d5c73a1efca2;p=quix0rs-apt-p2p.git diff --git a/apt_dht/apt_dht.py b/apt_dht/apt_dht.py index 3679685..94260bd 100644 --- a/apt_dht/apt_dht.py +++ b/apt_dht/apt_dht.py @@ -1,11 +1,11 @@ from binascii import b2a_hex from urlparse import urlunparse -import os, re +import os, re, sha -from twisted.internet import defer -from twisted.web2 import server, http, http_headers -from twisted.python import log +from twisted.internet import defer, reactor +from twisted.web2 import server, http, http_headers, static +from twisted.python import log, failure from twisted.python.filepath import FilePath from apt_dht_conf import config @@ -15,7 +15,10 @@ from MirrorManager import MirrorManager from CacheManager import CacheManager from Hash import HashObject from db import DB -from util import findMyIPAddr +from util import findMyIPAddr, compact + +DHT_PIECES = 4 +TORRENT_PIECES = 70 download_dir = 'cache' @@ -29,77 +32,141 @@ class AptDHT: self.dht = dht self.dht.loadConfig(config, config.get('DEFAULT', 'DHT')) self.dht.join().addCallbacks(self.joinComplete, self.joinError) - self.http_server = TopLevel(self.cache_dir.child(download_dir), self) - self.http_site = server.Site(self.http_server) + self.http_server = TopLevel(self.cache_dir.child(download_dir), self.db, self) + self.getHTTPFactory = self.http_server.getHTTPFactory self.peers = PeerManager() - self.mirrors = MirrorManager(self.cache_dir) - self.cache = CacheManager(self.cache_dir.child(download_dir), self.db, self) - self.my_addr = None - - def getSite(self): - return self.http_site + self.mirrors = MirrorManager(self.cache_dir, config.gettime('DEFAULT', 'UNLOAD_PACKAGES_CACHE')) + other_dirs = [FilePath(f) for f in config.getstringlist('DEFAULT', 'OTHER_DIRS')] + self.cache = CacheManager(self.cache_dir.child(download_dir), self.db, other_dirs, self) + self.my_contact = None def joinComplete(self, result): - self.my_addr = findMyIPAddr(result, config.getint(config.get('DEFAULT', 'DHT'), 'PORT')) - if not self.my_addr: + my_addr = findMyIPAddr(result, + config.getint(config.get('DEFAULT', 'DHT'), 'PORT'), + config.getboolean('DEFAULT', 'LOCAL_OK')) + if not my_addr: raise RuntimeError, "IP address for this machine could not be found" + self.my_contact = compact(my_addr, config.getint('DEFAULT', 'PORT')) + self.cache.scanDirectories() + reactor.callLater(60, self.refreshFiles) def joinError(self, failure): log.msg("joining DHT failed miserably") log.err(failure) raise RuntimeError, "IP address for this machine could not be found" - def check_freshness(self, path, modtime, resp): + def refreshFiles(self): + """Refresh any files in the DHT that are about to expire.""" + expireAfter = config.gettime('DEFAULT', 'KEY_REFRESH') + hashes = self.db.expiredHashes(expireAfter) + if len(hashes.keys()) > 0: + log.msg('Refreshing the keys of %d DHT values' % len(hashes.keys())) + self._refreshFiles(None, hashes) + + def _refreshFiles(self, result, hashes): + if result is not None: + log.msg('Storage resulted in: %r' % result) + + if hashes: + raw_hash = hashes.keys()[0] + self.db.refreshHash(raw_hash) + hash = HashObject(raw_hash, pieces = hashes[raw_hash]['pieces']) + del hashes[raw_hash] + storeDefer = self.store(hash) + storeDefer.addBoth(self._refreshFiles, hashes) + else: + reactor.callLater(60, self.refreshFiles) + + def check_freshness(self, req, path, modtime, resp): log.msg('Checking if %s is still fresh' % path) - d = self.peers.get([path], "HEAD", modtime) - d.addCallback(self.check_freshness_done, path, resp) + d = self.peers.get('', path, method = "HEAD", modtime = modtime) + d.addCallback(self.check_freshness_done, req, path, resp) return d - def check_freshness_done(self, resp, path, orig_resp): + def check_freshness_done(self, resp, req, path, orig_resp): if resp.code == 304: log.msg('Still fresh, returning: %s' % path) return orig_resp else: log.msg('Stale, need to redownload: %s' % path) - return self.get_resp(path) + return self.get_resp(req, path) - def get_resp(self, path): + def get_resp(self, req, path): d = defer.Deferred() log.msg('Trying to find hash for %s' % path) findDefer = self.mirrors.findHash(path) findDefer.addCallbacks(self.findHash_done, self.findHash_error, - callbackArgs=(path, d), errbackArgs=(path, d)) + callbackArgs=(req, path, d), errbackArgs=(req, path, d)) findDefer.addErrback(log.err) return d - def findHash_error(self, failure, path, d): + def findHash_error(self, failure, req, path, d): log.err(failure) - self.findHash_done(HashObject(), path, d) + self.findHash_done(HashObject(), req, path, d) - def findHash_done(self, hash, path, d): + def findHash_done(self, hash, req, path, d): if hash.expected() is None: log.msg('Hash for %s was not found' % path) self.lookupHash_done([], hash, path, d) else: log.msg('Found hash %s for %s' % (hash.hexexpected(), path)) - # Lookup hash from DHT - key = hash.normexpected(bits = config.getint(config.get('DEFAULT', 'DHT'), 'HASH_LENGTH')) - lookupDefer = self.dht.getValue(key) - lookupDefer.addCallback(self.lookupHash_done, hash, path, d) - def lookupHash_done(self, locations, hash, path, d): + # Lookup hash in cache + locations = self.db.lookupHash(hash.expected(), filesOnly = True) + self.getCachedFile(hash, req, path, d, locations) + + def getCachedFile(self, hash, req, path, d, locations): if not locations: + log.msg('Failed to return file from cache: %s' % path) + self.lookupHash(hash, path, d) + return + + # Get the first possible location from the list + file = locations.pop(0)['path'] + log.msg('Returning cached file: %s' % file.path) + + # Get it's response + resp = static.File(file.path).renderHTTP(req) + if isinstance(resp, defer.Deferred): + resp.addBoth(self._getCachedFile, hash, req, path, d, locations) + else: + self._getCachedFile(resp, hash, req, path, d, locations) + + def _getCachedFile(self, resp, hash, req, path, d, locations): + if isinstance(resp, failure.Failure): + log.msg('Got error trying to get cached file') + log.err() + # Try the next possible location + self.getCachedFile(hash, req, path, d, locations) + return + + log.msg('Cached response: %r' % resp) + + if resp.code >= 200 and resp.code < 400: + d.callback(resp) + else: + # Try the next possible location + self.getCachedFile(hash, req, path, d, locations) + + def lookupHash(self, hash, path, d): + log.msg('Looking up hash in DHT for file: %s' % path) + key = hash.normexpected(bits = config.getint(config.get('DEFAULT', 'DHT'), 'HASH_LENGTH')) + lookupDefer = self.dht.getValue(key) + lookupDefer.addCallback(self.lookupHash_done, hash, path, d) + + def lookupHash_done(self, values, hash, path, d): + if not values: log.msg('Peers for %s were not found' % path) - getDefer = self.peers.get([path]) + getDefer = self.peers.get(hash, path) getDefer.addCallback(self.cache.save_file, hash, path) getDefer.addErrback(self.cache.save_error, path) getDefer.addCallbacks(d.callback, d.errback) else: - log.msg('Found peers for %s: %r' % (path, locations)) + log.msg('Found peers for %s: %r' % (path, values)) # Download from the found peers - getDefer = self.peers.get(locations) + getDefer = self.peers.get(hash, path, values) getDefer.addCallback(self.check_response, hash, path) getDefer.addCallback(self.cache.save_file, hash, path) getDefer.addErrback(self.cache.save_error, path) @@ -108,21 +175,56 @@ class AptDHT: def check_response(self, response, hash, path): if response.code < 200 or response.code >= 300: log.msg('Download from peers failed, going to direct download: %s' % path) - getDefer = self.peers.get([path]) + getDefer = self.peers.get(hash, path) return getDefer return response - def new_cached_file(self, url, file_path, hash, urlpath): - self.mirrors.updatedFile(url, file_path) + def new_cached_file(self, file_path, hash, new_hash, url = None, forceDHT = False): + """Add a newly cached file to the appropriate places. + + If the file was downloaded, set url to the path it was downloaded for. + Doesn't add a file to the DHT unless a hash was found for it + (but does add it anyway if forceDHT is True). + """ + if url: + self.mirrors.updatedFile(url, file_path) - if self.my_addr and hash: - site = self.my_addr + ':' + str(config.getint('DEFAULT', 'PORT')) - full_path = urlunparse(('http', site, urlpath, None, None, None)) - key = hash.norm(bits = config.getint(config.get('DEFAULT', 'DHT'), 'HASH_LENGTH')) - storeDefer = self.dht.storeValue(key, full_path) - storeDefer.addCallback(self.store_done, full_path) - storeDefer.addErrback(log.err) + if self.my_contact and hash and new_hash and (hash.expected() is not None or forceDHT): + return self.store(hash) + return None + + def store(self, hash): + """Add a file to the DHT.""" + key = hash.norm(bits = config.getint(config.get('DEFAULT', 'DHT'), 'HASH_LENGTH')) + value = {'c': self.my_contact} + pieces = hash.pieceDigests() + if len(pieces) <= 1: + pass + elif len(pieces) <= DHT_PIECES: + value['t'] = {'t': ''.join(pieces)} + elif len(pieces) <= TORRENT_PIECES: + s = sha.new().update(''.join(pieces)) + value['h'] = s.digest() + else: + s = sha.new().update(''.join(pieces)) + value['l'] = s.digest() + storeDefer = self.dht.storeValue(key, value) + storeDefer.addCallback(self.store_done, hash) + return storeDefer + + def store_done(self, result, hash): + log.msg('Added %s to the DHT: %r' % (hash.hexdigest(), result)) + pieces = hash.pieceDigests() + if len(pieces) > DHT_PIECES and len(pieces) <= TORRENT_PIECES: + s = sha.new().update(''.join(pieces)) + key = s.digest() + value = {'t': ''.join(pieces)} + storeDefer = self.dht.storeValue(key, value) + storeDefer.addCallback(self.store_torrent_done, key) + return storeDefer + return result - def store_done(self, result, path): - log.msg('Added %s to the DHT: %r' % (path, result)) - \ No newline at end of file + def store_torrent_done(self, result, key): + log.msg('Added torrent string %s to the DHT: %r' % (b2ahex(key), result)) + return result + \ No newline at end of file