2017-01-10 21:15:35 +01:00
import collections
2017-01-19 19:48:12 +01:00
from threading import current_thread , enumerate as threadingEnumerate , RLock
2017-03-19 22:08:00 +01:00
import Queue
2017-01-10 21:15:35 +01:00
import time
from helper_sql import *
from singleton import Singleton
@Singleton
class Inventory ( collections . MutableMapping ) :
def __init__ ( self ) :
super ( self . __class__ , self ) . __init__ ( )
self . _inventory = { } #of objects (like msg payloads and pubkey payloads) Does not include protocol headers (the first 24 bytes of each packet).
self . numberOfInventoryLookupsPerformed = 0
self . _streams = collections . defaultdict ( set ) # key = streamNumer, value = a set which holds the inventory object hashes that we are aware of. This is used whenever we receive an inv message from a peer to check to see what items are new to us. We don't delete things out of it; instead, the singleCleaner thread clears and refills it every couple hours.
2017-01-15 19:21:24 +01:00
self . lock = RLock ( ) # Guarantees that two receiveDataThreads don't receive and process the same message concurrently (probably sent by a malicious individual)
self . InventoryItem = collections . namedtuple ( ' InventoryItem ' , ' type stream payload expires tag ' )
2017-01-10 21:15:35 +01:00
def __contains__ ( self , hash ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
self . numberOfInventoryLookupsPerformed + = 1
if hash in self . _inventory :
return True
return bool ( sqlQuery ( ' SELECT 1 FROM inventory WHERE hash=? ' , hash ) )
def __getitem__ ( self , hash ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
if hash in self . _inventory :
return self . _inventory [ hash ]
rows = sqlQuery ( ' SELECT objecttype, streamnumber, payload, expirestime, tag FROM inventory WHERE hash=? ' , hash )
if not rows :
raise KeyError ( hash )
2017-01-15 19:21:24 +01:00
return self . InventoryItem ( * rows [ 0 ] )
2017-01-10 21:15:35 +01:00
def __setitem__ ( self , hash , value ) :
2017-01-15 19:21:24 +01:00
with self . lock :
value = self . InventoryItem ( * value )
2017-01-10 21:15:35 +01:00
self . _inventory [ hash ] = value
self . _streams [ value . stream ] . add ( hash )
def __delitem__ ( self , hash ) :
raise NotImplementedError
def __iter__ ( self ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
hashes = self . _inventory . keys ( ) [ : ]
2017-01-19 20:04:45 +01:00
hashes + = ( x for x , in sqlQuery ( ' SELECT hash FROM inventory ' ) )
2017-01-10 21:15:35 +01:00
return hashes . __iter__ ( )
def __len__ ( self ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
return len ( self . _inventory ) + sqlQuery ( ' SELECT count(*) FROM inventory ' ) [ 0 ] [ 0 ]
def by_type_and_tag ( self , type , tag ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
values = [ value for value in self . _inventory . values ( ) if value . type == type and value . tag == tag ]
2017-01-15 19:21:24 +01:00
values + = ( self . InventoryItem ( * value ) for value in sqlQuery ( ' SELECT objecttype, streamnumber, payload, expirestime, tag FROM inventory WHERE objecttype=? AND tag=? ' , type , tag ) )
2017-01-10 21:15:35 +01:00
return values
def hashes_by_stream ( self , stream ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
return self . _streams [ stream ]
def unexpired_hashes_by_stream ( self , stream ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
t = int ( time . time ( ) )
2017-01-19 20:04:45 +01:00
hashes = [ x for x , value in self . _inventory . items ( ) if value . stream == stream and value . expires > t ]
2017-01-10 21:15:35 +01:00
hashes + = ( payload for payload , in sqlQuery ( ' SELECT hash FROM inventory WHERE streamnumber=? AND expirestime>? ' , stream , t ) )
return hashes
def flush ( self ) :
2017-01-15 19:21:24 +01:00
with self . lock : # If you use both the inventoryLock and the sqlLock, always use the inventoryLock OUTSIDE of the sqlLock.
2017-01-10 21:15:35 +01:00
with SqlBulkExecute ( ) as sql :
2017-01-19 20:04:45 +01:00
for objectHash , value in self . _inventory . items ( ) :
sql . execute ( ' INSERT INTO inventory VALUES (?, ?, ?, ?, ?, ?) ' , objectHash , * value )
2017-01-10 21:15:35 +01:00
self . _inventory . clear ( )
def clean ( self ) :
2017-01-15 19:21:24 +01:00
with self . lock :
2017-01-10 21:15:35 +01:00
sqlExecute ( ' DELETE FROM inventory WHERE expirestime<? ' , int ( time . time ( ) ) - ( 60 * 60 * 3 ) )
self . _streams . clear ( )
2017-01-19 20:04:45 +01:00
for objectHash , value in self . items ( ) :
self . _streams [ value . stream ] . add ( objectHash )
2017-01-15 19:21:24 +01:00
2017-03-19 22:08:00 +01:00
class PendingDownloadQueue ( Queue . Queue ) :
2017-01-19 19:48:12 +01:00
# keep a track of objects that have been advertised to us but we haven't downloaded them yet
2017-03-19 22:08:00 +01:00
def __init__ ( self , maxsize = 0 ) :
Queue . Queue . __init__ ( self , maxsize )
2017-01-15 22:21:19 +01:00
self . stopped = False
2017-03-19 22:08:00 +01:00
self . pendingSize = 0
def task_done ( self ) :
Queue . Queue . task_done ( self )
if self . pendingSize > 0 :
self . pendingSize - = 1
def get ( self , block = True , timeout = None ) :
retval = Queue . Queue . get ( self , block , timeout )
# no exception was raised
if not self . stopped :
self . pendingSize + = 1
return retval
@staticmethod
def totalSize ( ) :
size = 0
for thread in threadingEnumerate ( ) :
if thread . isAlive ( ) and hasattr ( thread , ' downloadQueue ' ) :
size + = thread . downloadQueue . qsize ( ) + thread . downloadQueue . pendingSize
return size
@staticmethod
def stop ( ) :
for thread in threadingEnumerate ( ) :
if thread . isAlive ( ) and hasattr ( thread , ' downloadQueue ' ) :
thread . downloadQueue . stopped = True
thread . downloadQueue . pendingSize = 0
2017-01-19 19:48:12 +01:00
class PendingUploadDeadlineException ( Exception ) :
pass
@Singleton
class PendingUpload ( object ) :
# keep a track of objects that we have created but haven't distributed yet
def __init__ ( self ) :
super ( self . __class__ , self ) . __init__ ( )
self . lock = RLock ( )
self . hashes = { }
# end by this time in any case
self . deadline = 0
self . maxLen = 0
2017-03-01 10:05:08 +01:00
# during shutdown, wait up to 20 seconds to finish uploading
self . shutdownWait = 20
# forget tracking objects after 60 seconds
self . objectWait = 60
# wait 10 seconds between clears
self . clearDelay = 10
self . lastCleared = time . time ( )
2017-01-19 19:48:12 +01:00
def add ( self , objectHash = None ) :
with self . lock :
# add a new object into existing thread lists
if objectHash :
if objectHash not in self . hashes :
2017-03-01 10:05:08 +01:00
self . hashes [ objectHash ] = { ' created ' : time . time ( ) , ' sendCount ' : 0 , ' peers ' : [ ] }
2017-01-19 19:48:12 +01:00
for thread in threadingEnumerate ( ) :
if thread . isAlive ( ) and hasattr ( thread , ' peer ' ) and \
2017-03-01 10:05:08 +01:00
thread . peer not in self . hashes [ objectHash ] [ ' peers ' ] :
self . hashes [ objectHash ] [ ' peers ' ] . append ( thread . peer )
2017-01-19 19:48:12 +01:00
# add all objects into the current thread
else :
2017-01-19 20:04:45 +01:00
for objectHash in self . hashes :
2017-03-01 10:05:08 +01:00
if current_thread ( ) . peer not in self . hashes [ objectHash ] [ ' peers ' ] :
self . hashes [ objectHash ] [ ' peers ' ] . append ( current_thread ( ) . peer )
2017-01-19 19:48:12 +01:00
def len ( self ) :
2017-03-01 10:05:08 +01:00
self . clearHashes ( )
2017-01-19 19:48:12 +01:00
with self . lock :
2017-03-01 10:05:08 +01:00
return sum ( 1
for x in self . hashes if ( self . hashes [ x ] [ ' created ' ] + self . objectWait < time . time ( ) or
self . hashes [ x ] [ ' sendCount ' ] == 0 ) )
2017-01-19 19:48:12 +01:00
def _progress ( self ) :
with self . lock :
2017-03-01 10:05:08 +01:00
return float ( sum ( len ( self . hashes [ x ] [ ' peers ' ] )
for x in self . hashes if ( self . hashes [ x ] [ ' created ' ] + self . objectWait < time . time ( ) ) or
self . hashes [ x ] [ ' sendCount ' ] == 0 ) )
2017-01-19 19:48:12 +01:00
2017-03-01 10:05:08 +01:00
def progress ( self , raiseDeadline = True ) :
2017-01-19 19:48:12 +01:00
if self . maxLen < self . _progress ( ) :
self . maxLen = self . _progress ( )
if self . deadline < time . time ( ) :
2017-03-01 10:05:08 +01:00
if self . deadline > 0 and raiseDeadline :
2017-01-19 19:48:12 +01:00
raise PendingUploadDeadlineException
self . deadline = time . time ( ) + 20
try :
return 1.0 - self . _progress ( ) / self . maxLen
except ZeroDivisionError :
return 1.0
2017-03-01 10:05:08 +01:00
def clearHashes ( self , objectHash = None ) :
if objectHash is None :
if self . lastCleared > time . time ( ) - self . clearDelay :
return
objects = self . hashes . keys ( )
else :
objects = objectHash ,
with self . lock :
for i in objects :
try :
if self . hashes [ i ] [ ' sendCount ' ] > 0 and (
len ( self . hashes [ i ] [ ' peers ' ] ) == 0 or
self . hashes [ i ] [ ' created ' ] + self . objectWait < time . time ( ) ) :
del self . hashes [ i ]
except KeyError :
pass
self . lastCleared = time . time ( )
def delete ( self , objectHash = None ) :
2017-01-19 19:48:12 +01:00
if not hasattr ( current_thread ( ) , ' peer ' ) :
return
2017-03-01 10:05:08 +01:00
if objectHash is None :
return
2017-01-19 19:48:12 +01:00
with self . lock :
2017-03-01 10:05:08 +01:00
try :
if objectHash in self . hashes and current_thread ( ) . peer in self . hashes [ objectHash ] [ ' peers ' ] :
self . hashes [ objectHash ] [ ' sendCount ' ] + = 1
self . hashes [ objectHash ] [ ' peers ' ] . remove ( current_thread ( ) . peer )
except KeyError :
2017-03-02 15:02:51 +01:00
pass
2017-03-01 10:05:08 +01:00
self . clearHashes ( objectHash )
2017-01-19 19:48:12 +01:00
def stop ( self ) :
with self . lock :
self . hashes = { }
def threadEnd ( self ) :
2017-03-01 10:05:08 +01:00
with self . lock :
for objectHash in self . hashes :
try :
if current_thread ( ) . peer in self . hashes [ objectHash ] [ ' peers ' ] :
self . hashes [ objectHash ] [ ' peers ' ] . remove ( current_thread ( ) . peer )
except KeyError :
pass
self . clearHashes ( )