2017-05-27 19:03:27 +02:00
|
|
|
from binascii import hexlify, unhexlify
|
|
|
|
from os import listdir, makedirs, path, remove, rmdir
|
|
|
|
import string
|
|
|
|
from threading import RLock
|
|
|
|
import time
|
|
|
|
|
|
|
|
from paths import lookupAppdataFolder
|
|
|
|
from storage import InventoryStorage, InventoryItem
|
|
|
|
|
2019-09-24 11:20:20 +02:00
|
|
|
|
2017-05-27 19:03:27 +02:00
|
|
|
class FilesystemInventory(InventoryStorage):
|
|
|
|
topDir = "inventory"
|
|
|
|
objectDir = "objects"
|
|
|
|
metadataFilename = "metadata"
|
|
|
|
dataFilename = "data"
|
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
super(self.__class__, self).__init__()
|
|
|
|
self.baseDir = path.join(lookupAppdataFolder(), FilesystemInventory.topDir)
|
|
|
|
for createDir in [self.baseDir, path.join(self.baseDir, "objects")]:
|
|
|
|
if path.exists(createDir):
|
|
|
|
if not path.isdir(createDir):
|
|
|
|
raise IOError("%s exists but it's not a directory" % (createDir))
|
|
|
|
else:
|
|
|
|
makedirs(createDir)
|
2019-09-24 11:20:20 +02:00
|
|
|
# Guarantees that two receiveDataThreads don't receive and process the same message
|
|
|
|
# concurrently (probably sent by a malicious individual)
|
|
|
|
self.lock = RLock()
|
2017-05-27 19:03:27 +02:00
|
|
|
self._inventory = {}
|
|
|
|
self._load()
|
|
|
|
|
|
|
|
def __contains__(self, hash):
|
|
|
|
retval = False
|
|
|
|
for streamDict in self._inventory.values():
|
|
|
|
if hash in streamDict:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
def __getitem__(self, hash):
|
|
|
|
for streamDict in self._inventory.values():
|
|
|
|
try:
|
|
|
|
retval = streamDict[hash]
|
|
|
|
except KeyError:
|
|
|
|
continue
|
|
|
|
if retval.payload is None:
|
|
|
|
retval = InventoryItem(retval.type, retval.stream, self.getData(hash), retval.expires, retval.tag)
|
|
|
|
return retval
|
|
|
|
raise KeyError(hash)
|
|
|
|
|
|
|
|
def __setitem__(self, hash, value):
|
|
|
|
with self.lock:
|
|
|
|
value = InventoryItem(*value)
|
|
|
|
try:
|
|
|
|
makedirs(path.join(self.baseDir, FilesystemInventory.objectDir, hexlify(hash)))
|
|
|
|
except OSError:
|
|
|
|
pass
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
with open(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hash),
|
|
|
|
FilesystemInventory.metadataFilename,
|
|
|
|
),
|
|
|
|
"w",
|
|
|
|
) as f:
|
2017-05-27 19:03:27 +02:00
|
|
|
f.write("%s,%s,%s,%s," % (value.type, value.stream, value.expires, hexlify(value.tag)))
|
2019-09-24 11:20:20 +02:00
|
|
|
with open(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hash),
|
|
|
|
FilesystemInventory.dataFilename,
|
|
|
|
),
|
|
|
|
"w",
|
|
|
|
) as f:
|
2017-05-27 19:03:27 +02:00
|
|
|
f.write(value.payload)
|
|
|
|
except IOError:
|
|
|
|
raise KeyError
|
|
|
|
try:
|
|
|
|
self._inventory[value.stream][hash] = value
|
|
|
|
except KeyError:
|
|
|
|
self._inventory[value.stream] = {}
|
|
|
|
self._inventory[value.stream][hash] = value
|
|
|
|
|
2017-06-21 12:17:40 +02:00
|
|
|
def delHashId(self, hash):
|
2017-05-27 19:03:27 +02:00
|
|
|
for stream in self._inventory.keys():
|
|
|
|
try:
|
|
|
|
del self._inventory[stream][hash]
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
with self.lock:
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
remove(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hash),
|
|
|
|
FilesystemInventory.metadataFilename))
|
2017-05-27 19:03:27 +02:00
|
|
|
except IOError:
|
|
|
|
pass
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
remove(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hash),
|
|
|
|
FilesystemInventory.dataFilename))
|
2017-05-27 19:03:27 +02:00
|
|
|
except IOError:
|
|
|
|
pass
|
|
|
|
try:
|
|
|
|
rmdir(path.join(self.baseDir, FilesystemInventory.objectDir, hexlify(hash)))
|
|
|
|
except IOError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
def __iter__(self):
|
|
|
|
elems = []
|
|
|
|
for streamDict in self._inventory.values():
|
2019-09-24 11:20:20 +02:00
|
|
|
elems.extend(streamDict.keys())
|
2017-05-27 19:03:27 +02:00
|
|
|
return elems.__iter__()
|
|
|
|
|
|
|
|
def __len__(self):
|
|
|
|
retval = 0
|
|
|
|
for streamDict in self._inventory.values():
|
|
|
|
retval += len(streamDict)
|
|
|
|
return retval
|
|
|
|
|
|
|
|
def _load(self):
|
|
|
|
newInventory = {}
|
|
|
|
for hashId in self.object_list():
|
|
|
|
try:
|
|
|
|
objectType, streamNumber, expiresTime, tag = self.getMetadata(hashId)
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
newInventory[streamNumber][hashId] = InventoryItem(
|
|
|
|
objectType, streamNumber, None, expiresTime, tag)
|
2017-05-27 19:03:27 +02:00
|
|
|
except KeyError:
|
|
|
|
newInventory[streamNumber] = {}
|
2019-09-24 11:20:20 +02:00
|
|
|
newInventory[streamNumber][hashId] = InventoryItem(
|
|
|
|
objectType, streamNumber, None, expiresTime, tag)
|
2017-05-27 19:03:27 +02:00
|
|
|
except KeyError:
|
|
|
|
print "error loading %s" % (hexlify(hashId))
|
|
|
|
pass
|
|
|
|
self._inventory = newInventory
|
2017-05-29 00:24:07 +02:00
|
|
|
# for i, v in self._inventory.items():
|
|
|
|
# print "loaded stream: %s, %i items" % (i, len(v))
|
2017-05-27 19:03:27 +02:00
|
|
|
|
|
|
|
def stream_list(self):
|
|
|
|
return self._inventory.keys()
|
|
|
|
|
|
|
|
def object_list(self):
|
|
|
|
return [unhexlify(x) for x in listdir(path.join(self.baseDir, FilesystemInventory.objectDir))]
|
|
|
|
|
|
|
|
def getData(self, hashId):
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
with open(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hashId),
|
|
|
|
FilesystemInventory.dataFilename,
|
|
|
|
),
|
|
|
|
"r",
|
|
|
|
) as f:
|
2017-05-27 19:03:27 +02:00
|
|
|
return f.read()
|
|
|
|
except IOError:
|
|
|
|
raise AttributeError
|
|
|
|
|
|
|
|
def getMetadata(self, hashId):
|
|
|
|
try:
|
2019-09-24 11:20:20 +02:00
|
|
|
with open(
|
|
|
|
path.join(
|
|
|
|
self.baseDir,
|
|
|
|
FilesystemInventory.objectDir,
|
|
|
|
hexlify(hashId),
|
|
|
|
FilesystemInventory.metadataFilename,
|
|
|
|
),
|
|
|
|
"r",
|
|
|
|
) as f:
|
2017-05-27 19:03:27 +02:00
|
|
|
objectType, streamNumber, expiresTime, tag, undef = string.split(f.read(), ",", 4)
|
|
|
|
return [int(objectType), int(streamNumber), int(expiresTime), unhexlify(tag)]
|
|
|
|
except IOError:
|
|
|
|
raise KeyError
|
|
|
|
|
|
|
|
def by_type_and_tag(self, objectType, tag):
|
2019-09-24 11:20:20 +02:00
|
|
|
"""Get a list of objects filtered by object type and tag"""
|
2017-05-27 19:03:27 +02:00
|
|
|
retval = []
|
|
|
|
for stream, streamDict in self._inventory:
|
|
|
|
for hashId, item in streamDict:
|
|
|
|
if item.type == objectType and item.tag == tag:
|
2019-09-24 11:20:20 +02:00
|
|
|
try:
|
2017-05-27 19:03:27 +02:00
|
|
|
if item.payload is None:
|
|
|
|
item.payload = self.getData(hashId)
|
|
|
|
except IOError:
|
|
|
|
continue
|
|
|
|
retval.append(InventoryItem(item.type, item.stream, item.payload, item.expires, item.tag))
|
|
|
|
return retval
|
|
|
|
|
|
|
|
def hashes_by_stream(self, stream):
|
|
|
|
try:
|
|
|
|
return self._inventory[stream].keys()
|
|
|
|
except KeyError:
|
|
|
|
return []
|
|
|
|
|
|
|
|
def unexpired_hashes_by_stream(self, stream):
|
|
|
|
t = int(time.time())
|
|
|
|
try:
|
|
|
|
return [x for x, value in self._inventory[stream].items() if value.expires > t]
|
|
|
|
except KeyError:
|
|
|
|
return []
|
|
|
|
|
|
|
|
def flush(self):
|
|
|
|
self._load()
|
|
|
|
|
|
|
|
def clean(self):
|
|
|
|
minTime = int(time.time()) - (60 * 60 * 30)
|
|
|
|
deletes = []
|
|
|
|
for stream, streamDict in self._inventory.items():
|
|
|
|
for hashId, item in streamDict.items():
|
|
|
|
if item.expires < minTime:
|
|
|
|
deletes.append(hashId)
|
|
|
|
for hashId in deletes:
|
2017-06-21 12:17:40 +02:00
|
|
|
self.delHashId(hashId)
|