From e4c2c1be16f6eea729014c5b48c426dc2fd9ed26 Mon Sep 17 00:00:00 2001 From: Lee Miller Date: Sun, 27 Aug 2023 01:55:05 +0300 Subject: [PATCH] Make load_data a static method in manager, use ascii while loading nodes csv. --- minode/main.py | 54 ---------------------------------------------- minode/manager.py | 55 +++++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 55 insertions(+), 54 deletions(-) diff --git a/minode/main.py b/minode/main.py index ace0781..7b52796 100644 --- a/minode/main.py +++ b/minode/main.py @@ -2,11 +2,9 @@ """Functions for starting the program""" import argparse import base64 -import csv import logging import multiprocessing import os -import pickle import signal import socket @@ -102,56 +100,6 @@ def parse_arguments(): # pylint: disable=too-many-branches,too-many-statements shared.i2p_transient = True -def load_data(): - """Loads initial nodes and data, stored in files between sessions""" - try: - with open( - os.path.join(shared.data_directory, 'objects.pickle'), 'br' - ) as src: - shared.objects = pickle.load(src) - except FileNotFoundError: - pass # first start - except Exception: - logging.warning( - 'Error while loading objects from disk.', exc_info=True) - - try: - with open( - os.path.join(shared.data_directory, 'nodes.pickle'), 'br' - ) as src: - shared.node_pool = pickle.load(src) - except FileNotFoundError: - pass - except Exception: - logging.warning('Error while loading nodes from disk.', exc_info=True) - - try: - with open( - os.path.join(shared.data_directory, 'i2p_nodes.pickle'), 'br' - ) as src: - shared.i2p_node_pool = pickle.load(src) - except FileNotFoundError: - pass - except Exception: - logging.warning('Error while loading nodes from disk.', exc_info=True) - - with open( - os.path.join(shared.source_directory, 'core_nodes.csv'), - 'r', newline='' - ) as src: - reader = csv.reader(src) - shared.core_nodes = {tuple(row) for row in reader} - shared.node_pool.update(shared.core_nodes) - - with open( - os.path.join(shared.source_directory, 'i2p_core_nodes.csv'), - 'r', newline='' - ) as f: - reader = csv.reader(f) - shared.i2p_core_nodes = {(row[0].encode(), 'i2p') for row in reader} - shared.i2p_node_pool.update(shared.i2p_core_nodes) - - def bootstrap_from_dns(): """Addes addresses of bootstrap servers to known nodes""" try: @@ -290,8 +238,6 @@ def main(): 'Error while creating data directory in: %s', shared.data_directory, exc_info=True) - load_data() - if shared.ip_enabled and not shared.trusted_peer: bootstrap_from_dns() diff --git a/minode/manager.py b/minode/manager.py index 0afce57..72dec49 100644 --- a/minode/manager.py +++ b/minode/manager.py @@ -1,6 +1,7 @@ # -*- coding: utf-8 -*- """The main thread, managing connections, nodes and objects""" import base64 +import csv import logging import os import pickle @@ -28,6 +29,7 @@ class Manager(threading.Thread): time.time() - 50 * 60 + random.uniform(-1, 1) * 300 # nosec def run(self): + self.load_data() self.clean_objects() while True: time.sleep(0.8) @@ -144,6 +146,59 @@ class Manager(threading.Thread): shared.connections.add(c) shared.hosts = hosts + @staticmethod + def load_data(): + """Loads initial nodes and data, stored in files between sessions""" + try: + with open( + os.path.join(shared.data_directory, 'objects.pickle'), 'br' + ) as src: + shared.objects = pickle.load(src) + except FileNotFoundError: + pass # first start + except Exception: + logging.warning( + 'Error while loading objects from disk.', exc_info=True) + + try: + with open( + os.path.join(shared.data_directory, 'nodes.pickle'), 'br' + ) as src: + shared.node_pool = pickle.load(src) + except FileNotFoundError: + pass + except Exception: + logging.warning( + 'Error while loading nodes from disk.', exc_info=True) + + try: + with open( + os.path.join(shared.data_directory, 'i2p_nodes.pickle'), 'br' + ) as src: + shared.i2p_node_pool = pickle.load(src) + except FileNotFoundError: + pass + except Exception: + logging.warning( + 'Error while loading nodes from disk.', exc_info=True) + + with open( + os.path.join(shared.source_directory, 'core_nodes.csv'), + 'r', newline='', encoding='ascii' + ) as src: + reader = csv.reader(src) + shared.core_nodes = {tuple(row) for row in reader} + shared.node_pool.update(shared.core_nodes) + + with open( + os.path.join(shared.source_directory, 'i2p_core_nodes.csv'), + 'r', newline='', encoding='ascii' + ) as f: + reader = csv.reader(f) + shared.i2p_core_nodes = { + (row[0].encode(), 'i2p') for row in reader} + shared.i2p_node_pool.update(shared.i2p_core_nodes) + @staticmethod def pickle_objects(): try: