mirror of
https://github.com/aljazceru/python-teos.git
synced 2025-12-17 22:24:23 +01:00
The setup of the logs and the parsing of the config file are closely related. The former need info from the later to be created, and needs to be setup only once per pisa instance. In the same way, the later need to only be loaded and validated once per pisa intance and contains info to setup the logs. Intead of setting up the logs in init and loading the config file in pisad, now both are dealt with in __init__
119 lines
4.6 KiB
Python
119 lines
4.6 KiB
Python
from getopt import getopt
|
|
from sys import argv, exit
|
|
from signal import signal, SIGINT, SIGQUIT, SIGTERM
|
|
|
|
from common.logger import Logger
|
|
from common.tools import setup_data_folder
|
|
|
|
from pisa import config, LOG_PREFIX
|
|
from pisa.api import API
|
|
from pisa.watcher import Watcher
|
|
from pisa.builder import Builder
|
|
from pisa.db_manager import DBManager
|
|
from pisa.chain_monitor import ChainMonitor
|
|
from pisa.block_processor import BlockProcessor
|
|
from pisa.tools import can_connect_to_bitcoind, in_correct_network
|
|
|
|
logger = Logger(actor="Daemon", log_name_prefix=LOG_PREFIX)
|
|
|
|
|
|
def handle_signals(signal_received, frame):
|
|
logger.info("Closing connection with appointments db")
|
|
db_manager.db.close()
|
|
chain_monitor.terminate = True
|
|
|
|
logger.info("Shutting down PISA")
|
|
exit(0)
|
|
|
|
|
|
def main():
|
|
global db_manager, chain_monitor
|
|
|
|
signal(SIGINT, handle_signals)
|
|
signal(SIGTERM, handle_signals)
|
|
signal(SIGQUIT, handle_signals)
|
|
|
|
logger.info("Starting PISA")
|
|
|
|
setup_data_folder(config.get("DATA_FOLDER"), logger)
|
|
db_manager = DBManager(config.get("DB_PATH"))
|
|
|
|
if not can_connect_to_bitcoind():
|
|
logger.error("Can't connect to bitcoind. Shutting down")
|
|
|
|
elif not in_correct_network(config.get("BTC_NETWORK")):
|
|
logger.error("bitcoind is running on a different network, check conf.py and bitcoin.conf. Shutting down")
|
|
|
|
else:
|
|
try:
|
|
# Create the chain monitor and start monitoring the chain
|
|
chain_monitor = ChainMonitor()
|
|
chain_monitor.monitor_chain()
|
|
|
|
watcher_appointments_data = db_manager.load_watcher_appointments()
|
|
responder_trackers_data = db_manager.load_responder_trackers()
|
|
|
|
with open(config.get("PISA_SECRET_KEY"), "rb") as key_file:
|
|
secret_key_der = key_file.read()
|
|
|
|
watcher = Watcher(db_manager, chain_monitor, secret_key_der, config)
|
|
chain_monitor.attach_watcher(watcher.block_queue, watcher.asleep)
|
|
chain_monitor.attach_responder(watcher.responder.block_queue, watcher.responder.asleep)
|
|
|
|
if len(watcher_appointments_data) == 0 and len(responder_trackers_data) == 0:
|
|
logger.info("Fresh bootstrap")
|
|
|
|
else:
|
|
logger.info("Bootstrapping from backed up data")
|
|
block_processor = BlockProcessor()
|
|
|
|
last_block_watcher = db_manager.load_last_block_hash_watcher()
|
|
last_block_responder = db_manager.load_last_block_hash_responder()
|
|
|
|
# FIXME: 32-reorgs-offline dropped txs are not used at this point.
|
|
last_common_ancestor_responder = None
|
|
missed_blocks_responder = None
|
|
|
|
# Build Responder with backed up data if found
|
|
if last_block_responder is not None:
|
|
last_common_ancestor_responder, dropped_txs_responder = block_processor.find_last_common_ancestor(
|
|
last_block_responder
|
|
)
|
|
missed_blocks_responder = block_processor.get_missed_blocks(last_common_ancestor_responder)
|
|
|
|
watcher.responder.trackers, watcher.responder.tx_tracker_map = Builder.build_trackers(
|
|
responder_trackers_data
|
|
)
|
|
watcher.responder.block_queue = Builder.build_block_queue(missed_blocks_responder)
|
|
|
|
# Build Watcher. If the blocks of both match we don't perform the search twice.
|
|
if last_block_watcher is not None:
|
|
if last_block_watcher == last_block_responder:
|
|
missed_blocks_watcher = missed_blocks_responder
|
|
else:
|
|
last_common_ancestor_watcher, dropped_txs_watcher = block_processor.find_last_common_ancestor(
|
|
last_block_watcher
|
|
)
|
|
missed_blocks_watcher = block_processor.get_missed_blocks(last_common_ancestor_watcher)
|
|
|
|
watcher.appointments, watcher.locator_uuid_map = Builder.build_appointments(
|
|
watcher_appointments_data
|
|
)
|
|
watcher.block_queue = Builder.build_block_queue(missed_blocks_watcher)
|
|
|
|
# Fire the API
|
|
API(watcher, config=config).start()
|
|
|
|
except Exception as e:
|
|
logger.error("An error occurred: {}. Shutting down".format(e))
|
|
exit(1)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
opts, _ = getopt(argv[1:], "", [""])
|
|
for opt, arg in opts:
|
|
# FIXME: Leaving this here for future option/arguments
|
|
pass
|
|
|
|
main()
|