mirror of
https://github.com/aljazceru/nutshell.git
synced 2025-12-20 18:44:20 +01:00
* clean up db * db: table lock * db.table_with_schema * fix encrypt.py * postgres nowait * add timeout to lock * melt quote state in db * kinda working * kinda working with postgres * remove dispose * getting there * porperly clean up db for tests * faster tests * configure connection pooling * try github with connection pool * invoice dispatcher does not lock db * fakewallet: pay_if_regtest waits * pay fakewallet invoices * add more * faster * slower * pay_if_regtest async * do not lock the invoice dispatcher * test: do I get disk I/O errors if we disable the invoice_callback_dispatcher? * fix fake so it workss without a callback dispatchert * test on github * readd tasks * refactor * increase time for lock invoice disatcher * try avoiding a race * remove task * github actions: test regtest with postgres * mint per module * no connection pool for testing * enable pool * do not resend paid event * reuse connection * close db connections * sessions * enable debug * dispose engine * disable connection pool for tests * enable connection pool for postgres only * clean up shutdown routine * remove wait for lightning fakewallet lightning invoice * cancel invoice listener tasks on shutdown * fakewallet conftest: decrease outgoing delay * delay payment and set postgres only if needed * disable fail fast for regtest * clean up regtest.yml * change order of tests_db.py * row-specific mint_quote locking * refactor * fix lock statement * refactor swap * refactor * remove psycopg2 * add connection string example to .env.example * remove unnecessary pay * shorter sleep in test_wallet_subscription_swap
110 lines
4.1 KiB
Python
110 lines
4.1 KiB
Python
import os
|
|
import re
|
|
import time
|
|
|
|
from loguru import logger
|
|
|
|
from ..core.db import COCKROACH, POSTGRES, SQLITE, Database
|
|
from ..core.settings import settings
|
|
|
|
|
|
async def backup_database(db: Database, version: int = 0) -> str:
|
|
# for postgres: use pg_dump
|
|
# for sqlite: use sqlite3
|
|
|
|
# skip backups if db_backup_path is None
|
|
# and if version is 0 (fresh database)
|
|
if not settings.db_backup_path or not version:
|
|
return ""
|
|
|
|
filename = f"backup_{db.name}_{int(time.time())}_v{version}"
|
|
try:
|
|
# create backup directory if it doesn't exist
|
|
os.makedirs(os.path.join(settings.db_backup_path), exist_ok=True)
|
|
except Exception as e:
|
|
logger.error(
|
|
f"Error creating backup directory: {e}. Run with BACKUP_DB_MIGRATION=False"
|
|
" to disable backups before database migrations."
|
|
)
|
|
raise e
|
|
filepath = os.path.join(settings.db_backup_path, filename)
|
|
|
|
if db.type == SQLITE:
|
|
filepath = f"{filepath}.sqlite3"
|
|
logger.info(f"Creating {db.type} backup of {db.name} db to {filepath}")
|
|
os.system(f"cp {db.path} {filepath}")
|
|
elif db.type in {POSTGRES, COCKROACH}:
|
|
filepath = f"{filepath}.dump"
|
|
logger.info(f"Creating {db.type} backup of {db.name} db to {filepath}")
|
|
os.system(f"pg_dump --dbname={db.db_location} --file={filepath}")
|
|
|
|
return filepath
|
|
|
|
|
|
async def migrate_databases(db: Database, migrations_module):
|
|
"""Creates the necessary databases if they don't exist already; or migrates them."""
|
|
|
|
async def set_migration_version(conn, db_name, version):
|
|
await conn.execute(
|
|
f"""
|
|
INSERT INTO {db.table_with_schema('dbversions')} (db, version) VALUES (:db, :version)
|
|
ON CONFLICT (db) DO UPDATE SET version = :version
|
|
""",
|
|
{"db": db_name, "version": version},
|
|
)
|
|
|
|
async def run_migration(db, migrations_module):
|
|
db_name = migrations_module.__name__.split(".")[-2]
|
|
# we first check whether any migration is needed and create a backup if so
|
|
migration_needed = False
|
|
for key, migrate in migrations_module.__dict__.items():
|
|
match = matcher.match(key)
|
|
if match:
|
|
version = int(match.group(1))
|
|
if version > current_versions.get(db_name, 0):
|
|
migration_needed = True
|
|
break
|
|
if migration_needed and settings.db_backup_path:
|
|
logger.debug(f"Creating backup of {db_name} db")
|
|
current_version = current_versions.get(db_name, 0)
|
|
await backup_database(db, current_version)
|
|
|
|
# then we run the migrations
|
|
for key, migrate in migrations_module.__dict__.items():
|
|
match = matcher.match(key)
|
|
if match:
|
|
version = int(match.group(1))
|
|
if version > current_versions.get(db_name, 0):
|
|
logger.debug(f"Migrating {db_name} db: {key}")
|
|
await migrate(db)
|
|
|
|
if db.schema is None:
|
|
await set_migration_version(db, db_name, version)
|
|
else:
|
|
async with db.connect() as conn:
|
|
await set_migration_version(conn, db_name, version)
|
|
|
|
async with db.connect() as conn: # type: ignore
|
|
exists = None
|
|
if conn.type == SQLITE:
|
|
exists = await conn.fetchone(
|
|
"SELECT * FROM sqlite_master WHERE type='table' AND"
|
|
f" name='{db.table_with_schema('dbversions')}'"
|
|
)
|
|
elif conn.type in {POSTGRES, COCKROACH}:
|
|
exists = await conn.fetchone(
|
|
"SELECT * FROM information_schema.tables WHERE table_name ="
|
|
f" '{db.table_with_schema('dbversions')}'"
|
|
)
|
|
|
|
if not exists:
|
|
await migrations_module.m000_create_migrations_table(conn)
|
|
|
|
result = await conn.execute(
|
|
f"SELECT * FROM {db.table_with_schema('dbversions')}"
|
|
)
|
|
rows = result.all()
|
|
current_versions = {row["db"]: row["version"] for row in rows}
|
|
matcher = re.compile(r"^m(\d\d\d)_")
|
|
await run_migration(db, migrations_module)
|