#include "db.h" #include #include #include #include #include #include #include #include #include #define DB_FILE "lightningd.sqlite3" #define NSEC_IN_SEC 1000000000 /* For testing, we want to catch fatal messages. */ #ifndef db_fatal #define db_fatal fatal #endif struct migration { const char *sql; void (*func)(struct lightningd *ld, struct db *db); }; void migrate_pr2342_feerate_per_channel(struct lightningd *ld, struct db *db); /* Do not reorder or remove elements from this array, it is used to * migrate existing databases from a previous state, based on the * string indices */ static struct migration dbmigrations[] = { { "CREATE TABLE version (version INTEGER)", NULL }, { "INSERT INTO version VALUES (1)", NULL }, { "CREATE TABLE outputs ( \ prev_out_tx CHAR(64), \ prev_out_index INTEGER, \ value INTEGER, \ type INTEGER, \ status INTEGER, \ keyindex INTEGER, \ PRIMARY KEY (prev_out_tx, prev_out_index) \ );", NULL }, { "CREATE TABLE vars (name VARCHAR(32), val VARCHAR(255), PRIMARY KEY (name));", NULL }, { "CREATE TABLE shachains ( \ id INTEGER, \ min_index INTEGER, \ num_valid INTEGER, \ PRIMARY KEY (id));", NULL }, { "CREATE TABLE shachain_known ( \ shachain_id INTEGER REFERENCES shachains(id) ON DELETE CASCADE, \ pos INTEGER, \ idx INTEGER, \ hash BLOB, \ PRIMARY KEY (shachain_id, pos));", NULL }, { "CREATE TABLE channels (" " id INTEGER," /* chan->id */ " peer_id INTEGER REFERENCES peers(id) ON DELETE CASCADE," " short_channel_id BLOB," " channel_config_local INTEGER," " channel_config_remote INTEGER," " state INTEGER," " funder INTEGER," " channel_flags INTEGER," " minimum_depth INTEGER," " next_index_local INTEGER," " next_index_remote INTEGER," " next_htlc_id INTEGER, " " funding_tx_id BLOB," " funding_tx_outnum INTEGER," " funding_satoshi INTEGER," " funding_locked_remote INTEGER," " push_msatoshi INTEGER," " msatoshi_local INTEGER," /* our_msatoshi */ /* START channel_info */ " fundingkey_remote BLOB," " revocation_basepoint_remote BLOB," " payment_basepoint_remote BLOB," " htlc_basepoint_remote BLOB," " delayed_payment_basepoint_remote BLOB," " per_commit_remote BLOB," " old_per_commit_remote BLOB," " local_feerate_per_kw INTEGER," " remote_feerate_per_kw INTEGER," /* END channel_info */ " shachain_remote_id INTEGER," " shutdown_scriptpubkey_remote BLOB," " shutdown_keyidx_local INTEGER," " last_sent_commit_state INTEGER," " last_sent_commit_id INTEGER," " last_tx BLOB," " last_sig BLOB," " closing_fee_received INTEGER," " closing_sig_received BLOB," " PRIMARY KEY (id)" ");", NULL }, { "CREATE TABLE peers (" " id INTEGER," " node_id BLOB UNIQUE," /* pubkey */ " address TEXT," " PRIMARY KEY (id)" ");", NULL }, { "CREATE TABLE channel_configs (" " id INTEGER," " dust_limit_satoshis INTEGER," " max_htlc_value_in_flight_msat INTEGER," " channel_reserve_satoshis INTEGER," " htlc_minimum_msat INTEGER," " to_self_delay INTEGER," " max_accepted_htlcs INTEGER," " PRIMARY KEY (id)" ");", NULL }, { "CREATE TABLE channel_htlcs (" " id INTEGER," " channel_id INTEGER REFERENCES channels(id) ON DELETE CASCADE," " channel_htlc_id INTEGER," " direction INTEGER," " origin_htlc INTEGER," " msatoshi INTEGER," " cltv_expiry INTEGER," " payment_hash BLOB," " payment_key BLOB," " routing_onion BLOB," " failuremsg BLOB," " malformed_onion INTEGER," " hstate INTEGER," " shared_secret BLOB," " PRIMARY KEY (id)," " UNIQUE (channel_id, channel_htlc_id, direction)" ");", NULL }, { "CREATE TABLE invoices (" " id INTEGER," " state INTEGER," " msatoshi INTEGER," " payment_hash BLOB," " payment_key BLOB," " label TEXT," " PRIMARY KEY (id)," " UNIQUE (label)," " UNIQUE (payment_hash)" ");", NULL }, { "CREATE TABLE payments (" " id INTEGER," " timestamp INTEGER," " status INTEGER," " payment_hash BLOB," " direction INTEGER," " destination BLOB," " msatoshi INTEGER," " PRIMARY KEY (id)," " UNIQUE (payment_hash)" ");", NULL }, /* Add expiry field to invoices (effectively infinite). */ { "ALTER TABLE invoices ADD expiry_time INTEGER;", NULL }, { "UPDATE invoices SET expiry_time=9223372036854775807;", NULL }, /* Add pay_index field to paid invoices (initially, same order as id). */ { "ALTER TABLE invoices ADD pay_index INTEGER;", NULL }, { "CREATE UNIQUE INDEX invoices_pay_index" " ON invoices(pay_index);", NULL }, { "UPDATE invoices SET pay_index=id WHERE state=1;", NULL }, /* only paid invoice */ /* Create next_pay_index variable (highest pay_index). */ { "INSERT OR REPLACE INTO vars(name, val)" " VALUES('next_pay_index', " " COALESCE((SELECT MAX(pay_index) FROM invoices WHERE state=1), 0) + 1" " );", NULL }, /* Create first_block field; initialize from channel id if any. * This fails for channels still awaiting lockin, but that only applies to * pre-release software, so it's forgivable. */ { "ALTER TABLE channels ADD first_blocknum INTEGER;", NULL }, { "UPDATE channels SET first_blocknum=CAST(short_channel_id AS INTEGER) WHERE short_channel_id IS NOT NULL;", NULL }, { "ALTER TABLE outputs ADD COLUMN channel_id INTEGER;", NULL }, { "ALTER TABLE outputs ADD COLUMN peer_id BLOB;", NULL }, { "ALTER TABLE outputs ADD COLUMN commitment_point BLOB;", NULL }, { "ALTER TABLE invoices ADD COLUMN msatoshi_received INTEGER;", NULL }, /* Normally impossible, so at least we'll know if databases are ancient. */ { "UPDATE invoices SET msatoshi_received=0 WHERE state=1;", NULL }, { "ALTER TABLE channels ADD COLUMN last_was_revoke INTEGER;", NULL }, /* We no longer record incoming payments: invoices cover that. * Without ALTER_TABLE DROP COLUMN support we need to do this by * rename & copy, which works because there are no triggers etc. */ { "ALTER TABLE payments RENAME TO temp_payments;", NULL }, { "CREATE TABLE payments (" " id INTEGER," " timestamp INTEGER," " status INTEGER," " payment_hash BLOB," " destination BLOB," " msatoshi INTEGER," " PRIMARY KEY (id)," " UNIQUE (payment_hash)" ");", NULL }, { "INSERT INTO payments SELECT id, timestamp, status, payment_hash, destination, msatoshi FROM temp_payments WHERE direction=1;", NULL }, { "DROP TABLE temp_payments;", NULL }, /* We need to keep the preimage in case they ask to pay again. */ { "ALTER TABLE payments ADD COLUMN payment_preimage BLOB;", NULL }, /* We need to keep the shared secrets to decode error returns. */ { "ALTER TABLE payments ADD COLUMN path_secrets BLOB;", NULL }, /* Create time-of-payment of invoice, default already-paid * invoices to current time. */ { "ALTER TABLE invoices ADD paid_timestamp INTEGER;", NULL }, { "UPDATE invoices" " SET paid_timestamp = strftime('%s', 'now')" " WHERE state = 1;", NULL }, /* We need to keep the route node pubkeys and short channel ids to * correctly mark routing failures. We separate short channel ids * because we cannot safely save them as blobs due to byteorder * concerns. */ { "ALTER TABLE payments ADD COLUMN route_nodes BLOB;", NULL }, { "ALTER TABLE payments ADD COLUMN route_channels TEXT;", NULL }, { "CREATE TABLE htlc_sigs (channelid INTEGER REFERENCES channels(id) ON DELETE CASCADE, signature BLOB);", NULL }, { "CREATE INDEX channel_idx ON htlc_sigs (channelid)", NULL }, /* Get rid of OPENINGD entries; we don't put them in db any more */ { "DELETE FROM channels WHERE state=1", NULL }, /* Keep track of db upgrades, for debugging */ { "CREATE TABLE db_upgrades (upgrade_from INTEGER, lightning_version TEXT);", NULL }, /* We used not to clean up peers when their channels were gone. */ { "DELETE FROM peers WHERE id NOT IN (SELECT peer_id FROM channels);", NULL }, /* The ONCHAIND_CHEATED/THEIR_UNILATERAL/OUR_UNILATERAL/MUTUAL are now one */ { "UPDATE channels SET STATE = 8 WHERE state > 8;", NULL }, /* Add bolt11 to invoices table*/ { "ALTER TABLE invoices ADD bolt11 TEXT;", NULL }, /* What do we think the head of the blockchain looks like? Used * primarily to track confirmations across restarts and making * sure we handle reorgs correctly. */ { "CREATE TABLE blocks (height INT, hash BLOB, prev_hash BLOB, UNIQUE(height));", NULL }, /* ON DELETE CASCADE would have been nice for confirmation_height, * so that we automatically delete outputs that fall off the * blockchain and then we rediscover them if they are included * again. However, we have the their_unilateral/to_us which we * can't simply recognize from the chain without additional * hints. So we just mark them as unconfirmed should the block * die. */ { "ALTER TABLE outputs ADD COLUMN confirmation_height INTEGER REFERENCES blocks(height) ON DELETE SET NULL;", NULL }, { "ALTER TABLE outputs ADD COLUMN spend_height INTEGER REFERENCES blocks(height) ON DELETE SET NULL;", NULL }, /* Create a covering index that covers both fields */ { "CREATE INDEX output_height_idx ON outputs (confirmation_height, spend_height);", NULL }, { "CREATE TABLE utxoset (" " txid BLOB," " outnum INT," " blockheight INT REFERENCES blocks(height) ON DELETE CASCADE," " spendheight INT REFERENCES blocks(height) ON DELETE SET NULL," " txindex INT," " scriptpubkey BLOB," " satoshis BIGINT," " PRIMARY KEY(txid, outnum));", NULL }, { "CREATE INDEX short_channel_id ON utxoset (blockheight, txindex, outnum)", NULL }, /* Necessary index for long rollbacks of the blockchain, otherwise we're * doing table scans for every block removed. */ { "CREATE INDEX utxoset_spend ON utxoset (spendheight)", NULL }, /* Assign key 0 to unassigned shutdown_keyidx_local. */ { "UPDATE channels SET shutdown_keyidx_local=0 WHERE shutdown_keyidx_local = -1;", NULL }, /* FIXME: We should rename shutdown_keyidx_local to final_key_index */ /* -- Payment routing failure information -- */ /* BLOB if failure was due to unparseable onion, NULL otherwise */ { "ALTER TABLE payments ADD failonionreply BLOB;", NULL }, /* 0 if we could theoretically retry, 1 if PERM fail at payee */ { "ALTER TABLE payments ADD faildestperm INTEGER;", NULL }, /* Contents of routing_failure (only if not unparseable onion) */ { "ALTER TABLE payments ADD failindex INTEGER;", NULL }, /* erring_index */ { "ALTER TABLE payments ADD failcode INTEGER;", NULL }, /* failcode */ { "ALTER TABLE payments ADD failnode BLOB;", NULL }, /* erring_node */ { "ALTER TABLE payments ADD failchannel BLOB;", NULL }, /* erring_channel */ { "ALTER TABLE payments ADD failupdate BLOB;", NULL }, /* channel_update - can be NULL*/ /* -- Payment routing failure information ends -- */ /* Delete route data for already succeeded or failed payments */ { "UPDATE payments" " SET path_secrets = NULL" " , route_nodes = NULL" " , route_channels = NULL" " WHERE status <> 0;", NULL }, /* PAYMENT_PENDING */ /* -- Routing statistics -- */ { "ALTER TABLE channels ADD in_payments_offered INTEGER;", NULL }, { "ALTER TABLE channels ADD in_payments_fulfilled INTEGER;", NULL }, { "ALTER TABLE channels ADD in_msatoshi_offered INTEGER;", NULL }, { "ALTER TABLE channels ADD in_msatoshi_fulfilled INTEGER;", NULL }, { "ALTER TABLE channels ADD out_payments_offered INTEGER;", NULL }, { "ALTER TABLE channels ADD out_payments_fulfilled INTEGER;", NULL }, { "ALTER TABLE channels ADD out_msatoshi_offered INTEGER;", NULL }, { "ALTER TABLE channels ADD out_msatoshi_fulfilled INTEGER;", NULL }, { "UPDATE channels" " SET in_payments_offered = 0, in_payments_fulfilled = 0" " , in_msatoshi_offered = 0, in_msatoshi_fulfilled = 0" " , out_payments_offered = 0, out_payments_fulfilled = 0" " , out_msatoshi_offered = 0, out_msatoshi_fulfilled = 0" " ;", NULL }, /* -- Routing statistics ends --*/ /* Record the msatoshi actually sent in a payment. */ { "ALTER TABLE payments ADD msatoshi_sent INTEGER;", NULL }, { "UPDATE payments SET msatoshi_sent = msatoshi;", NULL }, /* Delete dangling utxoset entries due to Issue #1280 */ { "DELETE FROM utxoset WHERE blockheight IN (" " SELECT DISTINCT(blockheight)" " FROM utxoset LEFT OUTER JOIN blocks on (blockheight == blocks.height) " " WHERE blocks.hash IS NULL" ");", NULL }, /* Record feerate range, to optimize onchaind grinding actual fees. */ { "ALTER TABLE channels ADD min_possible_feerate INTEGER;", NULL }, { "ALTER TABLE channels ADD max_possible_feerate INTEGER;", NULL }, /* https://bitcoinfees.github.io/#1d says Dec 17 peak was ~1M sat/kb * which is 250,000 sat/Sipa */ { "UPDATE channels SET min_possible_feerate=0, max_possible_feerate=250000;", NULL }, /* -- Min and max msatoshi_to_us -- */ { "ALTER TABLE channels ADD msatoshi_to_us_min INTEGER;", NULL }, { "ALTER TABLE channels ADD msatoshi_to_us_max INTEGER;", NULL }, { "UPDATE channels" " SET msatoshi_to_us_min = msatoshi_local" " , msatoshi_to_us_max = msatoshi_local" " ;", NULL }, /* -- Min and max msatoshi_to_us ends -- */ /* Transactions we are interested in. Either we sent them ourselves or we * are watching them. We don't cascade block height deletes so we don't * forget any of them by accident.*/ { "CREATE TABLE transactions (" " id BLOB" ", blockheight INTEGER REFERENCES blocks(height) ON DELETE SET NULL" ", txindex INTEGER" ", rawtx BLOB" ", PRIMARY KEY (id)" ");", NULL }, /* -- Detailed payment failure -- */ { "ALTER TABLE payments ADD faildetail TEXT;", NULL }, { "UPDATE payments" " SET faildetail = 'unspecified payment failure reason'" " WHERE status = 2;", NULL }, /* PAYMENT_FAILED */ /* -- Detailed payment faiure ends -- */ { "CREATE TABLE channeltxs (" /* The id serves as insertion order and short ID */ " id INTEGER" ", channel_id INTEGER REFERENCES channels(id) ON DELETE CASCADE" ", type INTEGER" ", transaction_id BLOB REFERENCES transactions(id) ON DELETE CASCADE" /* The input_num is only used by the txo_watch, 0 if txwatch */ ", input_num INTEGER" /* The height at which we sent the depth notice */ ", blockheight INTEGER REFERENCES blocks(height) ON DELETE CASCADE" ", PRIMARY KEY(id)" ");", NULL }, /* -- Set the correct rescan height for PR #1398 -- */ /* Delete blocks that are higher than our initial scan point, this is a * no-op if we don't have a channel. */ { "DELETE FROM blocks WHERE height > (SELECT MIN(first_blocknum) FROM channels);", NULL }, /* Now make sure we have the lower bound block with the first_blocknum * height. This may introduce a block with NULL height if we didn't have any * blocks, remove that in the next. */ { "INSERT OR IGNORE INTO blocks (height) VALUES ((SELECT MIN(first_blocknum) FROM channels));", NULL }, { "DELETE FROM blocks WHERE height IS NULL;", NULL }, /* -- End of PR #1398 -- */ { "ALTER TABLE invoices ADD description TEXT;", NULL }, /* FIXME: payments table 'description' is really a 'label' */ { "ALTER TABLE payments ADD description TEXT;", NULL }, /* future_per_commitment_point if other side proves we're out of date -- */ { "ALTER TABLE channels ADD future_per_commitment_point BLOB;", NULL }, /* last_sent_commit array fix */ { "ALTER TABLE channels ADD last_sent_commit BLOB;", NULL }, /* Stats table to track forwarded HTLCs. The values in the HTLCs * and their states are replicated here and the entries are not * deleted when the HTLC entries or the channel entries are * deleted to avoid unexpected drops in statistics. */ { "CREATE TABLE forwarded_payments (" " in_htlc_id INTEGER REFERENCES channel_htlcs(id) ON DELETE SET NULL" ", out_htlc_id INTEGER REFERENCES channel_htlcs(id) ON DELETE SET NULL" ", in_channel_scid INTEGER" ", out_channel_scid INTEGER" ", in_msatoshi INTEGER" ", out_msatoshi INTEGER" ", state INTEGER" ", UNIQUE(in_htlc_id, out_htlc_id)" ");", NULL }, /* Add a direction for failed payments. */ { "ALTER TABLE payments ADD faildirection INTEGER;", NULL }, /* erring_direction */ /* Fix dangling peers with no channels. */ { "DELETE FROM peers WHERE id NOT IN (SELECT peer_id FROM channels);", NULL }, { "ALTER TABLE outputs ADD scriptpubkey BLOB;", NULL }, /* Keep bolt11 string for payments. */ { "ALTER TABLE payments ADD bolt11 TEXT;", NULL }, /* PR #2342 feerate per channel */ { "ALTER TABLE channels ADD feerate_base INTEGER;", NULL }, { "ALTER TABLE channels ADD feerate_ppm INTEGER;", NULL }, { NULL, migrate_pr2342_feerate_per_channel }, { "ALTER TABLE channel_htlcs ADD received_time INTEGER", NULL }, { "ALTER TABLE forwarded_payments ADD received_time INTEGER", NULL }, { "ALTER TABLE forwarded_payments ADD resolved_time INTEGER", NULL }, { "ALTER TABLE channels ADD remote_upfront_shutdown_script BLOB;", NULL }, /* PR #2524: Add failcode into forward_payment */ { "ALTER TABLE forwarded_payments ADD failcode INTEGER;", NULL }, }; /* Leak tracking. */ #if DEVELOPER /* We need a global here, since caller has no context. Yuck! */ static struct list_head db_statements = LIST_HEAD_INIT(db_statements); struct db_statement { struct list_node list; sqlite3_stmt *stmt; const char *origin; }; static struct db_statement *find_statement(sqlite3_stmt *stmt) { struct db_statement *i; list_for_each(&db_statements, i, list) { if (i->stmt == stmt) return i; } return NULL; } void db_assert_no_outstanding_statements(void) { struct db_statement *dbstat; dbstat = list_top(&db_statements, struct db_statement, list); if (dbstat) db_fatal("Unfinalized statement %s", dbstat->origin); } static void dev_statement_start(sqlite3_stmt *stmt, const char *origin) { struct db_statement *dbstat = tal(NULL, struct db_statement); dbstat->stmt = stmt; dbstat->origin = origin; list_add(&db_statements, &dbstat->list); } static void dev_statement_end(sqlite3_stmt *stmt) { struct db_statement *dbstat = find_statement(stmt); list_del_from(&db_statements, &dbstat->list); tal_free(dbstat); } #else static void dev_statement_start(sqlite3_stmt *stmt, const char *origin) { } static void dev_statement_end(sqlite3_stmt *stmt) { } void db_assert_no_outstanding_statements(void) { } #endif #if !HAVE_SQLITE3_EXPANDED_SQL /* Prior to sqlite3 v3.14, we have to use tracing to dump statements */ static void trace_sqlite3(void *dbv, const char *stmt) { struct db *db = dbv; /* We get a "COMMIT;" after we've sent our changes. */ if (!db->changes) { assert(streq(stmt, "COMMIT;")); return; } tal_arr_expand(&db->changes, tal_strdup(db->changes, stmt)); } #endif void db_stmt_done(sqlite3_stmt *stmt) { dev_statement_end(stmt); sqlite3_finalize(stmt); } sqlite3_stmt *db_select_prepare_(const char *location, struct db *db, const char *query) { int err; sqlite3_stmt *stmt; const char *full_query = tal_fmt(db, "SELECT %s", query); assert(db->in_transaction); err = sqlite3_prepare_v2(db->sql, full_query, -1, &stmt, NULL); if (err != SQLITE_OK) db_fatal("%s: %s: %s", location, full_query, sqlite3_errmsg(db->sql)); dev_statement_start(stmt, location); tal_free(full_query); return stmt; } bool db_select_step_(const char *location, struct db *db, struct sqlite3_stmt *stmt) { int ret; ret = sqlite3_step(stmt); if (ret == SQLITE_ROW) return true; if (ret != SQLITE_DONE) db_fatal("%s: %s", location, sqlite3_errmsg(db->sql)); db_stmt_done(stmt); return false; } sqlite3_stmt *db_prepare_(const char *location, struct db *db, const char *query) { int err; sqlite3_stmt *stmt; assert(db->in_transaction); err = sqlite3_prepare_v2(db->sql, query, -1, &stmt, NULL); if (err != SQLITE_OK) db_fatal("%s: %s: %s", location, query, sqlite3_errmsg(db->sql)); dev_statement_start(stmt, location); return stmt; } void db_exec_prepared_(const char *caller, struct db *db, sqlite3_stmt *stmt) { assert(db->in_transaction); if (sqlite3_step(stmt) != SQLITE_DONE) db_fatal("%s: %s", caller, sqlite3_errmsg(db->sql)); #if HAVE_SQLITE3_EXPANDED_SQL char *expanded_sql; expanded_sql = sqlite3_expanded_sql(stmt); tal_arr_expand(&db->changes, tal_strdup(db->changes, expanded_sql)); sqlite3_free(expanded_sql); #endif db_stmt_done(stmt); } /* This one doesn't check if we're in a transaction. */ static void db_do_exec(const char *caller, struct db *db, const char *cmd) { char *errmsg; int err; err = sqlite3_exec(db->sql, cmd, NULL, NULL, &errmsg); if (err != SQLITE_OK) { db_fatal("%s:%s:%s:%s", caller, sqlite3_errstr(err), cmd, errmsg); /* Only reached in testing */ sqlite3_free(errmsg); } #if HAVE_SQLITE3_EXPANDED_SQL tal_arr_expand(&db->changes, tal_strdup(db->changes, cmd)); #endif } static void PRINTF_FMT(3, 4) db_exec(const char *caller, struct db *db, const char *fmt, ...) { va_list ap; char *cmd; assert(db->in_transaction); va_start(ap, fmt); cmd = tal_vfmt(db, fmt, ap); va_end(ap); db_do_exec(caller, db, cmd); tal_free(cmd); } /* This one can fail: returns NULL if so */ static sqlite3_stmt *db_query(const char *location, struct db *db, const char *query) { sqlite3_stmt *stmt; assert(db->in_transaction); /* Sets stmt to NULL if not SQLITE_OK */ sqlite3_prepare_v2(db->sql, query, -1, &stmt, NULL); if (stmt) dev_statement_start(stmt, location); return stmt; } sqlite3_stmt *PRINTF_FMT(3, 4) db_select_(const char *location, struct db *db, const char *fmt, ...) { va_list ap; char *query = tal_strdup(db, "SELECT "); sqlite3_stmt *stmt; assert(db->in_transaction); va_start(ap, fmt); tal_append_vfmt(&query, fmt, ap); va_end(ap); stmt = db_query(location, db, query); if (!stmt) db_fatal("%s:%s:%s", location, query, sqlite3_errmsg(db->sql)); tal_free(query); return stmt; } static void destroy_db(struct db *db) { db_assert_no_outstanding_statements(); sqlite3_close(db->sql); } /* We expect min changes (ie. BEGIN TRANSACTION): report if more. * Optionally add "final" at the end (ie. COMMIT). */ static void db_report_changes(struct db *db, const char *final, size_t min) { assert(db->changes); assert(tal_count(db->changes) >= min); if (tal_count(db->changes) > min) plugin_hook_db_sync(db, db->changes, final); db->changes = tal_free(db->changes); } static void db_prepare_for_changes(struct db *db) { assert(!db->changes); db->changes = tal_arr(db, const char *, 0); } void db_begin_transaction_(struct db *db, const char *location) { if (db->in_transaction) db_fatal("Already in transaction from %s", db->in_transaction); db_prepare_for_changes(db); db_do_exec(location, db, "BEGIN TRANSACTION;"); db->in_transaction = location; } void db_commit_transaction(struct db *db) { int err; char *errmsg; const char *cmd = "COMMIT;"; assert(db->in_transaction); db_assert_no_outstanding_statements(); /* We expect at least the BEGIN TRANSACTION */ db_report_changes(db, cmd, 1); err = sqlite3_exec(db->sql, cmd, NULL, NULL, &errmsg); if (err != SQLITE_OK) db_fatal("%s:%s:%s:%s", __func__, sqlite3_errstr(err), cmd, errmsg); db->in_transaction = NULL; } static void setup_open_db(struct db *db) { #if !HAVE_SQLITE3_EXPANDED_SQL sqlite3_trace(db->sql, trace_sqlite3, db); #endif /* This must be outside a transaction, so catch it */ assert(!db->in_transaction); db_prepare_for_changes(db); db_do_exec(__func__, db, "PRAGMA foreign_keys = ON;"); db_report_changes(db, NULL, 0); } /** * db_open - Open or create a sqlite3 database */ static struct db *db_open(const tal_t *ctx, char *filename) { int err; struct db *db; sqlite3 *sql; int flags = SQLITE_OPEN_READWRITE | SQLITE_OPEN_CREATE; err = sqlite3_open_v2(filename, &sql, flags, NULL); if (err != SQLITE_OK) { db_fatal("failed to open database %s: %s", filename, sqlite3_errstr(err)); } db = tal(ctx, struct db); db->filename = tal_strdup(db, filename); db->sql = sql; tal_add_destructor(db, destroy_db); db->in_transaction = NULL; db->changes = NULL; setup_open_db(db); return db; } /** * db_get_version - Determine the current DB schema version * * Will attempt to determine the current schema version of the * database @db by querying the `version` table. If the table does not * exist it'll return schema version -1, so that migration 0 is * applied, which should create the `version` table. */ static int db_get_version(struct db *db) { int res; sqlite3_stmt *stmt = db_query(__func__, db, "SELECT version FROM version LIMIT 1"); if (!stmt) return -1; if (!db_select_step(db, stmt)) return -1; res = sqlite3_column_int64(stmt, 0); db_stmt_done(stmt); return res; } /** * db_migrate - Apply all remaining migrations from the current version */ static void db_migrate(struct lightningd *ld, struct db *db, struct log *log) { /* Attempt to read the version from the database */ int current, orig, available; db_begin_transaction(db); orig = current = db_get_version(db); available = ARRAY_SIZE(dbmigrations) - 1; if (current == -1) log_info(log, "Creating database"); else if (available < current) db_fatal("Refusing to migrate down from version %u to %u", current, available); else if (current != available) log_info(log, "Updating database from version %u to %u", current, available); while (current < available) { current++; if (dbmigrations[current].sql) db_exec(__func__, db, "%s", dbmigrations[current].sql); if (dbmigrations[current].func) dbmigrations[current].func(ld, db); } /* Finally update the version number in the version table */ db_exec(__func__, db, "UPDATE version SET version=%d;", available); /* Annotate that we did upgrade, if any. */ if (current != orig) db_exec(__func__, db, "INSERT INTO db_upgrades VALUES (%i, '%s');", orig, version()); db_commit_transaction(db); } struct db *db_setup(const tal_t *ctx, struct lightningd *ld, struct log *log) { struct db *db = db_open(ctx, DB_FILE); db_migrate(ld, db, log); return db; } void db_close_for_fork(struct db *db) { /* https://www.sqlite.org/faq.html#q6 * * Under Unix, you should not carry an open SQLite database across a * fork() system call into the child process. */ if (sqlite3_close(db->sql) != SQLITE_OK) db_fatal("sqlite3_close: %s", sqlite3_errmsg(db->sql)); db->sql = NULL; } void db_reopen_after_fork(struct db *db) { int err = sqlite3_open_v2(db->filename, &db->sql, SQLITE_OPEN_READWRITE, NULL); if (err != SQLITE_OK) { db_fatal("failed to re-open database %s: %s", db->filename, sqlite3_errstr(err)); } setup_open_db(db); } s64 db_get_intvar(struct db *db, char *varname, s64 defval) { s64 res; sqlite3_stmt *stmt; const char *query; query = tal_fmt(db, "SELECT val FROM vars WHERE name='%s' LIMIT 1", varname); stmt = db_query(__func__, db, query); tal_free(query); if (!stmt) return defval; if (db_select_step(db, stmt)) { const unsigned char *stringvar = sqlite3_column_text(stmt, 0); res = atol((const char *)stringvar); db_stmt_done(stmt); } else res = defval; return res; } void db_set_intvar(struct db *db, char *varname, s64 val) { /* Attempt to update */ db_exec(__func__, db, "UPDATE vars SET val='%" PRId64 "' WHERE name='%s';", val, varname); if (sqlite3_changes(db->sql) == 0) db_exec( __func__, db, "INSERT INTO vars (name, val) VALUES ('%s', '%" PRId64 "');", varname, val); } void *sqlite3_column_arr_(const tal_t *ctx, sqlite3_stmt *stmt, int col, size_t bytes, const char *label, const char *caller) { size_t sourcelen = sqlite3_column_bytes(stmt, col); void *p; if (sqlite3_column_type(stmt, col) == SQLITE_NULL) return NULL; if (sourcelen % bytes != 0) db_fatal("%s: column size %zu not a multiple of %s (%zu)", caller, sourcelen, label, bytes); p = tal_arr_label(ctx, char, sourcelen, label); memcpy(p, sqlite3_column_blob(stmt, col), sourcelen); return p; } bool sqlite3_bind_short_channel_id(sqlite3_stmt *stmt, int col, const struct short_channel_id *id) { char *ser = short_channel_id_to_str(id, id); int err = sqlite3_bind_blob(stmt, col, ser, strlen(ser), SQLITE_TRANSIENT); tal_free(ser); return err == SQLITE_OK; } bool sqlite3_column_short_channel_id(sqlite3_stmt *stmt, int col, struct short_channel_id *dest) { const char *source = sqlite3_column_blob(stmt, col); size_t sourcelen = sqlite3_column_bytes(stmt, col); return short_channel_id_from_str(source, sourcelen, dest, true); } bool sqlite3_bind_short_channel_id_array(sqlite3_stmt *stmt, int col, const struct short_channel_id *id) { u8 *ser; size_t num; size_t i; /* Handle nulls early. */ if (!id) { int err = sqlite3_bind_null(stmt, col); return err == SQLITE_OK; } ser = tal_arr(NULL, u8, 0); num = tal_count(id); for (i = 0; i < num; ++i) towire_short_channel_id(&ser, &id[i]); int err = sqlite3_bind_blob(stmt, col, ser, tal_count(ser), SQLITE_TRANSIENT); tal_free(ser); return err == SQLITE_OK; } struct short_channel_id * sqlite3_column_short_channel_id_array(const tal_t *ctx, sqlite3_stmt *stmt, int col) { const u8 *ser; size_t len; struct short_channel_id *ret; /* Handle nulls early. */ if (sqlite3_column_type(stmt, col) == SQLITE_NULL) return NULL; ser = sqlite3_column_blob(stmt, col); len = sqlite3_column_bytes(stmt, col); ret = tal_arr(ctx, struct short_channel_id, 0); while (len != 0) { struct short_channel_id scid; fromwire_short_channel_id(&ser, &len, &scid); tal_arr_expand(&ret, scid); } return ret; } bool sqlite3_bind_tx(sqlite3_stmt *stmt, int col, const struct bitcoin_tx *tx) { u8 *ser = linearize_tx(NULL, tx); int err = sqlite3_bind_blob(stmt, col, ser, tal_count(ser), SQLITE_TRANSIENT); tal_free(ser); return err == SQLITE_OK; } struct bitcoin_tx *sqlite3_column_tx(const tal_t *ctx, sqlite3_stmt *stmt, int col) { const u8 *src = sqlite3_column_blob(stmt, col); size_t len = sqlite3_column_bytes(stmt, col); return pull_bitcoin_tx(ctx, &src, &len); } bool sqlite3_bind_signature(sqlite3_stmt *stmt, int col, const secp256k1_ecdsa_signature *sig) { bool ok; u8 buf[64]; ok = secp256k1_ecdsa_signature_serialize_compact(secp256k1_ctx, buf, sig) == 1; int err = sqlite3_bind_blob(stmt, col, buf, sizeof(buf), SQLITE_TRANSIENT); return ok && err == SQLITE_OK; } bool sqlite3_column_signature(sqlite3_stmt *stmt, int col, secp256k1_ecdsa_signature *sig) { assert(sqlite3_column_bytes(stmt, col) == 64); return secp256k1_ecdsa_signature_parse_compact( secp256k1_ctx, sig, sqlite3_column_blob(stmt, col)) == 1; } bool sqlite3_column_pubkey(sqlite3_stmt *stmt, int col, struct pubkey *dest) { assert(sqlite3_column_bytes(stmt, col) == PUBKEY_CMPR_LEN); return pubkey_from_der(sqlite3_column_blob(stmt, col), PUBKEY_CMPR_LEN, dest); } bool sqlite3_bind_pubkey(sqlite3_stmt *stmt, int col, const struct pubkey *pk) { u8 der[PUBKEY_CMPR_LEN]; pubkey_to_der(der, pk); int err = sqlite3_bind_blob(stmt, col, der, sizeof(der), SQLITE_TRANSIENT); return err == SQLITE_OK; } bool sqlite3_column_node_id(sqlite3_stmt *stmt, int col, struct node_id *dest) { assert(sqlite3_column_bytes(stmt, col) == sizeof(dest->k)); memcpy(dest->k, sqlite3_column_blob(stmt, col), sizeof(dest->k)); return node_id_valid(dest); } bool sqlite3_bind_node_id(sqlite3_stmt *stmt, int col, const struct node_id *id) { assert(node_id_valid(id)); int err = sqlite3_bind_blob(stmt, col, id->k, sizeof(id->k), SQLITE_TRANSIENT); return err == SQLITE_OK; } bool sqlite3_bind_pubkey_array(sqlite3_stmt *stmt, int col, const struct pubkey *pks) { size_t n; size_t i; u8 *ders; if (!pks) { int err = sqlite3_bind_null(stmt, col); return err == SQLITE_OK; } n = tal_count(pks); ders = tal_arr(NULL, u8, n * PUBKEY_CMPR_LEN); for (i = 0; i < n; ++i) pubkey_to_der(&ders[i * PUBKEY_CMPR_LEN], &pks[i]); int err = sqlite3_bind_blob(stmt, col, ders, tal_count(ders), SQLITE_TRANSIENT); tal_free(ders); return err == SQLITE_OK; } struct pubkey *sqlite3_column_pubkey_array(const tal_t *ctx, sqlite3_stmt *stmt, int col) { size_t i; size_t n; struct pubkey *ret; const u8 *ders; if (sqlite3_column_type(stmt, col) == SQLITE_NULL) return NULL; n = sqlite3_column_bytes(stmt, col) / PUBKEY_CMPR_LEN; assert(n * PUBKEY_CMPR_LEN == (size_t)sqlite3_column_bytes(stmt, col)); ret = tal_arr(ctx, struct pubkey, n); ders = sqlite3_column_blob(stmt, col); for (i = 0; i < n; ++i) { if (!pubkey_from_der(&ders[i * PUBKEY_CMPR_LEN], PUBKEY_CMPR_LEN, &ret[i])) return tal_free(ret); } return ret; } bool sqlite3_bind_node_id_array(sqlite3_stmt *stmt, int col, const struct node_id *ids) { size_t n; u8 *arr; if (!ids) { int err = sqlite3_bind_null(stmt, col); return err == SQLITE_OK; } /* Copy into contiguous array: ARM will add padding to struct node_id! */ n = tal_count(ids); arr = tal_arr(NULL, u8, n * sizeof(ids[0].k)); for (size_t i = 0; i < n; ++i) { assert(node_id_valid(&ids[i])); memcpy(arr + sizeof(ids[i].k) * i, ids[i].k, sizeof(ids[i].k)); } int err = sqlite3_bind_blob(stmt, col, arr, tal_count(arr), SQLITE_TRANSIENT); tal_free(arr); return err == SQLITE_OK; } struct node_id *sqlite3_column_node_id_array(const tal_t *ctx, sqlite3_stmt *stmt, int col) { size_t n; struct node_id *ret; const u8 *arr; if (sqlite3_column_type(stmt, col) == SQLITE_NULL) return NULL; n = sqlite3_column_bytes(stmt, col) / sizeof(ret->k); assert(n * sizeof(ret->k) == (size_t)sqlite3_column_bytes(stmt, col)); ret = tal_arr(ctx, struct node_id, n); arr = sqlite3_column_blob(stmt, col); for (size_t i = 0; i < n; i++) { memcpy(ret[i].k, arr + i * sizeof(ret[i].k), sizeof(ret[i].k)); if (!node_id_valid(&ret[i])) return tal_free(ret); } return ret; } bool sqlite3_column_preimage(sqlite3_stmt *stmt, int col, struct preimage *dest) { assert(sqlite3_column_bytes(stmt, col) == sizeof(struct preimage)); return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct preimage)); } bool sqlite3_bind_preimage(sqlite3_stmt *stmt, int col, const struct preimage *p) { int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct preimage), SQLITE_TRANSIENT); return err == SQLITE_OK; } bool sqlite3_column_sha256(sqlite3_stmt *stmt, int col, struct sha256 *dest) { assert(sqlite3_column_bytes(stmt, col) == sizeof(struct sha256)); return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct sha256)); } bool sqlite3_bind_sha256(sqlite3_stmt *stmt, int col, const struct sha256 *p) { int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct sha256), SQLITE_TRANSIENT); return err == SQLITE_OK; } bool sqlite3_column_sha256_double(sqlite3_stmt *stmt, int col, struct sha256_double *dest) { assert(sqlite3_column_bytes(stmt, col) == sizeof(struct sha256_double)); return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct sha256_double)); } struct secret *sqlite3_column_secrets(const tal_t *ctx, sqlite3_stmt *stmt, int col) { return sqlite3_column_arr(ctx, stmt, col, struct secret); } bool sqlite3_bind_sha256_double(sqlite3_stmt *stmt, int col, const struct sha256_double *p) { int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct sha256_double), SQLITE_TRANSIENT); return err == SQLITE_OK; } struct json_escaped *sqlite3_column_json_escaped(const tal_t *ctx, sqlite3_stmt *stmt, int col) { return json_escaped_string_(ctx, sqlite3_column_blob(stmt, col), sqlite3_column_bytes(stmt, col)); } bool sqlite3_bind_json_escaped(sqlite3_stmt *stmt, int col, const struct json_escaped *esc) { int err = sqlite3_bind_text(stmt, col, esc->s, strlen(esc->s), SQLITE_TRANSIENT); return err == SQLITE_OK; } struct amount_msat sqlite3_column_amount_msat(sqlite3_stmt *stmt, int col) { struct amount_msat msat; msat.millisatoshis = sqlite3_column_int64(stmt, col); /* Raw: low level function */ return msat; } struct amount_sat sqlite3_column_amount_sat(sqlite3_stmt *stmt, int col) { struct amount_sat sat; sat.satoshis = sqlite3_column_int64(stmt, col); /* Raw: low level function */ return sat; } void sqlite3_bind_amount_msat(sqlite3_stmt *stmt, int col, struct amount_msat msat) { sqlite3_bind_int64(stmt, col, msat.millisatoshis); /* Raw: low level function */ } void sqlite3_bind_amount_sat(sqlite3_stmt *stmt, int col, struct amount_sat sat) { sqlite3_bind_int64(stmt, col, sat.satoshis); /* Raw: low level function */ } /* Will apply the current config fee settings to all channels */ void migrate_pr2342_feerate_per_channel(struct lightningd *ld, struct db *db) { db_exec(__func__, db, "UPDATE channels SET feerate_base = %u, feerate_ppm = %u;", ld->config.fee_base, ld->config.fee_per_satoshi); } void sqlite3_bind_timeabs(sqlite3_stmt *stmt, int col, struct timeabs t) { u64 timestamp = t.ts.tv_nsec + (t.ts.tv_sec * NSEC_IN_SEC); sqlite3_bind_int64(stmt, col, timestamp); } struct timeabs sqlite3_column_timeabs(sqlite3_stmt *stmt, int col) { struct timeabs t; u64 timestamp = sqlite3_column_int64(stmt, col); t.ts.tv_sec = timestamp / NSEC_IN_SEC; t.ts.tv_nsec = timestamp % NSEC_IN_SEC; return t; }