You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

1180 lines
37 KiB

#include "db.h"
#include <ccan/array_size/array_size.h>
#include <ccan/tal/str/str.h>
#include <common/json_escaped.h>
#include <common/node_id.h>
#include <common/version.h>
#include <inttypes.h>
#include <lightningd/lightningd.h>
#include <lightningd/log.h>
#include <lightningd/plugin_hook.h>
#define DB_FILE "lightningd.sqlite3"
#define NSEC_IN_SEC 1000000000
/* For testing, we want to catch fatal messages. */
#ifndef db_fatal
#define db_fatal fatal
#endif
struct migration {
const char *sql;
void (*func)(struct lightningd *ld, struct db *db);
};
void migrate_pr2342_feerate_per_channel(struct lightningd *ld, struct db *db);
/* Do not reorder or remove elements from this array, it is used to
* migrate existing databases from a previous state, based on the
* string indices */
static struct migration dbmigrations[] = {
{ "CREATE TABLE version (version INTEGER)", NULL },
{ "INSERT INTO version VALUES (1)", NULL },
{ "CREATE TABLE outputs ( \
prev_out_tx CHAR(64), \
prev_out_index INTEGER, \
value INTEGER, \
type INTEGER, \
status INTEGER, \
keyindex INTEGER, \
PRIMARY KEY (prev_out_tx, prev_out_index) \
);", NULL },
{ "CREATE TABLE vars (name VARCHAR(32), val VARCHAR(255), PRIMARY KEY (name));", NULL },
{ "CREATE TABLE shachains ( \
id INTEGER, \
min_index INTEGER, \
num_valid INTEGER, \
PRIMARY KEY (id));", NULL },
{ "CREATE TABLE shachain_known ( \
shachain_id INTEGER REFERENCES shachains(id) ON DELETE CASCADE, \
pos INTEGER, \
idx INTEGER, \
hash BLOB, \
PRIMARY KEY (shachain_id, pos));", NULL },
{ "CREATE TABLE channels ("
" id INTEGER," /* chan->id */
" peer_id INTEGER REFERENCES peers(id) ON DELETE CASCADE,"
" short_channel_id BLOB,"
" channel_config_local INTEGER,"
" channel_config_remote INTEGER,"
" state INTEGER,"
" funder INTEGER,"
" channel_flags INTEGER,"
" minimum_depth INTEGER,"
" next_index_local INTEGER,"
" next_index_remote INTEGER,"
" next_htlc_id INTEGER, "
" funding_tx_id BLOB,"
" funding_tx_outnum INTEGER,"
" funding_satoshi INTEGER,"
" funding_locked_remote INTEGER,"
" push_msatoshi INTEGER,"
" msatoshi_local INTEGER," /* our_msatoshi */
/* START channel_info */
" fundingkey_remote BLOB,"
" revocation_basepoint_remote BLOB,"
" payment_basepoint_remote BLOB,"
" htlc_basepoint_remote BLOB,"
" delayed_payment_basepoint_remote BLOB,"
" per_commit_remote BLOB,"
" old_per_commit_remote BLOB,"
" local_feerate_per_kw INTEGER,"
" remote_feerate_per_kw INTEGER,"
/* END channel_info */
" shachain_remote_id INTEGER,"
" shutdown_scriptpubkey_remote BLOB,"
" shutdown_keyidx_local INTEGER,"
" last_sent_commit_state INTEGER,"
" last_sent_commit_id INTEGER,"
" last_tx BLOB,"
" last_sig BLOB,"
" closing_fee_received INTEGER,"
" closing_sig_received BLOB,"
" PRIMARY KEY (id)"
");", NULL },
{ "CREATE TABLE peers ("
" id INTEGER,"
" node_id BLOB UNIQUE," /* pubkey */
" address TEXT,"
" PRIMARY KEY (id)"
");", NULL },
{ "CREATE TABLE channel_configs ("
" id INTEGER,"
" dust_limit_satoshis INTEGER,"
" max_htlc_value_in_flight_msat INTEGER,"
" channel_reserve_satoshis INTEGER,"
" htlc_minimum_msat INTEGER,"
" to_self_delay INTEGER,"
" max_accepted_htlcs INTEGER,"
" PRIMARY KEY (id)"
");", NULL },
{ "CREATE TABLE channel_htlcs ("
" id INTEGER,"
" channel_id INTEGER REFERENCES channels(id) ON DELETE CASCADE,"
" channel_htlc_id INTEGER,"
" direction INTEGER,"
" origin_htlc INTEGER,"
" msatoshi INTEGER,"
" cltv_expiry INTEGER,"
" payment_hash BLOB,"
" payment_key BLOB,"
" routing_onion BLOB,"
" failuremsg BLOB,"
" malformed_onion INTEGER,"
" hstate INTEGER,"
" shared_secret BLOB,"
" PRIMARY KEY (id),"
" UNIQUE (channel_id, channel_htlc_id, direction)"
");", NULL },
{ "CREATE TABLE invoices ("
" id INTEGER,"
" state INTEGER,"
" msatoshi INTEGER,"
" payment_hash BLOB,"
" payment_key BLOB,"
" label TEXT,"
" PRIMARY KEY (id),"
" UNIQUE (label),"
" UNIQUE (payment_hash)"
");", NULL },
{ "CREATE TABLE payments ("
" id INTEGER,"
" timestamp INTEGER,"
" status INTEGER,"
" payment_hash BLOB,"
" direction INTEGER,"
" destination BLOB,"
" msatoshi INTEGER,"
" PRIMARY KEY (id),"
" UNIQUE (payment_hash)"
");", NULL },
/* Add expiry field to invoices (effectively infinite). */
{ "ALTER TABLE invoices ADD expiry_time INTEGER;", NULL },
{ "UPDATE invoices SET expiry_time=9223372036854775807;", NULL },
/* Add pay_index field to paid invoices (initially, same order as id). */
{ "ALTER TABLE invoices ADD pay_index INTEGER;", NULL },
{ "CREATE UNIQUE INDEX invoices_pay_index"
" ON invoices(pay_index);", NULL },
{ "UPDATE invoices SET pay_index=id WHERE state=1;", NULL }, /* only paid invoice */
/* Create next_pay_index variable (highest pay_index). */
{ "INSERT OR REPLACE INTO vars(name, val)"
" VALUES('next_pay_index', "
" COALESCE((SELECT MAX(pay_index) FROM invoices WHERE state=1), 0) + 1"
" );", NULL },
/* Create first_block field; initialize from channel id if any.
* This fails for channels still awaiting lockin, but that only applies to
* pre-release software, so it's forgivable. */
{ "ALTER TABLE channels ADD first_blocknum INTEGER;", NULL },
{ "UPDATE channels SET first_blocknum=CAST(short_channel_id AS INTEGER) WHERE short_channel_id IS NOT NULL;", NULL },
{ "ALTER TABLE outputs ADD COLUMN channel_id INTEGER;", NULL },
{ "ALTER TABLE outputs ADD COLUMN peer_id BLOB;", NULL },
{ "ALTER TABLE outputs ADD COLUMN commitment_point BLOB;", NULL },
{ "ALTER TABLE invoices ADD COLUMN msatoshi_received INTEGER;", NULL },
/* Normally impossible, so at least we'll know if databases are ancient. */
{ "UPDATE invoices SET msatoshi_received=0 WHERE state=1;", NULL },
{ "ALTER TABLE channels ADD COLUMN last_was_revoke INTEGER;", NULL },
/* We no longer record incoming payments: invoices cover that.
* Without ALTER_TABLE DROP COLUMN support we need to do this by
* rename & copy, which works because there are no triggers etc. */
{ "ALTER TABLE payments RENAME TO temp_payments;", NULL },
{ "CREATE TABLE payments ("
" id INTEGER,"
" timestamp INTEGER,"
" status INTEGER,"
" payment_hash BLOB,"
" destination BLOB,"
" msatoshi INTEGER,"
" PRIMARY KEY (id),"
" UNIQUE (payment_hash)"
");", NULL },
{ "INSERT INTO payments SELECT id, timestamp, status, payment_hash, destination, msatoshi FROM temp_payments WHERE direction=1;", NULL },
{ "DROP TABLE temp_payments;", NULL },
/* We need to keep the preimage in case they ask to pay again. */
{ "ALTER TABLE payments ADD COLUMN payment_preimage BLOB;", NULL },
/* We need to keep the shared secrets to decode error returns. */
{ "ALTER TABLE payments ADD COLUMN path_secrets BLOB;", NULL },
/* Create time-of-payment of invoice, default already-paid
* invoices to current time. */
{ "ALTER TABLE invoices ADD paid_timestamp INTEGER;", NULL },
{ "UPDATE invoices"
" SET paid_timestamp = strftime('%s', 'now')"
" WHERE state = 1;", NULL },
/* We need to keep the route node pubkeys and short channel ids to
* correctly mark routing failures. We separate short channel ids
* because we cannot safely save them as blobs due to byteorder
* concerns. */
{ "ALTER TABLE payments ADD COLUMN route_nodes BLOB;", NULL },
{ "ALTER TABLE payments ADD COLUMN route_channels TEXT;", NULL },
{ "CREATE TABLE htlc_sigs (channelid INTEGER REFERENCES channels(id) ON DELETE CASCADE, signature BLOB);", NULL },
{ "CREATE INDEX channel_idx ON htlc_sigs (channelid)", NULL },
/* Get rid of OPENINGD entries; we don't put them in db any more */
{ "DELETE FROM channels WHERE state=1", NULL },
/* Keep track of db upgrades, for debugging */
{ "CREATE TABLE db_upgrades (upgrade_from INTEGER, lightning_version TEXT);", NULL },
/* We used not to clean up peers when their channels were gone. */
{ "DELETE FROM peers WHERE id NOT IN (SELECT peer_id FROM channels);", NULL },
/* The ONCHAIND_CHEATED/THEIR_UNILATERAL/OUR_UNILATERAL/MUTUAL are now one */
{ "UPDATE channels SET STATE = 8 WHERE state > 8;", NULL },
/* Add bolt11 to invoices table*/
{ "ALTER TABLE invoices ADD bolt11 TEXT;", NULL },
/* What do we think the head of the blockchain looks like? Used
* primarily to track confirmations across restarts and making
* sure we handle reorgs correctly. */
{ "CREATE TABLE blocks (height INT, hash BLOB, prev_hash BLOB, UNIQUE(height));", NULL },
/* ON DELETE CASCADE would have been nice for confirmation_height,
* so that we automatically delete outputs that fall off the
* blockchain and then we rediscover them if they are included
* again. However, we have the their_unilateral/to_us which we
* can't simply recognize from the chain without additional
* hints. So we just mark them as unconfirmed should the block
* die. */
{ "ALTER TABLE outputs ADD COLUMN confirmation_height INTEGER REFERENCES blocks(height) ON DELETE SET NULL;", NULL },
{ "ALTER TABLE outputs ADD COLUMN spend_height INTEGER REFERENCES blocks(height) ON DELETE SET NULL;", NULL },
/* Create a covering index that covers both fields */
{ "CREATE INDEX output_height_idx ON outputs (confirmation_height, spend_height);", NULL },
{ "CREATE TABLE utxoset ("
" txid BLOB,"
" outnum INT,"
" blockheight INT REFERENCES blocks(height) ON DELETE CASCADE,"
" spendheight INT REFERENCES blocks(height) ON DELETE SET NULL,"
" txindex INT,"
" scriptpubkey BLOB,"
" satoshis BIGINT,"
" PRIMARY KEY(txid, outnum));", NULL },
{ "CREATE INDEX short_channel_id ON utxoset (blockheight, txindex, outnum)", NULL },
/* Necessary index for long rollbacks of the blockchain, otherwise we're
* doing table scans for every block removed. */
{ "CREATE INDEX utxoset_spend ON utxoset (spendheight)", NULL },
/* Assign key 0 to unassigned shutdown_keyidx_local. */
{ "UPDATE channels SET shutdown_keyidx_local=0 WHERE shutdown_keyidx_local = -1;", NULL },
/* FIXME: We should rename shutdown_keyidx_local to final_key_index */
/* -- Payment routing failure information -- */
/* BLOB if failure was due to unparseable onion, NULL otherwise */
{ "ALTER TABLE payments ADD failonionreply BLOB;", NULL },
/* 0 if we could theoretically retry, 1 if PERM fail at payee */
{ "ALTER TABLE payments ADD faildestperm INTEGER;", NULL },
/* Contents of routing_failure (only if not unparseable onion) */
{ "ALTER TABLE payments ADD failindex INTEGER;", NULL }, /* erring_index */
{ "ALTER TABLE payments ADD failcode INTEGER;", NULL }, /* failcode */
{ "ALTER TABLE payments ADD failnode BLOB;", NULL }, /* erring_node */
{ "ALTER TABLE payments ADD failchannel BLOB;", NULL }, /* erring_channel */
{ "ALTER TABLE payments ADD failupdate BLOB;", NULL }, /* channel_update - can be NULL*/
/* -- Payment routing failure information ends -- */
/* Delete route data for already succeeded or failed payments */
{ "UPDATE payments"
" SET path_secrets = NULL"
" , route_nodes = NULL"
" , route_channels = NULL"
" WHERE status <> 0;", NULL }, /* PAYMENT_PENDING */
/* -- Routing statistics -- */
{ "ALTER TABLE channels ADD in_payments_offered INTEGER;", NULL },
{ "ALTER TABLE channels ADD in_payments_fulfilled INTEGER;", NULL },
{ "ALTER TABLE channels ADD in_msatoshi_offered INTEGER;", NULL },
{ "ALTER TABLE channels ADD in_msatoshi_fulfilled INTEGER;", NULL },
{ "ALTER TABLE channels ADD out_payments_offered INTEGER;", NULL },
{ "ALTER TABLE channels ADD out_payments_fulfilled INTEGER;", NULL },
{ "ALTER TABLE channels ADD out_msatoshi_offered INTEGER;", NULL },
{ "ALTER TABLE channels ADD out_msatoshi_fulfilled INTEGER;", NULL },
{ "UPDATE channels"
" SET in_payments_offered = 0, in_payments_fulfilled = 0"
" , in_msatoshi_offered = 0, in_msatoshi_fulfilled = 0"
" , out_payments_offered = 0, out_payments_fulfilled = 0"
" , out_msatoshi_offered = 0, out_msatoshi_fulfilled = 0"
" ;", NULL },
/* -- Routing statistics ends --*/
/* Record the msatoshi actually sent in a payment. */
{ "ALTER TABLE payments ADD msatoshi_sent INTEGER;", NULL },
{ "UPDATE payments SET msatoshi_sent = msatoshi;", NULL },
/* Delete dangling utxoset entries due to Issue #1280 */
{ "DELETE FROM utxoset WHERE blockheight IN ("
" SELECT DISTINCT(blockheight)"
" FROM utxoset LEFT OUTER JOIN blocks on (blockheight == blocks.height) "
" WHERE blocks.hash IS NULL"
");", NULL },
/* Record feerate range, to optimize onchaind grinding actual fees. */
{ "ALTER TABLE channels ADD min_possible_feerate INTEGER;", NULL },
{ "ALTER TABLE channels ADD max_possible_feerate INTEGER;", NULL },
/* https://bitcoinfees.github.io/#1d says Dec 17 peak was ~1M sat/kb
* which is 250,000 sat/Sipa */
{ "UPDATE channels SET min_possible_feerate=0, max_possible_feerate=250000;", NULL },
/* -- Min and max msatoshi_to_us -- */
{ "ALTER TABLE channels ADD msatoshi_to_us_min INTEGER;", NULL },
{ "ALTER TABLE channels ADD msatoshi_to_us_max INTEGER;", NULL },
{ "UPDATE channels"
" SET msatoshi_to_us_min = msatoshi_local"
" , msatoshi_to_us_max = msatoshi_local"
" ;", NULL },
/* -- Min and max msatoshi_to_us ends -- */
/* Transactions we are interested in. Either we sent them ourselves or we
* are watching them. We don't cascade block height deletes so we don't
* forget any of them by accident.*/
{ "CREATE TABLE transactions ("
" id BLOB"
", blockheight INTEGER REFERENCES blocks(height) ON DELETE SET NULL"
", txindex INTEGER"
", rawtx BLOB"
", PRIMARY KEY (id)"
");", NULL },
/* -- Detailed payment failure -- */
{ "ALTER TABLE payments ADD faildetail TEXT;", NULL },
{ "UPDATE payments"
" SET faildetail = 'unspecified payment failure reason'"
" WHERE status = 2;", NULL }, /* PAYMENT_FAILED */
/* -- Detailed payment faiure ends -- */
{ "CREATE TABLE channeltxs ("
/* The id serves as insertion order and short ID */
" id INTEGER"
", channel_id INTEGER REFERENCES channels(id) ON DELETE CASCADE"
", type INTEGER"
", transaction_id BLOB REFERENCES transactions(id) ON DELETE CASCADE"
/* The input_num is only used by the txo_watch, 0 if txwatch */
", input_num INTEGER"
/* The height at which we sent the depth notice */
", blockheight INTEGER REFERENCES blocks(height) ON DELETE CASCADE"
", PRIMARY KEY(id)"
");", NULL },
/* -- Set the correct rescan height for PR #1398 -- */
/* Delete blocks that are higher than our initial scan point, this is a
* no-op if we don't have a channel. */
{ "DELETE FROM blocks WHERE height > (SELECT MIN(first_blocknum) FROM channels);", NULL },
/* Now make sure we have the lower bound block with the first_blocknum
* height. This may introduce a block with NULL height if we didn't have any
* blocks, remove that in the next. */
{ "INSERT OR IGNORE INTO blocks (height) VALUES ((SELECT MIN(first_blocknum) FROM channels));", NULL },
{ "DELETE FROM blocks WHERE height IS NULL;", NULL },
/* -- End of PR #1398 -- */
{ "ALTER TABLE invoices ADD description TEXT;", NULL },
/* FIXME: payments table 'description' is really a 'label' */
{ "ALTER TABLE payments ADD description TEXT;", NULL },
/* future_per_commitment_point if other side proves we're out of date -- */
{ "ALTER TABLE channels ADD future_per_commitment_point BLOB;", NULL },
/* last_sent_commit array fix */
{ "ALTER TABLE channels ADD last_sent_commit BLOB;", NULL },
/* Stats table to track forwarded HTLCs. The values in the HTLCs
* and their states are replicated here and the entries are not
* deleted when the HTLC entries or the channel entries are
* deleted to avoid unexpected drops in statistics. */
{ "CREATE TABLE forwarded_payments ("
" in_htlc_id INTEGER REFERENCES channel_htlcs(id) ON DELETE SET NULL"
", out_htlc_id INTEGER REFERENCES channel_htlcs(id) ON DELETE SET NULL"
", in_channel_scid INTEGER"
", out_channel_scid INTEGER"
", in_msatoshi INTEGER"
", out_msatoshi INTEGER"
", state INTEGER"
", UNIQUE(in_htlc_id, out_htlc_id)"
");", NULL },
/* Add a direction for failed payments. */
{ "ALTER TABLE payments ADD faildirection INTEGER;", NULL }, /* erring_direction */
/* Fix dangling peers with no channels. */
{ "DELETE FROM peers WHERE id NOT IN (SELECT peer_id FROM channels);", NULL },
{ "ALTER TABLE outputs ADD scriptpubkey BLOB;", NULL },
/* Keep bolt11 string for payments. */
{ "ALTER TABLE payments ADD bolt11 TEXT;", NULL },
/* PR #2342 feerate per channel */
{ "ALTER TABLE channels ADD feerate_base INTEGER;", NULL },
{ "ALTER TABLE channels ADD feerate_ppm INTEGER;", NULL },
{ NULL, migrate_pr2342_feerate_per_channel },
{ "ALTER TABLE channel_htlcs ADD received_time INTEGER", NULL },
{ "ALTER TABLE forwarded_payments ADD received_time INTEGER", NULL },
{ "ALTER TABLE forwarded_payments ADD resolved_time INTEGER", NULL },
{ "ALTER TABLE channels ADD remote_upfront_shutdown_script BLOB;", NULL },
/* PR #2524: Add failcode into forward_payment */
{ "ALTER TABLE forwarded_payments ADD failcode INTEGER;", NULL },
};
/* Leak tracking. */
#if DEVELOPER
/* We need a global here, since caller has no context. Yuck! */
static struct list_head db_statements = LIST_HEAD_INIT(db_statements);
struct db_statement {
struct list_node list;
sqlite3_stmt *stmt;
const char *origin;
};
static struct db_statement *find_statement(sqlite3_stmt *stmt)
{
struct db_statement *i;
list_for_each(&db_statements, i, list) {
if (i->stmt == stmt)
return i;
}
return NULL;
}
void db_assert_no_outstanding_statements(void)
{
struct db_statement *dbstat;
dbstat = list_top(&db_statements, struct db_statement, list);
if (dbstat)
db_fatal("Unfinalized statement %s", dbstat->origin);
}
static void dev_statement_start(sqlite3_stmt *stmt, const char *origin)
{
struct db_statement *dbstat = tal(NULL, struct db_statement);
dbstat->stmt = stmt;
dbstat->origin = origin;
list_add(&db_statements, &dbstat->list);
}
static void dev_statement_end(sqlite3_stmt *stmt)
{
struct db_statement *dbstat = find_statement(stmt);
list_del_from(&db_statements, &dbstat->list);
tal_free(dbstat);
}
#else
static void dev_statement_start(sqlite3_stmt *stmt, const char *origin)
{
}
static void dev_statement_end(sqlite3_stmt *stmt)
{
}
void db_assert_no_outstanding_statements(void)
{
}
#endif
#if !HAVE_SQLITE3_EXPANDED_SQL
/* Prior to sqlite3 v3.14, we have to use tracing to dump statements */
static void trace_sqlite3(void *dbv, const char *stmt)
{
struct db *db = dbv;
/* We get a "COMMIT;" after we've sent our changes. */
if (!db->changes) {
assert(streq(stmt, "COMMIT;"));
return;
}
tal_arr_expand(&db->changes, tal_strdup(db->changes, stmt));
}
#endif
void db_stmt_done(sqlite3_stmt *stmt)
{
dev_statement_end(stmt);
sqlite3_finalize(stmt);
}
sqlite3_stmt *db_select_prepare_(const char *location, struct db *db, const char *query)
{
int err;
sqlite3_stmt *stmt;
const char *full_query = tal_fmt(db, "SELECT %s", query);
assert(db->in_transaction);
err = sqlite3_prepare_v2(db->sql, full_query, -1, &stmt, NULL);
if (err != SQLITE_OK)
db_fatal("%s: %s: %s", location, full_query, sqlite3_errmsg(db->sql));
dev_statement_start(stmt, location);
tal_free(full_query);
return stmt;
}
bool db_select_step_(const char *location, struct db *db, struct sqlite3_stmt *stmt)
{
int ret;
ret = sqlite3_step(stmt);
if (ret == SQLITE_ROW)
return true;
if (ret != SQLITE_DONE)
db_fatal("%s: %s", location, sqlite3_errmsg(db->sql));
db_stmt_done(stmt);
return false;
}
sqlite3_stmt *db_prepare_(const char *location, struct db *db, const char *query)
{
int err;
sqlite3_stmt *stmt;
assert(db->in_transaction);
err = sqlite3_prepare_v2(db->sql, query, -1, &stmt, NULL);
if (err != SQLITE_OK)
db_fatal("%s: %s: %s", location, query, sqlite3_errmsg(db->sql));
dev_statement_start(stmt, location);
return stmt;
}
void db_exec_prepared_(const char *caller, struct db *db, sqlite3_stmt *stmt)
{
assert(db->in_transaction);
if (sqlite3_step(stmt) != SQLITE_DONE)
db_fatal("%s: %s", caller, sqlite3_errmsg(db->sql));
#if HAVE_SQLITE3_EXPANDED_SQL
char *expanded_sql;
expanded_sql = sqlite3_expanded_sql(stmt);
tal_arr_expand(&db->changes,
tal_strdup(db->changes, expanded_sql));
sqlite3_free(expanded_sql);
#endif
db_stmt_done(stmt);
}
/* This one doesn't check if we're in a transaction. */
static void db_do_exec(const char *caller, struct db *db, const char *cmd)
{
char *errmsg;
int err;
err = sqlite3_exec(db->sql, cmd, NULL, NULL, &errmsg);
if (err != SQLITE_OK) {
db_fatal("%s:%s:%s:%s", caller, sqlite3_errstr(err), cmd, errmsg);
/* Only reached in testing */
sqlite3_free(errmsg);
}
#if HAVE_SQLITE3_EXPANDED_SQL
tal_arr_expand(&db->changes, tal_strdup(db->changes, cmd));
#endif
}
static void PRINTF_FMT(3, 4)
db_exec(const char *caller, struct db *db, const char *fmt, ...)
{
va_list ap;
char *cmd;
assert(db->in_transaction);
va_start(ap, fmt);
cmd = tal_vfmt(db, fmt, ap);
va_end(ap);
db_do_exec(caller, db, cmd);
tal_free(cmd);
}
/* This one can fail: returns NULL if so */
static sqlite3_stmt *db_query(const char *location,
struct db *db, const char *query)
{
sqlite3_stmt *stmt;
assert(db->in_transaction);
/* Sets stmt to NULL if not SQLITE_OK */
sqlite3_prepare_v2(db->sql, query, -1, &stmt, NULL);
if (stmt)
dev_statement_start(stmt, location);
return stmt;
}
sqlite3_stmt *PRINTF_FMT(3, 4)
db_select_(const char *location, struct db *db, const char *fmt, ...)
{
va_list ap;
char *query = tal_strdup(db, "SELECT ");
sqlite3_stmt *stmt;
assert(db->in_transaction);
va_start(ap, fmt);
tal_append_vfmt(&query, fmt, ap);
va_end(ap);
stmt = db_query(location, db, query);
if (!stmt)
db_fatal("%s:%s:%s", location, query, sqlite3_errmsg(db->sql));
tal_free(query);
return stmt;
}
static void destroy_db(struct db *db)
{
db_assert_no_outstanding_statements();
sqlite3_close(db->sql);
}
/* We expect min changes (ie. BEGIN TRANSACTION): report if more.
* Optionally add "final" at the end (ie. COMMIT). */
static void db_report_changes(struct db *db, const char *final, size_t min)
{
assert(db->changes);
assert(tal_count(db->changes) >= min);
if (tal_count(db->changes) > min)
plugin_hook_db_sync(db, db->changes, final);
db->changes = tal_free(db->changes);
}
static void db_prepare_for_changes(struct db *db)
{
assert(!db->changes);
db->changes = tal_arr(db, const char *, 0);
}
void db_begin_transaction_(struct db *db, const char *location)
{
if (db->in_transaction)
db_fatal("Already in transaction from %s", db->in_transaction);
db_prepare_for_changes(db);
db_do_exec(location, db, "BEGIN TRANSACTION;");
db->in_transaction = location;
}
void db_commit_transaction(struct db *db)
{
int err;
char *errmsg;
const char *cmd = "COMMIT;";
assert(db->in_transaction);
db_assert_no_outstanding_statements();
/* We expect at least the BEGIN TRANSACTION */
db_report_changes(db, cmd, 1);
err = sqlite3_exec(db->sql, cmd, NULL, NULL, &errmsg);
if (err != SQLITE_OK)
db_fatal("%s:%s:%s:%s", __func__, sqlite3_errstr(err), cmd, errmsg);
db->in_transaction = NULL;
}
static void setup_open_db(struct db *db)
{
#if !HAVE_SQLITE3_EXPANDED_SQL
sqlite3_trace(db->sql, trace_sqlite3, db);
#endif
/* This must be outside a transaction, so catch it */
assert(!db->in_transaction);
db_prepare_for_changes(db);
db_do_exec(__func__, db, "PRAGMA foreign_keys = ON;");
db_report_changes(db, NULL, 0);
}
/**
* db_open - Open or create a sqlite3 database
*/
static struct db *db_open(const tal_t *ctx, char *filename)
{
int err;
struct db *db;
sqlite3 *sql;
int flags = SQLITE_OPEN_READWRITE | SQLITE_OPEN_CREATE;
err = sqlite3_open_v2(filename, &sql, flags, NULL);
if (err != SQLITE_OK) {
db_fatal("failed to open database %s: %s", filename,
sqlite3_errstr(err));
}
db = tal(ctx, struct db);
db->filename = tal_strdup(db, filename);
db->sql = sql;
tal_add_destructor(db, destroy_db);
db->in_transaction = NULL;
db->changes = NULL;
setup_open_db(db);
return db;
}
/**
* db_get_version - Determine the current DB schema version
*
* Will attempt to determine the current schema version of the
* database @db by querying the `version` table. If the table does not
* exist it'll return schema version -1, so that migration 0 is
* applied, which should create the `version` table.
*/
static int db_get_version(struct db *db)
{
int res;
sqlite3_stmt *stmt = db_query(__func__,
db, "SELECT version FROM version LIMIT 1");
if (!stmt)
return -1;
if (!db_select_step(db, stmt))
return -1;
res = sqlite3_column_int64(stmt, 0);
db_stmt_done(stmt);
return res;
}
/**
* db_migrate - Apply all remaining migrations from the current version
*/
static void db_migrate(struct lightningd *ld, struct db *db, struct log *log)
{
/* Attempt to read the version from the database */
int current, orig, available;
db_begin_transaction(db);
orig = current = db_get_version(db);
available = ARRAY_SIZE(dbmigrations) - 1;
if (current == -1)
log_info(log, "Creating database");
else if (available < current)
db_fatal("Refusing to migrate down from version %u to %u",
current, available);
else if (current != available)
log_info(log, "Updating database from version %u to %u",
current, available);
while (current < available) {
current++;
if (dbmigrations[current].sql)
db_exec(__func__, db, "%s", dbmigrations[current].sql);
if (dbmigrations[current].func)
dbmigrations[current].func(ld, db);
}
/* Finally update the version number in the version table */
db_exec(__func__, db, "UPDATE version SET version=%d;", available);
/* Annotate that we did upgrade, if any. */
if (current != orig)
db_exec(__func__, db,
"INSERT INTO db_upgrades VALUES (%i, '%s');",
orig, version());
db_commit_transaction(db);
}
struct db *db_setup(const tal_t *ctx, struct lightningd *ld, struct log *log)
{
struct db *db = db_open(ctx, DB_FILE);
db_migrate(ld, db, log);
return db;
}
void db_close_for_fork(struct db *db)
{
/* https://www.sqlite.org/faq.html#q6
*
* Under Unix, you should not carry an open SQLite database across a
* fork() system call into the child process. */
if (sqlite3_close(db->sql) != SQLITE_OK)
db_fatal("sqlite3_close: %s", sqlite3_errmsg(db->sql));
db->sql = NULL;
}
void db_reopen_after_fork(struct db *db)
{
int err = sqlite3_open_v2(db->filename, &db->sql,
SQLITE_OPEN_READWRITE, NULL);
if (err != SQLITE_OK) {
db_fatal("failed to re-open database %s: %s", db->filename,
sqlite3_errstr(err));
}
setup_open_db(db);
}
s64 db_get_intvar(struct db *db, char *varname, s64 defval)
{
s64 res;
sqlite3_stmt *stmt;
const char *query;
query = tal_fmt(db, "SELECT val FROM vars WHERE name='%s' LIMIT 1", varname);
stmt = db_query(__func__, db, query);
tal_free(query);
if (!stmt)
return defval;
if (db_select_step(db, stmt)) {
const unsigned char *stringvar = sqlite3_column_text(stmt, 0);
res = atol((const char *)stringvar);
db_stmt_done(stmt);
} else
res = defval;
return res;
}
void db_set_intvar(struct db *db, char *varname, s64 val)
{
/* Attempt to update */
db_exec(__func__, db,
"UPDATE vars SET val='%" PRId64 "' WHERE name='%s';", val,
varname);
if (sqlite3_changes(db->sql) == 0)
db_exec(
__func__, db,
"INSERT INTO vars (name, val) VALUES ('%s', '%" PRId64
"');",
varname, val);
}
void *sqlite3_column_arr_(const tal_t *ctx, sqlite3_stmt *stmt, int col,
size_t bytes, const char *label, const char *caller)
{
size_t sourcelen = sqlite3_column_bytes(stmt, col);
void *p;
if (sqlite3_column_type(stmt, col) == SQLITE_NULL)
return NULL;
if (sourcelen % bytes != 0)
db_fatal("%s: column size %zu not a multiple of %s (%zu)",
caller, sourcelen, label, bytes);
p = tal_arr_label(ctx, char, sourcelen, label);
memcpy(p, sqlite3_column_blob(stmt, col), sourcelen);
return p;
}
bool sqlite3_bind_short_channel_id(sqlite3_stmt *stmt, int col,
const struct short_channel_id *id)
{
char *ser = short_channel_id_to_str(id, id);
int err = sqlite3_bind_blob(stmt, col, ser, strlen(ser), SQLITE_TRANSIENT);
tal_free(ser);
return err == SQLITE_OK;
}
bool sqlite3_column_short_channel_id(sqlite3_stmt *stmt, int col,
struct short_channel_id *dest)
{
const char *source = sqlite3_column_blob(stmt, col);
size_t sourcelen = sqlite3_column_bytes(stmt, col);
return short_channel_id_from_str(source, sourcelen, dest, true);
}
bool sqlite3_bind_short_channel_id_array(sqlite3_stmt *stmt, int col,
const struct short_channel_id *id)
{
u8 *ser;
size_t num;
size_t i;
/* Handle nulls early. */
if (!id) {
int err = sqlite3_bind_null(stmt, col);
return err == SQLITE_OK;
}
ser = tal_arr(NULL, u8, 0);
num = tal_count(id);
for (i = 0; i < num; ++i)
towire_short_channel_id(&ser, &id[i]);
int err = sqlite3_bind_blob(stmt, col, ser, tal_count(ser), SQLITE_TRANSIENT);
tal_free(ser);
return err == SQLITE_OK;
}
struct short_channel_id *
sqlite3_column_short_channel_id_array(const tal_t *ctx,
sqlite3_stmt *stmt, int col)
{
const u8 *ser;
size_t len;
struct short_channel_id *ret;
/* Handle nulls early. */
if (sqlite3_column_type(stmt, col) == SQLITE_NULL)
return NULL;
ser = sqlite3_column_blob(stmt, col);
len = sqlite3_column_bytes(stmt, col);
ret = tal_arr(ctx, struct short_channel_id, 0);
while (len != 0) {
struct short_channel_id scid;
fromwire_short_channel_id(&ser, &len, &scid);
tal_arr_expand(&ret, scid);
}
return ret;
}
bool sqlite3_bind_tx(sqlite3_stmt *stmt, int col, const struct bitcoin_tx *tx)
{
u8 *ser = linearize_tx(NULL, tx);
int err = sqlite3_bind_blob(stmt, col, ser, tal_count(ser), SQLITE_TRANSIENT);
tal_free(ser);
return err == SQLITE_OK;
}
struct bitcoin_tx *sqlite3_column_tx(const tal_t *ctx, sqlite3_stmt *stmt,
int col)
{
const u8 *src = sqlite3_column_blob(stmt, col);
size_t len = sqlite3_column_bytes(stmt, col);
return pull_bitcoin_tx(ctx, &src, &len);
}
bool sqlite3_bind_signature(sqlite3_stmt *stmt, int col,
const secp256k1_ecdsa_signature *sig)
{
bool ok;
u8 buf[64];
ok = secp256k1_ecdsa_signature_serialize_compact(secp256k1_ctx, buf,
sig) == 1;
int err = sqlite3_bind_blob(stmt, col, buf, sizeof(buf), SQLITE_TRANSIENT);
return ok && err == SQLITE_OK;
}
bool sqlite3_column_signature(sqlite3_stmt *stmt, int col,
secp256k1_ecdsa_signature *sig)
{
assert(sqlite3_column_bytes(stmt, col) == 64);
return secp256k1_ecdsa_signature_parse_compact(
secp256k1_ctx, sig, sqlite3_column_blob(stmt, col)) == 1;
}
bool sqlite3_column_pubkey(sqlite3_stmt *stmt, int col, struct pubkey *dest)
{
assert(sqlite3_column_bytes(stmt, col) == PUBKEY_CMPR_LEN);
return pubkey_from_der(sqlite3_column_blob(stmt, col), PUBKEY_CMPR_LEN, dest);
}
bool sqlite3_bind_pubkey(sqlite3_stmt *stmt, int col, const struct pubkey *pk)
{
u8 der[PUBKEY_CMPR_LEN];
pubkey_to_der(der, pk);
int err = sqlite3_bind_blob(stmt, col, der, sizeof(der), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
bool sqlite3_column_node_id(sqlite3_stmt *stmt, int col, struct node_id *dest)
{
assert(sqlite3_column_bytes(stmt, col) == sizeof(dest->k));
memcpy(dest->k, sqlite3_column_blob(stmt, col), sizeof(dest->k));
return node_id_valid(dest);
}
bool sqlite3_bind_node_id(sqlite3_stmt *stmt, int col, const struct node_id *id)
{
assert(node_id_valid(id));
int err = sqlite3_bind_blob(stmt, col, id->k, sizeof(id->k), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
bool sqlite3_bind_pubkey_array(sqlite3_stmt *stmt, int col,
const struct pubkey *pks)
{
size_t n;
size_t i;
u8 *ders;
if (!pks) {
int err = sqlite3_bind_null(stmt, col);
return err == SQLITE_OK;
}
n = tal_count(pks);
ders = tal_arr(NULL, u8, n * PUBKEY_CMPR_LEN);
for (i = 0; i < n; ++i)
pubkey_to_der(&ders[i * PUBKEY_CMPR_LEN], &pks[i]);
int err = sqlite3_bind_blob(stmt, col, ders, tal_count(ders), SQLITE_TRANSIENT);
tal_free(ders);
return err == SQLITE_OK;
}
struct pubkey *sqlite3_column_pubkey_array(const tal_t *ctx,
sqlite3_stmt *stmt, int col)
{
size_t i;
size_t n;
struct pubkey *ret;
const u8 *ders;
if (sqlite3_column_type(stmt, col) == SQLITE_NULL)
return NULL;
n = sqlite3_column_bytes(stmt, col) / PUBKEY_CMPR_LEN;
assert(n * PUBKEY_CMPR_LEN == (size_t)sqlite3_column_bytes(stmt, col));
ret = tal_arr(ctx, struct pubkey, n);
ders = sqlite3_column_blob(stmt, col);
for (i = 0; i < n; ++i) {
if (!pubkey_from_der(&ders[i * PUBKEY_CMPR_LEN], PUBKEY_CMPR_LEN, &ret[i]))
return tal_free(ret);
}
return ret;
}
bool sqlite3_bind_node_id_array(sqlite3_stmt *stmt, int col,
const struct node_id *ids)
{
size_t n;
u8 *arr;
if (!ids) {
int err = sqlite3_bind_null(stmt, col);
return err == SQLITE_OK;
}
/* Copy into contiguous array: ARM will add padding to struct node_id! */
n = tal_count(ids);
arr = tal_arr(NULL, u8, n * sizeof(ids[0].k));
for (size_t i = 0; i < n; ++i) {
assert(node_id_valid(&ids[i]));
memcpy(arr + sizeof(ids[i].k) * i,
ids[i].k,
sizeof(ids[i].k));
}
int err = sqlite3_bind_blob(stmt, col, arr, tal_count(arr), SQLITE_TRANSIENT);
tal_free(arr);
return err == SQLITE_OK;
}
struct node_id *sqlite3_column_node_id_array(const tal_t *ctx,
sqlite3_stmt *stmt, int col)
{
size_t n;
struct node_id *ret;
const u8 *arr;
if (sqlite3_column_type(stmt, col) == SQLITE_NULL)
return NULL;
n = sqlite3_column_bytes(stmt, col) / sizeof(ret->k);
assert(n * sizeof(ret->k) == (size_t)sqlite3_column_bytes(stmt, col));
ret = tal_arr(ctx, struct node_id, n);
arr = sqlite3_column_blob(stmt, col);
for (size_t i = 0; i < n; i++) {
memcpy(ret[i].k, arr + i * sizeof(ret[i].k), sizeof(ret[i].k));
if (!node_id_valid(&ret[i]))
return tal_free(ret);
}
return ret;
}
bool sqlite3_column_preimage(sqlite3_stmt *stmt, int col, struct preimage *dest)
{
assert(sqlite3_column_bytes(stmt, col) == sizeof(struct preimage));
return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct preimage));
}
bool sqlite3_bind_preimage(sqlite3_stmt *stmt, int col, const struct preimage *p)
{
int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct preimage), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
bool sqlite3_column_sha256(sqlite3_stmt *stmt, int col, struct sha256 *dest)
{
assert(sqlite3_column_bytes(stmt, col) == sizeof(struct sha256));
return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct sha256));
}
bool sqlite3_bind_sha256(sqlite3_stmt *stmt, int col, const struct sha256 *p)
{
int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct sha256), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
bool sqlite3_column_sha256_double(sqlite3_stmt *stmt, int col, struct sha256_double *dest)
{
assert(sqlite3_column_bytes(stmt, col) == sizeof(struct sha256_double));
return memcpy(dest, sqlite3_column_blob(stmt, col), sizeof(struct sha256_double));
}
struct secret *sqlite3_column_secrets(const tal_t *ctx,
sqlite3_stmt *stmt, int col)
{
return sqlite3_column_arr(ctx, stmt, col, struct secret);
}
bool sqlite3_bind_sha256_double(sqlite3_stmt *stmt, int col, const struct sha256_double *p)
{
int err = sqlite3_bind_blob(stmt, col, p, sizeof(struct sha256_double), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
struct json_escaped *sqlite3_column_json_escaped(const tal_t *ctx,
sqlite3_stmt *stmt, int col)
{
return json_escaped_string_(ctx,
sqlite3_column_blob(stmt, col),
sqlite3_column_bytes(stmt, col));
}
bool sqlite3_bind_json_escaped(sqlite3_stmt *stmt, int col,
const struct json_escaped *esc)
{
int err = sqlite3_bind_text(stmt, col, esc->s, strlen(esc->s), SQLITE_TRANSIENT);
return err == SQLITE_OK;
}
struct amount_msat sqlite3_column_amount_msat(sqlite3_stmt *stmt, int col)
{
struct amount_msat msat;
msat.millisatoshis = sqlite3_column_int64(stmt, col); /* Raw: low level function */
return msat;
}
struct amount_sat sqlite3_column_amount_sat(sqlite3_stmt *stmt, int col)
{
struct amount_sat sat;
sat.satoshis = sqlite3_column_int64(stmt, col); /* Raw: low level function */
return sat;
}
void sqlite3_bind_amount_msat(sqlite3_stmt *stmt, int col,
struct amount_msat msat)
{
sqlite3_bind_int64(stmt, col, msat.millisatoshis); /* Raw: low level function */
}
void sqlite3_bind_amount_sat(sqlite3_stmt *stmt, int col,
struct amount_sat sat)
{
sqlite3_bind_int64(stmt, col, sat.satoshis); /* Raw: low level function */
}
/* Will apply the current config fee settings to all channels */
void migrate_pr2342_feerate_per_channel(struct lightningd *ld, struct db *db)
{
db_exec(__func__, db,
"UPDATE channels SET feerate_base = %u, feerate_ppm = %u;",
ld->config.fee_base,
ld->config.fee_per_satoshi);
}
void sqlite3_bind_timeabs(sqlite3_stmt *stmt, int col, struct timeabs t)
{
u64 timestamp = t.ts.tv_nsec + (t.ts.tv_sec * NSEC_IN_SEC);
sqlite3_bind_int64(stmt, col, timestamp);
}
struct timeabs sqlite3_column_timeabs(sqlite3_stmt *stmt, int col)
{
struct timeabs t;
u64 timestamp = sqlite3_column_int64(stmt, col);
t.ts.tv_sec = timestamp / NSEC_IN_SEC;
t.ts.tv_nsec = timestamp % NSEC_IN_SEC;
return t;
}