|
|
|
from collections import namedtuple
|
|
|
|
from decimal import Decimal
|
|
|
|
from fixtures import * # noqa: F401,F403
|
|
|
|
from fixtures import TEST_NETWORK
|
|
|
|
from flaky import flaky # noqa: F401
|
|
|
|
from pyln.client import RpcError, Millisatoshi
|
|
|
|
from utils import (
|
|
|
|
DEVELOPER, only_one, wait_for, sync_blockheight, VALGRIND, TIMEOUT,
|
|
|
|
SLOW_MACHINE, expected_peer_features, expected_node_features,
|
|
|
|
expected_channel_features,
|
|
|
|
check_coin_moves, first_channel_id, account_balance
|
|
|
|
)
|
|
|
|
from bitcoin.core import CMutableTransaction, CMutableTxOut
|
|
|
|
|
|
|
|
import binascii
|
|
|
|
import os
|
|
|
|
import pytest
|
|
|
|
import random
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import time
|
|
|
|
import unittest
|
|
|
|
|
|
|
|
|
|
|
|
def test_connect(node_factory):
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=False)
|
|
|
|
|
openingd: take peer before we're opening, wait for explicit funding msg.
Prior to this, lightningd would hand uninteresting peers back to connectd,
which would then return it to lightningd if it sent a non-gossip msg,
or if lightningd asked it to release the peer.
Now connectd hands the peer to lightningd once we've done the init
handshake, which hands it off to openingd.
This is a deep structural change, so we do the minimum here and cleanup
in the following patches.
Lightningd:
1. Remove peer_nongossip handling from connect_control and peer_control.
2. Remove list of outstanding fundchannel command; it was only needed to
find the race between us asking connectd to release the peer and it
reconnecting.
3. We can no longer tell if the remote end has started trying to fund a
channel (until it has succeeded): it's very transitory anyway so not
worth fixing.
4. We now always have a struct peer, and allocate an uncommitted_channel
for it, though it may never be used if neither end funds a channel.
5. We start funding on messages for openingd: we can get a funder_reply
or a fundee, or an error in response to our request to fund a channel.
so we handle all of them.
6. A new peer_start_openingd() is called after connectd hands us a peer.
7. json_fund_channel just looks through local peers; there are none
hidden in connectd any more.
8. We sometimes start a new openingd just to send an error message.
Openingd:
1. We always have information we need to accept them funding a channel (in
the init message).
2. We have to listen for three fds: peer, gossip and master, so we opencode
the poll.
3. We have an explicit message to start trying to fund a channel.
4. We can be told to send a message in our init message.
Testing:
1. We don't handle some things gracefully yet, so two tests are disabled.
2. 'hand_back_peer .*: now local again' from connectd is no longer a message,
openingd says 'Handed peer, entering loop' once its managing it.
3. peer['state'] used to be set to 'GOSSIPING' (otherwise this field doesn't
exist; 'state' is now per-channel. It doesn't exist at all now.
4. Some tests now need to turn on IO logging in openingd, not connectd.
5. There's a gap between connecting on one node and having connectd on
the peer hand over the connection to openingd. Our tests sometimes
checked getpeers() on the peer, and didn't see anything, so line_graph
needed updating.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
6 years ago
|
|
|
# These should be in openingd.
|
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['connected']
|
|
|
|
assert l2.rpc.getpeer(l1.info['id'])['connected']
|
|
|
|
assert len(l1.rpc.getpeer(l2.info['id'])['channels']) == 0
|
|
|
|
assert len(l2.rpc.getpeer(l1.info['id'])['channels']) == 0
|
|
|
|
|
|
|
|
# Reconnect should be a noop
|
|
|
|
ret = l1.rpc.connect(l2.info['id'], 'localhost', port=l2.port)
|
|
|
|
assert ret['id'] == l2.info['id']
|
|
|
|
|
|
|
|
ret = l2.rpc.connect(l1.info['id'], host='localhost', port=l1.port)
|
|
|
|
assert ret['id'] == l1.info['id']
|
|
|
|
|
|
|
|
# Should still only have one peer!
|
|
|
|
assert len(l1.rpc.listpeers()) == 1
|
|
|
|
assert len(l2.rpc.listpeers()) == 1
|
|
|
|
|
|
|
|
# Should get reasonable error if unknown addr for peer.
|
|
|
|
with pytest.raises(RpcError, match=r'Unable to connect, no address known'):
|
|
|
|
l1.rpc.connect('032cf15d1ad9c4a08d26eab1918f732d8ef8fdc6abb9640bf3db174372c491304e')
|
|
|
|
|
|
|
|
# Should get reasonable error if connection refuse.
|
|
|
|
with pytest.raises(RpcError, match=r'Connection establishment: Connection refused'):
|
|
|
|
l1.rpc.connect('032cf15d1ad9c4a08d26eab1918f732d8ef8fdc6abb9640bf3db174372c491304e', 'localhost', 1)
|
|
|
|
|
|
|
|
# Should get reasonable error if wrong key for peer.
|
|
|
|
with pytest.raises(RpcError, match=r'Cryptographic handshake: peer closed connection \(wrong key\?\)'):
|
|
|
|
l1.rpc.connect('032cf15d1ad9c4a08d26eab1918f732d8ef8fdc6abb9640bf3db174372c491304e', 'localhost', l2.port)
|
|
|
|
|
|
|
|
|
|
|
|
def test_connect_standard_addr(node_factory):
|
|
|
|
"""Test standard node@host:port address
|
|
|
|
"""
|
|
|
|
l1, l2, l3 = node_factory.get_nodes(3)
|
|
|
|
|
|
|
|
# node@host
|
|
|
|
ret = l1.rpc.connect("{}@{}".format(l2.info['id'], 'localhost'), port=l2.port)
|
|
|
|
assert ret['id'] == l2.info['id']
|
|
|
|
|
|
|
|
# node@host:port
|
|
|
|
ret = l1.rpc.connect("{}@localhost:{}".format(l3.info['id'], l3.port))
|
|
|
|
assert ret['id'] == l3.info['id']
|
|
|
|
|
|
|
|
# node@[ipv6]:port --- not supported by our CI
|
|
|
|
# ret = l1.rpc.connect("{}@[::1]:{}".format(l3.info['id'], l3.port))
|
|
|
|
# assert ret['id'] == l3.info['id']
|
|
|
|
|
|
|
|
|
|
|
|
def test_reconnect_channel_peers(node_factory, executor):
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
l2.restart()
|
|
|
|
|
|
|
|
# Should reconnect.
|
|
|
|
wait_for(lambda: only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['connected'])
|
|
|
|
wait_for(lambda: only_one(l2.rpc.listpeers(l1.info['id'])['peers'])['connected'])
|
|
|
|
# Connect command should succeed.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Stop l2 and wait for l1 to notice.
|
|
|
|
l2.stop()
|
|
|
|
wait_for(lambda: not only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['connected'])
|
|
|
|
|
|
|
|
# Now should fail.
|
|
|
|
with pytest.raises(RpcError, match=r'Connection refused'):
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Wait for exponential backoff to give us a 2 second window.
|
|
|
|
l1.daemon.wait_for_log('Will try reconnect in 2 seconds')
|
|
|
|
|
|
|
|
# It should now succeed when it restarts.
|
|
|
|
l2.start()
|
|
|
|
|
|
|
|
# Multiples should be fine!
|
|
|
|
fut1 = executor.submit(l1.rpc.connect, l2.info['id'], 'localhost', l2.port)
|
|
|
|
fut2 = executor.submit(l1.rpc.connect, l2.info['id'], 'localhost', l2.port)
|
|
|
|
fut3 = executor.submit(l1.rpc.connect, l2.info['id'], 'localhost', l2.port)
|
|
|
|
fut1.result(10)
|
|
|
|
fut2.result(10)
|
|
|
|
fut3.result(10)
|
|
|
|
|
|
|
|
|
|
|
|
def test_balance(node_factory):
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=True)
|
|
|
|
p1 = only_one(l1.rpc.getpeer(peer_id=l2.info['id'], level='info')['channels'])
|
|
|
|
p2 = only_one(l2.rpc.getpeer(l1.info['id'], 'info')['channels'])
|
|
|
|
assert p1['msatoshi_to_us'] == 10**6 * 1000
|
|
|
|
assert p1['msatoshi_total'] == 10**6 * 1000
|
|
|
|
assert p2['msatoshi_to_us'] == 0
|
|
|
|
assert p2['msatoshi_total'] == 10**6 * 1000
|
|
|
|
|
|
|
|
|
|
|
|
def test_bad_opening(node_factory):
|
|
|
|
# l1 asks for a too-long locktime
|
|
|
|
l1 = node_factory.get_node(options={'watchtime-blocks': 100})
|
|
|
|
l2 = node_factory.get_node(options={'max-locktime-blocks': 99})
|
|
|
|
ret = l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
assert ret['id'] == l2.info['id']
|
|
|
|
|
openingd: take peer before we're opening, wait for explicit funding msg.
Prior to this, lightningd would hand uninteresting peers back to connectd,
which would then return it to lightningd if it sent a non-gossip msg,
or if lightningd asked it to release the peer.
Now connectd hands the peer to lightningd once we've done the init
handshake, which hands it off to openingd.
This is a deep structural change, so we do the minimum here and cleanup
in the following patches.
Lightningd:
1. Remove peer_nongossip handling from connect_control and peer_control.
2. Remove list of outstanding fundchannel command; it was only needed to
find the race between us asking connectd to release the peer and it
reconnecting.
3. We can no longer tell if the remote end has started trying to fund a
channel (until it has succeeded): it's very transitory anyway so not
worth fixing.
4. We now always have a struct peer, and allocate an uncommitted_channel
for it, though it may never be used if neither end funds a channel.
5. We start funding on messages for openingd: we can get a funder_reply
or a fundee, or an error in response to our request to fund a channel.
so we handle all of them.
6. A new peer_start_openingd() is called after connectd hands us a peer.
7. json_fund_channel just looks through local peers; there are none
hidden in connectd any more.
8. We sometimes start a new openingd just to send an error message.
Openingd:
1. We always have information we need to accept them funding a channel (in
the init message).
2. We have to listen for three fds: peer, gossip and master, so we opencode
the poll.
3. We have an explicit message to start trying to fund a channel.
4. We can be told to send a message in our init message.
Testing:
1. We don't handle some things gracefully yet, so two tests are disabled.
2. 'hand_back_peer .*: now local again' from connectd is no longer a message,
openingd says 'Handed peer, entering loop' once its managing it.
3. peer['state'] used to be set to 'GOSSIPING' (otherwise this field doesn't
exist; 'state' is now per-channel. It doesn't exist at all now.
4. Some tests now need to turn on IO logging in openingd, not connectd.
5. There's a gap between connecting on one node and having connectd on
the peer hand over the connection to openingd. Our tests sometimes
checked getpeers() on the peer, and didn't see anything, so line_graph
needed updating.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
6 years ago
|
|
|
l1.daemon.wait_for_log('openingd-.*: Handed peer, entering loop')
|
|
|
|
l2.daemon.wait_for_log('openingd-.*: Handed peer, entering loop')
|
|
|
|
|
|
|
|
l1.fundwallet(10**6 + 1000000)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6)
|
|
|
|
|
|
|
|
l2.daemon.wait_for_log('to_self_delay 100 larger than 99')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "gossip without DEVELOPER=1 is slow")
|
|
|
|
@unittest.skipIf(TEST_NETWORK != 'regtest', "Fee computation and limits are network specific")
|
|
|
|
def test_opening_tiny_channel(node_factory):
|
|
|
|
# Test custom min-capacity-sat parameters
|
|
|
|
#
|
|
|
|
# o---> [l2] (1000) - old default (too little for reserves)
|
|
|
|
# /
|
|
|
|
# [l1]-----> [l3] (~6000) - technical minimal value that wont be rejected
|
|
|
|
# \
|
|
|
|
# o---> [l4] (~10000) - the current default
|
|
|
|
# \
|
|
|
|
# o-> [l5] (20000) - a node with a higher minimal value
|
|
|
|
#
|
|
|
|
# For each:
|
|
|
|
# 1. Try to establish channel 1sat smaller than min_capacity_sat
|
|
|
|
# 2. Try to establish channel exact min_capacity_sat
|
|
|
|
#
|
|
|
|
# BOLT2
|
|
|
|
# The receiving node MAY fail the channel if:
|
|
|
|
# - funding_satoshis is too small
|
|
|
|
# - it considers `feerate_per_kw` too small for timely processing or unreasonably large.
|
|
|
|
#
|
|
|
|
dustlimit = 546
|
|
|
|
reserves = 2 * dustlimit
|
|
|
|
min_commit_tx_fees = 5430
|
|
|
|
min_for_opener = min_commit_tx_fees + dustlimit + 1
|
|
|
|
|
|
|
|
l1_min_capacity = 1000 # 1k old default, too small but used at l1 to allow small incoming channels
|
|
|
|
l2_min_capacity = reserves # just enough to get past capacity filter
|
|
|
|
l3_min_capacity = min_for_opener # the absolute technical minimum
|
|
|
|
l4_min_capacity = 10000 # the current default
|
|
|
|
l5_min_capacity = 20000 # a server with more than default minimum
|
|
|
|
|
|
|
|
l1 = node_factory.get_node(options={'min-capacity-sat': l1_min_capacity})
|
|
|
|
l2 = node_factory.get_node(options={'min-capacity-sat': l2_min_capacity})
|
|
|
|
l3 = node_factory.get_node(options={'min-capacity-sat': l3_min_capacity})
|
|
|
|
l4 = node_factory.get_node(options={'min-capacity-sat': l4_min_capacity})
|
|
|
|
l5 = node_factory.get_node(options={'min-capacity-sat': l5_min_capacity})
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.connect(l3.info['id'], 'localhost', l3.port)
|
|
|
|
l1.rpc.connect(l4.info['id'], 'localhost', l4.port)
|
|
|
|
l1.rpc.connect(l5.info['id'], 'localhost', l5.port)
|
|
|
|
|
|
|
|
# Open channel with one less than reserves should be rejected at l2
|
|
|
|
with pytest.raises(RpcError, match=r'channel_reserve_satoshis .*sat and .*sat too large for funding .*sat'):
|
|
|
|
l1.fund_channel(l2, l2_min_capacity - 1)
|
|
|
|
# Open a channel with exactly the minimal amount for the fundee,
|
|
|
|
# This will raise an exception at l1, as the opener cannot afford fees for initial_commit_tx.
|
|
|
|
# Note: The old default of 1k sat is below the technical minimum when accounting for dust reserves and fees
|
|
|
|
# This is why this must fail, for this reason the default will be raised to 10k sat.
|
|
|
|
with pytest.raises(RpcError, match=r'Funder cannot afford fee on initial commitment transaction'):
|
|
|
|
l1.fund_channel(l2, l2_min_capacity)
|
|
|
|
|
|
|
|
# Open channel with one less than technical minimum should be rejected at l3
|
|
|
|
with pytest.raises(RpcError, match=r'channel capacity is .*sat, which is below .*sat'):
|
|
|
|
l1.fund_channel(l3, l3_min_capacity - 1)
|
|
|
|
# When amount technical minimum matches exactly, own initial_commit_tx fees can now be covered
|
|
|
|
l1.fund_channel(l3, l3_min_capacity)
|
|
|
|
|
|
|
|
# Open channel with one less than default 10k sats should be rejected at l4
|
|
|
|
with pytest.raises(RpcError, match=r'channel capacity is .*, which is below .*msat'):
|
|
|
|
l1.fund_channel(l4, l4_min_capacity - 1)
|
|
|
|
# This must be possible with enough capacity
|
|
|
|
l1.fund_channel(l4, l4_min_capacity)
|
|
|
|
|
|
|
|
# Open channel with less than minimum should be rejected at l5
|
|
|
|
with pytest.raises(RpcError, match=r'channel capacity is .*, which is below .*msat'):
|
|
|
|
l1.fund_channel(l5, l5_min_capacity - 1)
|
|
|
|
# bigger channels must not be affected
|
|
|
|
l1.fund_channel(l5, l5_min_capacity * 10)
|
|
|
|
|
|
|
|
|
|
|
|
def test_second_channel(node_factory):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
l3 = node_factory.get_node()
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.connect(l3.info['id'], 'localhost', l3.port)
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
l1.fund_channel(l3, 10**6)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_disconnect(node_factory):
|
|
|
|
# These should all make us fail
|
|
|
|
disconnects = ['-WIRE_INIT',
|
|
|
|
'@WIRE_INIT',
|
|
|
|
'+WIRE_INIT']
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects)
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Should have 3 connect fails.
|
|
|
|
for d in disconnects:
|
|
|
|
l1.daemon.wait_for_log('{}-.*Failed connected out'
|
|
|
|
.format(l2.info['id']))
|
|
|
|
|
|
|
|
# Should still only have one peer!
|
|
|
|
assert len(l1.rpc.listpeers()) == 1
|
|
|
|
assert len(l2.rpc.listpeers()) == 1
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_disconnect_opener(node_factory):
|
|
|
|
# Now error on opener side during channel open.
|
|
|
|
disconnects = ['-WIRE_OPEN_CHANNEL',
|
|
|
|
'@WIRE_OPEN_CHANNEL',
|
|
|
|
'+WIRE_OPEN_CHANNEL',
|
|
|
|
'-WIRE_FUNDING_CREATED',
|
|
|
|
'@WIRE_FUNDING_CREATED']
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects)
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
|
|
|
|
l1.fundwallet(2000000)
|
|
|
|
|
|
|
|
for d in disconnects:
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
assert l1.rpc.getpeer(l2.info['id']) is None
|
|
|
|
|
|
|
|
# This one will succeed.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
|
|
|
|
# Should still only have one peer!
|
|
|
|
assert len(l1.rpc.listpeers()) == 1
|
|
|
|
assert len(l2.rpc.listpeers()) == 1
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_disconnect_fundee(node_factory):
|
|
|
|
# Now error on fundee side during channel open.
|
|
|
|
disconnects = ['-WIRE_ACCEPT_CHANNEL',
|
|
|
|
'@WIRE_ACCEPT_CHANNEL',
|
|
|
|
'+WIRE_ACCEPT_CHANNEL']
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects)
|
|
|
|
|
|
|
|
l1.fundwallet(2000000)
|
|
|
|
|
|
|
|
for d in disconnects:
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
assert l1.rpc.getpeer(l2.info['id']) is None
|
|
|
|
|
|
|
|
# This one will succeed.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
|
|
|
|
# Should still only have one peer!
|
|
|
|
assert len(l1.rpc.listpeers()) == 1
|
|
|
|
assert len(l2.rpc.listpeers()) == 1
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_disconnect_half_signed(node_factory):
|
|
|
|
# Now, these are the corner cases. Fundee sends funding_signed,
|
|
|
|
# but opener doesn't receive it.
|
|
|
|
disconnects = ['@WIRE_FUNDING_SIGNED']
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects)
|
|
|
|
|
|
|
|
l1.fundwallet(2000000)
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
|
|
|
|
# Peer remembers, opener doesn't.
|
|
|
|
assert l1.rpc.getpeer(l2.info['id']) is None
|
|
|
|
assert l2.rpc.getpeer(l1.info['id'])['id'] == l1.info['id']
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_signed(node_factory):
|
|
|
|
# This will fail *after* both sides consider channel opening.
|
|
|
|
disconnects = ['+WIRE_FUNDING_SIGNED']
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
|
|
|
|
l1.fundwallet(2000000)
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
|
|
|
|
# They haven't forgotten each other.
|
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['id'] == l2.info['id']
|
|
|
|
assert l2.rpc.getpeer(l1.info['id'])['id'] == l1.info['id']
|
|
|
|
|
|
|
|
# Technically, this is async to fundchannel (and could reconnect first)
|
|
|
|
l1.daemon.wait_for_logs(['sendrawtx exit 0',
|
|
|
|
'Peer has reconnected, state CHANNELD_AWAITING_LOCKIN'])
|
|
|
|
|
|
|
|
l1.bitcoin.generate_block(6)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log(' to CHANNELD_NORMAL')
|
|
|
|
l2.daemon.wait_for_log(' to CHANNELD_NORMAL')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_openingd(node_factory):
|
|
|
|
# Openingd thinks we're still opening; opener reconnects..
|
|
|
|
disconnects = ['0WIRE_ACCEPT_CHANNEL']
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fundwallet(2000000)
|
|
|
|
|
|
|
|
# l2 closes on l1, l1 forgets.
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
assert l1.rpc.getpeer(l2.info['id']) is None
|
|
|
|
|
|
|
|
# Reconnect.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# We should get a message about reconnecting.
|
|
|
|
l2.daemon.wait_for_log('Killing openingd: Reconnected')
|
|
|
|
l2.daemon.wait_for_log('openingd.*Handed peer, entering loop')
|
|
|
|
|
|
|
|
# Should work fine.
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 20000)
|
|
|
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
|
|
|
|
|
|
|
l1.bitcoin.generate_block(3)
|
|
|
|
|
|
|
|
# Just to be sure, second openingd hand over to channeld. This log line is about channeld being started
|
|
|
|
l2.daemon.wait_for_log(r'channeld-chan#[0-9]: pid [0-9]+, msgfd [0-9]+')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_gossiping(node_factory):
|
|
|
|
# connectd thinks we're still gossiping; peer reconnects.
|
|
|
|
disconnects = ['0WIRE_PING']
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
# Make sure l2 knows about l1
|
|
|
|
wait_for(lambda: l2.rpc.listpeers(l1.info['id'])['peers'] != [])
|
|
|
|
|
|
|
|
l2.rpc.ping(l1.info['id'], 1, 65532)
|
|
|
|
wait_for(lambda: l1.rpc.listpeers(l2.info['id'])['peers'] == [])
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l2.daemon.wait_for_log('processing now old peer gone')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs dev-disconnect")
|
|
|
|
def test_reconnect_no_update(node_factory, executor):
|
|
|
|
"""
|
|
|
|
This tests if the `funding_locked` is sent if we receive a
|
|
|
|
`channel_reestablish` message with `next_commitment_number` == 1 and
|
|
|
|
our `next_commitment_number` == 1.
|
|
|
|
"""
|
|
|
|
disconnects = ["@WIRE_FUNDING_LOCKED", "@WIRE_SHUTDOWN"]
|
|
|
|
# Allow bad gossip because it might receive WIRE_CHANNEL_UPDATE before
|
|
|
|
# announcement before of the disconnection
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True, allow_bad_gossip=True)
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects, may_reconnect=True)
|
|
|
|
|
|
|
|
# For channeld reconnection
|
|
|
|
l1.rpc.connect(l2.info["id"], "localhost", l2.port)
|
|
|
|
fundchannel_exec = executor.submit(l1.fund_channel, l2, 10**6, False)
|
|
|
|
l1.daemon.wait_for_log(r"channeld.* Retransmitting funding_locked for channel")
|
|
|
|
l1.stop()
|
|
|
|
|
|
|
|
# For closingd reconnection
|
|
|
|
scid = fundchannel_exec.result()
|
|
|
|
l1.daemon.start()
|
|
|
|
executor.submit(l1.rpc.close, scid, 0)
|
|
|
|
l2.daemon.wait_for_log(r"closingd.* Retransmitting funding_locked for channel")
|
|
|
|
l1.daemon.wait_for_log(r"CLOSINGD_COMPLETE")
|
|
|
|
|
|
|
|
|
|
|
|
def test_connect_stresstest(node_factory, executor):
|
|
|
|
# This test is unreliable, but it's better than nothing.
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l3 = node_factory.get_node(may_reconnect=True)
|
|
|
|
|
|
|
|
# Hack l3 into a clone of l2, to stress reconnect code.
|
|
|
|
l3.stop()
|
|
|
|
shutil.copyfile(os.path.join(l2.daemon.lightning_dir, TEST_NETWORK, 'hsm_secret'),
|
|
|
|
os.path.join(l3.daemon.lightning_dir, TEST_NETWORK, 'hsm_secret'))
|
|
|
|
l3.start()
|
|
|
|
l3.info = l3.rpc.getinfo()
|
|
|
|
|
|
|
|
assert l3.info['id'] == l2.info['id']
|
|
|
|
|
|
|
|
# We fire off random connect/disconnect commands.
|
|
|
|
actions = [
|
|
|
|
(l2.rpc.connect, l1.info['id'], 'localhost', l1.port),
|
|
|
|
(l3.rpc.connect, l1.info['id'], 'localhost', l3.port),
|
|
|
|
(l1.rpc.connect, l2.info['id'], 'localhost', l2.port),
|
|
|
|
(l1.rpc.connect, l3.info['id'], 'localhost', l3.port),
|
|
|
|
(l1.rpc.disconnect, l2.info['id'])
|
|
|
|
]
|
|
|
|
args = [random.choice(actions) for _ in range(1000)]
|
|
|
|
|
|
|
|
# We get them all to connect to each other.
|
|
|
|
futs = []
|
|
|
|
for a in args:
|
|
|
|
futs.append(executor.submit(*a))
|
|
|
|
|
|
|
|
# We don't actually care if they fail, since some will.
|
|
|
|
successes = 0
|
|
|
|
failures = 0
|
|
|
|
for f in futs:
|
|
|
|
if f.exception():
|
|
|
|
failures += 1
|
|
|
|
else:
|
|
|
|
f.result()
|
|
|
|
successes += 1
|
|
|
|
|
|
|
|
assert successes > failures
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_normal(node_factory):
|
|
|
|
# Should reconnect fine even if locked message gets lost.
|
|
|
|
disconnects = ['-WIRE_FUNDING_LOCKED',
|
|
|
|
'@WIRE_FUNDING_LOCKED',
|
|
|
|
'+WIRE_FUNDING_LOCKED']
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_sender_add1(node_factory):
|
|
|
|
# Fail after add is OK, will cause payment failure though.
|
|
|
|
disconnects = ['-WIRE_UPDATE_ADD_HTLC-nocommit',
|
|
|
|
'+WIRE_UPDATE_ADD_HTLC-nocommit',
|
|
|
|
'@WIRE_UPDATE_ADD_HTLC-nocommit']
|
|
|
|
|
|
|
|
# Feerates identical so we don't get gratuitous commit to update them
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True,
|
|
|
|
feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
amt = 200000000
|
|
|
|
rhash = l2.rpc.invoice(amt, 'test_reconnect_sender_add1', 'desc')['payment_hash']
|
|
|
|
assert only_one(l2.rpc.listinvoices('test_reconnect_sender_add1')['invoices'])['status'] == 'unpaid'
|
|
|
|
|
|
|
|
route = [{'msatoshi': amt, 'id': l2.info['id'], 'delay': 5, 'channel': '1x1x1'}]
|
|
|
|
|
|
|
|
for i in range(0, len(disconnects)):
|
|
|
|
l1.rpc.sendpay(route, rhash)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.waitsendpay(rhash)
|
|
|
|
|
|
|
|
# Wait for reconnection.
|
|
|
|
l1.daemon.wait_for_log('Already have funding locked in')
|
|
|
|
|
|
|
|
# This will send commit, so will reconnect as required.
|
|
|
|
l1.rpc.sendpay(route, rhash)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_sender_add(node_factory):
|
|
|
|
disconnects = ['-WIRE_COMMITMENT_SIGNED',
|
|
|
|
'@WIRE_COMMITMENT_SIGNED',
|
|
|
|
'+WIRE_COMMITMENT_SIGNED',
|
|
|
|
'-WIRE_REVOKE_AND_ACK',
|
|
|
|
'@WIRE_REVOKE_AND_ACK',
|
|
|
|
'+WIRE_REVOKE_AND_ACK']
|
|
|
|
# Feerates identical so we don't get gratuitous commit to update them
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True,
|
|
|
|
feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
amt = 200000000
|
|
|
|
rhash = l2.rpc.invoice(amt, 'testpayment', 'desc')['payment_hash']
|
|
|
|
assert only_one(l2.rpc.listinvoices('testpayment')['invoices'])['status'] == 'unpaid'
|
|
|
|
|
|
|
|
route = [{'msatoshi': amt, 'id': l2.info['id'], 'delay': 5, 'channel': '1x1x1'}]
|
|
|
|
|
|
|
|
# This will send commit, so will reconnect as required.
|
|
|
|
l1.rpc.sendpay(route, rhash)
|
|
|
|
# Should have printed this for every reconnect.
|
|
|
|
for i in range(0, len(disconnects)):
|
|
|
|
l1.daemon.wait_for_log('Already have funding locked in')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_receiver_add(node_factory):
|
|
|
|
disconnects = ['-WIRE_COMMITMENT_SIGNED',
|
|
|
|
'@WIRE_COMMITMENT_SIGNED',
|
|
|
|
'+WIRE_COMMITMENT_SIGNED',
|
|
|
|
'-WIRE_REVOKE_AND_ACK',
|
|
|
|
'@WIRE_REVOKE_AND_ACK',
|
|
|
|
'+WIRE_REVOKE_AND_ACK']
|
|
|
|
# Feerates identical so we don't get gratuitous commit to update them
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True, feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
amt = 200000000
|
|
|
|
rhash = l2.rpc.invoice(amt, 'testpayment2', 'desc')['payment_hash']
|
|
|
|
assert only_one(l2.rpc.listinvoices('testpayment2')['invoices'])['status'] == 'unpaid'
|
|
|
|
|
|
|
|
route = [{'msatoshi': amt, 'id': l2.info['id'], 'delay': 5, 'channel': '1x1x1'}]
|
|
|
|
l1.rpc.sendpay(route, rhash)
|
|
|
|
for i in range(len(disconnects)):
|
|
|
|
l1.daemon.wait_for_log('Already have funding locked in')
|
|
|
|
assert only_one(l2.rpc.listinvoices('testpayment2')['invoices'])['status'] == 'paid'
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_receiver_fulfill(node_factory):
|
|
|
|
# Ordering matters: after +WIRE_UPDATE_FULFILL_HTLC, channeld
|
|
|
|
# will continue and try to send WIRE_COMMITMENT_SIGNED: if
|
|
|
|
# that's the next failure, it will do two in one run.
|
|
|
|
disconnects = ['@WIRE_UPDATE_FULFILL_HTLC',
|
|
|
|
'+WIRE_UPDATE_FULFILL_HTLC',
|
|
|
|
'-WIRE_UPDATE_FULFILL_HTLC',
|
|
|
|
'-WIRE_COMMITMENT_SIGNED',
|
|
|
|
'@WIRE_COMMITMENT_SIGNED',
|
|
|
|
'+WIRE_COMMITMENT_SIGNED',
|
|
|
|
'-WIRE_REVOKE_AND_ACK',
|
|
|
|
'@WIRE_REVOKE_AND_ACK',
|
|
|
|
'+WIRE_REVOKE_AND_ACK']
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
amt = 200000000
|
|
|
|
rhash = l2.rpc.invoice(amt, 'testpayment2', 'desc')['payment_hash']
|
|
|
|
assert only_one(l2.rpc.listinvoices('testpayment2')['invoices'])['status'] == 'unpaid'
|
|
|
|
|
|
|
|
route = [{'msatoshi': amt, 'id': l2.info['id'], 'delay': 5, 'channel': '1x1x1'}]
|
|
|
|
l1.rpc.sendpay(route, rhash)
|
|
|
|
for i in range(len(disconnects)):
|
|
|
|
l1.daemon.wait_for_log('Already have funding locked in')
|
|
|
|
assert only_one(l2.rpc.listinvoices('testpayment2')['invoices'])['status'] == 'paid'
|
|
|
|
|
|
|
|
|
|
|
|
@flaky
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_shutdown_reconnect(node_factory):
|
|
|
|
disconnects = ['-WIRE_SHUTDOWN',
|
|
|
|
'@WIRE_SHUTDOWN',
|
|
|
|
'+WIRE_SHUTDOWN']
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects,
|
|
|
|
may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
chan = l1.fund_channel(l2, 10**6)
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
|
|
|
|
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
|
|
|
|
|
|
|
# This should wait until we're closed.
|
|
|
|
l1.rpc.close(chan)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
|
|
|
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log(' to CLOSINGD_SIGEXCHANGE')
|
|
|
|
l2.daemon.wait_for_log(' to CLOSINGD_SIGEXCHANGE')
|
|
|
|
|
|
|
|
# And should put closing into mempool (happens async, so
|
|
|
|
# CLOSINGD_COMPLETE may come first).
|
|
|
|
l1.daemon.wait_for_logs(['sendrawtx exit 0', ' to CLOSINGD_COMPLETE'])
|
|
|
|
l2.daemon.wait_for_logs(['sendrawtx exit 0', ' to CLOSINGD_COMPLETE'])
|
|
|
|
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 1
|
|
|
|
|
|
|
|
|
|
|
|
@flaky
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_reconnect_remote_sends_no_sigs(node_factory):
|
|
|
|
"""We re-announce, even when remote node doesn't send its announcement_signatures on reconnect.
|
|
|
|
"""
|
|
|
|
l1, l2 = node_factory.line_graph(2, wait_for_announce=True, opts={'may_reconnect': True})
|
|
|
|
|
|
|
|
# When l1 restarts (with rescan=1), make it think it hasn't
|
|
|
|
# reached announce_depth, so it wont re-send announcement_signatures
|
|
|
|
def no_blocks_above(req):
|
|
|
|
if req['params'][0] > 107:
|
|
|
|
return {"result": None,
|
|
|
|
"error": {"code": -8, "message": "Block height out of range"}, "id": req['id']}
|
|
|
|
else:
|
|
|
|
return {'result': l1.bitcoin.rpc.getblockhash(req['params'][0]),
|
|
|
|
"error": None, 'id': req['id']}
|
|
|
|
|
|
|
|
l1.daemon.rpcproxy.mock_rpc('getblockhash', no_blocks_above)
|
|
|
|
l1.restart()
|
|
|
|
|
|
|
|
# l2 will now uses (REMOTE's) announcement_signatures it has stored
|
|
|
|
wait_for(lambda: only_one(l2.rpc.listpeers()['peers'][0]['channels'])['status'] == [
|
|
|
|
'CHANNELD_NORMAL:Reconnected, and reestablished.',
|
|
|
|
'CHANNELD_NORMAL:Funding transaction locked. Channel announced.'])
|
|
|
|
|
|
|
|
# But l2 still sends its own sigs on reconnect
|
|
|
|
l2.daemon.wait_for_logs([r'peer_out WIRE_ANNOUNCEMENT_SIGNATURES',
|
|
|
|
r'peer_out WIRE_ANNOUNCEMENT_SIGNATURES'])
|
|
|
|
|
|
|
|
# l1 only did send them the first time
|
|
|
|
assert(''.join(l1.daemon.logs).count(r'peer_out WIRE_ANNOUNCEMENT_SIGNATURES') == 1)
|
|
|
|
|
|
|
|
|
|
|
|
def test_shutdown_awaiting_lockin(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node(options={'funding-confirms': 3})
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fundwallet(10**6 + 1000000)
|
|
|
|
chanid = l1.rpc.fundchannel(l2.info['id'], 10**6)['channel_id']
|
|
|
|
|
|
|
|
# Technically, this is async to fundchannel.
|
|
|
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
l1.rpc.close(chanid)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
|
|
|
l2.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log('CHANNELD_SHUTTING_DOWN to CLOSINGD_SIGEXCHANGE')
|
|
|
|
l2.daemon.wait_for_log('CHANNELD_SHUTTING_DOWN to CLOSINGD_SIGEXCHANGE')
|
|
|
|
|
|
|
|
# And should put closing into mempool (happens async, so
|
|
|
|
# CLOSINGD_COMPLETE may come first).
|
|
|
|
l1.daemon.wait_for_logs(['sendrawtx exit 0', ' to CLOSINGD_COMPLETE'])
|
|
|
|
l2.daemon.wait_for_logs(['sendrawtx exit 0', ' to CLOSINGD_COMPLETE'])
|
|
|
|
assert bitcoind.rpc.getmempoolinfo()['size'] == 1
|
|
|
|
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
l1.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
l2.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
wait_for(lambda: l1.rpc.listpeers()['peers'] == [])
|
|
|
|
wait_for(lambda: l2.rpc.listpeers()['peers'] == [])
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_change(node_factory, bitcoind):
|
|
|
|
"""Add some funds, fund a channel, and make sure we remember the change
|
|
|
|
"""
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=False)
|
|
|
|
l1.fundwallet(10000000)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
sync_blockheight(bitcoind, [l1])
|
|
|
|
|
|
|
|
outputs = l1.db_query('SELECT value FROM outputs WHERE status=0;')
|
|
|
|
assert only_one(outputs)['value'] == 10000000
|
|
|
|
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 1000000)
|
|
|
|
outputs = {r['status']: r['value'] for r in l1.db_query(
|
|
|
|
'SELECT status, SUM(value) AS value FROM outputs GROUP BY status;')}
|
|
|
|
|
|
|
|
# The 10m out is spent and we have a change output of 9m-fee
|
|
|
|
assert outputs[0] > 8990000
|
|
|
|
assert outputs[2] == 10000000
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_all(node_factory, bitcoind):
|
|
|
|
"""Add some funds, fund a channel using all funds, make sure no funds remain
|
|
|
|
"""
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=False)
|
|
|
|
|
|
|
|
l1.fundwallet(0.1 * 10**8)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
sync_blockheight(bitcoind, [l1])
|
|
|
|
|
|
|
|
outputs = l1.db_query('SELECT value FROM outputs WHERE status=0;')
|
|
|
|
assert only_one(outputs)['value'] == 10000000
|
|
|
|
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], "all")
|
|
|
|
|
|
|
|
outputs = l1.db_query('SELECT value FROM outputs WHERE status=0;')
|
|
|
|
assert len(outputs) == 0
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_all_too_much(node_factory):
|
|
|
|
"""Add more than max possible funds, fund a channel using all funds we can.
|
|
|
|
"""
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=False)
|
|
|
|
|
|
|
|
l1.fundwallet(2**24 + 10000)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], "all")
|
|
|
|
|
|
|
|
assert only_one(l1.rpc.listfunds()['outputs'])['status'] == 'unconfirmed'
|
|
|
|
assert only_one(l1.rpc.listfunds()['channels'])['channel_total_sat'] == 2**24 - 1
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_fail(node_factory, bitcoind):
|
|
|
|
"""Add some funds, fund a channel without enough funds"""
|
|
|
|
# Previous runs with same bitcoind can leave funds!
|
|
|
|
max_locktime = 5 * 6 * 24
|
|
|
|
l1 = node_factory.get_node(random_hsm=True, options={'max-locktime-blocks': max_locktime})
|
|
|
|
l2 = node_factory.get_node(options={'watchtime-blocks': max_locktime + 1})
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
funds = 1000000
|
|
|
|
|
|
|
|
addr = l1.rpc.newaddr()['bech32']
|
|
|
|
l1.bitcoin.rpc.sendtoaddress(addr, funds / 10**8)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
# Wait for it to arrive.
|
|
|
|
wait_for(lambda: len(l1.rpc.listfunds()['outputs']) > 0)
|
|
|
|
|
|
|
|
# Fail because l1 dislikes l2's huge locktime.
|
|
|
|
with pytest.raises(RpcError, match=r'to_self_delay \d+ larger than \d+'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], int(funds / 10))
|
|
|
|
|
|
|
|
assert only_one(l1.rpc.listpeers()['peers'])['connected']
|
|
|
|
assert only_one(l2.rpc.listpeers()['peers'])['connected']
|
|
|
|
|
|
|
|
# Restart l2 without ridiculous locktime.
|
|
|
|
del l2.daemon.opts['watchtime-blocks']
|
|
|
|
l2.restart()
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# We don't have enough left to cover fees if we try to spend it all.
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot afford transaction'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], funds)
|
|
|
|
|
|
|
|
# Should still be connected.
|
|
|
|
assert only_one(l1.rpc.listpeers()['peers'])['connected']
|
|
|
|
l2.daemon.wait_for_log('openingd-.*: Handed peer, entering loop')
|
|
|
|
assert only_one(l2.rpc.listpeers()['peers'])['connected']
|
|
|
|
|
|
|
|
# This works.
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], int(funds / 10))
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_toolarge(node_factory, bitcoind):
|
|
|
|
"""Try to create a giant channel"""
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Send funds.
|
|
|
|
amount = 2**24
|
|
|
|
bitcoind.rpc.sendtoaddress(l1.rpc.newaddr()['bech32'], amount / 10**8 + 0.01)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
# Wait for it to arrive.
|
|
|
|
wait_for(lambda: len(l1.rpc.listfunds()['outputs']) > 0)
|
|
|
|
|
|
|
|
# Fail to open (too large)
|
|
|
|
with pytest.raises(RpcError, match=r'Amount exceeded 16777215'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], amount)
|
|
|
|
|
|
|
|
# This should work.
|
|
|
|
amount = amount - 1
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], amount)
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_push(node_factory, bitcoind, chainparams):
|
|
|
|
""" Try to push peer some sats """
|
|
|
|
# We track balances, to verify that accounting is ok.
|
|
|
|
coin_mvt_plugin = os.path.join(os.getcwd(), 'tests/plugins/coin_movements.py')
|
|
|
|
|
|
|
|
l1 = node_factory.get_node(options={'plugin': coin_mvt_plugin})
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Send funds.
|
|
|
|
amount = 2**24
|
|
|
|
push_sat = 20000
|
|
|
|
bitcoind.rpc.sendtoaddress(l1.rpc.newaddr()['bech32'], amount / 10**8 + 0.01)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
# Wait for it to arrive.
|
|
|
|
wait_for(lambda: len(l1.rpc.listfunds()['outputs']) > 0)
|
|
|
|
|
|
|
|
# Fail to open (try to push too much)
|
|
|
|
with pytest.raises(RpcError, match=r'Requested to push_msat of 20000000msat is greater than available funding amount 10000sat'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10000, push_msat=push_sat * 1000)
|
|
|
|
|
|
|
|
# This should work.
|
|
|
|
amount = amount - 1
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], amount, push_msat=push_sat * 1000)
|
|
|
|
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
sync_blockheight(bitcoind, [l1])
|
|
|
|
funds = only_one(l1.rpc.listfunds()['channels'])
|
|
|
|
assert funds['channel_sat'] + push_sat == funds['channel_total_sat']
|
|
|
|
|
|
|
|
chanid = first_channel_id(l2, l1)
|
|
|
|
l1.daemon.wait_for_log('coins account: {}'.format(chanid))
|
|
|
|
# give the file write a second
|
|
|
|
time.sleep(1)
|
|
|
|
channel_mvts = [
|
|
|
|
{'type': 'chain_mvt', 'credit': 0, 'debit': 20000000, 'tag': 'pushed'},
|
|
|
|
{'type': 'chain_mvt', 'credit': 16777215000, 'debit': 0, 'tag': 'deposit'},
|
|
|
|
]
|
|
|
|
check_coin_moves(l1, chanid, channel_mvts, chainparams)
|
|
|
|
assert account_balance(l1, chanid) == (amount - push_sat) * 1000
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_by_utxos(node_factory, bitcoind):
|
|
|
|
"""Fund a channel with specific utxos"""
|
|
|
|
l1, l2, l3 = node_factory.line_graph(3, fundchannel=False)
|
|
|
|
|
|
|
|
# Get 3 differents utxo
|
|
|
|
l1.fundwallet(0.01 * 10**8)
|
|
|
|
l1.fundwallet(0.01 * 10**8)
|
|
|
|
l1.fundwallet(0.01 * 10**8)
|
|
|
|
wait_for(lambda: len(l1.rpc.listfunds()["outputs"]) == 3)
|
|
|
|
|
|
|
|
utxos = [utxo["txid"] + ":" + str(utxo["output"]) for utxo in l1.rpc.listfunds()["outputs"]]
|
|
|
|
|
|
|
|
# Fund with utxos we don't own
|
|
|
|
with pytest.raises(RpcError, match=r"No matching utxo was found from the wallet"):
|
|
|
|
l3.rpc.fundchannel(l2.info["id"], int(0.01 * 10**8), utxos=utxos)
|
|
|
|
|
|
|
|
# Fund with an empty array
|
|
|
|
with pytest.raises(RpcError, match=r"Please specify an array of \\'txid:output_index\\', not \"*\""):
|
|
|
|
l1.rpc.fundchannel(l2.info["id"], int(0.01 * 10**8), utxos=[])
|
|
|
|
|
|
|
|
# Fund a channel from some of the utxos, without change
|
|
|
|
l1.rpc.fundchannel(l2.info["id"], "all", utxos=utxos[0:2])
|
|
|
|
|
|
|
|
# Fund a channel from the rest of utxos, with change
|
|
|
|
l1.rpc.connect(l3.info["id"], "localhost", l3.port)
|
|
|
|
l1.rpc.fundchannel(l3.info["id"], int(0.007 * 10**8), utxos=[utxos[2]])
|
|
|
|
|
|
|
|
# Fund another channel with already spent utxos
|
|
|
|
with pytest.raises(RpcError, match=r"No matching utxo was found from the wallet"):
|
|
|
|
l1.rpc.fundchannel(l3.info["id"], int(0.01 * 10**8), utxos=utxos)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs dev_forget_channel")
|
|
|
|
def test_funding_external_wallet_corners(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
|
|
|
|
amount = 2**24
|
|
|
|
l1.fundwallet(amount + 10000000)
|
|
|
|
|
|
|
|
amount = amount - 1
|
|
|
|
fake_txid = '929764844a8f9938b669a60a1d51a11c9e2613c7eb4776e4126f1f20c0a685c3'
|
|
|
|
fake_txout = 0
|
|
|
|
|
|
|
|
with pytest.raises(RpcError, match=r'Unknown peer'):
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount)
|
|
|
|
|
|
|
|
with pytest.raises(RpcError, match=r'Unknown peer'):
|
|
|
|
l1.rpc.fundchannel_complete(l2.info['id'], fake_txid, fake_txout)
|
|
|
|
|
|
|
|
# Should not be able to continue without being in progress.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
with pytest.raises(RpcError, match=r'No channel funding in progress.'):
|
|
|
|
l1.rpc.fundchannel_complete(l2.info['id'], fake_txid, fake_txout)
|
|
|
|
|
|
|
|
# Fail to open (too large)
|
|
|
|
with pytest.raises(RpcError, match=r'Amount exceeded 16777215'):
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount + 1)
|
|
|
|
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount)
|
|
|
|
with pytest.raises(RpcError, match=r'Already funding channel'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], amount)
|
|
|
|
|
|
|
|
l1.rpc.fundchannel_cancel(l2.info['id'])
|
|
|
|
# Should be able to 'restart' after canceling
|
|
|
|
amount2 = 1000000
|
|
|
|
funding_addr = l1.rpc.fundchannel_start(l2.info['id'], amount2)['funding_address']
|
|
|
|
|
|
|
|
addr = l1.rpc.newaddr()['bech32']
|
|
|
|
l1.bitcoin.rpc.sendtoaddress(addr, 0.1)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
wait_for(lambda: len(l1.rpc.listfunds()['outputs']) == 1)
|
|
|
|
# Create the funding transaction
|
|
|
|
prep = l1.rpc.txprepare([{funding_addr: amount2}])
|
|
|
|
decode = bitcoind.rpc.decoderawtransaction(prep['unsigned_tx'])
|
|
|
|
assert decode['txid'] == prep['txid']
|
|
|
|
|
|
|
|
# One output will be correct.
|
|
|
|
if decode['vout'][0]['value'] == Decimal('0.01000000'):
|
|
|
|
txout = 0
|
|
|
|
elif decode['vout'][1]['value'] == Decimal('0.01000000'):
|
|
|
|
txout = 1
|
|
|
|
else:
|
|
|
|
assert False
|
|
|
|
|
|
|
|
# Be sure fundchannel_complete is successful
|
|
|
|
assert l1.rpc.fundchannel_complete(l2.info['id'], prep['txid'], txout)['commitments_secured']
|
|
|
|
|
|
|
|
# Peer shouldn't be able to cancel channel
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot cancel channel that was initiated by peer'):
|
|
|
|
l2.rpc.fundchannel_cancel(l1.info['id'])
|
|
|
|
|
|
|
|
# We can cancel channel after fundchannel_complete
|
|
|
|
assert l1.rpc.fundchannel_cancel(l2.info['id'])['cancelled']
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount)['funding_address']
|
|
|
|
assert l1.rpc.fundchannel_complete(l2.info['id'], prep['txid'], txout)['commitments_secured']
|
|
|
|
|
|
|
|
# Check that can still cancel when peer is disconnected
|
|
|
|
l1.rpc.disconnect(l2.info['id'], force=True)
|
|
|
|
wait_for(lambda: not only_one(l1.rpc.listpeers()['peers'])['connected'])
|
|
|
|
assert l1.rpc.fundchannel_cancel(l2.info['id'])['cancelled']
|
|
|
|
assert len(l1.rpc.listpeers()['peers']) == 0
|
|
|
|
|
|
|
|
# l2 still has the channel open/waiting
|
|
|
|
wait_for(lambda: only_one(only_one(l2.rpc.listpeers()['peers'])['channels'])['state']
|
|
|
|
== 'CHANNELD_AWAITING_LOCKIN')
|
|
|
|
|
|
|
|
# on reconnect, channel should get destroyed
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.daemon.wait_for_log('Rejecting WIRE_CHANNEL_REESTABLISH for unknown channel_id')
|
|
|
|
wait_for(lambda: len(l1.rpc.listpeers()['peers']) == 0)
|
|
|
|
wait_for(lambda: len(l2.rpc.listpeers()['peers']) == 0)
|
|
|
|
|
|
|
|
# we have to connect again, because we got disconnected when everything errored
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount)['funding_address']
|
|
|
|
# A successful funding_complete will always have a commitments_secured that is true,
|
|
|
|
# otherwise it would have failed
|
|
|
|
assert l1.rpc.fundchannel_complete(l2.info['id'], prep['txid'], txout)['commitments_secured']
|
|
|
|
l1.rpc.txsend(prep['txid'])
|
|
|
|
with pytest.raises(RpcError, match=r'.* been broadcast.*'):
|
|
|
|
l1.rpc.fundchannel_cancel(l2.info['id'])
|
|
|
|
l1.rpc.close(l2.info['id'])
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_cancel_race(node_factory, bitcoind, executor):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
|
|
|
|
if VALGRIND or SLOW_MACHINE:
|
|
|
|
num = 5
|
|
|
|
else:
|
|
|
|
num = 100
|
|
|
|
|
|
|
|
nodes = node_factory.get_nodes(num)
|
|
|
|
|
|
|
|
# Speed up cleanup by not cleaning our test nodes: on my laptop, this goes
|
|
|
|
# from 214 to 15 seconds
|
|
|
|
node_factory.nodes = [l1]
|
|
|
|
|
|
|
|
num_complete = 0
|
|
|
|
num_cancel = 0
|
|
|
|
|
|
|
|
for count, n in enumerate(nodes):
|
|
|
|
l1.rpc.connect(n.info['id'], 'localhost', n.port)
|
|
|
|
l1.rpc.fundchannel_start(n.info['id'], "100000sat")
|
|
|
|
|
|
|
|
# We simply make up txids. And submit two of each at once.
|
|
|
|
completes = []
|
|
|
|
cancels = []
|
|
|
|
|
|
|
|
# Switch order around.
|
|
|
|
for i in range(4):
|
|
|
|
if (i + count) % 2 == 0:
|
|
|
|
completes.append(executor.submit(l1.rpc.fundchannel_complete, n.info['id'], "9f1844419d2f41532a57fb5ef038cacb602000f7f37b3dae68dc2d047c89048f", 0))
|
|
|
|
else:
|
|
|
|
cancels.append(executor.submit(l1.rpc.fundchannel_cancel, n.info['id']))
|
|
|
|
|
lightningd/opening_control.c: Remove 'Try fundchannel_cancel again' error.
Changelog-Changed: `fundchannel_cancel` will now succeed even when executed while a `fundchannel_complete` is ongoing; in that case, it will be considered as cancelling the funding *after* the `fundchannel_complete` succeeds.
Let me introduce the concept of "Sequential Consistency":
All operations on parallel processes form a single total order agreed upon by all processes.
So for example, suppose we have parallel invocations of `fundchannel_complete` and `fundchannel_cancel`:
+--[fundchannel_complete]-->
|
--[fundchannel_start]-+
|
+--[fundchannel_cancel]---->
What "Sequential Consistency" means is that the above parallel operations can be serialized as a single total order as:
--[fundchannel_start]--[fundchannel_complete]--[fundchannel_cancel]-->
Or:
--[fundchannel_start]--[fundchannel_cancel]--[fundchannel_complete]-->
In the first case, `fundchannel_complete` succeeds, and the `fundchannel_cancel` invocation also succeeds, sending an `error` to the peer to make them forget the chanel.
In the second case, `fundchannel_cancel` succeeds, and the succeeding `fundchannel_complete` invocation fails, since the funding is already cancelled and there is nothing to complete.
Note that in both cases, `fundchannel_cancel` **always** succeeds.
Unfortunately, prior to this commit, `fundchannel_cancel` could fail with a `Try fundchannel_cancel again` error if the `fundchannel_complete` is ongoing when the `fundchannel_cancel` is initiated.
This violates Sequential Consistency, as there is no single total order that would have caused `fundchannel_cancel` to fail.
This commit is a minimal patch which just reschedules `fundchannel_cancel` to occur after any `fundchannel_complete` that is ongoing.
5 years ago
|
|
|
# Only up to one should succeed.
|
|
|
|
success = False
|
|
|
|
for c in completes:
|
|
|
|
try:
|
|
|
|
c.result(TIMEOUT)
|
|
|
|
num_complete += 1
|
|
|
|
assert not success
|
|
|
|
success = True
|
|
|
|
except RpcError:
|
|
|
|
pass
|
|
|
|
|
lightningd/opening_control.c: Remove 'Try fundchannel_cancel again' error.
Changelog-Changed: `fundchannel_cancel` will now succeed even when executed while a `fundchannel_complete` is ongoing; in that case, it will be considered as cancelling the funding *after* the `fundchannel_complete` succeeds.
Let me introduce the concept of "Sequential Consistency":
All operations on parallel processes form a single total order agreed upon by all processes.
So for example, suppose we have parallel invocations of `fundchannel_complete` and `fundchannel_cancel`:
+--[fundchannel_complete]-->
|
--[fundchannel_start]-+
|
+--[fundchannel_cancel]---->
What "Sequential Consistency" means is that the above parallel operations can be serialized as a single total order as:
--[fundchannel_start]--[fundchannel_complete]--[fundchannel_cancel]-->
Or:
--[fundchannel_start]--[fundchannel_cancel]--[fundchannel_complete]-->
In the first case, `fundchannel_complete` succeeds, and the `fundchannel_cancel` invocation also succeeds, sending an `error` to the peer to make them forget the chanel.
In the second case, `fundchannel_cancel` succeeds, and the succeeding `fundchannel_complete` invocation fails, since the funding is already cancelled and there is nothing to complete.
Note that in both cases, `fundchannel_cancel` **always** succeeds.
Unfortunately, prior to this commit, `fundchannel_cancel` could fail with a `Try fundchannel_cancel again` error if the `fundchannel_complete` is ongoing when the `fundchannel_cancel` is initiated.
This violates Sequential Consistency, as there is no single total order that would have caused `fundchannel_cancel` to fail.
This commit is a minimal patch which just reschedules `fundchannel_cancel` to occur after any `fundchannel_complete` that is ongoing.
5 years ago
|
|
|
# At least one of these must succeed, regardless of whether
|
|
|
|
# the completes succeeded or not.
|
|
|
|
cancelled = False
|
|
|
|
for c in cancels:
|
|
|
|
try:
|
|
|
|
c.result(TIMEOUT)
|
|
|
|
cancelled = True
|
|
|
|
except RpcError:
|
|
|
|
pass
|
lightningd/opening_control.c: Remove 'Try fundchannel_cancel again' error.
Changelog-Changed: `fundchannel_cancel` will now succeed even when executed while a `fundchannel_complete` is ongoing; in that case, it will be considered as cancelling the funding *after* the `fundchannel_complete` succeeds.
Let me introduce the concept of "Sequential Consistency":
All operations on parallel processes form a single total order agreed upon by all processes.
So for example, suppose we have parallel invocations of `fundchannel_complete` and `fundchannel_cancel`:
+--[fundchannel_complete]-->
|
--[fundchannel_start]-+
|
+--[fundchannel_cancel]---->
What "Sequential Consistency" means is that the above parallel operations can be serialized as a single total order as:
--[fundchannel_start]--[fundchannel_complete]--[fundchannel_cancel]-->
Or:
--[fundchannel_start]--[fundchannel_cancel]--[fundchannel_complete]-->
In the first case, `fundchannel_complete` succeeds, and the `fundchannel_cancel` invocation also succeeds, sending an `error` to the peer to make them forget the chanel.
In the second case, `fundchannel_cancel` succeeds, and the succeeding `fundchannel_complete` invocation fails, since the funding is already cancelled and there is nothing to complete.
Note that in both cases, `fundchannel_cancel` **always** succeeds.
Unfortunately, prior to this commit, `fundchannel_cancel` could fail with a `Try fundchannel_cancel again` error if the `fundchannel_complete` is ongoing when the `fundchannel_cancel` is initiated.
This violates Sequential Consistency, as there is no single total order that would have caused `fundchannel_cancel` to fail.
This commit is a minimal patch which just reschedules `fundchannel_cancel` to occur after any `fundchannel_complete` that is ongoing.
5 years ago
|
|
|
# cancel always succeeds, as per Sequential Consistency.
|
|
|
|
# Either the cancel occurred before complete, in which
|
|
|
|
# case it prevents complete from succeeding, or it
|
|
|
|
# occurred after complete, in which case it errors the
|
|
|
|
# channel to force the remote to forget it.
|
|
|
|
assert cancelled
|
|
|
|
num_cancel += 1
|
|
|
|
|
|
|
|
print("Cancelled {} complete {}".format(num_cancel, num_complete))
|
lightningd/opening_control.c: Remove 'Try fundchannel_cancel again' error.
Changelog-Changed: `fundchannel_cancel` will now succeed even when executed while a `fundchannel_complete` is ongoing; in that case, it will be considered as cancelling the funding *after* the `fundchannel_complete` succeeds.
Let me introduce the concept of "Sequential Consistency":
All operations on parallel processes form a single total order agreed upon by all processes.
So for example, suppose we have parallel invocations of `fundchannel_complete` and `fundchannel_cancel`:
+--[fundchannel_complete]-->
|
--[fundchannel_start]-+
|
+--[fundchannel_cancel]---->
What "Sequential Consistency" means is that the above parallel operations can be serialized as a single total order as:
--[fundchannel_start]--[fundchannel_complete]--[fundchannel_cancel]-->
Or:
--[fundchannel_start]--[fundchannel_cancel]--[fundchannel_complete]-->
In the first case, `fundchannel_complete` succeeds, and the `fundchannel_cancel` invocation also succeeds, sending an `error` to the peer to make them forget the chanel.
In the second case, `fundchannel_cancel` succeeds, and the succeeding `fundchannel_complete` invocation fails, since the funding is already cancelled and there is nothing to complete.
Note that in both cases, `fundchannel_cancel` **always** succeeds.
Unfortunately, prior to this commit, `fundchannel_cancel` could fail with a `Try fundchannel_cancel again` error if the `fundchannel_complete` is ongoing when the `fundchannel_cancel` is initiated.
This violates Sequential Consistency, as there is no single total order that would have caused `fundchannel_cancel` to fail.
This commit is a minimal patch which just reschedules `fundchannel_cancel` to occur after any `fundchannel_complete` that is ongoing.
5 years ago
|
|
|
assert num_cancel == len(nodes)
|
|
|
|
|
|
|
|
# We should have raced at least once!
|
|
|
|
if not VALGRIND:
|
|
|
|
assert num_cancel > 0
|
|
|
|
assert num_complete > 0
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(TEST_NETWORK != 'regtest', "External wallet support doesn't work with elements yet.")
|
|
|
|
def test_funding_close_upfront(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
|
|
|
|
opts = {'plugin': os.path.join(os.getcwd(), 'tests/plugins/accepter_close_to.py')}
|
|
|
|
l2 = node_factory.get_node(options=opts)
|
|
|
|
|
|
|
|
# The 'accepter_close_to' plugin uses the channel funding amount to determine
|
|
|
|
# whether or not to include a 'close_to' address
|
|
|
|
amt_normal = 100007 # continues without returning a close_to
|
|
|
|
amt_addr = 100001 # returns valid regtest address
|
|
|
|
|
|
|
|
remote_valid_addr = 'bcrt1q7gtnxmlaly9vklvmfj06amfdef3rtnrdazdsvw'
|
|
|
|
|
|
|
|
def _fundchannel(l1, l2, amount, close_to):
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
assert(l1.rpc.listpeers()['peers'][0]['id'] == l2.info['id'])
|
|
|
|
|
|
|
|
resp = l1.rpc.fundchannel_start(l2.info['id'], amount, close_to=close_to)
|
|
|
|
address = resp['funding_address']
|
|
|
|
|
|
|
|
if close_to:
|
|
|
|
assert resp['close_to']
|
|
|
|
else:
|
|
|
|
assert 'close_to' not in resp
|
|
|
|
|
|
|
|
peer = l1.rpc.listpeers()['peers'][0]
|
|
|
|
# Peer should still be connected and in state waiting for funding_txid
|
|
|
|
assert peer['id'] == l2.info['id']
|
|
|
|
r = re.compile('Funding channel start: awaiting funding_txid with output to .*')
|
|
|
|
assert any(r.match(line) for line in peer['channels'][0]['status'])
|
|
|
|
assert 'OPENINGD' in peer['channels'][0]['state']
|
|
|
|
|
|
|
|
# 'Externally' fund the address from fundchannel_start
|
|
|
|
addr_scriptpubkey = bitcoind.rpc.getaddressinfo(address)['scriptPubKey']
|
|
|
|
txout = CMutableTxOut(amount, bytearray.fromhex(addr_scriptpubkey))
|
|
|
|
unfunded_tx = CMutableTransaction([], [txout])
|
|
|
|
hextx = binascii.hexlify(unfunded_tx.serialize()).decode('utf8')
|
|
|
|
|
|
|
|
funded_tx_obj = bitcoind.rpc.fundrawtransaction(hextx)
|
|
|
|
raw_funded_tx = funded_tx_obj['hex']
|
|
|
|
txid = bitcoind.rpc.decoderawtransaction(raw_funded_tx)['txid']
|
|
|
|
txout = 1 if funded_tx_obj['changepos'] == 0 else 0
|
|
|
|
|
|
|
|
assert l1.rpc.fundchannel_complete(l2.info['id'], txid, txout)['commitments_secured']
|
|
|
|
|
|
|
|
# Broadcast the transaction manually and confirm that channel locks in
|
|
|
|
signed_tx = bitcoind.rpc.signrawtransactionwithwallet(raw_funded_tx)['hex']
|
|
|
|
assert txid == bitcoind.rpc.decoderawtransaction(signed_tx)['txid']
|
|
|
|
|
|
|
|
bitcoind.rpc.sendrawtransaction(signed_tx)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
for node in [l1, l2]:
|
|
|
|
node.daemon.wait_for_log(r'State changed from CHANNELD_AWAITING_LOCKIN to CHANNELD_NORMAL')
|
|
|
|
channel = node.rpc.listpeers()['peers'][0]['channels'][-1]
|
|
|
|
assert amount * 1000 == channel['msatoshi_total']
|
|
|
|
|
|
|
|
# check that normal peer close works
|
|
|
|
_fundchannel(l1, l2, amt_normal, None)
|
|
|
|
assert l1.rpc.close(l2.info['id'])['type'] == 'mutual'
|
|
|
|
|
|
|
|
# check that you can provide a closing address upfront
|
|
|
|
addr = l1.rpc.newaddr()['bech32']
|
|
|
|
_fundchannel(l1, l2, amt_normal, addr)
|
|
|
|
# confirm that it appears in listpeers
|
|
|
|
assert addr == only_one(l1.rpc.listpeers()['peers'])['channels'][1]['close_to_addr']
|
|
|
|
resp = l1.rpc.close(l2.info['id'])
|
|
|
|
assert resp['type'] == 'mutual'
|
|
|
|
assert only_one(only_one(bitcoind.rpc.decoderawtransaction(resp['tx'])['vout'])['scriptPubKey']['addresses']) == addr
|
|
|
|
|
|
|
|
# check that passing in the same addr to close works
|
|
|
|
addr = bitcoind.rpc.getnewaddress()
|
|
|
|
_fundchannel(l1, l2, amt_normal, addr)
|
|
|
|
assert addr == only_one(l1.rpc.listpeers()['peers'])['channels'][2]['close_to_addr']
|
|
|
|
resp = l1.rpc.close(l2.info['id'], destination=addr)
|
|
|
|
assert resp['type'] == 'mutual'
|
|
|
|
assert only_one(only_one(bitcoind.rpc.decoderawtransaction(resp['tx'])['vout'])['scriptPubKey']['addresses']) == addr
|
|
|
|
|
|
|
|
# check that remote peer closing works as expected (and that remote's close_to works)
|
|
|
|
_fundchannel(l1, l2, amt_addr, addr)
|
|
|
|
# send some money to remote so that they have a closeout
|
|
|
|
l1.rpc.pay(l2.rpc.invoice((amt_addr // 2) * 1000, 'test_remote_close_to', 'desc')['bolt11'])
|
|
|
|
assert only_one(l2.rpc.listpeers()['peers'])['channels'][-1]['close_to_addr'] == remote_valid_addr
|
|
|
|
|
|
|
|
resp = l2.rpc.close(l1.info['id'])
|
|
|
|
assert resp['type'] == 'mutual'
|
|
|
|
vouts = bitcoind.rpc.decoderawtransaction(resp['tx'])['vout']
|
|
|
|
assert len(vouts) == 2
|
|
|
|
for vout in vouts:
|
|
|
|
assert only_one(vout['scriptPubKey']['addresses']) in [addr, remote_valid_addr]
|
|
|
|
|
|
|
|
# check that passing in a different addr to close causes an RPC error
|
|
|
|
addr2 = l1.rpc.newaddr()['bech32']
|
|
|
|
_fundchannel(l1, l2, amt_normal, addr)
|
|
|
|
with pytest.raises(RpcError, match=r'does not match previous shutdown script'):
|
|
|
|
l1.rpc.close(l2.info['id'], destination=addr2)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(TEST_NETWORK != 'regtest', "External wallet support doesn't work with elements yet.")
|
|
|
|
def test_funding_external_wallet(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node(options={'funding-confirms': 2})
|
|
|
|
l2 = node_factory.get_node(options={'funding-confirms': 2})
|
|
|
|
l3 = node_factory.get_node()
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
assert(l1.rpc.listpeers()['peers'][0]['id'] == l2.info['id'])
|
|
|
|
|
|
|
|
amount = 2**24 - 1
|
|
|
|
address = l1.rpc.fundchannel_start(l2.info['id'], amount)['funding_address']
|
|
|
|
assert len(address) > 0
|
|
|
|
|
|
|
|
peer = l1.rpc.listpeers()['peers'][0]
|
|
|
|
# Peer should still be connected and in state waiting for funding_txid
|
|
|
|
assert peer['id'] == l2.info['id']
|
|
|
|
r = re.compile('Funding channel start: awaiting funding_txid with output to .*')
|
|
|
|
assert any(r.match(line) for line in peer['channels'][0]['status'])
|
|
|
|
assert 'OPENINGD' in peer['channels'][0]['state']
|
|
|
|
|
|
|
|
# Trying to start a second funding should not work, it's in progress.
|
|
|
|
with pytest.raises(RpcError, match=r'Already funding channel'):
|
|
|
|
l1.rpc.fundchannel_start(l2.info['id'], amount)
|
|
|
|
|
|
|
|
# 'Externally' fund the address from fundchannel_start
|
|
|
|
addr_scriptpubkey = bitcoind.rpc.getaddressinfo(address)['scriptPubKey']
|
|
|
|
txout = CMutableTxOut(amount, bytearray.fromhex(addr_scriptpubkey))
|
|
|
|
unfunded_tx = CMutableTransaction([], [txout])
|
|
|
|
hextx = binascii.hexlify(unfunded_tx.serialize()).decode('utf8')
|
|
|
|
|
|
|
|
funded_tx_obj = bitcoind.rpc.fundrawtransaction(hextx)
|
|
|
|
raw_funded_tx = funded_tx_obj['hex']
|
|
|
|
txid = bitcoind.rpc.decoderawtransaction(raw_funded_tx)['txid']
|
|
|
|
txout = 1 if funded_tx_obj['changepos'] == 0 else 0
|
|
|
|
|
|
|
|
assert l1.rpc.fundchannel_complete(l2.info['id'], txid, txout)['commitments_secured']
|
|
|
|
|
|
|
|
# Broadcast the transaction manually
|
|
|
|
signed_tx = bitcoind.rpc.signrawtransactionwithwallet(raw_funded_tx)['hex']
|
|
|
|
assert txid == bitcoind.rpc.decoderawtransaction(signed_tx)['txid']
|
|
|
|
|
|
|
|
bitcoind.rpc.sendrawtransaction(signed_tx)
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log(r'Funding tx {} depth 1 of 2'.format(txid))
|
|
|
|
|
|
|
|
# Check that tx is broadcast by a third party can be catched.
|
|
|
|
# Only when the transaction (broadcast by a third pary) is onchain, we can catch it.
|
|
|
|
with pytest.raises(RpcError, match=r'.* been broadcast.*'):
|
|
|
|
l1.rpc.fundchannel_cancel(l2.info['id'])
|
|
|
|
|
|
|
|
# Confirm that channel locks in
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
for node in [l1, l2]:
|
|
|
|
node.daemon.wait_for_log(r'State changed from CHANNELD_AWAITING_LOCKIN to CHANNELD_NORMAL')
|
|
|
|
channel = node.rpc.listpeers()['peers'][0]['channels'][0]
|
|
|
|
assert amount * 1000 == channel['msatoshi_total']
|
|
|
|
|
|
|
|
# Test that we don't crash if peer disconnects after fundchannel_start
|
|
|
|
l2.connect(l3)
|
|
|
|
l2.rpc.fundchannel_start(l3.info["id"], amount)
|
|
|
|
l3.rpc.close(l2.info["id"])
|
|
|
|
|
|
|
|
|
|
|
|
def test_lockin_between_restart(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l2 = node_factory.get_node(options={'funding-confirms': 3},
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
l1.fundwallet(10**6 + 1000000)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6)['tx']
|
|
|
|
|
|
|
|
# l1 goes down.
|
|
|
|
l1.stop()
|
|
|
|
|
|
|
|
# Now 120 blocks go by...
|
|
|
|
bitcoind.generate_block(120)
|
|
|
|
|
|
|
|
# Restart
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
# All should be good.
|
|
|
|
l1.daemon.wait_for_log(' to CHANNELD_NORMAL')
|
|
|
|
l2.daemon.wait_for_log(' to CHANNELD_NORMAL')
|
|
|
|
|
|
|
|
|
|
|
|
def test_funding_while_offline(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
addr = l1.rpc.newaddr()['bech32']
|
|
|
|
sync_blockheight(bitcoind, [l1])
|
|
|
|
|
|
|
|
# l1 goes down.
|
|
|
|
l1.stop()
|
|
|
|
|
|
|
|
# We send funds
|
|
|
|
bitcoind.rpc.sendtoaddress(addr, (10**6 + 1000000) / 10**8)
|
|
|
|
|
|
|
|
# Now 120 blocks go by...
|
|
|
|
bitcoind.generate_block(120)
|
|
|
|
|
|
|
|
# Restart
|
|
|
|
l1.start()
|
|
|
|
sync_blockheight(bitcoind, [l1])
|
|
|
|
|
|
|
|
assert len(l1.rpc.listfunds()['outputs']) == 1
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_channel_persistence(node_factory, bitcoind, executor):
|
|
|
|
# Start two nodes and open a channel (to remember). l2 will
|
|
|
|
# mysteriously die while committing the first HTLC so we can
|
|
|
|
# check that HTLCs reloaded from the DB work.
|
|
|
|
# Feerates identical so we don't get gratuitous commit to update them
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True, feerates=(7500, 7500, 7500,
|
|
|
|
7500))
|
|
|
|
l2 = node_factory.get_node(disconnect=['=WIRE_COMMITMENT_SIGNED-nocommit'],
|
|
|
|
may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Neither node should have a channel open, they are just connected
|
|
|
|
for n in (l1, l2):
|
|
|
|
assert(n.db_query('SELECT COUNT(id) as count FROM channels;')[0]['count'] == 0)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 100000)
|
|
|
|
|
|
|
|
peers = l1.rpc.listpeers()['peers']
|
|
|
|
assert(only_one(peers[0]['channels'])['state'] == 'CHANNELD_NORMAL')
|
|
|
|
|
|
|
|
# Both nodes should now have exactly one channel in the database
|
|
|
|
for n in (l1, l2):
|
|
|
|
assert(n.db_query('SELECT COUNT(id) as count FROM channels;')[0]['count'] == 1)
|
|
|
|
|
|
|
|
# Fire off a sendpay request, it'll get interrupted by a restart
|
|
|
|
executor.submit(l1.pay, l2, 10000)
|
|
|
|
# Wait for it to be committed to, i.e., stored in the DB
|
|
|
|
l1.daemon.wait_for_log('peer_in WIRE_COMMITMENT_SIGNED')
|
|
|
|
|
|
|
|
# Stop l2, l1 will reattempt to connect
|
|
|
|
print("Killing l2 in mid HTLC")
|
|
|
|
l2.daemon.kill()
|
|
|
|
|
|
|
|
# Clear the disconnect and timer stop so we can proceed normally
|
|
|
|
del l2.daemon.opts['dev-disconnect']
|
|
|
|
|
|
|
|
# Wait for l1 to notice
|
|
|
|
wait_for(lambda: 'connected' not in only_one(l1.rpc.listpeers()['peers'][0]['channels']))
|
|
|
|
|
|
|
|
# Now restart l2 and it should reload peers/channels from the DB
|
|
|
|
l2.start()
|
|
|
|
wait_for(lambda: len(l2.rpc.listpeers()['peers']) == 1)
|
|
|
|
|
|
|
|
# Wait for the restored HTLC to finish
|
|
|
|
wait_for(lambda: only_one(l1.rpc.listpeers()['peers'][0]['channels'])['msatoshi_to_us'] == 99990000)
|
|
|
|
|
|
|
|
wait_for(lambda: len([p for p in l1.rpc.listpeers()['peers'] if p['connected']]))
|
|
|
|
wait_for(lambda: len([p for p in l2.rpc.listpeers()['peers'] if p['connected']]))
|
|
|
|
|
|
|
|
# Now make sure this is really functional by sending a payment
|
|
|
|
l1.pay(l2, 10000)
|
|
|
|
|
|
|
|
# L1 doesn't actually update msatoshi_to_us until it receives
|
|
|
|
# revoke_and_ack from L2, which can take a little bit.
|
|
|
|
wait_for(lambda: only_one(l1.rpc.listpeers()['peers'][0]['channels'])['msatoshi_to_us'] == 99980000)
|
|
|
|
assert only_one(l2.rpc.listpeers()['peers'][0]['channels'])['msatoshi_to_us'] == 20000
|
|
|
|
|
|
|
|
# Finally restart l1, and make sure it remembers
|
|
|
|
l1.restart()
|
|
|
|
assert only_one(l1.rpc.listpeers()['peers'][0]['channels'])['msatoshi_to_us'] == 99980000
|
|
|
|
|
|
|
|
# Now make sure l1 is watching for unilateral closes
|
|
|
|
l2.rpc.dev_fail(l1.info['id'])
|
|
|
|
l2.daemon.wait_for_log('Failing due to dev-fail command')
|
|
|
|
l2.wait_for_channel_onchain(l1.info['id'])
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
# L1 must notice.
|
|
|
|
l1.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "gossip without DEVELOPER=1 is slow")
|
|
|
|
def test_private_channel(node_factory):
|
|
|
|
l1, l2 = node_factory.line_graph(2, announce_channels=False, wait_for_announce=False)
|
|
|
|
l3, l4 = node_factory.line_graph(2, announce_channels=True, wait_for_announce=True)
|
|
|
|
|
|
|
|
assert l1.daemon.is_in_log('Will open private channel with node {}'.format(l2.info['id']))
|
|
|
|
assert not l2.daemon.is_in_log('Will open private channel with node {}'.format(l1.info['id']))
|
|
|
|
assert not l3.daemon.is_in_log('Will open private channel with node {}'.format(l4.info['id']))
|
|
|
|
|
|
|
|
l3.daemon.wait_for_log('Received node_announcement for node {}'.format(l4.info['id']))
|
|
|
|
l4.daemon.wait_for_log('Received node_announcement for node {}'.format(l3.info['id']))
|
|
|
|
|
|
|
|
assert not l1.daemon.is_in_log('Received node_announcement for node {}'.format(l2.info['id']))
|
|
|
|
assert not l2.daemon.is_in_log('Received node_announcement for node {}'.format(l1.info['id']))
|
|
|
|
|
|
|
|
# test for 'private' flag in rpc output
|
|
|
|
assert only_one(only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['channels'])['private']
|
|
|
|
# check non-private channel
|
|
|
|
assert not only_one(only_one(l4.rpc.listpeers(l3.info['id'])['peers'])['channels'])['private']
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "Too slow without --dev-fast-gossip")
|
|
|
|
def test_channel_reenable(node_factory):
|
|
|
|
l1, l2 = node_factory.line_graph(2, opts={'may_reconnect': True}, fundchannel=True, wait_for_announce=True)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log('Received node_announcement for node {}'.format(l2.info['id']))
|
|
|
|
l2.daemon.wait_for_log('Received node_announcement for node {}'.format(l1.info['id']))
|
|
|
|
|
|
|
|
# Both directions should be active before the restart
|
|
|
|
wait_for(lambda: [c['active'] for c in l1.rpc.listchannels()['channels']] == [True, True])
|
|
|
|
|
|
|
|
# Restart l2, will cause l1 to reconnect
|
|
|
|
l2.stop()
|
|
|
|
wait_for(lambda: [c['active'] for c in l1.rpc.listchannels()['channels']] == [False, False])
|
|
|
|
l2.start()
|
|
|
|
|
|
|
|
# Updates may be suppressed if redundant; just test results.
|
|
|
|
wait_for(lambda: [c['active'] for c in l1.rpc.listchannels()['channels']] == [True, True])
|
|
|
|
wait_for(lambda: [c['active'] for c in l2.rpc.listchannels()['channels']] == [True, True])
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_update_fee(node_factory, bitcoind):
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=True)
|
|
|
|
chanid = l1.get_channel_scid(l2)
|
|
|
|
|
|
|
|
# Make l1 send out feechange.
|
|
|
|
l1.set_feerates((14000, 11000, 7500, 3750))
|
|
|
|
|
|
|
|
# Now make sure an HTLC works.
|
|
|
|
# (First wait for route propagation.)
|
|
|
|
l1.wait_channel_active(chanid)
|
|
|
|
sync_blockheight(bitcoind, [l1, l2])
|
|
|
|
|
|
|
|
# Make payments.
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
# First payment causes fee update.
|
|
|
|
l2.daemon.wait_for_log('peer updated fee to 14000')
|
|
|
|
l2.pay(l1, 100000000)
|
|
|
|
|
|
|
|
# Now shutdown cleanly.
|
|
|
|
l1.rpc.close(chanid)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
|
|
|
l2.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
|
|
|
|
|
|
|
# And should put closing into mempool.
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
l2.wait_for_channel_onchain(l1.info['id'])
|
|
|
|
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
l1.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
l2.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
|
|
|
|
bitcoind.generate_block(99)
|
|
|
|
l1.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_fee_limits(node_factory, bitcoind):
|
|
|
|
l1, l2 = node_factory.line_graph(2, opts={'dev-max-fee-multiplier': 5, 'may_reconnect': True}, fundchannel=True)
|
|
|
|
|
|
|
|
# Kick off fee adjustment using HTLC.
|
|
|
|
l1.pay(l2, 1000)
|
|
|
|
|
|
|
|
# L1 asks for stupid low fee (will actually hit the floor of 253)
|
|
|
|
l1.stop()
|
|
|
|
l1.set_feerates((15, 15, 15, 15), False)
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log('Peer transient failure in CHANNELD_NORMAL: channeld: .*: update_fee 253 outside range 1875-75000')
|
|
|
|
# Make sure the resolution of this one doesn't interfere with the next!
|
|
|
|
# Note: may succeed, may fail with insufficient fee, depending on how
|
|
|
|
# bitcoind feels!
|
|
|
|
l2.daemon.wait_for_log('sendrawtx exit')
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
sync_blockheight(bitcoind, [l1, l2])
|
|
|
|
|
|
|
|
# Trying to open a channel with too low a fee-rate is denied
|
|
|
|
l4 = node_factory.get_node()
|
|
|
|
l1.rpc.connect(l4.info['id'], 'localhost', l4.port)
|
|
|
|
with pytest.raises(RpcError, match='They sent error .* feerate_per_kw 253 below minimum'):
|
|
|
|
l1.fund_channel(l4, 10**6)
|
|
|
|
|
|
|
|
# Restore to normal.
|
|
|
|
l1.stop()
|
|
|
|
l1.set_feerates((15000, 11000, 7500, 3750), False)
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
# Try with node which sets --ignore-fee-limits
|
|
|
|
l3 = node_factory.get_node(options={'ignore-fee-limits': 'true'}, may_reconnect=True)
|
|
|
|
l1.rpc.connect(l3.info['id'], 'localhost', l3.port)
|
|
|
|
chan = l1.fund_channel(l3, 10**6)
|
|
|
|
|
|
|
|
# Kick off fee adjustment using HTLC.
|
|
|
|
l1.pay(l3, 1000)
|
|
|
|
|
|
|
|
# Try stupid high fees
|
|
|
|
l1.stop()
|
|
|
|
l1.set_feerates((15000 * 10, 11000, 7500, 3750), False)
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
l3.daemon.wait_for_log('peer_in WIRE_UPDATE_FEE')
|
|
|
|
l3.daemon.wait_for_log('peer_in WIRE_COMMITMENT_SIGNED')
|
|
|
|
|
|
|
|
# We need to wait until both have committed and revoked the
|
|
|
|
# old state, otherwise we'll still try to commit with the old
|
|
|
|
# 15sat/byte fee
|
|
|
|
l1.daemon.wait_for_log('peer_out WIRE_REVOKE_AND_ACK')
|
|
|
|
|
|
|
|
# This should wait for close to complete
|
|
|
|
l1.rpc.close(chan)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_update_fee_reconnect(node_factory, bitcoind):
|
|
|
|
# Disconnect after commitsig for fee update.
|
|
|
|
disconnects = ['+WIRE_COMMITMENT_SIGNED*3']
|
|
|
|
# Feerates identical so we don't get gratuitous commit to update them
|
|
|
|
l1 = node_factory.get_node(disconnect=disconnects, may_reconnect=True,
|
|
|
|
feerates=(15000, 15000, 15000, 3750))
|
|
|
|
# We match l2's later feerate, so we agree on same closing tx for simplicity.
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True,
|
|
|
|
feerates=(14000, 15000, 14000, 3750))
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
chan = l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
# Make an HTLC just to get us to do feechanges.
|
|
|
|
l1.pay(l2, 1000)
|
|
|
|
|
|
|
|
# Make l1 send out feechange; triggers disconnect/reconnect.
|
|
|
|
# (Note: < 10% change, so no smoothing here!)
|
|
|
|
l1.set_feerates((14000, 14000, 14000, 3750))
|
|
|
|
l1.daemon.wait_for_log('Setting REMOTE feerate to 14000')
|
|
|
|
l2.daemon.wait_for_log('Setting LOCAL feerate to 14000')
|
|
|
|
l1.daemon.wait_for_log(r'dev_disconnect: \+WIRE_COMMITMENT_SIGNED')
|
|
|
|
|
|
|
|
# Wait for reconnect....
|
|
|
|
l1.daemon.wait_for_log('Feerate:.*LOCAL now 14000')
|
|
|
|
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
l2.pay(l1, 100000000)
|
|
|
|
|
|
|
|
# They should both have gotten commits with correct feerate.
|
|
|
|
assert l1.daemon.is_in_log('got commitsig [0-9]*: feerate 14000')
|
|
|
|
assert l2.daemon.is_in_log('got commitsig [0-9]*: feerate 14000')
|
|
|
|
|
|
|
|
# Now shutdown cleanly.
|
|
|
|
l1.rpc.close(chan)
|
|
|
|
|
|
|
|
# And should put closing into mempool.
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
l2.wait_for_channel_onchain(l1.info['id'])
|
|
|
|
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
l1.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
l2.daemon.wait_for_log(' to ONCHAIN')
|
|
|
|
|
|
|
|
bitcoind.generate_block(99)
|
|
|
|
l1.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "Too slow without --dev-bitcoind-poll")
|
|
|
|
def test_multiple_channels(node_factory):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
|
|
|
|
for i in range(3):
|
|
|
|
# FIXME: we shouldn't disconnect on close?
|
|
|
|
ret = l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
assert ret['id'] == l2.info['id']
|
|
|
|
|
openingd: take peer before we're opening, wait for explicit funding msg.
Prior to this, lightningd would hand uninteresting peers back to connectd,
which would then return it to lightningd if it sent a non-gossip msg,
or if lightningd asked it to release the peer.
Now connectd hands the peer to lightningd once we've done the init
handshake, which hands it off to openingd.
This is a deep structural change, so we do the minimum here and cleanup
in the following patches.
Lightningd:
1. Remove peer_nongossip handling from connect_control and peer_control.
2. Remove list of outstanding fundchannel command; it was only needed to
find the race between us asking connectd to release the peer and it
reconnecting.
3. We can no longer tell if the remote end has started trying to fund a
channel (until it has succeeded): it's very transitory anyway so not
worth fixing.
4. We now always have a struct peer, and allocate an uncommitted_channel
for it, though it may never be used if neither end funds a channel.
5. We start funding on messages for openingd: we can get a funder_reply
or a fundee, or an error in response to our request to fund a channel.
so we handle all of them.
6. A new peer_start_openingd() is called after connectd hands us a peer.
7. json_fund_channel just looks through local peers; there are none
hidden in connectd any more.
8. We sometimes start a new openingd just to send an error message.
Openingd:
1. We always have information we need to accept them funding a channel (in
the init message).
2. We have to listen for three fds: peer, gossip and master, so we opencode
the poll.
3. We have an explicit message to start trying to fund a channel.
4. We can be told to send a message in our init message.
Testing:
1. We don't handle some things gracefully yet, so two tests are disabled.
2. 'hand_back_peer .*: now local again' from connectd is no longer a message,
openingd says 'Handed peer, entering loop' once its managing it.
3. peer['state'] used to be set to 'GOSSIPING' (otherwise this field doesn't
exist; 'state' is now per-channel. It doesn't exist at all now.
4. Some tests now need to turn on IO logging in openingd, not connectd.
5. There's a gap between connecting on one node and having connectd on
the peer hand over the connection to openingd. Our tests sometimes
checked getpeers() on the peer, and didn't see anything, so line_graph
needed updating.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
6 years ago
|
|
|
l1.daemon.wait_for_log('openingd-.*: Handed peer, entering loop')
|
|
|
|
l2.daemon.wait_for_log('openingd-.*: Handed peer, entering loop')
|
|
|
|
chan = l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
l1.rpc.close(chan)
|
|
|
|
|
|
|
|
channels = only_one(l1.rpc.listpeers()['peers'])['channels']
|
|
|
|
assert len(channels) == 3
|
|
|
|
# Most in state ONCHAIN, last is CLOSINGD_COMPLETE
|
|
|
|
for i in range(len(channels) - 1):
|
|
|
|
assert channels[i]['state'] == 'ONCHAIN'
|
|
|
|
assert channels[-1]['state'] == 'CLOSINGD_COMPLETE'
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
|
|
|
def test_forget_channel(node_factory):
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
l1.fundwallet(10**6)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**5)
|
|
|
|
|
|
|
|
assert len(l1.rpc.listpeers()['peers']) == 1
|
|
|
|
|
|
|
|
# This should fail, the funding tx is in the mempool and may confirm
|
|
|
|
with pytest.raises(RpcError, match=r'Cowardly refusing to forget channel'):
|
|
|
|
l1.rpc.dev_forget_channel(l2.info['id'])
|
|
|
|
|
|
|
|
assert len(l1.rpc.listpeers()['peers']) == 1
|
|
|
|
|
|
|
|
# Forcing should work
|
|
|
|
l1.rpc.dev_forget_channel(l2.info['id'], True)
|
|
|
|
assert len(l1.rpc.listpeers()['peers']) == 0
|
|
|
|
|
|
|
|
# And restarting should keep that peer forgotten
|
|
|
|
l1.restart()
|
|
|
|
assert len(l1.rpc.listpeers()['peers']) == 0
|
|
|
|
|
|
|
|
# The entry in the channels table should still be there
|
|
|
|
assert l1.db_query("SELECT count(*) as c FROM channels;")[0]['c'] == 1
|
|
|
|
assert l2.db_query("SELECT count(*) as c FROM channels;")[0]['c'] == 1
|
|
|
|
|
|
|
|
|
|
|
|
def test_peerinfo(node_factory, bitcoind):
|
|
|
|
l1, l2 = node_factory.line_graph(2, fundchannel=False, opts={'may_reconnect': True})
|
|
|
|
lfeatures = expected_peer_features()
|
|
|
|
nfeatures = expected_node_features()
|
|
|
|
# Gossiping but no node announcement yet
|
openingd: take peer before we're opening, wait for explicit funding msg.
Prior to this, lightningd would hand uninteresting peers back to connectd,
which would then return it to lightningd if it sent a non-gossip msg,
or if lightningd asked it to release the peer.
Now connectd hands the peer to lightningd once we've done the init
handshake, which hands it off to openingd.
This is a deep structural change, so we do the minimum here and cleanup
in the following patches.
Lightningd:
1. Remove peer_nongossip handling from connect_control and peer_control.
2. Remove list of outstanding fundchannel command; it was only needed to
find the race between us asking connectd to release the peer and it
reconnecting.
3. We can no longer tell if the remote end has started trying to fund a
channel (until it has succeeded): it's very transitory anyway so not
worth fixing.
4. We now always have a struct peer, and allocate an uncommitted_channel
for it, though it may never be used if neither end funds a channel.
5. We start funding on messages for openingd: we can get a funder_reply
or a fundee, or an error in response to our request to fund a channel.
so we handle all of them.
6. A new peer_start_openingd() is called after connectd hands us a peer.
7. json_fund_channel just looks through local peers; there are none
hidden in connectd any more.
8. We sometimes start a new openingd just to send an error message.
Openingd:
1. We always have information we need to accept them funding a channel (in
the init message).
2. We have to listen for three fds: peer, gossip and master, so we opencode
the poll.
3. We have an explicit message to start trying to fund a channel.
4. We can be told to send a message in our init message.
Testing:
1. We don't handle some things gracefully yet, so two tests are disabled.
2. 'hand_back_peer .*: now local again' from connectd is no longer a message,
openingd says 'Handed peer, entering loop' once its managing it.
3. peer['state'] used to be set to 'GOSSIPING' (otherwise this field doesn't
exist; 'state' is now per-channel. It doesn't exist at all now.
4. Some tests now need to turn on IO logging in openingd, not connectd.
5. There's a gap between connecting on one node and having connectd on
the peer hand over the connection to openingd. Our tests sometimes
checked getpeers() on the peer, and didn't see anything, so line_graph
needed updating.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
6 years ago
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['connected']
|
|
|
|
assert len(l1.rpc.getpeer(l2.info['id'])['channels']) == 0
|
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['features'] == lfeatures
|
|
|
|
|
|
|
|
# Fund a channel to force a node announcement
|
|
|
|
chan = l1.fund_channel(l2, 10**6)
|
|
|
|
# Now proceed to funding-depth and do a full gossip round
|
|
|
|
bitcoind.generate_block(5)
|
|
|
|
l1.daemon.wait_for_logs(['Received node_announcement for node ' + l2.info['id']])
|
|
|
|
l2.daemon.wait_for_logs(['Received node_announcement for node ' + l1.info['id']])
|
|
|
|
|
|
|
|
# Should have announced the same features as told to peer.
|
|
|
|
nodes1 = l1.rpc.listnodes(l2.info['id'])['nodes']
|
|
|
|
nodes2 = l2.rpc.listnodes(l2.info['id'])['nodes']
|
|
|
|
peer1 = l1.rpc.getpeer(l2.info['id'])
|
|
|
|
peer2 = l2.rpc.getpeer(l1.info['id'])
|
|
|
|
# peer features != to node features now because of keysend, which adds a node feature
|
|
|
|
assert only_one(nodes1)['features'] == nfeatures
|
|
|
|
assert only_one(nodes2)['features'] == nfeatures
|
|
|
|
assert peer1['features'] == lfeatures
|
|
|
|
assert peer2['features'] == lfeatures
|
|
|
|
|
|
|
|
# If it reconnects after db load, it should know features.
|
|
|
|
l1.restart()
|
|
|
|
wait_for(lambda: l1.rpc.getpeer(l2.info['id'])['connected'])
|
|
|
|
wait_for(lambda: l2.rpc.getpeer(l1.info['id'])['connected'])
|
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['features'] == lfeatures
|
|
|
|
assert l2.rpc.getpeer(l1.info['id'])['features'] == lfeatures
|
|
|
|
|
|
|
|
# Close the channel to forget the peer
|
|
|
|
l1.rpc.close(chan)
|
|
|
|
|
|
|
|
wait_for(lambda: not only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['connected'])
|
|
|
|
wait_for(lambda: not only_one(l2.rpc.listpeers(l1.info['id'])['peers'])['connected'])
|
|
|
|
|
|
|
|
# Make sure close tx hits mempool before we mine blocks.
|
|
|
|
bitcoind.generate_block(100, wait_for_mempool=1)
|
|
|
|
l1.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
# The only channel was closed, everybody should have forgotten the nodes
|
|
|
|
assert l1.rpc.listnodes()['nodes'] == []
|
|
|
|
assert l2.rpc.listnodes()['nodes'] == []
|
|
|
|
|
|
|
|
|
|
|
|
def test_disconnectpeer(node_factory, bitcoind):
|
|
|
|
l1, l2, l3 = node_factory.get_nodes(3, opts={'may_reconnect': False})
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.connect(l3.info['id'], 'localhost', l3.port)
|
|
|
|
|
|
|
|
# Gossiping
|
openingd: take peer before we're opening, wait for explicit funding msg.
Prior to this, lightningd would hand uninteresting peers back to connectd,
which would then return it to lightningd if it sent a non-gossip msg,
or if lightningd asked it to release the peer.
Now connectd hands the peer to lightningd once we've done the init
handshake, which hands it off to openingd.
This is a deep structural change, so we do the minimum here and cleanup
in the following patches.
Lightningd:
1. Remove peer_nongossip handling from connect_control and peer_control.
2. Remove list of outstanding fundchannel command; it was only needed to
find the race between us asking connectd to release the peer and it
reconnecting.
3. We can no longer tell if the remote end has started trying to fund a
channel (until it has succeeded): it's very transitory anyway so not
worth fixing.
4. We now always have a struct peer, and allocate an uncommitted_channel
for it, though it may never be used if neither end funds a channel.
5. We start funding on messages for openingd: we can get a funder_reply
or a fundee, or an error in response to our request to fund a channel.
so we handle all of them.
6. A new peer_start_openingd() is called after connectd hands us a peer.
7. json_fund_channel just looks through local peers; there are none
hidden in connectd any more.
8. We sometimes start a new openingd just to send an error message.
Openingd:
1. We always have information we need to accept them funding a channel (in
the init message).
2. We have to listen for three fds: peer, gossip and master, so we opencode
the poll.
3. We have an explicit message to start trying to fund a channel.
4. We can be told to send a message in our init message.
Testing:
1. We don't handle some things gracefully yet, so two tests are disabled.
2. 'hand_back_peer .*: now local again' from connectd is no longer a message,
openingd says 'Handed peer, entering loop' once its managing it.
3. peer['state'] used to be set to 'GOSSIPING' (otherwise this field doesn't
exist; 'state' is now per-channel. It doesn't exist at all now.
4. Some tests now need to turn on IO logging in openingd, not connectd.
5. There's a gap between connecting on one node and having connectd on
the peer hand over the connection to openingd. Our tests sometimes
checked getpeers() on the peer, and didn't see anything, so line_graph
needed updating.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
6 years ago
|
|
|
assert l1.rpc.getpeer(l2.info['id'])['connected']
|
|
|
|
assert len(l1.rpc.getpeer(l2.info['id'])['channels']) == 0
|
|
|
|
assert l1.rpc.getpeer(l3.info['id'])['connected']
|
|
|
|
assert len(l1.rpc.getpeer(l3.info['id'])['channels']) == 0
|
|
|
|
wait_for(lambda: l2.rpc.getpeer(l1.info['id']) is not None)
|
|
|
|
|
|
|
|
# Disconnect l2 from l1
|
|
|
|
l1.rpc.disconnect(l2.info['id'])
|
|
|
|
|
|
|
|
# Make sure listpeers no longer returns the disconnected node
|
|
|
|
assert l1.rpc.getpeer(l2.info['id']) is None
|
|
|
|
wait_for(lambda: l2.rpc.getpeer(l1.info['id']) is None)
|
|
|
|
|
|
|
|
# Make sure you cannot disconnect after disconnecting
|
|
|
|
with pytest.raises(RpcError, match=r'Peer not connected'):
|
|
|
|
l1.rpc.disconnect(l2.info['id'])
|
|
|
|
with pytest.raises(RpcError, match=r'Peer not connected'):
|
|
|
|
l2.rpc.disconnect(l1.info['id'])
|
|
|
|
|
|
|
|
# Fund channel l1 -> l3
|
|
|
|
l1.fund_channel(l3, 10**6)
|
|
|
|
bitcoind.generate_block(5)
|
|
|
|
|
|
|
|
# disconnecting a non gossiping peer results in error
|
|
|
|
with pytest.raises(RpcError, match=r'Peer is in state CHANNELD_NORMAL'):
|
|
|
|
l1.rpc.disconnect(l3.info['id'])
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs --dev-max-funding-unconfirmed-blocks")
|
|
|
|
def test_fundee_forget_funding_tx_unconfirmed(node_factory, bitcoind):
|
|
|
|
"""Test that fundee will forget the channel if
|
|
|
|
the funding tx has been unconfirmed for too long.
|
|
|
|
"""
|
|
|
|
# Keep this low (default is 2016), since everything
|
|
|
|
# is much slower in VALGRIND mode and wait_for_log
|
|
|
|
# could time out before lightningd processes all the
|
|
|
|
# blocks.
|
|
|
|
blocks = 200
|
|
|
|
# opener
|
|
|
|
l1 = node_factory.get_node()
|
|
|
|
# peer
|
|
|
|
l2 = node_factory.get_node(options={"dev-max-funding-unconfirmed-blocks": blocks})
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Give opener some funds.
|
|
|
|
l1.fundwallet(10**7)
|
|
|
|
# Let blocks settle.
|
|
|
|
time.sleep(1)
|
|
|
|
|
|
|
|
def mock_sendrawtransaction(r):
|
|
|
|
return {'id': r['id'], 'error': {'code': 100, 'message': 'sendrawtransaction disabled'}}
|
|
|
|
|
|
|
|
# Prevent opener from broadcasting funding tx (any tx really).
|
|
|
|
l1.daemon.rpcproxy.mock_rpc('sendrawtransaction', mock_sendrawtransaction)
|
|
|
|
|
|
|
|
# Fund the channel.
|
|
|
|
# The process will complete, but opener will be unable
|
|
|
|
# to broadcast and confirm funding tx.
|
|
|
|
with pytest.raises(RpcError, match=r'sendrawtransaction disabled'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6)
|
|
|
|
|
|
|
|
# Generate blocks until unconfirmed.
|
|
|
|
bitcoind.generate_block(blocks)
|
|
|
|
|
|
|
|
# fundee will forget channel!
|
|
|
|
l2.daemon.wait_for_log('Forgetting channel: It has been {} blocks'.format(blocks))
|
|
|
|
|
|
|
|
# fundee will also forget and disconnect from peer.
|
|
|
|
assert len(l2.rpc.listpeers(l1.info['id'])['peers']) == 0
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs dev_fail")
|
|
|
|
def test_no_fee_estimate(node_factory, bitcoind, executor):
|
|
|
|
l1 = node_factory.get_node(start=False)
|
|
|
|
|
|
|
|
# Fail any fee estimation requests until we allow them further down
|
|
|
|
l1.daemon.rpcproxy.mock_rpc('estimatesmartfee', {
|
|
|
|
'error': {"errors": ["Insufficient data or no feerate found"], "blocks": 0}
|
|
|
|
})
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
|
|
|
|
# Can't fund a channel.
|
|
|
|
l1.fundwallet(10**7)
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot estimate fees'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6)
|
|
|
|
|
|
|
|
# Can't withdraw either.
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot estimate fees'):
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 'all')
|
|
|
|
|
|
|
|
# Can't use feerate names, either.
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot estimate fees'):
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 'all', 'urgent')
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 'all', 'normal')
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 'all', 'slow')
|
|
|
|
|
|
|
|
with pytest.raises(RpcError, match=r'Cannot estimate fees'):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6, 'urgent')
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6, 'normal')
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6, 'slow')
|
|
|
|
|
|
|
|
# Can with manual feerate.
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 10000, '1500perkb')
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6, '2000perkw', minconf=0)
|
|
|
|
|
|
|
|
# Make sure we clean up cahnnel for later attempt.
|
|
|
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
|
|
|
l1.rpc.dev_fail(l2.info['id'])
|
|
|
|
l1.daemon.wait_for_log('Failing due to dev-fail command')
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
bitcoind.generate_block(6)
|
|
|
|
wait_for(lambda: only_one(l1.rpc.getpeer(l2.info['id'])['channels'])['state'] == 'ONCHAIN')
|
|
|
|
wait_for(lambda: only_one(l2.rpc.getpeer(l1.info['id'])['channels'])['state'] == 'ONCHAIN')
|
|
|
|
|
|
|
|
# But can accept incoming connections.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l2.fund_channel(l1, 10**6)
|
|
|
|
|
|
|
|
# Can do HTLCs.
|
|
|
|
l2.pay(l1, 10**5)
|
|
|
|
|
|
|
|
# Can do mutual close.
|
|
|
|
l1.rpc.close(l2.info['id'])
|
|
|
|
wait_for(lambda: len(bitcoind.rpc.getrawmempool()) > 0)
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
|
|
|
|
# Can do unilateral close.
|
|
|
|
l2.rpc.connect(l1.info['id'], 'localhost', l1.port)
|
|
|
|
l2.fund_channel(l1, 10**6)
|
|
|
|
l2.pay(l1, 10**9 // 2)
|
|
|
|
l1.rpc.dev_fail(l2.info['id'])
|
|
|
|
l1.daemon.wait_for_log('Failing due to dev-fail command')
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
bitcoind.generate_block(5)
|
|
|
|
wait_for(lambda: len(bitcoind.rpc.getrawmempool()) > 0)
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
|
|
|
|
# Start estimatesmartfee.
|
|
|
|
l1.set_feerates((15000, 11000, 7500, 3750), True)
|
|
|
|
|
|
|
|
# Can now fund a channel (as a test, use slow feerate).
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**6, 'slow')
|
|
|
|
|
|
|
|
# Can withdraw (use urgent feerate). `minconf` may be needed depending on
|
|
|
|
# the previous `fundchannel` selecting all confirmed outputs.
|
|
|
|
l1.rpc.withdraw(l2.rpc.newaddr()['bech32'], 'all', 'urgent', minconf=0)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs --dev-disconnect")
|
|
|
|
def test_opener_feerate_reconnect(node_factory, bitcoind):
|
|
|
|
# l1 updates fees, then reconnect so l2 retransmits commitment_signed.
|
|
|
|
disconnects = ['-WIRE_COMMITMENT_SIGNED*3']
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True,
|
|
|
|
feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(disconnect=disconnects, may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
# Need a payment otherwise it won't update fee.
|
|
|
|
l1.pay(l2, 10**9 // 2)
|
|
|
|
|
|
|
|
# create fee update, causing disconnect.
|
|
|
|
l1.set_feerates((15000, 11000, 7500, 3750))
|
|
|
|
l2.daemon.wait_for_log(r'dev_disconnect: \-WIRE_COMMITMENT_SIGNED')
|
|
|
|
|
|
|
|
# Wait until they reconnect.
|
|
|
|
l1.daemon.wait_for_log('Peer transient failure in CHANNELD_NORMAL')
|
|
|
|
wait_for(lambda: l1.rpc.getpeer(l2.info['id'])['connected'])
|
|
|
|
|
|
|
|
# Should work normally.
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
|
|
|
|
|
|
|
|
def test_opener_simple_reconnect(node_factory, bitcoind):
|
|
|
|
"""Sanity check that reconnection works with completely unused channels"""
|
|
|
|
# Set fees even so it doesn't send any commitments.
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True,
|
|
|
|
feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True)
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
|
|
|
|
l1.rpc.disconnect(l2.info['id'], True)
|
|
|
|
|
|
|
|
# Wait until they reconnect.
|
|
|
|
wait_for(lambda: l1.rpc.getpeer(l2.info['id'])['connected'])
|
|
|
|
|
|
|
|
# Should work normally.
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(os.getenv('TEST_DB_PROVIDER', 'sqlite3') != 'sqlite3', "sqlite3-specific DB rollback")
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs LIGHTNINGD_DEV_LOG_IO")
|
|
|
|
def test_dataloss_protection(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True, options={'log-level': 'io'},
|
|
|
|
feerates=(7500, 7500, 7500, 7500))
|
|
|
|
l2 = node_factory.get_node(may_reconnect=True, options={'log-level': 'io'},
|
|
|
|
feerates=(7500, 7500, 7500, 7500), allow_broken_log=True)
|
|
|
|
|
|
|
|
lf = expected_peer_features()
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
# l1 should send out WIRE_INIT (0010)
|
|
|
|
l1.daemon.wait_for_log(r"\[OUT\] 0010.*"
|
|
|
|
# lflen
|
|
|
|
+ format(len(lf) // 2, '04x')
|
|
|
|
+ lf)
|
|
|
|
|
|
|
|
l1.fund_channel(l2, 10**6)
|
|
|
|
l2.stop()
|
|
|
|
|
|
|
|
# Save copy of the db.
|
|
|
|
dbpath = os.path.join(l2.daemon.lightning_dir, TEST_NETWORK, "lightningd.sqlite3")
|
|
|
|
orig_db = open(dbpath, "rb").read()
|
|
|
|
l2.start()
|
|
|
|
|
|
|
|
# l1 should have sent WIRE_CHANNEL_REESTABLISH with extra fields.
|
|
|
|
l1.daemon.wait_for_log(r"\[OUT\] 0088"
|
|
|
|
# channel_id
|
|
|
|
"[0-9a-f]{64}"
|
|
|
|
# next_local_commitment_number
|
|
|
|
"0000000000000001"
|
|
|
|
# next_remote_revocation_number
|
|
|
|
"0000000000000000"
|
|
|
|
# your_last_per_commitment_secret (funding_depth may
|
|
|
|
# trigger a fee-update and commit, hence this may not
|
|
|
|
# be zero)
|
|
|
|
"[0-9a-f]{64}"
|
|
|
|
# my_current_per_commitment_point
|
|
|
|
"0[23][0-9a-f]{64}'$")
|
|
|
|
|
|
|
|
# After an htlc, we should get different results (two more commits)
|
|
|
|
l1.pay(l2, 200000000)
|
|
|
|
|
|
|
|
# Make sure both sides consider it completely settled (has received both
|
|
|
|
# REVOKE_AND_ACK)
|
|
|
|
l1.daemon.wait_for_logs([r"\[IN\] 0085"] * 2)
|
|
|
|
l2.daemon.wait_for_logs([r"\[IN\] 0085"] * 2)
|
|
|
|
|
|
|
|
l2.restart()
|
|
|
|
|
|
|
|
# l1 should have sent WIRE_CHANNEL_REESTABLISH with extra fields.
|
|
|
|
l1.daemon.wait_for_log(r"\[OUT\] 0088"
|
|
|
|
# channel_id
|
|
|
|
"[0-9a-f]{64}"
|
|
|
|
# next_local_commitment_number
|
|
|
|
"000000000000000[1-9]"
|
|
|
|
# next_remote_revocation_number
|
|
|
|
"000000000000000[1-9]"
|
|
|
|
# your_last_per_commitment_secret
|
|
|
|
"[0-9a-f]{64}"
|
|
|
|
# my_current_per_commitment_point
|
|
|
|
"0[23][0-9a-f]{64}'$")
|
|
|
|
|
|
|
|
# Now, move l2 back in time.
|
|
|
|
l2.stop()
|
|
|
|
# Overwrite with OLD db.
|
|
|
|
open(dbpath, "wb").write(orig_db)
|
|
|
|
l2.start()
|
|
|
|
|
|
|
|
# l2 should freak out!
|
|
|
|
l2.daemon.wait_for_log("Peer permanent failure in CHANNELD_NORMAL: Awaiting unilateral close")
|
|
|
|
|
|
|
|
# l1 should drop to chain.
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
|
|
|
|
# l2 must NOT drop to chain.
|
|
|
|
l2.daemon.wait_for_log("Cannot broadcast our commitment tx: they have a future one")
|
|
|
|
assert not l2.daemon.is_in_log('sendrawtx exit 0')
|
|
|
|
|
|
|
|
closetxid = only_one(bitcoind.rpc.getrawmempool(False))
|
|
|
|
|
|
|
|
# l2 should still recover something!
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
|
|
|
|
l2.daemon.wait_for_log("ERROR: Unknown commitment #[0-9], recovering our funds!")
|
|
|
|
|
|
|
|
# Restarting l2, and it should remember from db.
|
|
|
|
l2.restart()
|
|
|
|
|
|
|
|
l2.daemon.wait_for_log("ERROR: Unknown commitment #[0-9], recovering our funds!")
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
# l2 should have it in wallet.
|
|
|
|
assert (closetxid, "confirmed") in set([(o['txid'], o['status']) for o in l2.rpc.listfunds()['outputs']])
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs dev_disconnect")
|
|
|
|
def test_restart_multi_htlc_rexmit(node_factory, bitcoind, executor):
|
|
|
|
# l1 disables commit timer once we send first htlc, dies on commit
|
|
|
|
disconnects = ['=WIRE_UPDATE_ADD_HTLC-nocommit',
|
|
|
|
'-WIRE_COMMITMENT_SIGNED']
|
|
|
|
l1, l2 = node_factory.line_graph(2, opts=[{'disconnect': disconnects,
|
|
|
|
'may_reconnect': True},
|
|
|
|
{'may_reconnect': True}])
|
|
|
|
|
|
|
|
executor.submit(l1.pay, l2, 20000)
|
|
|
|
executor.submit(l1.pay, l2, 30000)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_logs(['peer_out WIRE_UPDATE_ADD_HTLC'] * 2)
|
|
|
|
l1.rpc.dev_reenable_commit(l2.info['id'])
|
|
|
|
l1.daemon.wait_for_log('dev_disconnect: -WIRE_COMMITMENT_SIGNED')
|
|
|
|
|
|
|
|
# This will make it reconnect
|
|
|
|
l1.stop()
|
|
|
|
# Clear the disconnect so we can proceed normally
|
|
|
|
del l1.daemon.opts['dev-disconnect']
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
# Payments will fail due to restart, but we can see results in listsendpays.
|
|
|
|
print(l1.rpc.listsendpays())
|
|
|
|
wait_for(lambda: [p['status'] for p in l1.rpc.listsendpays()['payments']] == ['complete', 'complete'])
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "needs dev-disconnect")
|
|
|
|
def test_fulfill_incoming_first(node_factory, bitcoind):
|
|
|
|
"""Test that we handle the case where we completely resolve incoming htlc
|
|
|
|
before fulfilled outgoing htlc"""
|
|
|
|
|
|
|
|
# We agree on fee change first, then add HTLC, then remove; stop after remove.
|
|
|
|
disconnects = ['+WIRE_COMMITMENT_SIGNED*3']
|
|
|
|
# We manually reconnect l2 & l3, after 100 blocks; hence allowing manual
|
|
|
|
# reconnect, but disabling auto connect, and massive cltv so 2/3 doesn't
|
|
|
|
# time out.
|
|
|
|
l1, l2, l3 = node_factory.line_graph(3, opts=[{'disable-mpp': None},
|
|
|
|
{'may_reconnect': True,
|
|
|
|
'dev-no-reconnect': None},
|
|
|
|
{'may_reconnect': True,
|
|
|
|
'dev-no-reconnect': None,
|
|
|
|
'disconnect': disconnects,
|
|
|
|
'cltv-final': 200}],
|
|
|
|
wait_for_announce=True)
|
|
|
|
|
|
|
|
# This succeeds.
|
|
|
|
l1.rpc.pay(l3.rpc.invoice(200000000, 'test_fulfill_incoming_first', 'desc')['bolt11'])
|
|
|
|
|
|
|
|
# l1 can shutdown, fine.
|
|
|
|
l1.rpc.close(l2.info['id'])
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
# Now, l2 should restore from DB fine, even though outgoing HTLC no longer
|
|
|
|
# has an incoming.
|
|
|
|
l2.restart()
|
|
|
|
|
|
|
|
# Manually reconnect l2->l3.
|
|
|
|
l2.rpc.connect(l3.info['id'], 'localhost', l3.port)
|
|
|
|
|
|
|
|
# Fulfill should be retransmitted OK (ignored result).
|
|
|
|
l2.rpc.close(l3.info['id'])
|
|
|
|
l2.wait_for_channel_onchain(l3.info['id'])
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
l2.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
l3.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "gossip without DEVELOPER=1 is slow")
|
|
|
|
def test_restart_many_payments(node_factory, bitcoind):
|
|
|
|
l1 = node_factory.get_node(may_reconnect=True)
|
|
|
|
|
|
|
|
# On my laptop, these take 74 seconds and 44 seconds (with restart commented out)
|
|
|
|
if VALGRIND:
|
|
|
|
num = 2
|
|
|
|
else:
|
|
|
|
num = 5
|
|
|
|
|
|
|
|
# Nodes with channels into the main node
|
|
|
|
innodes = node_factory.get_nodes(num, opts={'may_reconnect': True})
|
|
|
|
inchans = []
|
|
|
|
for n in innodes:
|
|
|
|
n.rpc.connect(l1.info['id'], 'localhost', l1.port)
|
|
|
|
inchans.append(n.fund_channel(l1, 10**6, False))
|
|
|
|
|
|
|
|
# Nodes with channels out of the main node
|
|
|
|
outnodes = node_factory.get_nodes(len(innodes), opts={'may_reconnect': True})
|
|
|
|
outchans = []
|
|
|
|
for n in outnodes:
|
|
|
|
n.rpc.connect(l1.info['id'], 'localhost', l1.port)
|
|
|
|
outchans.append(l1.fund_channel(n, 10**6, False))
|
|
|
|
|
|
|
|
# Make sure they're all announced.
|
|
|
|
bitcoind.generate_block(5)
|
|
|
|
|
|
|
|
# We wait for each node to see each dir active, and its own
|
|
|
|
# channel CHANNELD_NORMAL
|
|
|
|
logs = ([r'update for channel {}/0 now ACTIVE'.format(scid)
|
|
|
|
for scid in inchans + outchans]
|
|
|
|
+ [r'update for channel {}/1 now ACTIVE'.format(scid)
|
|
|
|
for scid in inchans + outchans]
|
|
|
|
+ ['to CHANNELD_NORMAL'])
|
|
|
|
|
|
|
|
# Now do all the waiting at once: if !DEVELOPER, this can be *very* slow!
|
|
|
|
for n in innodes + outnodes:
|
|
|
|
n.daemon.wait_for_logs(logs)
|
|
|
|
|
|
|
|
# Manually create routes, get invoices
|
|
|
|
Payment = namedtuple('Payment', ['innode', 'route', 'payment_hash'])
|
|
|
|
|
|
|
|
to_pay = []
|
|
|
|
for i in range(len(innodes)):
|
|
|
|
# This one will cause WIRE_INCORRECT_CLTV_EXPIRY from l1.
|
|
|
|
route = [{'msatoshi': 100001001,
|
|
|
|
'id': l1.info['id'],
|
|
|
|
'delay': 10,
|
|
|
|
'channel': inchans[i]},
|
|
|
|
{'msatoshi': 100000000,
|
|
|
|
'id': outnodes[i].info['id'],
|
|
|
|
'delay': 5,
|
|
|
|
'channel': outchans[i]}]
|
|
|
|
payment_hash = outnodes[i].rpc.invoice(100000000, "invoice", "invoice")['payment_hash']
|
|
|
|
to_pay.append(Payment(innodes[i], route, payment_hash))
|
|
|
|
|
|
|
|
# This one should be routed through to the outnode.
|
|
|
|
route = [{'msatoshi': 100001001,
|
|
|
|
'id': l1.info['id'],
|
|
|
|
'delay': 11,
|
|
|
|
'channel': inchans[i]},
|
|
|
|
{'msatoshi': 100000000,
|
|
|
|
'id': outnodes[i].info['id'],
|
|
|
|
'delay': 5,
|
|
|
|
'channel': outchans[i]}]
|
|
|
|
payment_hash = outnodes[i].rpc.invoice(100000000, "invoice2", "invoice2")['payment_hash']
|
|
|
|
to_pay.append(Payment(innodes[i], route, payment_hash))
|
|
|
|
|
|
|
|
# sendpay is async.
|
|
|
|
for p in to_pay:
|
|
|
|
p.innode.rpc.sendpay(p.route, p.payment_hash)
|
|
|
|
|
|
|
|
# Now restart l1 while traffic is flowing...
|
|
|
|
l1.restart()
|
|
|
|
|
|
|
|
# Wait for them to finish.
|
|
|
|
for n in innodes:
|
|
|
|
wait_for(lambda: 'pending' not in [p['status'] for p in n.rpc.listsendpays()['payments']])
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "need dev-disconnect")
|
|
|
|
def test_fail_unconfirmed(node_factory, bitcoind, executor):
|
|
|
|
"""Test that if we crash with an unconfirmed connection to a known
|
|
|
|
peer, we don't have a dangling peer in db"""
|
|
|
|
# = is a NOOP disconnect, but sets up file.
|
|
|
|
l1 = node_factory.get_node(disconnect=['=WIRE_OPEN_CHANNEL'])
|
|
|
|
l2 = node_factory.get_node()
|
|
|
|
|
|
|
|
# First one, we close by mutual agreement.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fund_channel(l2, 200000, wait_for_active=True)
|
|
|
|
l1.rpc.close(l2.info['id'])
|
|
|
|
|
|
|
|
# Make sure it's closed
|
|
|
|
l1.wait_for_channel_onchain(l2.info['id'])
|
|
|
|
bitcoind.generate_block(1)
|
|
|
|
l1.daemon.wait_for_log('State changed from CLOSINGD_COMPLETE to FUNDING_SPEND_SEEN')
|
|
|
|
|
|
|
|
l1.stop()
|
|
|
|
# Mangle disconnect file so this time it blackholes....
|
|
|
|
with open(l1.daemon.disconnect_file, "w") as f:
|
|
|
|
f.write("0WIRE_OPEN_CHANNEL\n")
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
# Now we establish a new channel, which gets stuck.
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fundwallet(10**7)
|
|
|
|
executor.submit(l1.rpc.fundchannel, l2.info['id'], 100000)
|
|
|
|
|
|
|
|
l1.daemon.wait_for_log("dev_disconnect")
|
|
|
|
|
|
|
|
# Now complete old channel.
|
|
|
|
bitcoind.generate_block(100)
|
|
|
|
l1.daemon.wait_for_log('onchaind complete, forgetting peer')
|
|
|
|
|
|
|
|
# And crash l1, which is stuck.
|
|
|
|
l1.daemon.kill()
|
|
|
|
|
|
|
|
# Now, restart and see if it can connect OK.
|
|
|
|
del l1.daemon.opts['dev-disconnect']
|
|
|
|
l1.start()
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.port)
|
|
|
|
l1.fund_channel(l2, 200000, wait_for_active=True)
|
|
|
|
|
|
|
|
|
|
|
|
def test_change_chaining(node_factory, bitcoind):
|
|
|
|
"""Test change chaining of unconfirmed fundings
|
|
|
|
|
|
|
|
Change chaining is the case where one transaction is broadcast but not
|
|
|
|
confirmed yet and we already build a followup on top of the change. If the
|
|
|
|
first transaction doesn't confirm we may end up creating a series of
|
|
|
|
unconfirmable transactions. This is why we generally disallow chaining.
|
|
|
|
|
|
|
|
"""
|
|
|
|
l1, l2, l3 = node_factory.get_nodes(3)
|
|
|
|
l1.fundwallet(10**8) # This will create an output with 1 confirmation
|
|
|
|
|
|
|
|
# Now fund a channel from l1 to l2, that should succeed, with minconf=1 but not before
|
|
|
|
l1.connect(l2)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**7, minconf=2)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 10**7) # Defaults to minconf=1
|
|
|
|
|
|
|
|
# We don't have confirmed outputs anymore, so this should fail without minconf=0
|
|
|
|
l1.connect(l3)
|
|
|
|
with pytest.raises(RpcError):
|
|
|
|
l1.rpc.fundchannel(l3.info['id'], 10**7) # Defaults to minconf=1
|
|
|
|
l1.rpc.fundchannel(l3.info['id'], 10**7, minconf=0)
|
|
|
|
|
|
|
|
|
|
|
|
def test_feerate_spam(node_factory, chainparams):
|
|
|
|
l1, l2 = node_factory.line_graph(2)
|
|
|
|
|
|
|
|
# We constrain the value the opener has at its disposal so we get the
|
|
|
|
# REMOTE feerate we are looking for below. This may be fragile and depends
|
|
|
|
# on the transactions we generate.
|
|
|
|
slack = 45000000 if not chainparams['elements'] else 68000000
|
|
|
|
|
|
|
|
# Pay almost everything to l2.
|
|
|
|
l1.pay(l2, 10**9 - slack)
|
|
|
|
|
|
|
|
# It will send this once (may have happened before line_graph's wait)
|
|
|
|
wait_for(lambda: l1.daemon.is_in_log('Setting REMOTE feerate to 15000'))
|
|
|
|
wait_for(lambda: l1.daemon.is_in_log('peer_out WIRE_UPDATE_FEE'))
|
|
|
|
|
|
|
|
# Now change feerates to something l1 can't afford.
|
|
|
|
l1.set_feerates((100000, 100000, 100000, 100000))
|
|
|
|
|
|
|
|
# It will raise as far as it can (48000)
|
|
|
|
l1.daemon.wait_for_log('Setting REMOTE feerate to 48000')
|
|
|
|
l1.daemon.wait_for_log('peer_out WIRE_UPDATE_FEE')
|
|
|
|
|
|
|
|
# But it won't do it again once it's at max.
|
|
|
|
with pytest.raises(TimeoutError):
|
|
|
|
l1.daemon.wait_for_log('peer_out WIRE_UPDATE_FEE', timeout=5)
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "need dev-feerate")
|
|
|
|
def test_feerate_stress(node_factory, executor):
|
|
|
|
# Third node makes HTLC traffic less predictable.
|
|
|
|
l1, l2, l3 = node_factory.line_graph(3, opts={'commit-time': 100,
|
|
|
|
'may_reconnect': True})
|
|
|
|
|
|
|
|
l1.pay(l2, 10**9 // 2)
|
|
|
|
scid12 = l1.get_channel_scid(l2)
|
|
|
|
scid23 = l2.get_channel_scid(l3)
|
|
|
|
|
|
|
|
routel1l3 = [{'msatoshi': '10002msat', 'id': l2.info['id'], 'delay': 11, 'channel': scid12},
|
|
|
|
{'msatoshi': '10000msat', 'id': l3.info['id'], 'delay': 5, 'channel': scid23}]
|
|
|
|
routel2l1 = [{'msatoshi': '10000msat', 'id': l1.info['id'], 'delay': 5, 'channel': scid12}]
|
|
|
|
|
|
|
|
rate = 1875
|
|
|
|
NUM_ATTEMPTS = 25
|
|
|
|
l1done = 0
|
|
|
|
l2done = 0
|
|
|
|
prev_log = 0
|
|
|
|
while l1done < NUM_ATTEMPTS and l2done < NUM_ATTEMPTS:
|
|
|
|
try:
|
|
|
|
r = random.randrange(6)
|
|
|
|
if r == 5:
|
|
|
|
l1.rpc.sendpay(routel1l3, "{:064x}".format(l1done))
|
|
|
|
l1done += 1
|
|
|
|
elif r == 4:
|
|
|
|
l2.rpc.sendpay(routel2l1, "{:064x}".format(l2done))
|
|
|
|
l2done += 1
|
|
|
|
elif r > 0:
|
|
|
|
l1.rpc.call('dev-feerate', [l2.info['id'], rate])
|
|
|
|
rate += 5
|
|
|
|
else:
|
|
|
|
l2.rpc.disconnect(l1.info['id'], True)
|
|
|
|
time.sleep(1)
|
|
|
|
except RpcError:
|
|
|
|
time.sleep(0.01)
|
|
|
|
assert not l1.daemon.is_in_log('Bad.*signature', start=prev_log)
|
|
|
|
prev_log = len(l1.daemon.logs)
|
|
|
|
|
|
|
|
# Make sure it's reconnected, and wait for last payment.
|
|
|
|
wait_for(lambda: l1.rpc.getpeer(l2.info['id'])['connected'])
|
|
|
|
# We can get TEMPORARY_CHANNEL_FAILURE due to disconnect, too.
|
|
|
|
with pytest.raises(RpcError, match='WIRE_INCORRECT_OR_UNKNOWN_PAYMENT_DETAILS|WIRE_TEMPORARY_CHANNEL_FAILURE'):
|
|
|
|
l1.rpc.waitsendpay("{:064x}".format(l1done - 1))
|
|
|
|
with pytest.raises(RpcError, match='WIRE_INCORRECT_OR_UNKNOWN_PAYMENT_DETAILS|WIRE_TEMPORARY_CHANNEL_FAILURE'):
|
|
|
|
l2.rpc.waitsendpay("{:064x}".format(l2done - 1))
|
|
|
|
l1.rpc.call('dev-feerate', [l2.info['id'], rate - 5])
|
|
|
|
assert not l1.daemon.is_in_log('Bad.*signature')
|
|
|
|
assert not l2.daemon.is_in_log('Bad.*signature')
|
|
|
|
|
|
|
|
|
|
|
|
@unittest.skipIf(not DEVELOPER, "need dev_disconnect")
|
|
|
|
def test_pay_disconnect_stress(node_factory, executor):
|
|
|
|
"""Expose race in htlc restoration in channeld: 50% chance of failure"""
|
|
|
|
if SLOW_MACHINE and VALGRIND:
|
|
|
|
NUM_RUNS = 2
|
|
|
|
else:
|
|
|
|
NUM_RUNS = 5
|
|
|
|
for i in range(NUM_RUNS):
|
|
|
|
l1, l2 = node_factory.line_graph(2, opts=[{'may_reconnect': True},
|
|
|
|
{'may_reconnect': True,
|
|
|
|
'disconnect': ['=WIRE_UPDATE_ADD_HTLC',
|
|
|
|
'-WIRE_COMMITMENT_SIGNED']}])
|
|
|
|
|
|
|
|
scid12 = l1.get_channel_scid(l2)
|
|
|
|
routel2l1 = [{'msatoshi': '10000msat', 'id': l1.info['id'], 'delay': 5, 'channel': scid12}]
|
|
|
|
|
|
|
|
# Get invoice from l1 to pay.
|
|
|
|
payhash1 = l1.rpc.invoice(10000, "invoice", "invoice")['payment_hash']
|
|
|
|
|
|
|
|
# Start balancing payment.
|
|
|
|
fut = executor.submit(l1.pay, l2, 10**9 // 2)
|
|
|
|
|
|
|
|
# As soon as reverse payment is accepted, reconnect.
|
|
|
|
while True:
|
|
|
|
l2.rpc.sendpay(routel2l1, payhash1)
|
|
|
|
try:
|
|
|
|
# This will usually fail with Capacity exceeded
|
|
|
|
l2.rpc.waitsendpay(payhash1, timeout=TIMEOUT)
|
|
|
|
break
|
|
|
|
except RpcError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
fut.result()
|
|
|
|
|
|
|
|
|
|
|
|
def test_wumbo_channels(node_factory, bitcoind):
|
|
|
|
l1, l2, l3 = node_factory.get_nodes(3,
|
|
|
|
opts=[{'large-channels': None},
|
|
|
|
{'large-channels': None},
|
|
|
|
{}])
|
|
|
|
conn = l1.rpc.connect(l2.info['id'], 'localhost', port=l2.port)
|
|
|
|
assert conn['features'] == expected_peer_features(wumbo_channels=True)
|
|
|
|
assert only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['features'] == expected_peer_features(wumbo_channels=True)
|
|
|
|
|
|
|
|
# Now, can we open a giant channel?
|
|
|
|
l1.fundwallet(1 << 26)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 1 << 24)
|
|
|
|
|
|
|
|
# Get that mined, and announced.
|
|
|
|
bitcoind.generate_block(6, wait_for_mempool=1)
|
|
|
|
|
|
|
|
# Connect l3, get gossip.
|
|
|
|
l3.rpc.connect(l1.info['id'], 'localhost', port=l1.port)
|
|
|
|
wait_for(lambda: len(l3.rpc.listnodes(l1.info['id'])['nodes']) == 1)
|
|
|
|
wait_for(lambda: 'features' in only_one(l3.rpc.listnodes(l1.info['id'])['nodes']))
|
|
|
|
|
|
|
|
# Make sure channel capacity is what we expected.
|
|
|
|
assert ([c['amount_msat'] for c in l3.rpc.listchannels()['channels']]
|
|
|
|
== [Millisatoshi(str(1 << 24) + "sat")] * 2)
|
|
|
|
|
|
|
|
# Make sure channel features are right from channel_announcement
|
|
|
|
assert ([c['features'] for c in l3.rpc.listchannels()['channels']]
|
|
|
|
== [expected_channel_features(wumbo_channels=True)] * 2)
|
|
|
|
|
|
|
|
# Make sure we can't open a wumbo channel if we don't agree.
|
|
|
|
with pytest.raises(RpcError, match='Amount exceeded'):
|
|
|
|
l1.rpc.fundchannel(l3.info['id'], 1 << 24)
|
|
|
|
|
|
|
|
# But we can open and announce a normal one.
|
|
|
|
l1.rpc.fundchannel(l3.info['id'], 'all')
|
|
|
|
bitcoind.generate_block(6, wait_for_mempool=1)
|
|
|
|
wait_for(lambda: l1.channel_state(l3) == 'CHANNELD_NORMAL')
|
|
|
|
|
|
|
|
# Make sure l2 sees correct size.
|
|
|
|
wait_for(lambda: [c['amount_msat'] for c in l2.rpc.listchannels(l1.get_channel_scid(l3))['channels']]
|
|
|
|
== [Millisatoshi(str((1 << 24) - 1) + "sat")] * 2)
|
|
|
|
|
|
|
|
# Make sure 'all' works with wumbo peers.
|
|
|
|
l1.rpc.close(l2.info['id'])
|
|
|
|
bitcoind.generate_block(1, wait_for_mempool=1)
|
|
|
|
wait_for(lambda: l1.channel_state(l2) == 'ONCHAIN')
|
|
|
|
|
|
|
|
l1.rpc.connect(l2.info['id'], 'localhost', port=l2.port)
|
|
|
|
l1.rpc.fundchannel(l2.info['id'], 'all')
|
|
|
|
bitcoind.generate_block(1, wait_for_mempool=1)
|
|
|
|
wait_for(lambda: 'CHANNELD_NORMAL' in [c['state'] for c in only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['channels']])
|
|
|
|
|
|
|
|
# Exact amount depends on fees, but it will be wumbo!
|
|
|
|
amount = [c['funding_msat'][l1.info['id']] for c in only_one(l1.rpc.listpeers(l2.info['id'])['peers'])['channels'] if c['state'] == 'CHANNELD_NORMAL'][0]
|
|
|
|
assert Millisatoshi(amount) > Millisatoshi(str((1 << 24) - 1) + "sat")
|