mirror of
https://github.com/aljazceru/lightning.git
synced 2025-12-20 07:34:24 +01:00
test_lightningd: Update close to use scid, add test for peer id and full cid.
This commit is contained in:
committed by
Christian Decker
parent
e588737511
commit
b1ffcc3b52
@@ -1263,10 +1263,25 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
|
|
||||||
l2.daemon.wait_for_log('to_self_delay 100 larger than 99')
|
l2.daemon.wait_for_log('to_self_delay 100 larger than 99')
|
||||||
|
|
||||||
|
def test_closing_id(self):
|
||||||
|
"""Test closing using peer ID and full channel ID"""
|
||||||
|
l1, l2 = self.connect()
|
||||||
|
|
||||||
|
# Close by full channel ID.
|
||||||
|
self.fund_channel(l1, l2, 10**6)
|
||||||
|
cid = l2.rpc.listpeers()['peers'][0]['channels'][0]['channel_id']
|
||||||
|
l2.rpc.close(cid)
|
||||||
|
|
||||||
|
# Close by peer ID.
|
||||||
|
l1.rpc.connect(l2.info['id'])
|
||||||
|
self.fund_channel(l2, l1, 10**6)
|
||||||
|
pid = l1.info['id']
|
||||||
|
l2.rpc.close(pid)
|
||||||
|
|
||||||
def test_closing(self):
|
def test_closing(self):
|
||||||
l1, l2 = self.connect()
|
l1, l2 = self.connect()
|
||||||
|
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
self.pay(l1, l2, 200000000)
|
self.pay(l1, l2, 200000000)
|
||||||
|
|
||||||
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
||||||
@@ -1290,7 +1305,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# This should return with an error, then close.
|
# This should return with an error, then close.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chan, False, 0)
|
||||||
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
|
|
||||||
@@ -1334,7 +1349,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
def test_closing_while_disconnected(self):
|
def test_closing_while_disconnected(self):
|
||||||
l1, l2 = self.connect(may_reconnect=True)
|
l1, l2 = self.connect(may_reconnect=True)
|
||||||
|
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
self.pay(l1, l2, 200000000)
|
self.pay(l1, l2, 200000000)
|
||||||
|
|
||||||
l2.stop()
|
l2.stop()
|
||||||
@@ -1342,7 +1357,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# The close should still be triggered afterwards.
|
# The close should still be triggered afterwards.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chan, False, 0)
|
||||||
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
|
|
||||||
l2.daemon.start()
|
l2.daemon.start()
|
||||||
@@ -1426,7 +1441,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
peers.append(p)
|
peers.append(p)
|
||||||
|
|
||||||
for p in peers:
|
for p in peers:
|
||||||
l1.rpc.fundchannel(p.info['id'], 10**6)
|
p.channel = l1.rpc.fundchannel(p.info['id'], 10**6)['channel_id']
|
||||||
# Technically, this is async to fundchannel returning.
|
# Technically, this is async to fundchannel returning.
|
||||||
l1.daemon.wait_for_log('sendrawtx exit 0')
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
||||||
|
|
||||||
@@ -1439,17 +1454,16 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
if p.amount != 0:
|
if p.amount != 0:
|
||||||
self.pay(l1, p, 100000000)
|
self.pay(l1, p, 100000000)
|
||||||
|
|
||||||
# Now close
|
# Now close all channels
|
||||||
# All closes occur in parallel, and on Travis,
|
# All closes occur in parallel, and on Travis,
|
||||||
# ALL those lightningd are running on a single core,
|
# ALL those lightningd are running on a single core,
|
||||||
# so increase the timeout so that this test will pass
|
# so increase the timeout so that this test will pass
|
||||||
# when valgrind is enabled.
|
# when valgrind is enabled.
|
||||||
# (close timeout defaults to 30 as of this writing, more
|
# (close timeout defaults to 30 as of this writing)
|
||||||
# than double the default)
|
closes = [self.executor.submit(l1.rpc.close, p.channel, False, 90) for p in peers]
|
||||||
closes = [self.executor.submit(l1.rpc.close, p.info['id'], False, 72) for p in peers]
|
|
||||||
|
|
||||||
for c in closes:
|
for c in closes:
|
||||||
c.result(72)
|
c.result(90)
|
||||||
|
|
||||||
# close does *not* wait for the sendrawtransaction, so do that!
|
# close does *not* wait for the sendrawtransaction, so do that!
|
||||||
# Note that since they disagree on the ideal fee, they may conflict
|
# Note that since they disagree on the ideal fee, they may conflict
|
||||||
@@ -3325,7 +3339,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l2 = self.node_factory.get_node(may_reconnect=True)
|
l2 = self.node_factory.get_node(may_reconnect=True)
|
||||||
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
||||||
|
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
self.pay(l1, l2, 200000000)
|
self.pay(l1, l2, 200000000)
|
||||||
|
|
||||||
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
||||||
@@ -3333,7 +3347,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# This should return with an error, then close.
|
# This should return with an error, then close.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chan, False, 0)
|
||||||
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
|
|
||||||
@@ -3352,7 +3366,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
|
|
||||||
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
||||||
self.give_funds(l1, 10**6 + 1000000)
|
self.give_funds(l1, 10**6 + 1000000)
|
||||||
l1.rpc.fundchannel(l2.info['id'], 10**6)
|
chanid = l1.rpc.fundchannel(l2.info['id'], 10**6)['channel_id']
|
||||||
|
|
||||||
# Technically, this is async to fundchannel.
|
# Technically, this is async to fundchannel.
|
||||||
l1.daemon.wait_for_log('sendrawtx exit 0')
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
||||||
@@ -3361,7 +3375,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# This should return with an error, then close.
|
# This should return with an error, then close.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chanid, False, 0)
|
||||||
l1.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
l1.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
||||||
l2.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
l2.daemon.wait_for_log('CHANNELD_AWAITING_LOCKIN to CHANNELD_SHUTTING_DOWN')
|
||||||
|
|
||||||
@@ -3391,7 +3405,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l2 = self.node_factory.get_node(may_reconnect=True)
|
l2 = self.node_factory.get_node(may_reconnect=True)
|
||||||
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
||||||
|
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
self.pay(l1, l2, 200000000)
|
self.pay(l1, l2, 200000000)
|
||||||
|
|
||||||
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
assert l1.bitcoin.rpc.getmempoolinfo()['size'] == 0
|
||||||
@@ -3399,7 +3413,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# This should return with an error, then close.
|
# This should return with an error, then close.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chan, False, 0)
|
||||||
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l1.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
l2.daemon.wait_for_log(' to CHANNELD_SHUTTING_DOWN')
|
||||||
|
|
||||||
@@ -4079,7 +4093,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# Now shutdown cleanly.
|
# Now shutdown cleanly.
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chanid, False, 0)
|
||||||
l1.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
l1.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
||||||
l2.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
l2.daemon.wait_for_log(' to CLOSINGD_COMPLETE')
|
||||||
|
|
||||||
@@ -4098,7 +4112,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
||||||
def test_update_all_fees(self):
|
def test_update_all_fees(self):
|
||||||
l1, l2 = self.connect()
|
l1, l2 = self.connect()
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
|
|
||||||
# Set all fees as positional parameters.
|
# Set all fees as positional parameters.
|
||||||
l1.rpc.dev_setfees('12345', '6789', '123')
|
l1.rpc.dev_setfees('12345', '6789', '123')
|
||||||
@@ -4124,7 +4138,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l1.daemon.wait_for_log('dev-setfees: fees now 21098/7654/321')
|
l1.daemon.wait_for_log('dev-setfees: fees now 21098/7654/321')
|
||||||
|
|
||||||
# This should return finish closing.
|
# This should return finish closing.
|
||||||
l1.rpc.close(l2.info['id'])
|
l1.rpc.close(chan)
|
||||||
|
|
||||||
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
||||||
def test_fee_limits(self):
|
def test_fee_limits(self):
|
||||||
@@ -4148,7 +4162,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l3 = self.node_factory.get_node(options={'ignore-fee-limits': 'true'})
|
l3 = self.node_factory.get_node(options={'ignore-fee-limits': 'true'})
|
||||||
l1.rpc.connect(l3.info['id'], 'localhost', l3.info['port'])
|
l1.rpc.connect(l3.info['id'], 'localhost', l3.info['port'])
|
||||||
|
|
||||||
self.fund_channel(l1, l3, 10**6)
|
chan = self.fund_channel(l1, l3, 10**6)
|
||||||
|
|
||||||
# Try stupid high fees
|
# Try stupid high fees
|
||||||
l1.rpc.dev_setfees(15000 * 10)
|
l1.rpc.dev_setfees(15000 * 10)
|
||||||
@@ -4162,7 +4176,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l1.daemon.wait_for_log('peer_out WIRE_REVOKE_AND_ACK')
|
l1.daemon.wait_for_log('peer_out WIRE_REVOKE_AND_ACK')
|
||||||
|
|
||||||
# This should wait for close to complete
|
# This should wait for close to complete
|
||||||
l1.rpc.close(l3.info['id'])
|
l1.rpc.close(chan)
|
||||||
|
|
||||||
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
@unittest.skipIf(not DEVELOPER, "needs DEVELOPER=1")
|
||||||
def test_update_fee_reconnect(self):
|
def test_update_fee_reconnect(self):
|
||||||
@@ -4172,7 +4186,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
l2 = self.node_factory.get_node(may_reconnect=True)
|
l2 = self.node_factory.get_node(may_reconnect=True)
|
||||||
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
l1.rpc.connect(l2.info['id'], 'localhost', l2.info['port'])
|
||||||
|
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
|
|
||||||
# Make l1 send out feechange; triggers disconnect/reconnect.
|
# Make l1 send out feechange; triggers disconnect/reconnect.
|
||||||
l1.rpc.dev_setfees('14000')
|
l1.rpc.dev_setfees('14000')
|
||||||
@@ -4191,7 +4205,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
assert l2.daemon.is_in_log('got commitsig [0-9]*: feerate 14000')
|
assert l2.daemon.is_in_log('got commitsig [0-9]*: feerate 14000')
|
||||||
|
|
||||||
# Now shutdown cleanly.
|
# Now shutdown cleanly.
|
||||||
l1.rpc.close(l2.info['id'])
|
l1.rpc.close(chan)
|
||||||
|
|
||||||
# And should put closing into mempool.
|
# And should put closing into mempool.
|
||||||
l1.daemon.wait_for_log('sendrawtx exit 0')
|
l1.daemon.wait_for_log('sendrawtx exit 0')
|
||||||
@@ -4323,9 +4337,9 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
|
|
||||||
l1.daemon.wait_for_log('Handing back peer .* to master')
|
l1.daemon.wait_for_log('Handing back peer .* to master')
|
||||||
l2.daemon.wait_for_log('Handing back peer .* to master')
|
l2.daemon.wait_for_log('Handing back peer .* to master')
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
|
|
||||||
l1.rpc.close(l2.info['id'])
|
l1.rpc.close(chan)
|
||||||
|
|
||||||
channels = l1.rpc.listpeers()['peers'][0]['channels']
|
channels = l1.rpc.listpeers()['peers'][0]['channels']
|
||||||
assert len(channels) == 3
|
assert len(channels) == 3
|
||||||
@@ -4458,7 +4472,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
assert 'color' not in l1.rpc.getpeer(l2.info['id'])
|
assert 'color' not in l1.rpc.getpeer(l2.info['id'])
|
||||||
|
|
||||||
# Fund a channel to force a node announcement
|
# Fund a channel to force a node announcement
|
||||||
self.fund_channel(l1, l2, 10**6)
|
chan = self.fund_channel(l1, l2, 10**6)
|
||||||
# Now proceed to funding-depth and do a full gossip round
|
# Now proceed to funding-depth and do a full gossip round
|
||||||
bitcoind.generate_block(5)
|
bitcoind.generate_block(5)
|
||||||
l1.daemon.wait_for_logs(['Received node_announcement for node ' + l2.info['id']])
|
l1.daemon.wait_for_logs(['Received node_announcement for node ' + l2.info['id']])
|
||||||
@@ -4470,7 +4484,7 @@ class LightningDTests(BaseLightningDTests):
|
|||||||
# Close the channel to forget the peer
|
# Close the channel to forget the peer
|
||||||
self.assertRaisesRegex(ValueError,
|
self.assertRaisesRegex(ValueError,
|
||||||
"Channel close negotiation not finished",
|
"Channel close negotiation not finished",
|
||||||
l1.rpc.close, l2.info['id'], False, 0)
|
l1.rpc.close, chan, False, 0)
|
||||||
l1.daemon.wait_for_log('Forgetting remote peer')
|
l1.daemon.wait_for_log('Forgetting remote peer')
|
||||||
bitcoind.generate_block(100)
|
bitcoind.generate_block(100)
|
||||||
l1.daemon.wait_for_log('WIRE_ONCHAIN_ALL_IRREVOCABLY_RESOLVED')
|
l1.daemon.wait_for_log('WIRE_ONCHAIN_ALL_IRREVOCABLY_RESOLVED')
|
||||||
|
|||||||
Reference in New Issue
Block a user