X-Git-Url: http://git.bitcoin.ninja/index.cgi?a=blobdiff_plain;f=src%2Ftest%2Fjava%2Forg%2Fldk%2FHumanObjectPeerTest.java;h=e719779f3ba56e2bf15975dceb6a8c36507f815f;hb=57c69192cf720e784e837610444e1c762ab0a6fa;hp=781d4d2c9870ada248bdba729dc2bf6b802058d9;hpb=55fbcecedad3fd98420db299c2177fb3f40d607e;p=ldk-java diff --git a/src/test/java/org/ldk/HumanObjectPeerTest.java b/src/test/java/org/ldk/HumanObjectPeerTest.java index 781d4d2c..e719779f 100644 --- a/src/test/java/org/ldk/HumanObjectPeerTest.java +++ b/src/test/java/org/ldk/HumanObjectPeerTest.java @@ -1,20 +1,42 @@ package org.ldk; import org.bitcoinj.core.*; -import org.bitcoinj.core.Transaction; import org.bitcoinj.script.Script; import org.junit.jupiter.api.Test; +import org.ldk.batteries.ChannelManagerConstructor; +import org.ldk.batteries.NioPeerHandler; +import org.ldk.enums.LDKCurrency; import org.ldk.enums.LDKNetwork; import org.ldk.impl.bindings; import org.ldk.structs.*; import org.ldk.util.TwoTuple; +import java.io.IOException; import java.lang.ref.WeakReference; -import java.util.Arrays; -import java.util.HashMap; -import java.util.LinkedList; +import java.net.InetSocketAddress; +import java.util.*; class HumanObjectPeerTestInstance { + private final boolean nice_close; + private final boolean use_km_wrapper; + private final boolean use_manual_watch; + private final boolean reload_peers; + private final boolean break_cross_peer_refs; + private final boolean use_nio_peer_handler; + private final boolean use_filter; + private final boolean use_chan_manager_constructor; + + HumanObjectPeerTestInstance(boolean nice_close, boolean use_km_wrapper, boolean use_manual_watch, boolean reload_peers, boolean break_cross_peer_refs, boolean use_nio_peer_handler, boolean use_filter, boolean use_chan_manager_constructor) { + this.nice_close = nice_close; + this.use_km_wrapper = use_km_wrapper; + this.use_manual_watch = use_manual_watch; + this.reload_peers = reload_peers; + this.break_cross_peer_refs = break_cross_peer_refs; + this.use_nio_peer_handler = use_nio_peer_handler; + this.use_filter = use_filter; + this.use_chan_manager_constructor = use_chan_manager_constructor; + } + class Peer { KeysInterface manual_keysif(KeysInterface underlying_if) { return KeysInterface.new_impl(new KeysInterface.KeysInterfaceInterface() { @@ -23,9 +45,10 @@ class HumanObjectPeerTestInstance { @Override public byte[] get_shutdown_pubkey() { return underlying_if.get_shutdown_pubkey(); } @Override - public ChannelKeys get_channel_keys(boolean inbound, long channel_value_satoshis) { - ChannelKeys underlying_ck = underlying_if.get_channel_keys(inbound, channel_value_satoshis); - ChannelKeys.ChannelKeysInterface cki = new ChannelKeys.ChannelKeysInterface() { + public Sign get_channel_signer(boolean inbound, long channel_value_satoshis) { + Sign underlying_ck = underlying_if.get_channel_signer(inbound, channel_value_satoshis); + // TODO: Expose the underlying signer from a Sign + /*BaseSign.BaseSignInterface si = new BaseSign.BaseSignInterface() { @Override public byte[] get_per_commitment_point(long idx) { return underlying_ck.get_per_commitment_point(idx); @@ -37,8 +60,8 @@ class HumanObjectPeerTestInstance { } @Override - public TwoTuple key_derivation_params() { - return new TwoTuple((long)0, (long)1); + public byte[] channel_keys_id() { + return new byte[32]; } @Override @@ -47,13 +70,8 @@ class HumanObjectPeerTestInstance { } @Override - public Result_SignatureNoneZ sign_holder_commitment(HolderCommitmentTransaction holder_commitment_tx) { - return underlying_ck.sign_holder_commitment(holder_commitment_tx); - } - - @Override - public Result_CVec_SignatureZNoneZ sign_holder_commitment_htlc_transactions(HolderCommitmentTransaction holder_commitment_tx) { - return underlying_ck.sign_holder_commitment_htlc_transactions(holder_commitment_tx); + public Result_C2Tuple_SignatureCVec_SignatureZZNoneZ sign_holder_commitment_and_htlcs(HolderCommitmentTransaction holder_commitment_tx) { + return underlying_ck.sign_holder_commitment_and_htlcs(holder_commitment_tx); } @Override @@ -85,12 +103,13 @@ class HumanObjectPeerTestInstance { public byte[] write() { return underlying_ck.write(); } - }; - ChannelKeys resp = ChannelKeys.new_impl(cki, underlying_ck.get_pubkeys()); - must_free_objs.add(new WeakReference<>(cki)); - must_free_objs.add(new WeakReference<>(resp)); + };*/ + //Sign resp = Sign.new_impl(si, underlying_ck.get_pubkeys()); + //must_free_objs.add(new WeakReference<>(si)); + //must_free_objs.add(new WeakReference<>(resp)); must_free_objs.add(new WeakReference<>(underlying_ck)); - return resp; + //return resp; + return underlying_ck; } @Override @@ -99,19 +118,24 @@ class HumanObjectPeerTestInstance { } @Override - public Result_ChanKeySignerDecodeErrorZ read_chan_signer(byte[] reader) { + public Result_SignDecodeErrorZ read_chan_signer(byte[] reader) { return underlying_if.read_chan_signer(reader); } + + @Override + public Result_RecoverableSignatureNoneZ sign_invoice(byte[] invoice_preimage) { + return underlying_if.sign_invoice(invoice_preimage); + } }); } Watch get_manual_watch() { - return Watch.new_impl(new Watch.WatchInterface() { + Watch.WatchInterface watch_impl = new Watch.WatchInterface() { public Result_NoneChannelMonitorUpdateErrZ watch_channel(OutPoint funding_txo, ChannelMonitor monitor) { synchronized (monitors) { assert monitors.put(Arrays.toString(funding_txo.get_txid()), monitor) == null; } - return new Result_NoneChannelMonitorUpdateErrZ.Result_NoneChannelMonitorUpdateErrZ_OK(); + return Result_NoneChannelMonitorUpdateErrZ.ok(); } public Result_NoneChannelMonitorUpdateErrZ update_channel(OutPoint funding_txo, ChannelMonitorUpdate update) { @@ -121,7 +145,7 @@ class HumanObjectPeerTestInstance { Result_NoneMonitorUpdateErrorZ update_res = monitors.get(txid).update_monitor(update, tx_broadcaster, fee_estimator, logger); assert update_res instanceof Result_NoneMonitorUpdateErrorZ.Result_NoneMonitorUpdateErrorZ_OK; } - return new Result_NoneChannelMonitorUpdateErrZ.Result_NoneChannelMonitorUpdateErrZ_OK(); + return Result_NoneChannelMonitorUpdateErrZ.ok(); } @Override @@ -134,46 +158,100 @@ class HumanObjectPeerTestInstance { } return new MonitorEvent[0]; } - }); + }; + Watch watch = Watch.new_impl(watch_impl); + must_free_objs.add(new WeakReference<>(watch_impl)); + must_free_objs.add(new WeakReference<>(watch)); + return watch; } + NioPeerHandler nio_peer_handler; + short nio_port; + final byte seed; final Logger logger; final FeeEstimator fee_estimator; final BroadcasterInterface tx_broadcaster; + final KeysManager explicit_keys_manager; final KeysInterface keys_interface; final ChainMonitor chain_monitor; - final ChannelManager chan_manager; - final EventsProvider chan_manager_events; final NetGraphMsgHandler router; - final PeerManager peer_manager; + final Watch chain_watch; + final HashSet filter_additions; + final Filter filter; + ChannelManager chan_manager; + PeerManager peer_manager; final HashMap monitors; // Wow I forgot just how terrible Java is - we can't put a byte array here. byte[] node_id; final LinkedList broadcast_set = new LinkedList<>(); + final LinkedList pending_manager_events = new LinkedList<>(); + ChannelManagerConstructor constructor = null; + GcCheck obj = new GcCheck(); + + private TwoTuple test_mon_roundtrip(ChannelMonitor mon) { + // Because get_funding_txo() returns an OutPoint in a tuple that is a reference to an OutPoint inside the + // ChannelMonitor, its a good test to ensure that the OutPoint isn't freed (or is cloned) before the + // ChannelMonitor is. This used to be broken. + Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ roundtrip_monitor = UtilMethods.BlockHashChannelMonitorZ_read(mon.write(), keys_interface); + assert roundtrip_monitor instanceof Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ.Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ_OK; + TwoTuple funding_txo = ((Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ.Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ_OK) roundtrip_monitor).res.b.get_funding_txo(); + System.gc(); System.runFinalization(); // Give the GC a chance to run. + return funding_txo; + } - Peer(byte seed, boolean use_km_wrapper, boolean use_manual_watch) { + private Peer(Object _dummy, byte seed) { logger = Logger.new_impl((String arg) -> System.out.println(seed + ": " + arg)); fee_estimator = FeeEstimator.new_impl((confirmation_target -> 253)); tx_broadcaster = BroadcasterInterface.new_impl(tx -> { broadcast_set.add(tx); }); - this.monitors = new HashMap<>(); + monitors = new HashMap<>(); + this.seed = seed; Persist persister = Persist.new_impl(new Persist.PersistInterface() { @Override public Result_NoneChannelMonitorUpdateErrZ persist_new_channel(OutPoint id, ChannelMonitor data) { - return new Result_NoneChannelMonitorUpdateErrZ.Result_NoneChannelMonitorUpdateErrZ_OK(); + synchronized (monitors) { + String key = Arrays.toString(id.to_channel_id()); + assert monitors.put(key, data) == null; + TwoTuple res = test_mon_roundtrip(data); + assert Arrays.equals(res.a.get_txid(), id.get_txid()); + assert res.a.get_index() == id.get_index(); + } + return Result_NoneChannelMonitorUpdateErrZ.ok(); } @Override public Result_NoneChannelMonitorUpdateErrZ update_persisted_channel(OutPoint id, ChannelMonitorUpdate update, ChannelMonitor data) { - return new Result_NoneChannelMonitorUpdateErrZ.Result_NoneChannelMonitorUpdateErrZ_OK(); + synchronized (monitors) { + String key = Arrays.toString(id.to_channel_id()); + assert monitors.put(key, data) != null; + TwoTuple res = test_mon_roundtrip(data); + assert Arrays.equals(res.a.get_txid(), id.get_txid()); + assert res.a.get_index() == id.get_index(); + } + return Result_NoneChannelMonitorUpdateErrZ.ok(); } }); - Watch chain_watch; + + filter_additions = new HashSet<>(); + if (use_filter) { + this.filter = Filter.new_impl(new Filter.FilterInterface() { + @Override public void register_tx(byte[] txid, byte[] script_pubkey) { + filter_additions.add(Arrays.toString(txid)); + } + @Override public Option_C2Tuple_usizeTransactionZZ register_output(WatchedOutput output) { + filter_additions.add(Arrays.toString(output.get_outpoint().get_txid()) + ":" + output.get_outpoint().get_index()); + return Option_C2Tuple_usizeTransactionZZ.none(); + } + }); + } else { + this.filter = null; + } + if (use_manual_watch) { chain_watch = get_manual_watch(); chain_monitor = null; } else { - chain_monitor = ChainMonitor.constructor_new(null, tx_broadcaster, logger, fee_estimator, persister); + chain_monitor = ChainMonitor.of(filter, tx_broadcaster, logger, fee_estimator, persister); chain_watch = chain_monitor.as_Watch(); } @@ -181,39 +259,165 @@ class HumanObjectPeerTestInstance { for (byte i = 0; i < 32; i++) { key_seed[i] = (byte) (i ^ seed); } + KeysManager keys = KeysManager.of(key_seed, System.currentTimeMillis() / 1000, (int) (System.currentTimeMillis() * 1000)); if (use_km_wrapper) { - KeysManager underlying = KeysManager.constructor_new(key_seed, LDKNetwork.LDKNetwork_Bitcoin, System.currentTimeMillis() / 1000, (int) (System.currentTimeMillis() * 1000) & 0xffffffff); - this.keys_interface = manual_keysif(underlying.as_KeysInterface()); + this.keys_interface = manual_keysif(keys.as_KeysInterface()); + this.explicit_keys_manager = null; } else { - KeysManager keys = KeysManager.constructor_new(key_seed, LDKNetwork.LDKNetwork_Bitcoin, System.currentTimeMillis() / 1000, (int) (System.currentTimeMillis() * 1000) & 0xffffffff); this.keys_interface = keys.as_KeysInterface(); + this.explicit_keys_manager = keys; + } + this.router = NetGraphMsgHandler.of(new byte[32], null, logger); + } + private void bind_nio() { + if (!use_nio_peer_handler) return; + try { this.nio_peer_handler = new NioPeerHandler(peer_manager); } catch (IOException e) { assert false; } + for (short i = 10_000; true; i++) { + try { + nio_peer_handler.bind_listener(new InetSocketAddress("127.0.0.1", i)); + nio_port = i; + break; + } catch (IOException e) { assert i < 10_500; } + } + } + Peer(byte seed) { + this(null, seed); + if (use_chan_manager_constructor) { + try { + this.constructor = new ChannelManagerConstructor(LDKNetwork.LDKNetwork_Bitcoin, UserConfig.with_default(), new byte[32], 0, + this.keys_interface, this.fee_estimator, this.chain_monitor, this.tx_broadcaster, this.logger); + constructor.chain_sync_completed(new ChannelManagerConstructor.ChannelManagerPersister() { + @Override public void handle_events(Event[] events) { + synchronized (pending_manager_events) { + pending_manager_events.addAll(Arrays.asList(events)); + pending_manager_events.notifyAll(); + } + } + @Override public void persist_manager(byte[] channel_manager_bytes) { } + }); + this.chan_manager = constructor.channel_manager; + must_free_objs.add(new WeakReference<>(this.chan_manager)); + } catch (ChannelManagerConstructor.InvalidSerializedDataException e) { + assert false; + } + } else { + this.chan_manager = ChannelManager.of(this.fee_estimator, chain_watch, tx_broadcaster, logger, this.keys_interface, UserConfig.with_default(), LDKNetwork.LDKNetwork_Bitcoin, BestBlock.of(new byte[32], 0)); } - this.chan_manager = ChannelManager.constructor_new(LDKNetwork.LDKNetwork_Bitcoin, FeeEstimator.new_impl(confirmation_target -> 0), chain_watch, tx_broadcaster, logger, this.keys_interface, UserConfig.constructor_default(), 1); + this.node_id = chan_manager.get_our_node_id(); - this.chan_manager_events = chan_manager.as_EventsProvider(); - this.router = NetGraphMsgHandler.constructor_new(new byte[32], null, logger); byte[] random_data = new byte[32]; for (byte i = 0; i < 32; i++) { random_data[i] = (byte) ((i ^ seed) ^ 0xf0); } - this.peer_manager = PeerManager.constructor_new(chan_manager.as_ChannelMessageHandler(), router.as_RoutingMessageHandler(), keys_interface.get_node_secret(), random_data, logger); + this.peer_manager = PeerManager.of(chan_manager.as_ChannelMessageHandler(), router.as_RoutingMessageHandler(), keys_interface.get_node_secret(), random_data, logger); + bind_nio(); System.gc(); } + byte[] hexStringToByteArray(String s) { + int len = s.length(); + byte[] data = new byte[len / 2]; + for (int i = 0; i < len; i += 2) { + data[i / 2] = (byte) ((Character.digit(s.charAt(i), 16) << 4) + + Character.digit(s.charAt(i+1), 16)); + } + return data; + } + + + Object ptr_to; + Peer(Peer orig) { + this(null, orig.seed); + if (use_chan_manager_constructor) { + byte[][] monitors = {orig.monitors.values().stream().iterator().next().write()}; + byte[] serialized = orig.chan_manager.write(); + try { + this.constructor = new ChannelManagerConstructor(serialized, monitors, this.keys_interface, + this.fee_estimator, this.chain_monitor, this.filter, this.tx_broadcaster, this.logger); + constructor.chain_sync_completed(new ChannelManagerConstructor.ChannelManagerPersister() { + @Override public void handle_events(Event[] events) { + synchronized (pending_manager_events) { + pending_manager_events.addAll(Arrays.asList(events)); + pending_manager_events.notifyAll(); + } + } + @Override public void persist_manager(byte[] channel_manager_bytes) { } + }); + this.chan_manager = constructor.channel_manager; + must_free_objs.add(new WeakReference<>(this.chan_manager)); + // If we are using a ChannelManagerConstructor, we may have pending events waiting on the old peer + // which have been removed from the ChannelManager but which we still need to handle. + this.pending_manager_events.addAll(orig.pending_manager_events); + if (!this.pending_manager_events.isEmpty()) { + // However, this implies cross_reload_ref_pollution + cross_reload_ref_pollution = true; + } + } catch (ChannelManagerConstructor.InvalidSerializedDataException e) { + assert false; + } + } else { + ChannelMonitor[] monitors = new ChannelMonitor[1]; + assert orig.monitors.size() == 1; + if (!break_cross_peer_refs) { + monitors[0] = orig.monitors.values().stream().iterator().next(); + } else { + byte[] serialized = orig.monitors.values().stream().iterator().next().write(); + Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ res = + UtilMethods.BlockHashChannelMonitorZ_read(serialized, this.keys_interface); + assert res instanceof Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ.Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ_OK; + monitors[0] = ((Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ.Result_C2Tuple_BlockHashChannelMonitorZDecodeErrorZ_OK) res).res.b; + } + byte[] serialized = orig.chan_manager.write(); + Result_C2Tuple_BlockHashChannelManagerZDecodeErrorZ read_res = + UtilMethods.BlockHashChannelManagerZ_read(serialized, this.keys_interface, this.fee_estimator, this.chain_watch, this.tx_broadcaster, this.logger, UserConfig.with_default(), monitors); + assert read_res instanceof Result_C2Tuple_BlockHashChannelManagerZDecodeErrorZ.Result_C2Tuple_BlockHashChannelManagerZDecodeErrorZ_OK; + this.chan_manager = ((Result_C2Tuple_BlockHashChannelManagerZDecodeErrorZ.Result_C2Tuple_BlockHashChannelManagerZDecodeErrorZ_OK) read_res).res.b; + this.chain_watch.watch_channel(monitors[0].get_funding_txo().a, monitors[0]); + if (!break_cross_peer_refs && (use_manual_watch || use_km_wrapper)) { + // When we pass monitors[0] into chain_watch.watch_channel we create a reference from the new Peer to a + // field in the old peer, preventing freeing of the original Peer until the new Peer is freed. Thus, we + // shouldn't bother waiting for the original to be freed later on. + cross_reload_ref_pollution = true; + } + } + this.node_id = chan_manager.get_our_node_id(); + + if (cross_reload_ref_pollution) { + // This really, really needs to be handled at the bindings layer, but its rather complicated - + // ChannelSigners can be cloned and passed around without java being involved, resulting in them being + // owned by both one or more ChannelMonitors and a ChannelManager, with only one having proper pointers + // to the ChannelSigner. Ideally, the ChannelSigner would have a global reference to the Java + // implementation class, but that results in circular references. Instead, we need some ability to, + // while cloning ChannelSigners, add new references in the calling Java struct (ie ChannelMonitor) to + // the ChannelSigner. + this.ptr_to = orig.chan_manager; + } + + byte[] random_data = new byte[32]; + for (byte i = 0; i < 32; i++) { + random_data[i] = (byte) ((i ^ seed) ^ 0xf0); + } + this.peer_manager = PeerManager.of(chan_manager.as_ChannelMessageHandler(), router.as_RoutingMessageHandler(), keys_interface.get_node_secret(), random_data, logger); + bind_nio(); + } + TwoTuple[]>[] connect_block(Block b, int height, long expected_monitor_update_len) { byte[] header = Arrays.copyOfRange(b.bitcoinSerialize(), 0, 80); TwoTuple[] txn; if (b.hasTransactions()) { assert b.getTransactions().size() == 1; - TwoTuple txp = new TwoTuple<>((long) 1, b.getTransactions().get(0).bitcoinSerialize()); + TwoTuple txp = new TwoTuple<>((long) 0, b.getTransactions().get(0).bitcoinSerialize()); txn = new TwoTuple[]{txp}; } else txn = new TwoTuple[0]; - chan_manager.block_connected(header, txn, height); if (chain_monitor != null) { - chain_monitor.block_connected(header, txn, height); + chan_manager.as_Listen().block_connected(b.bitcoinSerialize(), height); + chain_monitor.as_Listen().block_connected(b.bitcoinSerialize(), height); } else { + chan_manager.as_Confirm().transactions_confirmed(header, txn, height); + chan_manager.as_Confirm().best_block_updated(header, height); + // Connect manually if we aren't using a ChainMonitor and are implementing Watch ourselves synchronized (monitors) { assert monitors.size() == 1; for (ChannelMonitor mon : monitors.values()) { @@ -226,15 +430,67 @@ class HumanObjectPeerTestInstance { return null; } + Event[] get_monitor_events(int expected_len) { + if (use_chan_manager_constructor) { + while (true) { + synchronized (this.pending_manager_events) { + if (expected_len != 0 && this.pending_manager_events.size() != expected_len) { + break; + } + } + try { Thread.sleep(500); } catch (InterruptedException e) { assert false; } + break; + } + synchronized (this.pending_manager_events) { + Event[] res = this.pending_manager_events.toArray(new Event[0]); + this.pending_manager_events.clear(); + assert res.length == expected_len; + return res; + } + } else if (chain_monitor != null) { + return chain_monitor.as_EventsProvider().get_and_clear_pending_events(); + } else { + synchronized (monitors) { + assert monitors.size() == 1; + for (ChannelMonitor mon : monitors.values()) { + Event[] res = mon.get_and_clear_pending_events(); + assert res.length == expected_len; + return res; + } + return null; + } + } + } + + Event[] get_manager_events(int expected_len) { + Event[] res = new Event[0]; + if (use_chan_manager_constructor) { + while (res.length < expected_len) { + synchronized (this.pending_manager_events) { + res = this.pending_manager_events.toArray(res); + assert res.length == expected_len || res.length == 0; // We don't handle partial results + this.pending_manager_events.clear(); + if (res.length < expected_len) { + try { this.pending_manager_events.wait(); } catch (InterruptedException e) { assert false; } + } + } + } + } else { + res = this.chan_manager.as_EventsProvider().get_and_clear_pending_events(); + } + assert res.length == expected_len; + return res; + } + Route get_route(byte[] dest_node, ChannelDetails[] our_chans) { try (LockedNetworkGraph netgraph = this.router.read_locked_graph()) { NetworkGraph graph = netgraph.graph(); - long res = bindings.get_route(this.node_id, graph._test_only_get_ptr(), dest_node, new long[]{our_chans[0]._test_only_get_ptr()}, - new long[0], 1000, 42, this.logger._test_only_get_ptr()); + long res = bindings.get_route(this.node_id, graph._test_only_get_ptr(), dest_node, 0L, new long[]{our_chans[0]._test_only_get_ptr()}, + new long[0], 1000000, 42, this.logger._test_only_get_ptr()); assert bindings.LDKCResult_RouteLightningErrorZ_result_ok(res); byte[] serialized_route = bindings.Route_write(bindings.LDKCResult_RouteLightningErrorZ_get_ok(res)); must_free_objs.add(new WeakReference<>(serialized_route)); - Result_RouteDecodeErrorZ copy = Route.constructor_read(serialized_route); + Result_RouteDecodeErrorZ copy = Route.read(serialized_route); assert copy instanceof Result_RouteDecodeErrorZ.Result_RouteDecodeErrorZ_OK; bindings.CResult_RouteLightningErrorZ_free(res); return ((Result_RouteDecodeErrorZ.Result_RouteDecodeErrorZ_OK) copy).res; @@ -242,10 +498,11 @@ class HumanObjectPeerTestInstance { } } - class DescriptorHolder { SocketDescriptor val; } + static class DescriptorHolder { SocketDescriptor val; } boolean running = false; final LinkedList runqueue = new LinkedList(); + boolean ran = false; Thread t = new Thread(() -> { while (true) { try { @@ -267,17 +524,34 @@ class HumanObjectPeerTestInstance { } } }); - void wait_events_processed() { - while (true) { + void wait_events_processed(Peer peer1, Peer peer2) { + if (use_nio_peer_handler) { + peer1.nio_peer_handler.check_events(); + peer2.nio_peer_handler.check_events(); + try { Thread.sleep(400); } catch (InterruptedException e) { assert false; } + } else { synchronized (runqueue) { - if (runqueue.isEmpty() && !running) break; - try { runqueue.wait(); } catch (InterruptedException e) { assert false; } + ran = false; + } + while (true) { + peer1.peer_manager.process_events(); + peer2.peer_manager.process_events(); + synchronized (runqueue) { + if (runqueue.isEmpty() && !running) { + if (ran) { + ran = false; + continue; + } else { break; } + } + try { runqueue.wait(); } catch (InterruptedException e) { assert false; } + } } } } void do_read_event(PeerManager pm, SocketDescriptor descriptor, byte[] data) { if (!t.isAlive()) t.start(); synchronized (runqueue) { + ran = true; runqueue.add(() -> { Result_boolPeerHandleErrorZ res = pm.read_event(descriptor, data); assert res instanceof Result_boolPeerHandleErrorZ.Result_boolPeerHandleErrorZ_OK; @@ -287,56 +561,59 @@ class HumanObjectPeerTestInstance { must_free_objs.add(new WeakReference<>(data)); } - void do_test_message_handler(boolean nice_close, boolean use_km_wrapper, boolean use_manual_watch) throws InterruptedException { - GcCheck obj = new GcCheck(); - Peer peer1 = new Peer((byte) 1, use_km_wrapper, use_manual_watch); - Peer peer2 = new Peer((byte) 2, use_km_wrapper, use_manual_watch); - - DescriptorHolder descriptor1 = new DescriptorHolder(); - DescriptorHolder descriptor1ref = descriptor1; - SocketDescriptor descriptor2 = SocketDescriptor.new_impl(new SocketDescriptor.SocketDescriptorInterface() { - @Override - public long send_data(byte[] data, boolean resume_read) { - do_read_event(peer1.peer_manager, descriptor1ref.val, data); - return data.length; - } + void connect_peers(final Peer peer1, final Peer peer2) { + if (use_nio_peer_handler) { + try { + peer1.nio_peer_handler.connect(peer2.chan_manager.get_our_node_id(), new InetSocketAddress("127.0.0.1", peer2.nio_port), 100); + } catch (IOException e) { assert false; } + } else { + DescriptorHolder descriptor1 = new DescriptorHolder(); + DescriptorHolder descriptor1ref = descriptor1; + SocketDescriptor descriptor2 = SocketDescriptor.new_impl(new SocketDescriptor.SocketDescriptorInterface() { + @Override + public long send_data(byte[] data, boolean resume_read) { + do_read_event(peer1.peer_manager, descriptor1ref.val, data); + return data.length; + } - @Override public void disconnect_socket() { assert false; } - @Override public boolean eq(SocketDescriptor other_arg) { return other_arg.hash() == 2; } - @Override public long hash() { return 2; } - }); + @Override public void disconnect_socket() { assert false; } + @Override public boolean eq(SocketDescriptor other_arg) { return other_arg.hash() == 2; } + @Override public long hash() { return 2; } + }); - descriptor1.val = SocketDescriptor.new_impl(new SocketDescriptor.SocketDescriptorInterface() { - @Override - public long send_data(byte[] data, boolean resume_read) { - do_read_event(peer2.peer_manager, descriptor2, data); - return data.length; - } + descriptor1.val = SocketDescriptor.new_impl(new SocketDescriptor.SocketDescriptorInterface() { + @Override + public long send_data(byte[] data, boolean resume_read) { + do_read_event(peer2.peer_manager, descriptor2, data); + return data.length; + } - @Override public void disconnect_socket() { assert false; } - @Override public boolean eq(SocketDescriptor other_arg) { return other_arg.hash() == 1; } - @Override public long hash() { return 1; } - }); + @Override public void disconnect_socket() { assert false; } + @Override public boolean eq(SocketDescriptor other_arg) { return other_arg.hash() == 1; } + @Override public long hash() { return 1; } + }); + + Result_CVec_u8ZPeerHandleErrorZ conn_res = peer1.peer_manager.new_outbound_connection(peer2.node_id, descriptor1.val); + assert conn_res instanceof Result_CVec_u8ZPeerHandleErrorZ.Result_CVec_u8ZPeerHandleErrorZ_OK; - Result_CVec_u8ZPeerHandleErrorZ conn_res = peer1.peer_manager.new_outbound_connection(peer2.node_id, descriptor1.val); - assert conn_res instanceof Result_CVec_u8ZPeerHandleErrorZ.Result_CVec_u8ZPeerHandleErrorZ_OK; + Result_NonePeerHandleErrorZ inbound_conn_res = peer2.peer_manager.new_inbound_connection(descriptor2); + assert inbound_conn_res instanceof Result_NonePeerHandleErrorZ.Result_NonePeerHandleErrorZ_OK; + do_read_event(peer2.peer_manager, descriptor2, ((Result_CVec_u8ZPeerHandleErrorZ.Result_CVec_u8ZPeerHandleErrorZ_OK) conn_res).res); + } + } - Result_NonePeerHandleErrorZ inbound_conn_res = peer2.peer_manager.new_inbound_connection(descriptor2); - assert inbound_conn_res instanceof Result_NonePeerHandleErrorZ.Result_NonePeerHandleErrorZ_OK; - do_read_event(peer2.peer_manager, descriptor2, ((Result_CVec_u8ZPeerHandleErrorZ.Result_CVec_u8ZPeerHandleErrorZ_OK) conn_res).res); + TestState do_test_message_handler() throws InterruptedException { + Peer peer1 = new Peer((byte) 1); + Peer peer2 = new Peer((byte) 2); - wait_events_processed(); + connect_peers(peer1, peer2); + wait_events_processed(peer1, peer2); Result_NoneAPIErrorZ cc_res = peer1.chan_manager.create_channel(peer2.node_id, 10000, 1000, 42, null); assert cc_res instanceof Result_NoneAPIErrorZ.Result_NoneAPIErrorZ_OK; + wait_events_processed(peer1, peer2); - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - - Event[] events = peer1.chan_manager_events.get_and_clear_pending_events(); - assert events.length == 1; + Event[] events = peer1.get_manager_events(1); assert events[0] instanceof Event.FundingGenerationReady; assert ((Event.FundingGenerationReady) events[0]).channel_value_satoshis == 10000; assert ((Event.FundingGenerationReady) events[0]).user_channel_id == 42; @@ -351,17 +628,13 @@ class HumanObjectPeerTestInstance { funding.getInputs().get(0).setWitness(new TransactionWitness(2)); // Make sure we don't complain about lack of witness funding.getInput(0).getWitness().setPush(0, new byte[]{0x1}); funding.addOutput(Coin.SATOSHI.multiply(10000), new Script(funding_spk)); - peer1.chan_manager.funding_transaction_generated(chan_id, OutPoint.constructor_new(funding.getTxId().getReversedBytes(), (short) 0)); - - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); + Result_NoneAPIErrorZ funding_res = peer1.chan_manager.funding_transaction_generated(chan_id, funding.bitcoinSerialize()); + assert funding_res instanceof Result_NoneAPIErrorZ.Result_NoneAPIErrorZ_OK; + wait_events_processed(peer1, peer2); - events = peer1.chan_manager_events.get_and_clear_pending_events(); - assert events.length == 1; - assert events[0] instanceof Event.FundingBroadcastSafe; - assert ((Event.FundingBroadcastSafe) events[0]).user_channel_id == 42; + assert peer1.broadcast_set.size() == 1; + assert Arrays.equals(peer1.broadcast_set.get(0), funding.bitcoinSerialize()); + peer1.broadcast_set.clear(); Block b = new Block(bitcoinj_net, 2, Sha256Hash.ZERO_HASH, Sha256Hash.ZERO_HASH, 42, 0, 0, Arrays.asList(new Transaction[]{funding})); peer1.connect_block(b, 1, 0); @@ -372,124 +645,207 @@ class HumanObjectPeerTestInstance { peer1.connect_block(b, height, 0); peer2.connect_block(b, height, 0); } - - peer1.peer_manager.process_events(); - peer2.peer_manager.process_events(); - wait_events_processed(); + wait_events_processed(peer1, peer2); peer1.chan_manager.list_channels(); - ChannelDetails[] peer1_chans = peer1.chan_manager.list_channels(); - ChannelDetails[] peer2_chans = peer2.chan_manager.list_channels(); + ChannelDetails[] peer1_chans = peer1.chan_manager.list_usable_channels(); + ChannelDetails[] peer2_chans = peer2.chan_manager.list_usable_channels(); assert peer1_chans.length == 1; assert peer2_chans.length == 1; assert peer1_chans[0].get_channel_value_satoshis() == 10000; - assert peer1_chans[0].get_is_live(); + assert peer1_chans[0].get_is_usable(); + Option_u64Z short_chan_id = peer1_chans[0].get_short_channel_id(); + assert short_chan_id instanceof Option_u64Z.Some; + assert ((Option_u64Z.Some)short_chan_id).some == (1L << 40); // 0th output in the 0th transaction in the 1st block assert Arrays.equals(peer1_chans[0].get_channel_id(), funding.getTxId().getReversedBytes()); assert Arrays.equals(peer2_chans[0].get_channel_id(), funding.getTxId().getReversedBytes()); - byte[] payment_preimage = new byte[32]; - for (int i = 0; i < 32; i++) payment_preimage[i] = (byte) (i ^ 0x0f); - byte[] payment_hash = Sha256Hash.hash(payment_preimage); + Result_InvoiceSignOrCreationErrorZ invoice = UtilMethods.invoice_from_channelmanager(peer2.chan_manager, peer2.keys_interface, LDKCurrency.LDKCurrency_Bitcoin, Option_u64Z.none(), "Invoice Description"); + assert invoice instanceof Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK; + System.out.println("Got invoice: " + ((Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK) invoice).res.to_str()); + Result_InvoiceNoneZ parsed_invoice = Invoice.from_str(((Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK) invoice).res.to_str()); + assert parsed_invoice instanceof Result_InvoiceNoneZ.Result_InvoiceNoneZ_OK; + assert Arrays.equals(((Result_InvoiceNoneZ.Result_InvoiceNoneZ_OK) parsed_invoice).res.payment_hash(), ((Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK) invoice).res.payment_hash()); + SignedRawInvoice signed_raw = ((Result_InvoiceNoneZ.Result_InvoiceNoneZ_OK) parsed_invoice).res.into_signed_raw(); + RawInvoice raw_invoice = signed_raw.raw_invoice(); + byte[] desc_hash = raw_invoice.hash(); + Description raw_invoice_description = raw_invoice.description(); + String description_string = raw_invoice_description.into_inner(); + assert description_string.equals("Invoice Description"); + byte[] payment_hash = ((Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK) invoice).res.payment_hash(); + byte[] payment_secret = ((Result_InvoiceSignOrCreationErrorZ.Result_InvoiceSignOrCreationErrorZ_OK) invoice).res.payment_secret(); + Route route = peer1.get_route(peer2.node_id, peer1_chans); - Result_NonePaymentSendFailureZ payment_res = peer1.chan_manager.send_payment(route, payment_hash, new byte[32]); + Result_NonePaymentSendFailureZ payment_res = peer1.chan_manager.send_payment(route, payment_hash, payment_secret); assert payment_res instanceof Result_NonePaymentSendFailureZ.Result_NonePaymentSendFailureZ_OK; + wait_events_processed(peer1, peer2); + + RouteHop[][] hops = new RouteHop[1][1]; + byte[] hop_pubkey = new byte[33]; + hop_pubkey[0] = 3; + hop_pubkey[1] = 42; + hops[0][0] = RouteHop.of(hop_pubkey, NodeFeatures.known(), 42, ChannelFeatures.known(), 100, 0); + Route r2 = Route.of(hops); + payment_res = peer1.chan_manager.send_payment(r2, payment_hash, payment_secret); + assert payment_res instanceof Result_NonePaymentSendFailureZ.Result_NonePaymentSendFailureZ_Err; + + if (!use_chan_manager_constructor) { + peer1.get_monitor_events(0); + peer2.get_monitor_events(0); + } else { + // The events are combined across manager + monitors but peer1 still has no events + } - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - peer1.peer_manager.process_events(); - wait_events_processed(); + if (reload_peers) { + if (use_nio_peer_handler) { + peer1.nio_peer_handler.interrupt(); + peer2.nio_peer_handler.interrupt(); + } + if (use_chan_manager_constructor) { + peer1.constructor.interrupt(); + peer2.constructor.interrupt(); + } + WeakReference op1 = new WeakReference(peer1); + peer1 = new Peer(peer1); + peer2 = new Peer(peer2); + return new TestState(op1, peer1, peer2, b.getHash()); + } + return new TestState(null, peer1, peer2, b.getHash()); + } - { - peer1.chan_manager.write(); - //ChannelManager. + boolean cross_reload_ref_pollution = false; + class TestState { + private final WeakReference ref_block; + private final Peer peer1; + private final Peer peer2; + public Sha256Hash best_blockhash; + + public TestState(WeakReference ref_block, Peer peer1, Peer peer2, Sha256Hash best_blockhash) { + this.ref_block = ref_block; + this.peer1 = peer1; + this.peer2 = peer2; + this.best_blockhash = best_blockhash; + } + } + void do_test_message_handler_b(TestState state) { + GcCheck obj = new GcCheck(); + if (state.ref_block != null) { + // Ensure the original peers get freed before we move on. Note that we have to be in a different function + // scope to do so as the (at least current OpenJDK) JRE won't release anything created in the same scope. + while (!cross_reload_ref_pollution && state.ref_block.get() != null) { + System.gc(); + System.runFinalization(); + } + connect_peers(state.peer1, state.peer2); } + wait_events_processed(state.peer1, state.peer2); - events = peer2.chan_manager_events.get_and_clear_pending_events(); - assert events.length == 1; + Event[] events = state.peer2.get_manager_events(1); assert events[0] instanceof Event.PendingHTLCsForwardable; - peer2.chan_manager.process_pending_htlc_forwards(); + state.peer2.chan_manager.process_pending_htlc_forwards(); - events = peer2.chan_manager_events.get_and_clear_pending_events(); - assert events.length == 1; + events = state.peer2.get_manager_events(1); assert events[0] instanceof Event.PaymentReceived; - peer2.chan_manager.claim_funds(payment_preimage, new byte[32], ((Event.PaymentReceived) events[0]).amt); + byte[] payment_preimage = ((Event.PaymentReceived)events[0]).payment_preimage; + assert !Arrays.equals(payment_preimage, new byte[32]); + state.peer2.chan_manager.claim_funds(payment_preimage); + wait_events_processed(state.peer1, state.peer2); - peer2.peer_manager.process_events(); - wait_events_processed(); - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - - events = peer1.chan_manager_events.get_and_clear_pending_events(); - assert events.length == 1; + events = state.peer1.get_manager_events(1); assert events[0] instanceof Event.PaymentSent; assert Arrays.equals(((Event.PaymentSent) events[0]).payment_preimage, payment_preimage); + wait_events_processed(state.peer1, state.peer2); + + ChannelDetails[] peer1_chans = state.peer1.chan_manager.list_channels(); if (nice_close) { - Result_NoneAPIErrorZ close_res = peer1.chan_manager.close_channel(peer1_chans[0].get_channel_id()); + Result_NoneAPIErrorZ close_res = state.peer1.chan_manager.close_channel(peer1_chans[0].get_channel_id()); assert close_res instanceof Result_NoneAPIErrorZ.Result_NoneAPIErrorZ_OK; + wait_events_processed(state.peer1, state.peer2); - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - - assert peer1.broadcast_set.size() == 1; - assert peer2.broadcast_set.size() == 1; + assert state.peer1.broadcast_set.size() == 1; + assert state.peer2.broadcast_set.size() == 1; } else { - peer1.chan_manager.force_close_all_channels(); - - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - peer1.peer_manager.process_events(); - wait_events_processed(); - peer2.peer_manager.process_events(); - wait_events_processed(); - - assert peer1.broadcast_set.size() == 1; - assert peer2.broadcast_set.size() == 0; - - Transaction tx = new Transaction(bitcoinj_net, peer1.broadcast_set.getFirst()); - b = new Block(bitcoinj_net, 2, b.getHash(), Sha256Hash.ZERO_HASH, 42, 0, 0, + state.peer1.chan_manager.force_close_all_channels(); + wait_events_processed(state.peer1, state.peer2); + + assert state.peer1.broadcast_set.size() == 1; + assert state.peer2.broadcast_set.size() == 1; + + NetworkParameters bitcoinj_net = NetworkParameters.fromID(NetworkParameters.ID_MAINNET); + Transaction tx = new Transaction(bitcoinj_net, state.peer1.broadcast_set.getFirst()); + Block b = new Block(bitcoinj_net, 2, state.best_blockhash, Sha256Hash.ZERO_HASH, 42, 0, 0, Arrays.asList(new Transaction[]{tx})); - TwoTuple[]>[] watch_outputs = peer2.connect_block(b, 1, 1); + TwoTuple[]>[] watch_outputs = state.peer2.connect_block(b, 10, 1); if (watch_outputs != null) { // We only process watch_outputs manually when we use a manually-build Watch impl assert watch_outputs.length == 1; assert Arrays.equals(watch_outputs[0].a, tx.getTxId().getReversedBytes()); - assert watch_outputs[0].b.length == 1; + assert watch_outputs[0].b.length == 2; + assert watch_outputs[0].b[0].a == 0; + assert watch_outputs[0].b[1].a == 1; + } + + for (int i = 11; i < 21; i++) { + b = new Block(bitcoinj_net, 2, b.getHash(), Sha256Hash.ZERO_HASH, 42, 0, 0, new ArrayList<>()); + state.peer2.connect_block(b, i, 0); } - // This used to be buggy and double-free, so go ahead and fetch them! - for (ChannelMonitor mon : peer2.monitors.values()) { - byte[][] txn = mon.get_latest_holder_commitment_txn(peer2.logger); + Event[] broadcastable_event = state.peer2.get_monitor_events(1); + for (ChannelMonitor mon : state.peer2.monitors.values()) { + // This used to be buggy and double-free, so go ahead and fetch them! + byte[][] txn = mon.get_latest_holder_commitment_txn(state.peer2.logger); } + assert broadcastable_event.length == 1; + assert broadcastable_event[0] instanceof Event.SpendableOutputs; + if (state.peer2.explicit_keys_manager != null) { + Result_TransactionNoneZ tx_res = state.peer2.explicit_keys_manager.spend_spendable_outputs(((Event.SpendableOutputs) broadcastable_event[0]).outputs, new TxOut[0], new byte[] {0x00}, 253); + assert tx_res instanceof Result_TransactionNoneZ.Result_TransactionNoneZ_OK; + Transaction built_tx = new Transaction(bitcoinj_net, ((Result_TransactionNoneZ.Result_TransactionNoneZ_OK) tx_res).res); + assert built_tx.getOutputs().size() == 1; + assert Arrays.equals(built_tx.getOutput(0).getScriptBytes(), new byte[]{0x00}); + } + } + + if (use_nio_peer_handler) { + state.peer1.peer_manager.disconnect_by_node_id(state.peer2.chan_manager.get_our_node_id(), false); + wait_events_processed(state.peer1, state.peer2); + assert state.peer1.peer_manager.get_peer_node_ids().length == 0; + assert state.peer2.peer_manager.get_peer_node_ids().length == 0; + state.peer1.nio_peer_handler.interrupt(); + state.peer2.nio_peer_handler.interrupt(); + } + + state.peer1.get_monitor_events(0); + state.peer2.get_monitor_events(0); + + if (use_chan_manager_constructor) { + state.peer1.constructor.interrupt(); + state.peer2.constructor.interrupt(); } } java.util.LinkedList> must_free_objs = new java.util.LinkedList(); - boolean gc_ran = false; + int gc_count = 0; + int gc_exp_count = 0; class GcCheck { + GcCheck() { gc_exp_count += 1; } @Override protected void finalize() throws Throwable { - gc_ran = true; + gc_count += 1; super.finalize(); } } } public class HumanObjectPeerTest { - void do_test(boolean nice_close, boolean use_km_wrapper, boolean use_manual_watch) throws InterruptedException { - HumanObjectPeerTestInstance instance = new HumanObjectPeerTestInstance(); - instance.do_test_message_handler(nice_close, use_km_wrapper, use_manual_watch); - while (!instance.gc_ran) { + HumanObjectPeerTestInstance do_test_run(boolean nice_close, boolean use_km_wrapper, boolean use_manual_watch, boolean reload_peers, boolean break_cross_peer_refs, boolean nio_peer_handler, boolean use_chan_manager_constructor) throws InterruptedException { + HumanObjectPeerTestInstance instance = new HumanObjectPeerTestInstance(nice_close, use_km_wrapper, use_manual_watch, reload_peers, break_cross_peer_refs, nio_peer_handler, !nio_peer_handler, use_chan_manager_constructor); + HumanObjectPeerTestInstance.TestState state = instance.do_test_message_handler(); + instance.do_test_message_handler_b(state); + return instance; + } + void do_test(boolean nice_close, boolean use_km_wrapper, boolean use_manual_watch, boolean reload_peers, boolean break_cross_peer_refs, boolean nio_peer_handler, boolean use_chan_manager_constructor) throws InterruptedException { + HumanObjectPeerTestInstance instance = do_test_run(nice_close, use_km_wrapper, use_manual_watch, reload_peers, break_cross_peer_refs, nio_peer_handler, use_chan_manager_constructor); + while (instance.gc_count != instance.gc_exp_count) { System.gc(); System.runFinalization(); } @@ -497,19 +853,25 @@ public class HumanObjectPeerTest { assert o.get() == null; } @Test - public void test_message_handler_force_close() throws InterruptedException { - do_test(false, false, false); - } - @Test - public void test_message_handler_nice_close() throws InterruptedException { - do_test(true, false, false); - } - @Test - public void test_message_handler_nice_close_wrapper() throws InterruptedException { - do_test(true, true, true); - } - @Test - public void test_message_handler_force_close_wrapper() throws InterruptedException { - do_test(false, true, true); + public void test_message_handler() throws InterruptedException { + for (int i = 0; i < (1 << 7) - 1; i++) { + boolean nice_close = (i & (1 << 0)) != 0; + boolean use_km_wrapper = (i & (1 << 1)) != 0; + boolean use_manual_watch = (i & (1 << 2)) != 0; + boolean reload_peers = (i & (1 << 3)) != 0; + boolean break_cross_refs = (i & (1 << 4)) != 0; + boolean nio_peer_handler = (i & (1 << 5)) != 0; + boolean use_chan_manager_constructor = (i & (1 << 6)) != 0; + if (break_cross_refs && !reload_peers) { + // There are no cross refs to break without reloading peers. + continue; + } + if (use_chan_manager_constructor && use_manual_watch) { + // ChannelManagerConstructor requires a ChainMonitor as the Watch + continue; + } + System.err.println("Running test with flags " + i); + do_test(nice_close, use_km_wrapper, use_manual_watch, reload_peers, break_cross_refs, nio_peer_handler, use_chan_manager_constructor); + } } }