Merge pull request #1159 from jkczyz/2021-11-network-gossip
authorMatt Corallo <649246+TheBlueMatt@users.noreply.github.com>
Thu, 2 Jun 2022 23:53:32 +0000 (16:53 -0700)
committerGitHub <noreply@github.com>
Thu, 2 Jun 2022 23:53:32 +0000 (16:53 -0700)
Rename network-related types

28 files changed:
ARCH.md
fuzz/src/full_stack.rs
fuzz/src/process_network_graph.rs
fuzz/src/router.rs
lightning-background-processor/src/lib.rs
lightning-invoice/src/de.rs
lightning-invoice/src/lib.rs
lightning-invoice/src/payment.rs
lightning-invoice/src/utils.rs
lightning-invoice/tests/ser_de.rs
lightning-rapid-gossip-sync/src/lib.rs
lightning-rapid-gossip-sync/src/processing.rs
lightning/src/ln/chanmon_update_fail_tests.rs
lightning/src/ln/channelmanager.rs
lightning/src/ln/functional_test_utils.rs
lightning/src/ln/functional_tests.rs
lightning/src/ln/mod.rs
lightning/src/ln/onion_route_tests.rs
lightning/src/ln/onion_utils.rs
lightning/src/ln/peer_handler.rs
lightning/src/ln/priv_short_conf_tests.rs
lightning/src/routing/gossip.rs [new file with mode: 0644]
lightning/src/routing/mod.rs
lightning/src/routing/network_graph.rs [deleted file]
lightning/src/routing/router.rs
lightning/src/routing/scoring.rs
lightning/src/util/events.rs
lightning/src/util/persist.rs

diff --git a/ARCH.md b/ARCH.md
index 5b9304cce009f8d9b3eb1a7fcff3721fc3efb094..9efccd9c9ddcec9d9812e0d0eb420975b3a0e98c 100644 (file)
--- a/ARCH.md
+++ b/ARCH.md
@@ -11,10 +11,10 @@ receive `ChannelMonitorUpdate`s from `ChannelManager` and persist them to disk b
 channel steps forward.
 
 There are two additional important structures that you may use either on the same device
-as the `ChannelManager` or on a separate one. `NetGraphMsgHandler` handles receiving channel
-and node announcements, which are then used to calculate routes by `get_route` for sending payments.
-`PeerManager` handles the authenticated and encrypted communication protocol,
-monitoring for liveness of peers, routing messages to `ChannelManager` and `NetGraphMsgHandler`
+as the `ChannelManager` or on a separate one. `P2PGossipSync` handles receiving channel
+and node announcements, which are then used to calculate routes by `find_route` for sending
+payments. `PeerManager` handles the authenticated and encrypted communication protocol,
+monitoring for liveness of peers, routing messages to `ChannelManager` and `P2PGossipSync`
 instances directly, and receiving messages from them via the `EventsProvider` interface.
 
 These structs communicate with each other using a public API, so that you can easily add
@@ -56,7 +56,7 @@ At a high level, some of the common interfaces fit together as follows:
  |                    -----------------          \   /  ----------------
  |                            ^                   \ /          |
 (as RoutingMessageHandler)    |                    v           v
-  \                   ----------------------   ---------   -----------------
-   -----------------> | NetGraphMsgHandler |   | Event |   | chain::Filter |
-                      ----------------------   ---------   -----------------
+  \                   -----------------        ---------   -----------------
+   -----------------> | P2PGossipSync |        | Event |   | chain::Filter |
+                      -----------------        ---------   -----------------
 ```
index 00b123b84dd1942da7f20d1be5fdf10f2cd8a1fd..a59d78e1c54cc35507b2ec4d0f3a66fb0753a01a 100644 (file)
@@ -37,7 +37,7 @@ use lightning::ln::channelmanager::{ChainParameters, ChannelManager};
 use lightning::ln::peer_handler::{MessageHandler,PeerManager,SocketDescriptor,IgnoringMessageHandler};
 use lightning::ln::msgs::DecodeError;
 use lightning::ln::script::ShutdownScript;
-use lightning::routing::network_graph::{NetGraphMsgHandler, NetworkGraph};
+use lightning::routing::gossip::{P2PGossipSync, NetworkGraph};
 use lightning::routing::router::{find_route, PaymentParameters, RouteParameters};
 use lightning::routing::scoring::FixedPenaltyScorer;
 use lightning::util::config::UserConfig;
@@ -163,7 +163,7 @@ type ChannelMan = ChannelManager<
        EnforcingSigner,
        Arc<chainmonitor::ChainMonitor<EnforcingSigner, Arc<dyn chain::Filter>, Arc<TestBroadcaster>, Arc<FuzzEstimator>, Arc<dyn Logger>, Arc<TestPersister>>>,
        Arc<TestBroadcaster>, Arc<KeyProvider>, Arc<FuzzEstimator>, Arc<dyn Logger>>;
-type PeerMan<'a> = PeerManager<Peer<'a>, Arc<ChannelMan>, Arc<NetGraphMsgHandler<Arc<NetworkGraph>, Arc<dyn chain::Access>, Arc<dyn Logger>>>, Arc<dyn Logger>, IgnoringMessageHandler>;
+type PeerMan<'a> = PeerManager<Peer<'a>, Arc<ChannelMan>, Arc<P2PGossipSync<Arc<NetworkGraph>, Arc<dyn chain::Access>, Arc<dyn Logger>>>, Arc<dyn Logger>, IgnoringMessageHandler>;
 
 struct MoneyLossDetector<'a> {
        manager: Arc<ChannelMan>,
@@ -396,13 +396,13 @@ pub fn do_test(data: &[u8], logger: &Arc<dyn Logger>) {
        keys_manager.counter.fetch_sub(1, Ordering::AcqRel);
        let our_id = PublicKey::from_secret_key(&Secp256k1::signing_only(), &keys_manager.get_node_secret(Recipient::Node).unwrap());
        let network_graph = Arc::new(NetworkGraph::new(genesis_block(network).block_hash()));
-       let net_graph_msg_handler = Arc::new(NetGraphMsgHandler::new(Arc::clone(&network_graph), None, Arc::clone(&logger)));
+       let gossip_sync = Arc::new(P2PGossipSync::new(Arc::clone(&network_graph), None, Arc::clone(&logger)));
        let scorer = FixedPenaltyScorer::with_penalty(0);
 
        let peers = RefCell::new([false; 256]);
        let mut loss_detector = MoneyLossDetector::new(&peers, channelmanager.clone(), monitor.clone(), PeerManager::new(MessageHandler {
                chan_handler: channelmanager.clone(),
-               route_handler: net_graph_msg_handler.clone(),
+               route_handler: gossip_sync.clone(),
        }, our_network_key, &[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 15, 0], Arc::clone(&logger), IgnoringMessageHandler{}));
 
        let mut should_forward = false;
index a71ae0e221b294cb302960d34bc4f6e966fa8124..118862569d039b47ea967c79207c463bbbe837f6 100644 (file)
@@ -5,7 +5,7 @@ use utils::test_logger;
 /// Actual fuzz test, method signature and name are fixed
 fn do_test(data: &[u8]) {
        let block_hash = bitcoin::BlockHash::default();
-       let network_graph = lightning::routing::network_graph::NetworkGraph::new(block_hash);
+       let network_graph = lightning::routing::gossip::NetworkGraph::new(block_hash);
        let rapid_sync = RapidGossipSync::new(&network_graph);
        let _ = rapid_sync.update_network_graph(data);
 }
index bb6ba2c6e51cb0a2ac65d86023279c08954532b9..ec12ff7e144e4708670ccc3aea79cd91d1201247 100644 (file)
@@ -16,11 +16,11 @@ use lightning::chain::transaction::OutPoint;
 use lightning::ln::channelmanager::{ChannelDetails, ChannelCounterparty};
 use lightning::ln::features::InitFeatures;
 use lightning::ln::msgs;
+use lightning::routing::gossip::{NetworkGraph, RoutingFees};
 use lightning::routing::router::{find_route, PaymentParameters, RouteHint, RouteHintHop, RouteParameters};
 use lightning::routing::scoring::FixedPenaltyScorer;
 use lightning::util::logger::Logger;
 use lightning::util::ser::Readable;
-use lightning::routing::network_graph::{NetworkGraph, RoutingFees};
 
 use bitcoin::hashes::Hash;
 use bitcoin::secp256k1::PublicKey;
@@ -196,7 +196,7 @@ pub fn do_test<Out: test_logger::Output>(data: &[u8], out: Out) {
                        },
                        4 => {
                                let short_channel_id = slice_to_be64(get_slice!(8));
-                               net_graph.close_channel_from_update(short_channel_id, false);
+                               net_graph.channel_failed(short_channel_id, false);
                        },
                        _ if node_pks.is_empty() => {},
                        _ => {
index 0fab3e61a4af982fb49f14d89865133af29dad45..2e947a7041afdd92c4c8ac86c053c6781de24f26 100644 (file)
@@ -18,7 +18,7 @@ use lightning::chain::keysinterface::{Sign, KeysInterface};
 use lightning::ln::channelmanager::ChannelManager;
 use lightning::ln::msgs::{ChannelMessageHandler, RoutingMessageHandler};
 use lightning::ln::peer_handler::{CustomMessageHandler, PeerManager, SocketDescriptor};
-use lightning::routing::network_graph::{NetworkGraph, NetGraphMsgHandler};
+use lightning::routing::gossip::{NetworkGraph, P2PGossipSync};
 use lightning::routing::scoring::WriteableScore;
 use lightning::util::events::{Event, EventHandler, EventsProvider};
 use lightning::util::logger::Logger;
@@ -40,7 +40,7 @@ use std::ops::Deref;
 ///   [`ChannelManager`] persistence should be done in the background.
 /// * Calling [`ChannelManager::timer_tick_occurred`] and [`PeerManager::timer_tick_occurred`]
 ///   at the appropriate intervals.
-/// * Calling [`NetworkGraph::remove_stale_channels`] (if a [`NetGraphMsgHandler`] is provided to
+/// * Calling [`NetworkGraph::remove_stale_channels`] (if a [`P2PGossipSync`] is provided to
 ///   [`BackgroundProcessor::start`]).
 ///
 /// It will also call [`PeerManager::process_events`] periodically though this shouldn't be relied
@@ -93,26 +93,26 @@ const FIRST_NETWORK_PRUNE_TIMER: u64 = 1;
 /// Decorates an [`EventHandler`] with common functionality provided by standard [`EventHandler`]s.
 struct DecoratingEventHandler<
        E: EventHandler,
-       N: Deref<Target = NetGraphMsgHandler<G, A, L>>,
+       P: Deref<Target = P2PGossipSync<G, A, L>>,
        G: Deref<Target = NetworkGraph>,
        A: Deref,
        L: Deref,
 >
 where A::Target: chain::Access, L::Target: Logger {
        event_handler: E,
-       net_graph_msg_handler: Option<N>,
+       p2p_gossip_sync: Option<P>,
 }
 
 impl<
        E: EventHandler,
-       N: Deref<Target = NetGraphMsgHandler<G, A, L>>,
+       P: Deref<Target = P2PGossipSync<G, A, L>>,
        G: Deref<Target = NetworkGraph>,
        A: Deref,
        L: Deref,
-> EventHandler for DecoratingEventHandler<E, N, G, A, L>
+> EventHandler for DecoratingEventHandler<E, P, G, A, L>
 where A::Target: chain::Access, L::Target: Logger {
        fn handle_event(&self, event: &Event) {
-               if let Some(event_handler) = &self.net_graph_msg_handler {
+               if let Some(event_handler) = &self.p2p_gossip_sync {
                        event_handler.handle_event(event);
                }
                self.event_handler.handle_event(event);
@@ -147,7 +147,7 @@ impl BackgroundProcessor {
        /// `event_handler` is responsible for handling events that users should be notified of (e.g.,
        /// payment failed). [`BackgroundProcessor`] may decorate the given [`EventHandler`] with common
        /// functionality implemented by other handlers.
-       /// * [`NetGraphMsgHandler`] if given will update the [`NetworkGraph`] based on payment failures.
+       /// * [`P2PGossipSync`] if given will update the [`NetworkGraph`] based on payment failures.
        ///
        /// # Rapid Gossip Sync
        ///
@@ -162,8 +162,8 @@ impl BackgroundProcessor {
        /// [`ChannelManager::write`]: lightning::ln::channelmanager::ChannelManager#impl-Writeable
        /// [`Persister::persist_manager`]: lightning::util::persist::Persister::persist_manager
        /// [`Persister::persist_graph`]: lightning::util::persist::Persister::persist_graph
-       /// [`NetworkGraph`]: lightning::routing::network_graph::NetworkGraph
-       /// [`NetworkGraph::write`]: lightning::routing::network_graph::NetworkGraph#impl-Writeable
+       /// [`NetworkGraph`]: lightning::routing::gossip::NetworkGraph
+       /// [`NetworkGraph::write`]: lightning::routing::gossip::NetworkGraph#impl-Writeable
        pub fn start<
                'a,
                Signer: 'static + Sign,
@@ -183,7 +183,7 @@ impl BackgroundProcessor {
                PS: 'static + Deref + Send,
                M: 'static + Deref<Target = ChainMonitor<Signer, CF, T, F, L, P>> + Send + Sync,
                CM: 'static + Deref<Target = ChannelManager<Signer, CW, T, K, F, L>> + Send + Sync,
-               NG: 'static + Deref<Target = NetGraphMsgHandler<G, CA, L>> + Send + Sync,
+               PGS: 'static + Deref<Target = P2PGossipSync<G, CA, L>> + Send + Sync,
                UMH: 'static + Deref + Send + Sync,
                PM: 'static + Deref<Target = PeerManager<Descriptor, CMH, RMH, L, UMH>> + Send + Sync,
                S: 'static + Deref<Target = SC> + Send + Sync,
@@ -191,7 +191,7 @@ impl BackgroundProcessor {
                RGS: 'static + Deref<Target = RapidGossipSync<G>> + Send
        >(
                persister: PS, event_handler: EH, chain_monitor: M, channel_manager: CM,
-               net_graph_msg_handler: Option<NG>, peer_manager: PM, logger: L, scorer: Option<S>,
+               p2p_gossip_sync: Option<PGS>, peer_manager: PM, logger: L, scorer: Option<S>,
                rapid_gossip_sync: Option<RGS>
        ) -> Self
        where
@@ -211,7 +211,7 @@ impl BackgroundProcessor {
                let stop_thread = Arc::new(AtomicBool::new(false));
                let stop_thread_clone = stop_thread.clone();
                let handle = thread::spawn(move || -> Result<(), std::io::Error> {
-                       let event_handler = DecoratingEventHandler { event_handler, net_graph_msg_handler: net_graph_msg_handler.as_ref().map(|t| t.deref()) };
+                       let event_handler = DecoratingEventHandler { event_handler, p2p_gossip_sync: p2p_gossip_sync.as_ref().map(|t| t.deref()) };
 
                        log_trace!(logger, "Calling ChannelManager's timer_tick_occurred on startup");
                        channel_manager.timer_tick_occurred();
@@ -298,7 +298,7 @@ impl BackgroundProcessor {
                                                                None
                                                        }
                                                },
-                                               None => net_graph_msg_handler.as_ref().map(|handler| handler.network_graph())
+                                               None => p2p_gossip_sync.as_ref().map(|sync| sync.network_graph())
                                        };
 
                                        if let Some(network_graph_reference) = graph_to_prune {
@@ -337,8 +337,8 @@ impl BackgroundProcessor {
                        }
 
                        // Persist NetworkGraph on exit
-                       if let Some(ref handler) = net_graph_msg_handler {
-                               persister.persist_graph(handler.network_graph())?;
+                       if let Some(ref gossip_sync) = p2p_gossip_sync {
+                               persister.persist_graph(gossip_sync.network_graph())?;
                        }
 
                        Ok(())
@@ -408,7 +408,7 @@ mod tests {
        use lightning::ln::features::{ChannelFeatures, InitFeatures};
        use lightning::ln::msgs::{ChannelMessageHandler, Init};
        use lightning::ln::peer_handler::{PeerManager, MessageHandler, SocketDescriptor, IgnoringMessageHandler};
-       use lightning::routing::network_graph::{NetworkGraph, NetGraphMsgHandler};
+       use lightning::routing::gossip::{NetworkGraph, P2PGossipSync};
        use lightning::util::config::UserConfig;
        use lightning::util::events::{Event, MessageSendEventsProvider, MessageSendEvent};
        use lightning::util::ser::Writeable;
@@ -442,7 +442,7 @@ mod tests {
 
        struct Node {
                node: Arc<SimpleArcChannelManager<ChainMonitor, test_utils::TestBroadcaster, test_utils::TestFeeEstimator, test_utils::TestLogger>>,
-               net_graph_msg_handler: Option<Arc<NetGraphMsgHandler<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>>>,
+               p2p_gossip_sync: Option<Arc<P2PGossipSync<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>>>,
                peer_manager: Arc<PeerManager<TestDescriptor, Arc<test_utils::TestChannelMessageHandler>, Arc<test_utils::TestRoutingMessageHandler>, Arc<test_utils::TestLogger>, IgnoringMessageHandler>>,
                chain_monitor: Arc<ChainMonitor>,
                persister: Arc<FilesystemPersister>,
@@ -547,12 +547,12 @@ mod tests {
                        let params = ChainParameters { network, best_block };
                        let manager = Arc::new(ChannelManager::new(fee_estimator.clone(), chain_monitor.clone(), tx_broadcaster.clone(), logger.clone(), keys_manager.clone(), UserConfig::default(), params));
                        let network_graph = Arc::new(NetworkGraph::new(genesis_block.header.block_hash()));
-                       let net_graph_msg_handler = Some(Arc::new(NetGraphMsgHandler::new(network_graph.clone(), Some(chain_source.clone()), logger.clone())));
+                       let p2p_gossip_sync = Some(Arc::new(P2PGossipSync::new(network_graph.clone(), Some(chain_source.clone()), logger.clone())));
                        let msg_handler = MessageHandler { chan_handler: Arc::new(test_utils::TestChannelMessageHandler::new()), route_handler: Arc::new(test_utils::TestRoutingMessageHandler::new() )};
                        let peer_manager = Arc::new(PeerManager::new(msg_handler, keys_manager.get_node_secret(Recipient::Node).unwrap(), &seed, logger.clone(), IgnoringMessageHandler{}));
                        let scorer = Arc::new(Mutex::new(test_utils::TestScorer::with_penalty(0)));
                        let rapid_gossip_sync = None;
-                       let node = Node { node: manager, net_graph_msg_handler, peer_manager, chain_monitor, persister, tx_broadcaster, network_graph, logger, best_block, scorer, rapid_gossip_sync };
+                       let node = Node { node: manager, p2p_gossip_sync, peer_manager, chain_monitor, persister, tx_broadcaster, network_graph, logger, best_block, scorer, rapid_gossip_sync };
                        nodes.push(node);
                }
 
@@ -650,7 +650,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                macro_rules! check_persisted_data {
                        ($node: expr, $filepath: expr) => {
@@ -695,7 +695,7 @@ mod tests {
 
                // Check network graph is persisted
                let filepath = get_full_filepath("test_background_processor_persister_0".to_string(), "network_graph".to_string());
-               if let Some(ref handler) = nodes[0].net_graph_msg_handler {
+               if let Some(ref handler) = nodes[0].p2p_gossip_sync {
                        let network_graph = handler.network_graph();
                        check_persisted_data!(network_graph, filepath.clone());
                }
@@ -715,7 +715,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
                loop {
                        let log_entries = nodes[0].logger.lines.lock().unwrap();
                        let desired_log = "Calling ChannelManager's timer_tick_occurred".to_string();
@@ -738,7 +738,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir).with_manager_error(std::io::ErrorKind::Other, "test"));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
                match bg_processor.join() {
                        Ok(_) => panic!("Expected error persisting manager"),
                        Err(e) => {
@@ -755,7 +755,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir).with_graph_error(std::io::ErrorKind::Other, "test"));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                match bg_processor.stop() {
                        Ok(_) => panic!("Expected error persisting network graph"),
@@ -773,7 +773,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir).with_scorer_error(std::io::ErrorKind::Other, "test"));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                match bg_processor.stop() {
                        Ok(_) => panic!("Expected error persisting scorer"),
@@ -796,7 +796,7 @@ mod tests {
                let event_handler = move |event: &Event| {
                        sender.send(handle_funding_generation_ready!(event, channel_value)).unwrap();
                };
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                // Open a channel and check that the FundingGenerationReady event was handled.
                begin_open_channel!(nodes[0], nodes[1], channel_value);
@@ -821,7 +821,7 @@ mod tests {
                let (sender, receiver) = std::sync::mpsc::sync_channel(1);
                let event_handler = move |event: &Event| sender.send(event.clone()).unwrap();
                let persister = Arc::new(Persister::new(data_dir));
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                // Force close the channel and check that the SpendableOutputs event was handled.
                nodes[0].node.force_close_channel(&nodes[0].node.list_channels()[0].channel_id, &nodes[1].node.get_our_node_id()).unwrap();
@@ -845,7 +845,7 @@ mod tests {
                let data_dir = nodes[0].persister.get_data_dir();
                let persister = Arc::new(Persister::new(data_dir));
                let event_handler = |_: &_| {};
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(),  nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
 
                loop {
                        let log_entries = nodes[0].logger.lines.lock().unwrap();
@@ -874,7 +874,7 @@ mod tests {
                assert_eq!(network_graph.read_only().channels().len(), 1);
 
                let event_handler = |_: &_| {};
-               let background_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), Some(rapid_sync.clone()));
+               let background_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), Some(rapid_sync.clone()));
 
                loop {
                        let log_entries = nodes[0].logger.lines.lock().unwrap();
@@ -928,7 +928,7 @@ mod tests {
                let router = DefaultRouter::new(Arc::clone(&nodes[0].network_graph), Arc::clone(&nodes[0].logger), random_seed_bytes);
                let invoice_payer = Arc::new(InvoicePayer::new(Arc::clone(&nodes[0].node), router, Arc::clone(&nodes[0].scorer), Arc::clone(&nodes[0].logger), |_: &_| {}, Retry::Attempts(2)));
                let event_handler = Arc::clone(&invoice_payer);
-               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].net_graph_msg_handler.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
+               let bg_processor = BackgroundProcessor::start(persister, event_handler, nodes[0].chain_monitor.clone(), nodes[0].node.clone(), nodes[0].p2p_gossip_sync.clone(), nodes[0].peer_manager.clone(), nodes[0].logger.clone(), Some(nodes[0].scorer.clone()), nodes[0].rapid_gossip_sync.clone());
                assert!(bg_processor.stop().is_ok());
        }
 }
index e9b639c101307dccd7e229caee1303c77a0bbca6..b4eb04a3d21666369fb6956ca026aebcf8e44496 100644 (file)
@@ -13,7 +13,7 @@ use bitcoin_hashes::Hash;
 use bitcoin_hashes::sha256;
 use crate::prelude::*;
 use lightning::ln::PaymentSecret;
-use lightning::routing::network_graph::RoutingFees;
+use lightning::routing::gossip::RoutingFees;
 use lightning::routing::router::{RouteHint, RouteHintHop};
 
 use num_traits::{CheckedAdd, CheckedMul};
@@ -909,7 +909,7 @@ mod test {
 
        #[test]
        fn test_parse_route() {
-               use lightning::routing::network_graph::RoutingFees;
+               use lightning::routing::gossip::RoutingFees;
                use lightning::routing::router::{RouteHint, RouteHintHop};
                use ::PrivateRoute;
                use bech32::FromBase32;
index 008d3344b55c5f2becae9cfd553d1c3f3f84dc2a..609b33b2d50d941f74585addc5cdd7565fb348fb 100644 (file)
@@ -45,7 +45,7 @@ use bitcoin_hashes::sha256;
 use lightning::ln::PaymentSecret;
 use lightning::ln::features::InvoiceFeatures;
 #[cfg(any(doc, test))]
-use lightning::routing::network_graph::RoutingFees;
+use lightning::routing::gossip::RoutingFees;
 use lightning::routing::router::RouteHint;
 use lightning::util::invoice::construct_invoice_preimage;
 
index c095f74dd16e2e0dc9c3c41dacf7122c75e179a9..e672b89c919c4aefaa6cb1c7bb6074bf54059941 100644 (file)
@@ -38,7 +38,7 @@
 //! # use lightning::ln::{PaymentHash, PaymentPreimage, PaymentSecret};
 //! # use lightning::ln::channelmanager::{ChannelDetails, PaymentId, PaymentSendFailure};
 //! # use lightning::ln::msgs::LightningError;
-//! # use lightning::routing::network_graph::NodeId;
+//! # use lightning::routing::gossip::NodeId;
 //! # use lightning::routing::router::{Route, RouteHop, RouteParameters};
 //! # use lightning::routing::scoring::{ChannelUsage, Score};
 //! # use lightning::util::events::{Event, EventHandler, EventsProvider};
@@ -602,7 +602,7 @@ mod tests {
        use lightning::ln::features::{ChannelFeatures, NodeFeatures, InitFeatures};
        use lightning::ln::functional_test_utils::*;
        use lightning::ln::msgs::{ChannelMessageHandler, ErrorAction, LightningError};
-       use lightning::routing::network_graph::NodeId;
+       use lightning::routing::gossip::NodeId;
        use lightning::routing::router::{PaymentParameters, Route, RouteHop};
        use lightning::routing::scoring::ChannelUsage;
        use lightning::util::test_utils::TestLogger;
index 9f6691564ca64399b9e3299096783248187bc043..04cfb3f147099181b37ed8e6dee70d989f983233 100644 (file)
@@ -15,9 +15,9 @@ use lightning::ln::channelmanager::{ChannelDetails, ChannelManager, PaymentId, P
 use lightning::ln::channelmanager::{PhantomRouteHints, MIN_CLTV_EXPIRY_DELTA};
 use lightning::ln::inbound_payment::{create, create_from_hash, ExpandedKey};
 use lightning::ln::msgs::LightningError;
-use lightning::routing::scoring::Score;
-use lightning::routing::network_graph::{NetworkGraph, RoutingFees};
+use lightning::routing::gossip::{NetworkGraph, RoutingFees};
 use lightning::routing::router::{Route, RouteHint, RouteHintHop, RouteParameters, find_route};
+use lightning::routing::scoring::Score;
 use lightning::util::logger::Logger;
 use secp256k1::PublicKey;
 use core::ops::Deref;
index 1d9c481513dd98944514fd25ea2f7601b2e935a2..ffac702ce1a99582f42ae75e2e26cf7af98f8f6d 100644 (file)
@@ -9,8 +9,8 @@ use bitcoin_hashes::hex::FromHex;
 use bitcoin_hashes::{sha256, Hash};
 use bech32::u5;
 use lightning::ln::PaymentSecret;
+use lightning::routing::gossip::RoutingFees;
 use lightning::routing::router::{RouteHint, RouteHintHop};
-use lightning::routing::network_graph::RoutingFees;
 use lightning_invoice::*;
 use secp256k1::PublicKey;
 use secp256k1::ecdsa::{RecoverableSignature, RecoveryId};
index e2e7807e398b48b327b63f6b5fec49a5724464df..278cede397411720415b89d20a6c9ccaecb1160d 100644 (file)
@@ -29,7 +29,7 @@
 //! ```
 //! use bitcoin::blockdata::constants::genesis_block;
 //! use bitcoin::Network;
-//! use lightning::routing::network_graph::NetworkGraph;
+//! use lightning::routing::gossip::NetworkGraph;
 //! use lightning_rapid_gossip_sync::RapidGossipSync;
 //!
 //! let block_hash = genesis_block(Network::Bitcoin).header.block_hash();
@@ -62,7 +62,7 @@ use std::fs::File;
 use std::ops::Deref;
 use std::sync::atomic::{AtomicBool, Ordering};
 
-use lightning::routing::network_graph::NetworkGraph;
+use lightning::routing::gossip::NetworkGraph;
 
 use crate::error::GraphSyncError;
 
@@ -127,7 +127,7 @@ mod tests {
        use bitcoin::Network;
 
        use lightning::ln::msgs::DecodeError;
-       use lightning::routing::network_graph::NetworkGraph;
+       use lightning::routing::gossip::NetworkGraph;
        use crate::RapidGossipSync;
 
        #[test]
@@ -253,7 +253,7 @@ pub mod bench {
        use bitcoin::Network;
 
        use lightning::ln::msgs::DecodeError;
-       use lightning::routing::network_graph::NetworkGraph;
+       use lightning::routing::gossip::NetworkGraph;
 
        use crate::RapidGossipSync;
 
@@ -271,6 +271,7 @@ pub mod bench {
                                        println!("{}", error_string);
                                        return;
                                }
+                               #[cfg(require_route_graph_test)]
                                panic!("{}", error_string);
                        }
                        assert!(sync_result.is_ok())
index 6ffc6f58ea88bf6adbf1fc9e7fbd2184a0ed78e4..21c1ce29a99142f427dd8f9947f8fe0c94cfbcfa 100644 (file)
@@ -10,7 +10,7 @@ use bitcoin::secp256k1::PublicKey;
 use lightning::ln::msgs::{
        DecodeError, ErrorAction, LightningError, OptionalField, UnsignedChannelUpdate,
 };
-use lightning::routing::network_graph::NetworkGraph;
+use lightning::routing::gossip::NetworkGraph;
 use lightning::util::ser::{BigSize, Readable};
 
 use crate::error::GraphSyncError;
@@ -235,7 +235,7 @@ mod tests {
        use bitcoin::Network;
 
        use lightning::ln::msgs::DecodeError;
-       use lightning::routing::network_graph::NetworkGraph;
+       use lightning::routing::gossip::NetworkGraph;
 
        use crate::error::GraphSyncError;
        use crate::RapidGossipSync;
index 770fe8139413446bcc498086d00572db696f7dbc..495b507426c19bd7f31d6709da30db79c13f0dd9 100644 (file)
@@ -1901,9 +1901,9 @@ fn do_during_funding_monitor_fail(confirm_a_first: bool, restore_b_before_conf:
                (channel_id, create_chan_between_nodes_with_value_b(&nodes[1], &nodes[0], &channel_ready))
        };
        for node in nodes.iter() {
-               assert!(node.net_graph_msg_handler.handle_channel_announcement(&announcement).unwrap());
-               node.net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
+               assert!(node.gossip_sync.handle_channel_announcement(&announcement).unwrap());
+               node.gossip_sync.handle_channel_update(&as_update).unwrap();
+               node.gossip_sync.handle_channel_update(&bs_update).unwrap();
        }
 
        send_payment(&nodes[0], &[&nodes[1]], 8000000);
index 6eeed58b5fd195d7ffebaecebbed791c2de3ffdc..c787cedfad3a7da3a47e719c842689ee94f62545 100644 (file)
 //! responsible for tracking which channels are open, HTLCs are in flight and reestablishing those
 //! upon reconnect to the relevant peer(s).
 //!
-//! It does not manage routing logic (see routing::router::get_route for that) nor does it manage constructing
+//! It does not manage routing logic (see [`find_route`] for that) nor does it manage constructing
 //! on-chain transactions (it only monitors the chain to watch for any force-closes that might
 //! imply it needs to fail HTLCs/payments/channels it manages).
 //!
+//! [`find_route`]: crate::routing::router::find_route
 
 use bitcoin::blockdata::block::BlockHeader;
 use bitcoin::blockdata::transaction::Transaction;
@@ -1777,12 +1778,14 @@ impl<Signer: Sign, M: Deref, T: Deref, K: Deref, F: Deref, L: Deref> ChannelMana
                self.list_channels_with_filter(|_| true)
        }
 
-       /// Gets the list of usable channels, in random order. Useful as an argument to
-       /// get_route to ensure non-announced channels are used.
+       /// Gets the list of usable channels, in random order. Useful as an argument to [`find_route`]
+       /// to ensure non-announced channels are used.
        ///
        /// These are guaranteed to have their [`ChannelDetails::is_usable`] value set to true, see the
        /// documentation for [`ChannelDetails::is_usable`] for more info on exactly what the criteria
        /// are.
+       ///
+       /// [`find_route`]: crate::routing::router::find_route
        pub fn list_usable_channels(&self) -> Vec<ChannelDetails> {
                // Note we use is_live here instead of usable which leads to somewhat confused
                // internal/external nomenclature, but that's ok cause that's probably what the user
@@ -3765,7 +3768,7 @@ impl<Signer: Sign, M: Deref, T: Deref, K: Deref, F: Deref, L: Deref> ChannelMana
                                                        .. } => {
                                                // we get a fail_malformed_htlc from the first hop
                                                // TODO: We'd like to generate a NetworkUpdate for temporary
-                                               // failures here, but that would be insufficient as get_route
+                                               // failures here, but that would be insufficient as find_route
                                                // generally ignores its view of our own channels as we provide them via
                                                // ChannelDetails.
                                                // TODO: For non-temporary failures, we really should be closing the
@@ -7584,7 +7587,7 @@ pub mod bench {
        use ln::features::{InitFeatures, InvoiceFeatures};
        use ln::functional_test_utils::*;
        use ln::msgs::{ChannelMessageHandler, Init};
-       use routing::network_graph::NetworkGraph;
+       use routing::gossip::NetworkGraph;
        use routing::router::{PaymentParameters, get_route};
        use util::test_utils;
        use util::config::UserConfig;
index 8929a9774381e9342f20790e094e51c64b36d959..594f02fc52e040e4981c312c8b2cf0e93078d08d 100644 (file)
@@ -15,7 +15,7 @@ use chain::channelmonitor::ChannelMonitor;
 use chain::transaction::OutPoint;
 use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
 use ln::channelmanager::{ChainParameters, ChannelManager, ChannelManagerReadArgs, RAACommitmentOrder, PaymentSendFailure, PaymentId, MIN_CLTV_EXPIRY_DELTA};
-use routing::network_graph::{NetGraphMsgHandler, NetworkGraph};
+use routing::gossip::{P2PGossipSync, NetworkGraph};
 use routing::router::{PaymentParameters, Route, get_route};
 use ln::features::{InitFeatures, InvoiceFeatures};
 use ln::msgs;
@@ -279,7 +279,7 @@ pub struct Node<'a, 'b: 'a, 'c: 'b> {
        pub keys_manager: &'b test_utils::TestKeysInterface,
        pub node: &'a ChannelManager<EnforcingSigner, &'b TestChainMonitor<'c>, &'c test_utils::TestBroadcaster, &'b test_utils::TestKeysInterface, &'c test_utils::TestFeeEstimator, &'c test_utils::TestLogger>,
        pub network_graph: &'c NetworkGraph,
-       pub net_graph_msg_handler: NetGraphMsgHandler<&'c NetworkGraph, &'c test_utils::TestChainSource, &'c test_utils::TestLogger>,
+       pub gossip_sync: P2PGossipSync<&'c NetworkGraph, &'c test_utils::TestChainSource, &'c test_utils::TestLogger>,
        pub node_seed: [u8; 32],
        pub network_payment_count: Rc<RefCell<u8>>,
        pub network_chan_count: Rc<RefCell<u32>>,
@@ -313,13 +313,13 @@ impl<'a, 'b, 'c> Drop for Node<'a, 'b, 'c> {
                                self.network_graph.write(&mut w).unwrap();
                                let network_graph_deser = <NetworkGraph>::read(&mut io::Cursor::new(&w.0)).unwrap();
                                assert!(network_graph_deser == *self.network_graph);
-                               let net_graph_msg_handler = NetGraphMsgHandler::new(
+                               let gossip_sync = P2PGossipSync::new(
                                        &network_graph_deser, Some(self.chain_source), self.logger
                                );
                                let mut chan_progress = 0;
                                loop {
-                                       let orig_announcements = self.net_graph_msg_handler.get_next_channel_announcements(chan_progress, 255);
-                                       let deserialized_announcements = net_graph_msg_handler.get_next_channel_announcements(chan_progress, 255);
+                                       let orig_announcements = self.gossip_sync.get_next_channel_announcements(chan_progress, 255);
+                                       let deserialized_announcements = gossip_sync.get_next_channel_announcements(chan_progress, 255);
                                        assert!(orig_announcements == deserialized_announcements);
                                        chan_progress = match orig_announcements.last() {
                                                Some(announcement) => announcement.0.contents.short_channel_id + 1,
@@ -328,8 +328,8 @@ impl<'a, 'b, 'c> Drop for Node<'a, 'b, 'c> {
                                }
                                let mut node_progress = None;
                                loop {
-                                       let orig_announcements = self.net_graph_msg_handler.get_next_node_announcements(node_progress.as_ref(), 255);
-                                       let deserialized_announcements = net_graph_msg_handler.get_next_node_announcements(node_progress.as_ref(), 255);
+                                       let orig_announcements = self.gossip_sync.get_next_node_announcements(node_progress.as_ref(), 255);
+                                       let deserialized_announcements = gossip_sync.get_next_node_announcements(node_progress.as_ref(), 255);
                                        assert!(orig_announcements == deserialized_announcements);
                                        node_progress = match orig_announcements.last() {
                                                Some(announcement) => Some(announcement.contents.node_id),
@@ -876,11 +876,11 @@ pub fn update_nodes_with_chan_announce<'a, 'b, 'c, 'd>(nodes: &'a Vec<Node<'b, '
        };
 
        for node in nodes {
-               assert!(node.net_graph_msg_handler.handle_channel_announcement(ann).unwrap());
-               node.net_graph_msg_handler.handle_channel_update(upd_1).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(upd_2).unwrap();
-               node.net_graph_msg_handler.handle_node_announcement(&a_node_announcement).unwrap();
-               node.net_graph_msg_handler.handle_node_announcement(&b_node_announcement).unwrap();
+               assert!(node.gossip_sync.handle_channel_announcement(ann).unwrap());
+               node.gossip_sync.handle_channel_update(upd_1).unwrap();
+               node.gossip_sync.handle_channel_update(upd_2).unwrap();
+               node.gossip_sync.handle_node_announcement(&a_node_announcement).unwrap();
+               node.gossip_sync.handle_node_announcement(&b_node_announcement).unwrap();
 
                // Note that channel_updates are also delivered to ChannelManagers to ensure we have
                // forwarding info for local channels even if its not accepted in the network graph.
@@ -1506,13 +1506,13 @@ macro_rules! expect_payment_failed_conditions {
 
                                if let Some(chan_closed) = $conditions.expected_blamed_chan_closed {
                                        match network_update {
-                                               &Some($crate::routing::network_graph::NetworkUpdate::ChannelUpdateMessage { ref msg }) if !chan_closed => {
+                                               &Some($crate::routing::gossip::NetworkUpdate::ChannelUpdateMessage { ref msg }) if !chan_closed => {
                                                        if let Some(scid) = $conditions.expected_blamed_scid {
                                                                assert_eq!(msg.contents.short_channel_id, scid);
                                                        }
                                                        assert_eq!(msg.contents.flags & 2, 0);
                                                },
-                                               &Some($crate::routing::network_graph::NetworkUpdate::ChannelClosed { short_channel_id, is_permanent }) if chan_closed => {
+                                               &Some($crate::routing::gossip::NetworkUpdate::ChannelFailure { short_channel_id, is_permanent }) if chan_closed => {
                                                        if let Some(scid) = $conditions.expected_blamed_scid {
                                                                assert_eq!(short_channel_id, scid);
                                                        }
@@ -1992,11 +1992,11 @@ pub fn create_network<'a, 'b: 'a, 'c: 'b>(node_count: usize, cfgs: &'b Vec<NodeC
        let connect_style = Rc::new(RefCell::new(ConnectStyle::random_style()));
 
        for i in 0..node_count {
-               let net_graph_msg_handler = NetGraphMsgHandler::new(cfgs[i].network_graph, None, cfgs[i].logger);
+               let gossip_sync = P2PGossipSync::new(cfgs[i].network_graph, None, cfgs[i].logger);
                nodes.push(Node{
                        chain_source: cfgs[i].chain_source, tx_broadcaster: cfgs[i].tx_broadcaster,
                        chain_monitor: &cfgs[i].chain_monitor, keys_manager: &cfgs[i].keys_manager,
-                       node: &chan_mgrs[i], network_graph: &cfgs[i].network_graph, net_graph_msg_handler,
+                       node: &chan_mgrs[i], network_graph: &cfgs[i].network_graph, gossip_sync,
                        node_seed: cfgs[i].node_seed, network_chan_count: chan_count.clone(),
                        network_payment_count: payment_count.clone(), logger: cfgs[i].logger,
                        blocks: Arc::clone(&cfgs[i].tx_broadcaster.blocks),
@@ -2160,8 +2160,8 @@ pub fn handle_announce_close_broadcast_events<'a, 'b, 'c>(nodes: &Vec<Node<'a, '
        }
 
        for node in nodes {
-               node.net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
+               node.gossip_sync.handle_channel_update(&as_update).unwrap();
+               node.gossip_sync.handle_channel_update(&bs_update).unwrap();
        }
 }
 
index d1dbad4720b814db3fb89476e0e8e8b7cee22193..c7365452d67e5a7d80fb2661f48bcfb7adefbc67 100644 (file)
@@ -2347,8 +2347,8 @@ fn channel_monitor_network_test() {
                check_preimage_claim(&nodes[4], &node_txn);
                (close_chan_update_1, close_chan_update_2)
        };
-       nodes[3].net_graph_msg_handler.handle_channel_update(&close_chan_update_2).unwrap();
-       nodes[4].net_graph_msg_handler.handle_channel_update(&close_chan_update_1).unwrap();
+       nodes[3].gossip_sync.handle_channel_update(&close_chan_update_2).unwrap();
+       nodes[4].gossip_sync.handle_channel_update(&close_chan_update_1).unwrap();
        assert_eq!(nodes[3].node.list_channels().len(), 0);
        assert_eq!(nodes[4].node.list_channels().len(), 0);
 
@@ -3978,9 +3978,9 @@ fn test_funding_peer_disconnect() {
        };
 
        // Provide the channel announcement and public updates to the network graph
-       nodes[0].net_graph_msg_handler.handle_channel_announcement(&chan_announcement).unwrap();
-       nodes[0].net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
-       nodes[0].net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
+       nodes[0].gossip_sync.handle_channel_announcement(&chan_announcement).unwrap();
+       nodes[0].gossip_sync.handle_channel_update(&bs_update).unwrap();
+       nodes[0].gossip_sync.handle_channel_update(&as_update).unwrap();
 
        let (route, _, _, _) = get_route_and_payment_hash!(nodes[0], nodes[1], 1000000);
        let payment_preimage = send_along_route(&nodes[0], route, &[&nodes[1]], 1000000).0;
@@ -4421,9 +4421,9 @@ fn test_no_txn_manager_serialize_deserialize() {
        let (channel_ready, _) = create_chan_between_nodes_with_value_confirm(&nodes[0], &nodes[1], &tx);
        let (announcement, as_update, bs_update) = create_chan_between_nodes_with_value_b(&nodes[0], &nodes[1], &channel_ready);
        for node in nodes.iter() {
-               assert!(node.net_graph_msg_handler.handle_channel_announcement(&announcement).unwrap());
-               node.net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
+               assert!(node.gossip_sync.handle_channel_announcement(&announcement).unwrap());
+               node.gossip_sync.handle_channel_update(&as_update).unwrap();
+               node.gossip_sync.handle_channel_update(&bs_update).unwrap();
        }
 
        send_payment(&nodes[0], &[&nodes[1]], 1000000);
@@ -4541,9 +4541,9 @@ fn test_manager_serialize_deserialize_events() {
        let (channel_ready, _) = create_chan_between_nodes_with_value_confirm(&nodes[0], &nodes[1], &tx);
        let (announcement, as_update, bs_update) = create_chan_between_nodes_with_value_b(&nodes[0], &nodes[1], &channel_ready);
        for node in nodes.iter() {
-               assert!(node.net_graph_msg_handler.handle_channel_announcement(&announcement).unwrap());
-               node.net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
+               assert!(node.gossip_sync.handle_channel_announcement(&announcement).unwrap());
+               node.gossip_sync.handle_channel_update(&as_update).unwrap();
+               node.gossip_sync.handle_channel_update(&bs_update).unwrap();
        }
 
        send_payment(&nodes[0], &[&nodes[1]], 1000000);
index 1e5f49f07df6a0cc5056e6ea26545cfac33d04ec..9522e83d18d1cbfa02ddcbe9d0b7ef91ba2d4925 100644 (file)
@@ -9,14 +9,17 @@
 
 //! High level lightning structs and impls live here.
 //!
-//! You probably want to create a channelmanager::ChannelManager, and a routing::NetGraphMsgHandler first.
+//! You probably want to create a [`ChannelManager`], and a [`P2PGossipSync`] first.
 //! Then, you probably want to pass them both on to a peer_handler::PeerManager and use that to
 //! create/manage connections and call get_and_clear_pending_events after each action, handling
 //! them appropriately.
 //!
-//! When you want to open/close a channel or send a payment, call into your ChannelManager and when
-//! you want to learn things about the network topology (eg get a route for sending a payment),
-//! call into your NetGraphMsgHandler.
+//! When you want to open/close a channel or send a payment, call into your [`ChannelManager`] and
+//! when you want to learn things about the network topology (eg get a route for sending a payment),
+//! call into your [`P2PGossipSync`].
+//!
+//! [`ChannelManager`]: channelmanager::ChannelManager
+//! [`P2PGossipSync`]: crate::routing::gossip::P2PGossipSync
 
 #[cfg(any(test, feature = "_test_utils"))]
 #[macro_use]
index 802cd3acab45bd4592a7f80cdb982cfadeb7ed52..eabd2222078b5a73c7456db15d65ff453a45ab05 100644 (file)
@@ -16,7 +16,7 @@ use chain::keysinterface::{KeysInterface, Recipient};
 use ln::{PaymentHash, PaymentSecret};
 use ln::channelmanager::{HTLCForwardInfo, CLTV_FAR_FAR_AWAY, MIN_CLTV_EXPIRY_DELTA, PendingHTLCInfo, PendingHTLCRouting};
 use ln::onion_utils;
-use routing::network_graph::{NetworkUpdate, RoutingFees, NodeId};
+use routing::gossip::{NetworkUpdate, RoutingFees, NodeId};
 use routing::router::{get_route, PaymentParameters, Route, RouteHint, RouteHintHop};
 use ln::features::{InitFeatures, InvoiceFeatures, NodeFeatures};
 use ln::msgs;
@@ -177,8 +177,8 @@ fn run_onion_failure_test_with_fail_intercept<F1,F2,F3>(_name: &str, test_case:
                                                        panic!("channel_update not found!");
                                                }
                                        },
-                                       &NetworkUpdate::ChannelClosed { ref short_channel_id, ref is_permanent } => {
-                                               if let NetworkUpdate::ChannelClosed { short_channel_id: ref expected_short_channel_id, is_permanent: ref expected_is_permanent } = expected_channel_update.unwrap() {
+                                       &NetworkUpdate::ChannelFailure { ref short_channel_id, ref is_permanent } => {
+                                               if let NetworkUpdate::ChannelFailure { short_channel_id: ref expected_short_channel_id, is_permanent: ref expected_is_permanent } = expected_channel_update.unwrap() {
                                                        assert!(*short_channel_id == *expected_short_channel_id);
                                                        assert!(*is_permanent == *expected_is_permanent);
                                                } else {
@@ -283,7 +283,7 @@ fn test_fee_failures() {
        let short_channel_id = channels[0].0.contents.short_channel_id;
        run_onion_failure_test("fee_insufficient", 0, &nodes, &route, &payment_hash, &payment_secret, |msg| {
                msg.amount_msat -= 1;
-       }, || {}, true, Some(UPDATE|12), Some(NetworkUpdate::ChannelClosed { short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, || {}, true, Some(UPDATE|12), Some(NetworkUpdate::ChannelFailure { short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        // In an earlier version, we spuriously failed to forward payments if the expected feerate
        // changed between the channel open and the payment.
@@ -343,7 +343,7 @@ fn test_onion_failure() {
                // describing a length-1 TLV payload, which is obviously bogus.
                new_payloads[0].data[0] = 1;
                msg.onion_routing_packet = onion_utils::construct_onion_packet_bogus_hopdata(new_payloads, onion_keys, [0; 32], &payment_hash);
-       }, ||{}, true, Some(PERM|22), Some(NetworkUpdate::ChannelClosed{short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, ||{}, true, Some(PERM|22), Some(NetworkUpdate::ChannelFailure{short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        // final node failure
        let short_channel_id = channels[1].0.contents.short_channel_id;
@@ -360,7 +360,7 @@ fn test_onion_failure() {
                // length-1 TLV payload, which is obviously bogus.
                new_payloads[1].data[0] = 1;
                msg.onion_routing_packet = onion_utils::construct_onion_packet_bogus_hopdata(new_payloads, onion_keys, [0; 32], &payment_hash);
-       }, ||{}, false, Some(PERM|22), Some(NetworkUpdate::ChannelClosed{short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, ||{}, false, Some(PERM|22), Some(NetworkUpdate::ChannelFailure{short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        // the following three with run_onion_failure_test_with_fail_intercept() test only the origin node
        // receiving simulated fail messages
@@ -471,7 +471,7 @@ fn test_onion_failure() {
                let onion_keys = onion_utils::construct_onion_keys(&Secp256k1::new(), &route.paths[0], &session_priv).unwrap();
                msg.reason = onion_utils::build_first_hop_failure_packet(onion_keys[0].shared_secret.as_ref(), PERM|8, &[0;0]);
                // short_channel_id from the processing node
-       }, ||{}, true, Some(PERM|8), Some(NetworkUpdate::ChannelClosed{short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, ||{}, true, Some(PERM|8), Some(NetworkUpdate::ChannelFailure{short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        let short_channel_id = channels[1].0.contents.short_channel_id;
        run_onion_failure_test_with_fail_intercept("required_channel_feature_missing", 100, &nodes, &route, &payment_hash, &payment_secret, |msg| {
@@ -481,13 +481,13 @@ fn test_onion_failure() {
                let onion_keys = onion_utils::construct_onion_keys(&Secp256k1::new(), &route.paths[0], &session_priv).unwrap();
                msg.reason = onion_utils::build_first_hop_failure_packet(onion_keys[0].shared_secret.as_ref(), PERM|9, &[0;0]);
                // short_channel_id from the processing node
-       }, ||{}, true, Some(PERM|9), Some(NetworkUpdate::ChannelClosed{short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, ||{}, true, Some(PERM|9), Some(NetworkUpdate::ChannelFailure{short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        let mut bogus_route = route.clone();
        bogus_route.paths[0][1].short_channel_id -= 1;
        let short_channel_id = bogus_route.paths[0][1].short_channel_id;
        run_onion_failure_test("unknown_next_peer", 0, &nodes, &bogus_route, &payment_hash, &payment_secret, |_| {}, ||{}, true, Some(PERM|10),
-         Some(NetworkUpdate::ChannelClosed{short_channel_id, is_permanent:true}), Some(short_channel_id));
+         Some(NetworkUpdate::ChannelFailure{short_channel_id, is_permanent:true}), Some(short_channel_id));
 
        let short_channel_id = channels[1].0.contents.short_channel_id;
        let amt_to_forward = nodes[1].node.channel_state.lock().unwrap().by_id.get(&channels[1].2).unwrap().get_counterparty_htlc_minimum_msat() - 1;
@@ -511,13 +511,13 @@ fn test_onion_failure() {
        let short_channel_id = channels[0].0.contents.short_channel_id;
        run_onion_failure_test("fee_insufficient", 0, &nodes, &route, &payment_hash, &payment_secret, |msg| {
                msg.amount_msat -= 1;
-       }, || {}, true, Some(UPDATE|12), Some(NetworkUpdate::ChannelClosed { short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, || {}, true, Some(UPDATE|12), Some(NetworkUpdate::ChannelFailure { short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        let short_channel_id = channels[0].0.contents.short_channel_id;
        run_onion_failure_test("incorrect_cltv_expiry", 0, &nodes, &route, &payment_hash, &payment_secret, |msg| {
                // need to violate: cltv_expiry - cltv_expiry_delta >= outgoing_cltv_value
                msg.cltv_expiry -= 1;
-       }, || {}, true, Some(UPDATE|13), Some(NetworkUpdate::ChannelClosed { short_channel_id, is_permanent: true}), Some(short_channel_id));
+       }, || {}, true, Some(UPDATE|13), Some(NetworkUpdate::ChannelFailure { short_channel_id, is_permanent: true}), Some(short_channel_id));
 
        let short_channel_id = channels[1].0.contents.short_channel_id;
        run_onion_failure_test("expiry_too_soon", 0, &nodes, &route, &payment_hash, &payment_secret, |msg| {
index 9e4ae058f8d517bdfb93abb021d02a7eb588e8a5..1df374d7e007b8784d7235c0a216fea4679d408e 100644 (file)
@@ -11,7 +11,7 @@ use ln::{PaymentHash, PaymentPreimage, PaymentSecret};
 use ln::channelmanager::HTLCSource;
 use ln::msgs;
 use ln::wire::Encode;
-use routing::network_graph::NetworkUpdate;
+use routing::gossip::NetworkUpdate;
 use routing::router::RouteHop;
 use util::chacha20::{ChaCha20, ChaChaReader};
 use util::errors::{self, APIError};
@@ -395,7 +395,7 @@ pub(super) fn process_onion_failure<T: secp256k1::Signing, L: Deref>(secp_ctx: &
                                                }
                                                else if error_code & PERM == PERM {
                                                        if !payment_failed {
-                                                               network_update = Some(NetworkUpdate::ChannelClosed {
+                                                               network_update = Some(NetworkUpdate::ChannelFailure {
                                                                        short_channel_id: failing_route_hop.short_channel_id,
                                                                        is_permanent: true,
                                                                });
@@ -440,7 +440,7 @@ pub(super) fn process_onion_failure<T: secp256k1::Signing, L: Deref>(secp_ctx: &
                                                                                if is_chan_update_invalid {
                                                                                        // This probably indicates the node which forwarded
                                                                                        // to the node in question corrupted something.
-                                                                                       network_update = Some(NetworkUpdate::ChannelClosed {
+                                                                                       network_update = Some(NetworkUpdate::ChannelFailure {
                                                                                                short_channel_id: route_hop.short_channel_id,
                                                                                                is_permanent: true,
                                                                                        });
index 68a7ef952702b31436ea18ff27bb83f3b67b3fe3..d86cddb13dbc66162c52af7128f0c3c180688532 100644 (file)
@@ -12,8 +12,8 @@
 //! Instead of actually servicing sockets ourselves we require that you implement the
 //! SocketDescriptor interface and use that to receive actions which you should perform on the
 //! socket, and call into PeerManager with bytes read from the socket. The PeerManager will then
-//! call into the provided message handlers (probably a ChannelManager and NetGraphmsgHandler) with messages
-//! they should handle, and encoding/sending response messages.
+//! call into the provided message handlers (probably a ChannelManager and P2PGossipSync) with
+//! messages they should handle, and encoding/sending response messages.
 
 use bitcoin::secp256k1::{self, Secp256k1, SecretKey, PublicKey};
 
@@ -25,10 +25,10 @@ use util::ser::{VecWriter, Writeable, Writer};
 use ln::peer_channel_encryptor::{PeerChannelEncryptor,NextNoiseStep};
 use ln::wire;
 use ln::wire::Encode;
+use routing::gossip::{NetworkGraph, P2PGossipSync};
 use util::atomic_counter::AtomicCounter;
 use util::events::{MessageSendEvent, MessageSendEventsProvider};
 use util::logger::Logger;
-use routing::network_graph::{NetworkGraph, NetGraphMsgHandler};
 
 use prelude::*;
 use io;
@@ -208,10 +208,9 @@ pub struct MessageHandler<CM: Deref, RM: Deref> where
        /// [`ChannelManager`]: crate::ln::channelmanager::ChannelManager
        pub chan_handler: CM,
        /// A message handler which handles messages updating our knowledge of the network channel
-       /// graph. Usually this is just a [`NetGraphMsgHandler`] object or an
-       /// [`IgnoringMessageHandler`].
+       /// graph. Usually this is just a [`P2PGossipSync`] object or an [`IgnoringMessageHandler`].
        ///
-       /// [`NetGraphMsgHandler`]: crate::routing::network_graph::NetGraphMsgHandler
+       /// [`P2PGossipSync`]: crate::routing::gossip::P2PGossipSync
        pub route_handler: RM,
 }
 
@@ -388,7 +387,7 @@ impl Peer {
 /// issues such as overly long function definitions.
 ///
 /// (C-not exported) as Arcs don't make sense in bindings
-pub type SimpleArcPeerManager<SD, M, T, F, C, L> = PeerManager<SD, Arc<SimpleArcChannelManager<M, T, F, L>>, Arc<NetGraphMsgHandler<Arc<NetworkGraph>, Arc<C>, Arc<L>>>, Arc<L>, Arc<IgnoringMessageHandler>>;
+pub type SimpleArcPeerManager<SD, M, T, F, C, L> = PeerManager<SD, Arc<SimpleArcChannelManager<M, T, F, L>>, Arc<P2PGossipSync<Arc<NetworkGraph>, Arc<C>, Arc<L>>>, Arc<L>, Arc<IgnoringMessageHandler>>;
 
 /// SimpleRefPeerManager is a type alias for a PeerManager reference, and is the reference
 /// counterpart to the SimpleArcPeerManager type alias. Use this type by default when you don't
@@ -398,7 +397,7 @@ pub type SimpleArcPeerManager<SD, M, T, F, C, L> = PeerManager<SD, Arc<SimpleArc
 /// helps with issues such as long function definitions.
 ///
 /// (C-not exported) as Arcs don't make sense in bindings
-pub type SimpleRefPeerManager<'a, 'b, 'c, 'd, 'e, 'f, 'g, 'h, SD, M, T, F, C, L> = PeerManager<SD, SimpleRefChannelManager<'a, 'b, 'c, 'd, 'e, M, T, F, L>, &'e NetGraphMsgHandler<&'g NetworkGraph, &'h C, &'f L>, &'f L, IgnoringMessageHandler>;
+pub type SimpleRefPeerManager<'a, 'b, 'c, 'd, 'e, 'f, 'g, 'h, SD, M, T, F, C, L> = PeerManager<SD, SimpleRefChannelManager<'a, 'b, 'c, 'd, 'e, M, T, F, L>, &'e P2PGossipSync<&'g NetworkGraph, &'h C, &'f L>, &'f L, IgnoringMessageHandler>;
 
 /// A PeerManager manages a set of peers, described by their [`SocketDescriptor`] and marshalls
 /// socket events into messages which it passes on to its [`MessageHandler`].
index bc430fe52837c8a2282e6ea2bf929b0b52568a17..051703a7fcc04b2121a82c207cb18bdd99829cb9 100644 (file)
@@ -15,7 +15,7 @@ use chain::{ChannelMonitorUpdateErr, Watch};
 use chain::channelmonitor::ChannelMonitor;
 use chain::keysinterface::{Recipient, KeysInterface};
 use ln::channelmanager::{ChannelManager, ChannelManagerReadArgs, MIN_CLTV_EXPIRY_DELTA};
-use routing::network_graph::RoutingFees;
+use routing::gossip::RoutingFees;
 use routing::router::{PaymentParameters, RouteHint, RouteHintHop};
 use ln::features::{InitFeatures, InvoiceFeatures, ChannelTypeFeatures};
 use ln::msgs;
@@ -236,9 +236,9 @@ fn do_test_1_conf_open(connect_style: ConnectStyle) {
        assert_eq!(announcement, bs_announcement);
 
        for node in nodes {
-               assert!(node.net_graph_msg_handler.handle_channel_announcement(&announcement).unwrap());
-               node.net_graph_msg_handler.handle_channel_update(&as_update).unwrap();
-               node.net_graph_msg_handler.handle_channel_update(&bs_update).unwrap();
+               assert!(node.gossip_sync.handle_channel_announcement(&announcement).unwrap());
+               node.gossip_sync.handle_channel_update(&as_update).unwrap();
+               node.gossip_sync.handle_channel_update(&bs_update).unwrap();
        }
 }
 #[test]
diff --git a/lightning/src/routing/gossip.rs b/lightning/src/routing/gossip.rs
new file mode 100644 (file)
index 0000000..52cda71
--- /dev/null
@@ -0,0 +1,2771 @@
+// This file is Copyright its original authors, visible in version control
+// history.
+//
+// This file is licensed under the Apache License, Version 2.0 <LICENSE-APACHE
+// or http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
+// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your option.
+// You may not use this file except in accordance with one or both of these
+// licenses.
+
+//! The top-level network map tracking logic lives here.
+
+use bitcoin::secp256k1::constants::PUBLIC_KEY_SIZE;
+use bitcoin::secp256k1::PublicKey;
+use bitcoin::secp256k1::Secp256k1;
+use bitcoin::secp256k1;
+
+use bitcoin::hashes::sha256d::Hash as Sha256dHash;
+use bitcoin::hashes::Hash;
+use bitcoin::blockdata::script::Builder;
+use bitcoin::blockdata::transaction::TxOut;
+use bitcoin::blockdata::opcodes;
+use bitcoin::hash_types::BlockHash;
+
+use chain;
+use chain::Access;
+use ln::features::{ChannelFeatures, NodeFeatures};
+use ln::msgs::{DecodeError, ErrorAction, Init, LightningError, RoutingMessageHandler, NetAddress, MAX_VALUE_MSAT};
+use ln::msgs::{ChannelAnnouncement, ChannelUpdate, NodeAnnouncement, OptionalField, GossipTimestampFilter};
+use ln::msgs::{QueryChannelRange, ReplyChannelRange, QueryShortChannelIds, ReplyShortChannelIdsEnd};
+use ln::msgs;
+use util::ser::{Writeable, Readable, Writer};
+use util::logger::{Logger, Level};
+use util::events::{Event, EventHandler, MessageSendEvent, MessageSendEventsProvider};
+use util::scid_utils::{block_from_scid, scid_from_parts, MAX_SCID_BLOCK};
+
+use io;
+use prelude::*;
+use alloc::collections::{BTreeMap, btree_map::Entry as BtreeEntry};
+use core::{cmp, fmt};
+use sync::{RwLock, RwLockReadGuard};
+use core::sync::atomic::{AtomicUsize, Ordering};
+use sync::Mutex;
+use core::ops::Deref;
+use bitcoin::hashes::hex::ToHex;
+
+#[cfg(feature = "std")]
+use std::time::{SystemTime, UNIX_EPOCH};
+
+/// We remove stale channel directional info two weeks after the last update, per BOLT 7's
+/// suggestion.
+const STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS: u64 = 60 * 60 * 24 * 14;
+
+/// The maximum number of extra bytes which we do not understand in a gossip message before we will
+/// refuse to relay the message.
+const MAX_EXCESS_BYTES_FOR_RELAY: usize = 1024;
+
+/// Maximum number of short_channel_ids that will be encoded in one gossip reply message.
+/// This value ensures a reply fits within the 65k payload limit and is consistent with other implementations.
+const MAX_SCIDS_PER_REPLY: usize = 8000;
+
+/// Represents the compressed public key of a node
+#[derive(Clone, Copy)]
+pub struct NodeId([u8; PUBLIC_KEY_SIZE]);
+
+impl NodeId {
+       /// Create a new NodeId from a public key
+       pub fn from_pubkey(pubkey: &PublicKey) -> Self {
+               NodeId(pubkey.serialize())
+       }
+
+       /// Get the public key slice from this NodeId
+       pub fn as_slice(&self) -> &[u8] {
+               &self.0
+       }
+}
+
+impl fmt::Debug for NodeId {
+       fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+               write!(f, "NodeId({})", log_bytes!(self.0))
+       }
+}
+
+impl core::hash::Hash for NodeId {
+       fn hash<H: core::hash::Hasher>(&self, hasher: &mut H) {
+               self.0.hash(hasher);
+       }
+}
+
+impl Eq for NodeId {}
+
+impl PartialEq for NodeId {
+       fn eq(&self, other: &Self) -> bool {
+               self.0[..] == other.0[..]
+       }
+}
+
+impl cmp::PartialOrd for NodeId {
+       fn partial_cmp(&self, other: &Self) -> Option<cmp::Ordering> {
+               Some(self.cmp(other))
+       }
+}
+
+impl Ord for NodeId {
+       fn cmp(&self, other: &Self) -> cmp::Ordering {
+               self.0[..].cmp(&other.0[..])
+       }
+}
+
+impl Writeable for NodeId {
+       fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
+               writer.write_all(&self.0)?;
+               Ok(())
+       }
+}
+
+impl Readable for NodeId {
+       fn read<R: io::Read>(reader: &mut R) -> Result<Self, DecodeError> {
+               let mut buf = [0; PUBLIC_KEY_SIZE];
+               reader.read_exact(&mut buf)?;
+               Ok(Self(buf))
+       }
+}
+
+/// Represents the network as nodes and channels between them
+pub struct NetworkGraph {
+       last_rapid_gossip_sync_timestamp: Mutex<Option<u32>>,
+       genesis_hash: BlockHash,
+       // Lock order: channels -> nodes
+       channels: RwLock<BTreeMap<u64, ChannelInfo>>,
+       nodes: RwLock<BTreeMap<NodeId, NodeInfo>>,
+}
+
+impl Clone for NetworkGraph {
+       fn clone(&self) -> Self {
+               let channels = self.channels.read().unwrap();
+               let nodes = self.nodes.read().unwrap();
+               let last_rapid_gossip_sync_timestamp = self.get_last_rapid_gossip_sync_timestamp();
+               Self {
+                       genesis_hash: self.genesis_hash.clone(),
+                       channels: RwLock::new(channels.clone()),
+                       nodes: RwLock::new(nodes.clone()),
+                       last_rapid_gossip_sync_timestamp: Mutex::new(last_rapid_gossip_sync_timestamp)
+               }
+       }
+}
+
+/// A read-only view of [`NetworkGraph`].
+pub struct ReadOnlyNetworkGraph<'a> {
+       channels: RwLockReadGuard<'a, BTreeMap<u64, ChannelInfo>>,
+       nodes: RwLockReadGuard<'a, BTreeMap<NodeId, NodeInfo>>,
+}
+
+/// Update to the [`NetworkGraph`] based on payment failure information conveyed via the Onion
+/// return packet by a node along the route. See [BOLT #4] for details.
+///
+/// [BOLT #4]: https://github.com/lightning/bolts/blob/master/04-onion-routing.md
+#[derive(Clone, Debug, PartialEq)]
+pub enum NetworkUpdate {
+       /// An error indicating a `channel_update` messages should be applied via
+       /// [`NetworkGraph::update_channel`].
+       ChannelUpdateMessage {
+               /// The update to apply via [`NetworkGraph::update_channel`].
+               msg: ChannelUpdate,
+       },
+       /// An error indicating that a channel failed to route a payment, which should be applied via
+       /// [`NetworkGraph::channel_failed`].
+       ChannelFailure {
+               /// The short channel id of the closed channel.
+               short_channel_id: u64,
+               /// Whether the channel should be permanently removed or temporarily disabled until a new
+               /// `channel_update` message is received.
+               is_permanent: bool,
+       },
+       /// An error indicating that a node failed to route a payment, which should be applied via
+       /// [`NetworkGraph::node_failed`].
+       NodeFailure {
+               /// The node id of the failed node.
+               node_id: PublicKey,
+               /// Whether the node should be permanently removed from consideration or can be restored
+               /// when a new `channel_update` message is received.
+               is_permanent: bool,
+       }
+}
+
+impl_writeable_tlv_based_enum_upgradable!(NetworkUpdate,
+       (0, ChannelUpdateMessage) => {
+               (0, msg, required),
+       },
+       (2, ChannelFailure) => {
+               (0, short_channel_id, required),
+               (2, is_permanent, required),
+       },
+       (4, NodeFailure) => {
+               (0, node_id, required),
+               (2, is_permanent, required),
+       },
+);
+
+impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> EventHandler for P2PGossipSync<G, C, L>
+where C::Target: chain::Access, L::Target: Logger {
+       fn handle_event(&self, event: &Event) {
+               if let Event::PaymentPathFailed { payment_hash: _, rejected_by_dest: _, network_update, .. } = event {
+                       if let Some(network_update) = network_update {
+                               self.handle_network_update(network_update);
+                       }
+               }
+       }
+}
+
+/// Receives and validates network updates from peers,
+/// stores authentic and relevant data as a network graph.
+/// This network graph is then used for routing payments.
+/// Provides interface to help with initial routing sync by
+/// serving historical announcements.
+///
+/// Serves as an [`EventHandler`] for applying updates from [`Event::PaymentPathFailed`] to the
+/// [`NetworkGraph`].
+pub struct P2PGossipSync<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref>
+where C::Target: chain::Access, L::Target: Logger
+{
+       secp_ctx: Secp256k1<secp256k1::VerifyOnly>,
+       network_graph: G,
+       chain_access: Option<C>,
+       full_syncs_requested: AtomicUsize,
+       pending_events: Mutex<Vec<MessageSendEvent>>,
+       logger: L,
+}
+
+impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> P2PGossipSync<G, C, L>
+where C::Target: chain::Access, L::Target: Logger
+{
+       /// Creates a new tracker of the actual state of the network of channels and nodes,
+       /// assuming an existing Network Graph.
+       /// Chain monitor is used to make sure announced channels exist on-chain,
+       /// channel data is correct, and that the announcement is signed with
+       /// channel owners' keys.
+       pub fn new(network_graph: G, chain_access: Option<C>, logger: L) -> Self {
+               P2PGossipSync {
+                       secp_ctx: Secp256k1::verification_only(),
+                       network_graph,
+                       full_syncs_requested: AtomicUsize::new(0),
+                       chain_access,
+                       pending_events: Mutex::new(vec![]),
+                       logger,
+               }
+       }
+
+       /// Adds a provider used to check new announcements. Does not affect
+       /// existing announcements unless they are updated.
+       /// Add, update or remove the provider would replace the current one.
+       pub fn add_chain_access(&mut self, chain_access: Option<C>) {
+               self.chain_access = chain_access;
+       }
+
+       /// Gets a reference to the underlying [`NetworkGraph`] which was provided in
+       /// [`P2PGossipSync::new`].
+       ///
+       /// (C-not exported) as bindings don't support a reference-to-a-reference yet
+       pub fn network_graph(&self) -> &G {
+               &self.network_graph
+       }
+
+       /// Returns true when a full routing table sync should be performed with a peer.
+       fn should_request_full_sync(&self, _node_id: &PublicKey) -> bool {
+               //TODO: Determine whether to request a full sync based on the network map.
+               const FULL_SYNCS_TO_REQUEST: usize = 5;
+               if self.full_syncs_requested.load(Ordering::Acquire) < FULL_SYNCS_TO_REQUEST {
+                       self.full_syncs_requested.fetch_add(1, Ordering::AcqRel);
+                       true
+               } else {
+                       false
+               }
+       }
+
+       /// Applies changes to the [`NetworkGraph`] from the given update.
+       fn handle_network_update(&self, update: &NetworkUpdate) {
+               match *update {
+                       NetworkUpdate::ChannelUpdateMessage { ref msg } => {
+                               let short_channel_id = msg.contents.short_channel_id;
+                               let is_enabled = msg.contents.flags & (1 << 1) != (1 << 1);
+                               let status = if is_enabled { "enabled" } else { "disabled" };
+                               log_debug!(self.logger, "Updating channel with channel_update from a payment failure. Channel {} is {}.", short_channel_id, status);
+                               let _ = self.network_graph.update_channel(msg, &self.secp_ctx);
+                       },
+                       NetworkUpdate::ChannelFailure { short_channel_id, is_permanent } => {
+                               let action = if is_permanent { "Removing" } else { "Disabling" };
+                               log_debug!(self.logger, "{} channel graph entry for {} due to a payment failure.", action, short_channel_id);
+                               self.network_graph.channel_failed(short_channel_id, is_permanent);
+                       },
+                       NetworkUpdate::NodeFailure { ref node_id, is_permanent } => {
+                               let action = if is_permanent { "Removing" } else { "Disabling" };
+                               log_debug!(self.logger, "{} node graph entry for {} due to a payment failure.", action, node_id);
+                               self.network_graph.node_failed(node_id, is_permanent);
+                       },
+               }
+       }
+}
+
+macro_rules! secp_verify_sig {
+       ( $secp_ctx: expr, $msg: expr, $sig: expr, $pubkey: expr, $msg_type: expr ) => {
+               match $secp_ctx.verify_ecdsa($msg, $sig, $pubkey) {
+                       Ok(_) => {},
+                       Err(_) => {
+                               return Err(LightningError {
+                                       err: format!("Invalid signature on {} message", $msg_type),
+                                       action: ErrorAction::SendWarningMessage {
+                                               msg: msgs::WarningMessage {
+                                                       channel_id: [0; 32],
+                                                       data: format!("Invalid signature on {} message", $msg_type),
+                                               },
+                                               log_level: Level::Trace,
+                                       },
+                               });
+                       },
+               }
+       };
+}
+
+impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> RoutingMessageHandler for P2PGossipSync<G, C, L>
+where C::Target: chain::Access, L::Target: Logger
+{
+       fn handle_node_announcement(&self, msg: &msgs::NodeAnnouncement) -> Result<bool, LightningError> {
+               self.network_graph.update_node_from_announcement(msg, &self.secp_ctx)?;
+               Ok(msg.contents.excess_data.len() <=  MAX_EXCESS_BYTES_FOR_RELAY &&
+                  msg.contents.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
+                  msg.contents.excess_data.len() + msg.contents.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
+       }
+
+       fn handle_channel_announcement(&self, msg: &msgs::ChannelAnnouncement) -> Result<bool, LightningError> {
+               self.network_graph.update_channel_from_announcement(msg, &self.chain_access, &self.secp_ctx)?;
+               log_gossip!(self.logger, "Added channel_announcement for {}{}", msg.contents.short_channel_id, if !msg.contents.excess_data.is_empty() { " with excess uninterpreted data!" } else { "" });
+               Ok(msg.contents.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
+       }
+
+       fn handle_channel_update(&self, msg: &msgs::ChannelUpdate) -> Result<bool, LightningError> {
+               self.network_graph.update_channel(msg, &self.secp_ctx)?;
+               Ok(msg.contents.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
+       }
+
+       fn get_next_channel_announcements(&self, starting_point: u64, batch_amount: u8) -> Vec<(ChannelAnnouncement, Option<ChannelUpdate>, Option<ChannelUpdate>)> {
+               let mut result = Vec::with_capacity(batch_amount as usize);
+               let channels = self.network_graph.channels.read().unwrap();
+               let mut iter = channels.range(starting_point..);
+               while result.len() < batch_amount as usize {
+                       if let Some((_, ref chan)) = iter.next() {
+                               if chan.announcement_message.is_some() {
+                                       let chan_announcement = chan.announcement_message.clone().unwrap();
+                                       let mut one_to_two_announcement: Option<msgs::ChannelUpdate> = None;
+                                       let mut two_to_one_announcement: Option<msgs::ChannelUpdate> = None;
+                                       if let Some(one_to_two) = chan.one_to_two.as_ref() {
+                                               one_to_two_announcement = one_to_two.last_update_message.clone();
+                                       }
+                                       if let Some(two_to_one) = chan.two_to_one.as_ref() {
+                                               two_to_one_announcement = two_to_one.last_update_message.clone();
+                                       }
+                                       result.push((chan_announcement, one_to_two_announcement, two_to_one_announcement));
+                               } else {
+                                       // TODO: We may end up sending un-announced channel_updates if we are sending
+                                       // initial sync data while receiving announce/updates for this channel.
+                               }
+                       } else {
+                               return result;
+                       }
+               }
+               result
+       }
+
+       fn get_next_node_announcements(&self, starting_point: Option<&PublicKey>, batch_amount: u8) -> Vec<NodeAnnouncement> {
+               let mut result = Vec::with_capacity(batch_amount as usize);
+               let nodes = self.network_graph.nodes.read().unwrap();
+               let mut iter = if let Some(pubkey) = starting_point {
+                               let mut iter = nodes.range(NodeId::from_pubkey(pubkey)..);
+                               iter.next();
+                               iter
+                       } else {
+                               nodes.range::<NodeId, _>(..)
+                       };
+               while result.len() < batch_amount as usize {
+                       if let Some((_, ref node)) = iter.next() {
+                               if let Some(node_info) = node.announcement_info.as_ref() {
+                                       if node_info.announcement_message.is_some() {
+                                               result.push(node_info.announcement_message.clone().unwrap());
+                                       }
+                               }
+                       } else {
+                               return result;
+                       }
+               }
+               result
+       }
+
+       /// Initiates a stateless sync of routing gossip information with a peer
+       /// using gossip_queries. The default strategy used by this implementation
+       /// is to sync the full block range with several peers.
+       ///
+       /// We should expect one or more reply_channel_range messages in response
+       /// to our query_channel_range. Each reply will enqueue a query_scid message
+       /// to request gossip messages for each channel. The sync is considered complete
+       /// when the final reply_scids_end message is received, though we are not
+       /// tracking this directly.
+       fn peer_connected(&self, their_node_id: &PublicKey, init_msg: &Init) {
+               // We will only perform a sync with peers that support gossip_queries.
+               if !init_msg.features.supports_gossip_queries() {
+                       return ();
+               }
+
+               // The lightning network's gossip sync system is completely broken in numerous ways.
+               //
+               // Given no broadly-available set-reconciliation protocol, the only reasonable approach is
+               // to do a full sync from the first few peers we connect to, and then receive gossip
+               // updates from all our peers normally.
+               //
+               // Originally, we could simply tell a peer to dump us the entire gossip table on startup,
+               // wasting lots of bandwidth but ensuring we have the full network graph. After the initial
+               // dump peers would always send gossip and we'd stay up-to-date with whatever our peer has
+               // seen.
+               //
+               // In order to reduce the bandwidth waste, "gossip queries" were introduced, allowing you
+               // to ask for the SCIDs of all channels in your peer's routing graph, and then only request
+               // channel data which you are missing. Except there was no way at all to identify which
+               // `channel_update`s you were missing, so you still had to request everything, just in a
+               // very complicated way with some queries instead of just getting the dump.
+               //
+               // Later, an option was added to fetch the latest timestamps of the `channel_update`s to
+               // make efficient sync possible, however it has yet to be implemented in lnd, which makes
+               // relying on it useless.
+               //
+               // After gossip queries were introduced, support for receiving a full gossip table dump on
+               // connection was removed from several nodes, making it impossible to get a full sync
+               // without using the "gossip queries" messages.
+               //
+               // Once you opt into "gossip queries" the only way to receive any gossip updates that a
+               // peer receives after you connect, you must send a `gossip_timestamp_filter` message. This
+               // message, as the name implies, tells the peer to not forward any gossip messages with a
+               // timestamp older than a given value (not the time the peer received the filter, but the
+               // timestamp in the update message, which is often hours behind when the peer received the
+               // message).
+               //
+               // Obnoxiously, `gossip_timestamp_filter` isn't *just* a filter, but its also a request for
+               // your peer to send you the full routing graph (subject to the filter). Thus, in order to
+               // tell a peer to send you any updates as it sees them, you have to also ask for the full
+               // routing graph to be synced. If you set a timestamp filter near the current time, peers
+               // will simply not forward any new updates they see to you which were generated some time
+               // ago (which is not uncommon). If you instead set a timestamp filter near 0 (or two weeks
+               // ago), you will always get the full routing graph from all your peers.
+               //
+               // Most lightning nodes today opt to simply turn off receiving gossip data which only
+               // propagated some time after it was generated, and, worse, often disable gossiping with
+               // several peers after their first connection. The second behavior can cause gossip to not
+               // propagate fully if there are cuts in the gossiping subgraph.
+               //
+               // In an attempt to cut a middle ground between always fetching the full graph from all of
+               // our peers and never receiving gossip from peers at all, we send all of our peers a
+               // `gossip_timestamp_filter`, with the filter time set either two weeks ago or an hour ago.
+               //
+               // For no-std builds, we bury our head in the sand and do a full sync on each connection.
+               let should_request_full_sync = self.should_request_full_sync(&their_node_id);
+               #[allow(unused_mut, unused_assignments)]
+               let mut gossip_start_time = 0;
+               #[cfg(feature = "std")]
+               {
+                       gossip_start_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+                       if should_request_full_sync {
+                               gossip_start_time -= 60 * 60 * 24 * 7 * 2; // 2 weeks ago
+                       } else {
+                               gossip_start_time -= 60 * 60; // an hour ago
+                       }
+               }
+
+               let mut pending_events = self.pending_events.lock().unwrap();
+               pending_events.push(MessageSendEvent::SendGossipTimestampFilter {
+                       node_id: their_node_id.clone(),
+                       msg: GossipTimestampFilter {
+                               chain_hash: self.network_graph.genesis_hash,
+                               first_timestamp: gossip_start_time as u32, // 2106 issue!
+                               timestamp_range: u32::max_value(),
+                       },
+               });
+       }
+
+       fn handle_reply_channel_range(&self, _their_node_id: &PublicKey, _msg: ReplyChannelRange) -> Result<(), LightningError> {
+               // We don't make queries, so should never receive replies. If, in the future, the set
+               // reconciliation extensions to gossip queries become broadly supported, we should revert
+               // this code to its state pre-0.0.106.
+               Ok(())
+       }
+
+       fn handle_reply_short_channel_ids_end(&self, _their_node_id: &PublicKey, _msg: ReplyShortChannelIdsEnd) -> Result<(), LightningError> {
+               // We don't make queries, so should never receive replies. If, in the future, the set
+               // reconciliation extensions to gossip queries become broadly supported, we should revert
+               // this code to its state pre-0.0.106.
+               Ok(())
+       }
+
+       /// Processes a query from a peer by finding announced/public channels whose funding UTXOs
+       /// are in the specified block range. Due to message size limits, large range
+       /// queries may result in several reply messages. This implementation enqueues
+       /// all reply messages into pending events. Each message will allocate just under 65KiB. A full
+       /// sync of the public routing table with 128k channels will generated 16 messages and allocate ~1MB.
+       /// Logic can be changed to reduce allocation if/when a full sync of the routing table impacts
+       /// memory constrained systems.
+       fn handle_query_channel_range(&self, their_node_id: &PublicKey, msg: QueryChannelRange) -> Result<(), LightningError> {
+               log_debug!(self.logger, "Handling query_channel_range peer={}, first_blocknum={}, number_of_blocks={}", log_pubkey!(their_node_id), msg.first_blocknum, msg.number_of_blocks);
+
+               let inclusive_start_scid = scid_from_parts(msg.first_blocknum as u64, 0, 0);
+
+               // We might receive valid queries with end_blocknum that would overflow SCID conversion.
+               // If so, we manually cap the ending block to avoid this overflow.
+               let exclusive_end_scid = scid_from_parts(cmp::min(msg.end_blocknum() as u64, MAX_SCID_BLOCK), 0, 0);
+
+               // Per spec, we must reply to a query. Send an empty message when things are invalid.
+               if msg.chain_hash != self.network_graph.genesis_hash || inclusive_start_scid.is_err() || exclusive_end_scid.is_err() || msg.number_of_blocks == 0 {
+                       let mut pending_events = self.pending_events.lock().unwrap();
+                       pending_events.push(MessageSendEvent::SendReplyChannelRange {
+                               node_id: their_node_id.clone(),
+                               msg: ReplyChannelRange {
+                                       chain_hash: msg.chain_hash.clone(),
+                                       first_blocknum: msg.first_blocknum,
+                                       number_of_blocks: msg.number_of_blocks,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![],
+                               }
+                       });
+                       return Err(LightningError {
+                               err: String::from("query_channel_range could not be processed"),
+                               action: ErrorAction::IgnoreError,
+                       });
+               }
+
+               // Creates channel batches. We are not checking if the channel is routable
+               // (has at least one update). A peer may still want to know the channel
+               // exists even if its not yet routable.
+               let mut batches: Vec<Vec<u64>> = vec![Vec::with_capacity(MAX_SCIDS_PER_REPLY)];
+               let channels = self.network_graph.channels.read().unwrap();
+               for (_, ref chan) in channels.range(inclusive_start_scid.unwrap()..exclusive_end_scid.unwrap()) {
+                       if let Some(chan_announcement) = &chan.announcement_message {
+                               // Construct a new batch if last one is full
+                               if batches.last().unwrap().len() == batches.last().unwrap().capacity() {
+                                       batches.push(Vec::with_capacity(MAX_SCIDS_PER_REPLY));
+                               }
+
+                               let batch = batches.last_mut().unwrap();
+                               batch.push(chan_announcement.contents.short_channel_id);
+                       }
+               }
+               drop(channels);
+
+               let mut pending_events = self.pending_events.lock().unwrap();
+               let batch_count = batches.len();
+               let mut prev_batch_endblock = msg.first_blocknum;
+               for (batch_index, batch) in batches.into_iter().enumerate() {
+                       // Per spec, the initial `first_blocknum` needs to be <= the query's `first_blocknum`
+                       // and subsequent `first_blocknum`s must be >= the prior reply's `first_blocknum`.
+                       //
+                       // Additionally, c-lightning versions < 0.10 require that the `first_blocknum` of each
+                       // reply is >= the previous reply's `first_blocknum` and either exactly the previous
+                       // reply's `first_blocknum + number_of_blocks` or exactly one greater. This is a
+                       // significant diversion from the requirements set by the spec, and, in case of blocks
+                       // with no channel opens (e.g. empty blocks), requires that we use the previous value
+                       // and *not* derive the first_blocknum from the actual first block of the reply.
+                       let first_blocknum = prev_batch_endblock;
+
+                       // Each message carries the number of blocks (from the `first_blocknum`) its contents
+                       // fit in. Though there is no requirement that we use exactly the number of blocks its
+                       // contents are from, except for the bogus requirements c-lightning enforces, above.
+                       //
+                       // Per spec, the last end block (ie `first_blocknum + number_of_blocks`) needs to be
+                       // >= the query's end block. Thus, for the last reply, we calculate the difference
+                       // between the query's end block and the start of the reply.
+                       //
+                       // Overflow safe since end_blocknum=msg.first_block_num+msg.number_of_blocks and
+                       // first_blocknum will be either msg.first_blocknum or a higher block height.
+                       let (sync_complete, number_of_blocks) = if batch_index == batch_count-1 {
+                               (true, msg.end_blocknum() - first_blocknum)
+                       }
+                       // Prior replies should use the number of blocks that fit into the reply. Overflow
+                       // safe since first_blocknum is always <= last SCID's block.
+                       else {
+                               (false, block_from_scid(batch.last().unwrap()) - first_blocknum)
+                       };
+
+                       prev_batch_endblock = first_blocknum + number_of_blocks;
+
+                       pending_events.push(MessageSendEvent::SendReplyChannelRange {
+                               node_id: their_node_id.clone(),
+                               msg: ReplyChannelRange {
+                                       chain_hash: msg.chain_hash.clone(),
+                                       first_blocknum,
+                                       number_of_blocks,
+                                       sync_complete,
+                                       short_channel_ids: batch,
+                               }
+                       });
+               }
+
+               Ok(())
+       }
+
+       fn handle_query_short_channel_ids(&self, _their_node_id: &PublicKey, _msg: QueryShortChannelIds) -> Result<(), LightningError> {
+               // TODO
+               Err(LightningError {
+                       err: String::from("Not implemented"),
+                       action: ErrorAction::IgnoreError,
+               })
+       }
+}
+
+impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> MessageSendEventsProvider for P2PGossipSync<G, C, L>
+where
+       C::Target: chain::Access,
+       L::Target: Logger,
+{
+       fn get_and_clear_pending_msg_events(&self) -> Vec<MessageSendEvent> {
+               let mut ret = Vec::new();
+               let mut pending_events = self.pending_events.lock().unwrap();
+               core::mem::swap(&mut ret, &mut pending_events);
+               ret
+       }
+}
+
+#[derive(Clone, Debug, PartialEq)]
+/// Details about one direction of a channel as received within a [`ChannelUpdate`].
+pub struct ChannelUpdateInfo {
+       /// When the last update to the channel direction was issued.
+       /// Value is opaque, as set in the announcement.
+       pub last_update: u32,
+       /// Whether the channel can be currently used for payments (in this one direction).
+       pub enabled: bool,
+       /// The difference in CLTV values that you must have when routing through this channel.
+       pub cltv_expiry_delta: u16,
+       /// The minimum value, which must be relayed to the next hop via the channel
+       pub htlc_minimum_msat: u64,
+       /// The maximum value which may be relayed to the next hop via the channel.
+       pub htlc_maximum_msat: Option<u64>,
+       /// Fees charged when the channel is used for routing
+       pub fees: RoutingFees,
+       /// Most recent update for the channel received from the network
+       /// Mostly redundant with the data we store in fields explicitly.
+       /// Everything else is useful only for sending out for initial routing sync.
+       /// Not stored if contains excess data to prevent DoS.
+       pub last_update_message: Option<ChannelUpdate>,
+}
+
+impl fmt::Display for ChannelUpdateInfo {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               write!(f, "last_update {}, enabled {}, cltv_expiry_delta {}, htlc_minimum_msat {}, fees {:?}", self.last_update, self.enabled, self.cltv_expiry_delta, self.htlc_minimum_msat, self.fees)?;
+               Ok(())
+       }
+}
+
+impl_writeable_tlv_based!(ChannelUpdateInfo, {
+       (0, last_update, required),
+       (2, enabled, required),
+       (4, cltv_expiry_delta, required),
+       (6, htlc_minimum_msat, required),
+       (8, htlc_maximum_msat, required),
+       (10, fees, required),
+       (12, last_update_message, required),
+});
+
+#[derive(Clone, Debug, PartialEq)]
+/// Details about a channel (both directions).
+/// Received within a channel announcement.
+pub struct ChannelInfo {
+       /// Protocol features of a channel communicated during its announcement
+       pub features: ChannelFeatures,
+       /// Source node of the first direction of a channel
+       pub node_one: NodeId,
+       /// Details about the first direction of a channel
+       pub one_to_two: Option<ChannelUpdateInfo>,
+       /// Source node of the second direction of a channel
+       pub node_two: NodeId,
+       /// Details about the second direction of a channel
+       pub two_to_one: Option<ChannelUpdateInfo>,
+       /// The channel capacity as seen on-chain, if chain lookup is available.
+       pub capacity_sats: Option<u64>,
+       /// An initial announcement of the channel
+       /// Mostly redundant with the data we store in fields explicitly.
+       /// Everything else is useful only for sending out for initial routing sync.
+       /// Not stored if contains excess data to prevent DoS.
+       pub announcement_message: Option<ChannelAnnouncement>,
+       /// The timestamp when we received the announcement, if we are running with feature = "std"
+       /// (which we can probably assume we are - no-std environments probably won't have a full
+       /// network graph in memory!).
+       announcement_received_time: u64,
+}
+
+impl ChannelInfo {
+       /// Returns a [`DirectedChannelInfo`] for the channel directed to the given `target` from a
+       /// returned `source`, or `None` if `target` is not one of the channel's counterparties.
+       pub fn as_directed_to(&self, target: &NodeId) -> Option<(DirectedChannelInfo, &NodeId)> {
+               let (direction, source) = {
+                       if target == &self.node_one {
+                               (self.two_to_one.as_ref(), &self.node_two)
+                       } else if target == &self.node_two {
+                               (self.one_to_two.as_ref(), &self.node_one)
+                       } else {
+                               return None;
+                       }
+               };
+               Some((DirectedChannelInfo::new(self, direction), source))
+       }
+
+       /// Returns a [`DirectedChannelInfo`] for the channel directed from the given `source` to a
+       /// returned `target`, or `None` if `source` is not one of the channel's counterparties.
+       pub fn as_directed_from(&self, source: &NodeId) -> Option<(DirectedChannelInfo, &NodeId)> {
+               let (direction, target) = {
+                       if source == &self.node_one {
+                               (self.one_to_two.as_ref(), &self.node_two)
+                       } else if source == &self.node_two {
+                               (self.two_to_one.as_ref(), &self.node_one)
+                       } else {
+                               return None;
+                       }
+               };
+               Some((DirectedChannelInfo::new(self, direction), target))
+       }
+
+       /// Returns a [`ChannelUpdateInfo`] based on the direction implied by the channel_flag.
+       pub fn get_directional_info(&self, channel_flags: u8) -> Option<&ChannelUpdateInfo> {
+               let direction = channel_flags & 1u8;
+               if direction == 0 {
+                       self.one_to_two.as_ref()
+               } else {
+                       self.two_to_one.as_ref()
+               }
+       }
+}
+
+impl fmt::Display for ChannelInfo {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               write!(f, "features: {}, node_one: {}, one_to_two: {:?}, node_two: {}, two_to_one: {:?}",
+                  log_bytes!(self.features.encode()), log_bytes!(self.node_one.as_slice()), self.one_to_two, log_bytes!(self.node_two.as_slice()), self.two_to_one)?;
+               Ok(())
+       }
+}
+
+impl_writeable_tlv_based!(ChannelInfo, {
+       (0, features, required),
+       (1, announcement_received_time, (default_value, 0)),
+       (2, node_one, required),
+       (4, one_to_two, required),
+       (6, node_two, required),
+       (8, two_to_one, required),
+       (10, capacity_sats, required),
+       (12, announcement_message, required),
+});
+
+/// A wrapper around [`ChannelInfo`] representing information about the channel as directed from a
+/// source node to a target node.
+#[derive(Clone)]
+pub struct DirectedChannelInfo<'a> {
+       channel: &'a ChannelInfo,
+       direction: Option<&'a ChannelUpdateInfo>,
+       htlc_maximum_msat: u64,
+       effective_capacity: EffectiveCapacity,
+}
+
+impl<'a> DirectedChannelInfo<'a> {
+       #[inline]
+       fn new(channel: &'a ChannelInfo, direction: Option<&'a ChannelUpdateInfo>) -> Self {
+               let htlc_maximum_msat = direction.and_then(|direction| direction.htlc_maximum_msat);
+               let capacity_msat = channel.capacity_sats.map(|capacity_sats| capacity_sats * 1000);
+
+               let (htlc_maximum_msat, effective_capacity) = match (htlc_maximum_msat, capacity_msat) {
+                       (Some(amount_msat), Some(capacity_msat)) => {
+                               let htlc_maximum_msat = cmp::min(amount_msat, capacity_msat);
+                               (htlc_maximum_msat, EffectiveCapacity::Total { capacity_msat })
+                       },
+                       (Some(amount_msat), None) => {
+                               (amount_msat, EffectiveCapacity::MaximumHTLC { amount_msat })
+                       },
+                       (None, Some(capacity_msat)) => {
+                               (capacity_msat, EffectiveCapacity::Total { capacity_msat })
+                       },
+                       (None, None) => (EffectiveCapacity::Unknown.as_msat(), EffectiveCapacity::Unknown),
+               };
+
+               Self {
+                       channel, direction, htlc_maximum_msat, effective_capacity
+               }
+       }
+
+       /// Returns information for the channel.
+       pub fn channel(&self) -> &'a ChannelInfo { self.channel }
+
+       /// Returns information for the direction.
+       pub fn direction(&self) -> Option<&'a ChannelUpdateInfo> { self.direction }
+
+       /// Returns the maximum HTLC amount allowed over the channel in the direction.
+       pub fn htlc_maximum_msat(&self) -> u64 {
+               self.htlc_maximum_msat
+       }
+
+       /// Returns the [`EffectiveCapacity`] of the channel in the direction.
+       ///
+       /// This is either the total capacity from the funding transaction, if known, or the
+       /// `htlc_maximum_msat` for the direction as advertised by the gossip network, if known,
+       /// otherwise.
+       pub fn effective_capacity(&self) -> EffectiveCapacity {
+               self.effective_capacity
+       }
+
+       /// Returns `Some` if [`ChannelUpdateInfo`] is available in the direction.
+       pub(super) fn with_update(self) -> Option<DirectedChannelInfoWithUpdate<'a>> {
+               match self.direction {
+                       Some(_) => Some(DirectedChannelInfoWithUpdate { inner: self }),
+                       None => None,
+               }
+       }
+}
+
+impl<'a> fmt::Debug for DirectedChannelInfo<'a> {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               f.debug_struct("DirectedChannelInfo")
+                       .field("channel", &self.channel)
+                       .finish()
+       }
+}
+
+/// A [`DirectedChannelInfo`] with [`ChannelUpdateInfo`] available in its direction.
+#[derive(Clone)]
+pub(super) struct DirectedChannelInfoWithUpdate<'a> {
+       inner: DirectedChannelInfo<'a>,
+}
+
+impl<'a> DirectedChannelInfoWithUpdate<'a> {
+       /// Returns information for the channel.
+       #[inline]
+       pub(super) fn channel(&self) -> &'a ChannelInfo { &self.inner.channel }
+
+       /// Returns information for the direction.
+       #[inline]
+       pub(super) fn direction(&self) -> &'a ChannelUpdateInfo { self.inner.direction.unwrap() }
+
+       /// Returns the [`EffectiveCapacity`] of the channel in the direction.
+       #[inline]
+       pub(super) fn effective_capacity(&self) -> EffectiveCapacity { self.inner.effective_capacity() }
+
+       /// Returns the maximum HTLC amount allowed over the channel in the direction.
+       #[inline]
+       pub(super) fn htlc_maximum_msat(&self) -> u64 { self.inner.htlc_maximum_msat() }
+}
+
+impl<'a> fmt::Debug for DirectedChannelInfoWithUpdate<'a> {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               self.inner.fmt(f)
+       }
+}
+
+/// The effective capacity of a channel for routing purposes.
+///
+/// While this may be smaller than the actual channel capacity, amounts greater than
+/// [`Self::as_msat`] should not be routed through the channel.
+#[derive(Clone, Copy)]
+pub enum EffectiveCapacity {
+       /// The available liquidity in the channel known from being a channel counterparty, and thus a
+       /// direct hop.
+       ExactLiquidity {
+               /// Either the inbound or outbound liquidity depending on the direction, denominated in
+               /// millisatoshi.
+               liquidity_msat: u64,
+       },
+       /// The maximum HTLC amount in one direction as advertised on the gossip network.
+       MaximumHTLC {
+               /// The maximum HTLC amount denominated in millisatoshi.
+               amount_msat: u64,
+       },
+       /// The total capacity of the channel as determined by the funding transaction.
+       Total {
+               /// The funding amount denominated in millisatoshi.
+               capacity_msat: u64,
+       },
+       /// A capacity sufficient to route any payment, typically used for private channels provided by
+       /// an invoice.
+       Infinite,
+       /// A capacity that is unknown possibly because either the chain state is unavailable to know
+       /// the total capacity or the `htlc_maximum_msat` was not advertised on the gossip network.
+       Unknown,
+}
+
+/// The presumed channel capacity denominated in millisatoshi for [`EffectiveCapacity::Unknown`] to
+/// use when making routing decisions.
+pub const UNKNOWN_CHANNEL_CAPACITY_MSAT: u64 = 250_000 * 1000;
+
+impl EffectiveCapacity {
+       /// Returns the effective capacity denominated in millisatoshi.
+       pub fn as_msat(&self) -> u64 {
+               match self {
+                       EffectiveCapacity::ExactLiquidity { liquidity_msat } => *liquidity_msat,
+                       EffectiveCapacity::MaximumHTLC { amount_msat } => *amount_msat,
+                       EffectiveCapacity::Total { capacity_msat } => *capacity_msat,
+                       EffectiveCapacity::Infinite => u64::max_value(),
+                       EffectiveCapacity::Unknown => UNKNOWN_CHANNEL_CAPACITY_MSAT,
+               }
+       }
+}
+
+/// Fees for routing via a given channel or a node
+#[derive(Eq, PartialEq, Copy, Clone, Debug, Hash)]
+pub struct RoutingFees {
+       /// Flat routing fee in satoshis
+       pub base_msat: u32,
+       /// Liquidity-based routing fee in millionths of a routed amount.
+       /// In other words, 10000 is 1%.
+       pub proportional_millionths: u32,
+}
+
+impl_writeable_tlv_based!(RoutingFees, {
+       (0, base_msat, required),
+       (2, proportional_millionths, required)
+});
+
+#[derive(Clone, Debug, PartialEq)]
+/// Information received in the latest node_announcement from this node.
+pub struct NodeAnnouncementInfo {
+       /// Protocol features the node announced support for
+       pub features: NodeFeatures,
+       /// When the last known update to the node state was issued.
+       /// Value is opaque, as set in the announcement.
+       pub last_update: u32,
+       /// Color assigned to the node
+       pub rgb: [u8; 3],
+       /// Moniker assigned to the node.
+       /// May be invalid or malicious (eg control chars),
+       /// should not be exposed to the user.
+       pub alias: [u8; 32],
+       /// Internet-level addresses via which one can connect to the node
+       pub addresses: Vec<NetAddress>,
+       /// An initial announcement of the node
+       /// Mostly redundant with the data we store in fields explicitly.
+       /// Everything else is useful only for sending out for initial routing sync.
+       /// Not stored if contains excess data to prevent DoS.
+       pub announcement_message: Option<NodeAnnouncement>
+}
+
+impl_writeable_tlv_based!(NodeAnnouncementInfo, {
+       (0, features, required),
+       (2, last_update, required),
+       (4, rgb, required),
+       (6, alias, required),
+       (8, announcement_message, option),
+       (10, addresses, vec_type),
+});
+
+#[derive(Clone, Debug, PartialEq)]
+/// Details about a node in the network, known from the network announcement.
+pub struct NodeInfo {
+       /// All valid channels a node has announced
+       pub channels: Vec<u64>,
+       /// Lowest fees enabling routing via any of the enabled, known channels to a node.
+       /// The two fields (flat and proportional fee) are independent,
+       /// meaning they don't have to refer to the same channel.
+       pub lowest_inbound_channel_fees: Option<RoutingFees>,
+       /// More information about a node from node_announcement.
+       /// Optional because we store a Node entry after learning about it from
+       /// a channel announcement, but before receiving a node announcement.
+       pub announcement_info: Option<NodeAnnouncementInfo>
+}
+
+impl fmt::Display for NodeInfo {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               write!(f, "lowest_inbound_channel_fees: {:?}, channels: {:?}, announcement_info: {:?}",
+                  self.lowest_inbound_channel_fees, &self.channels[..], self.announcement_info)?;
+               Ok(())
+       }
+}
+
+impl_writeable_tlv_based!(NodeInfo, {
+       (0, lowest_inbound_channel_fees, option),
+       (2, announcement_info, option),
+       (4, channels, vec_type),
+});
+
+const SERIALIZATION_VERSION: u8 = 1;
+const MIN_SERIALIZATION_VERSION: u8 = 1;
+
+impl Writeable for NetworkGraph {
+       fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
+               write_ver_prefix!(writer, SERIALIZATION_VERSION, MIN_SERIALIZATION_VERSION);
+
+               self.genesis_hash.write(writer)?;
+               let channels = self.channels.read().unwrap();
+               (channels.len() as u64).write(writer)?;
+               for (ref chan_id, ref chan_info) in channels.iter() {
+                       (*chan_id).write(writer)?;
+                       chan_info.write(writer)?;
+               }
+               let nodes = self.nodes.read().unwrap();
+               (nodes.len() as u64).write(writer)?;
+               for (ref node_id, ref node_info) in nodes.iter() {
+                       node_id.write(writer)?;
+                       node_info.write(writer)?;
+               }
+
+               let last_rapid_gossip_sync_timestamp = self.get_last_rapid_gossip_sync_timestamp();
+               write_tlv_fields!(writer, {
+                       (1, last_rapid_gossip_sync_timestamp, option),
+               });
+               Ok(())
+       }
+}
+
+impl Readable for NetworkGraph {
+       fn read<R: io::Read>(reader: &mut R) -> Result<NetworkGraph, DecodeError> {
+               let _ver = read_ver_prefix!(reader, SERIALIZATION_VERSION);
+
+               let genesis_hash: BlockHash = Readable::read(reader)?;
+               let channels_count: u64 = Readable::read(reader)?;
+               let mut channels = BTreeMap::new();
+               for _ in 0..channels_count {
+                       let chan_id: u64 = Readable::read(reader)?;
+                       let chan_info = Readable::read(reader)?;
+                       channels.insert(chan_id, chan_info);
+               }
+               let nodes_count: u64 = Readable::read(reader)?;
+               let mut nodes = BTreeMap::new();
+               for _ in 0..nodes_count {
+                       let node_id = Readable::read(reader)?;
+                       let node_info = Readable::read(reader)?;
+                       nodes.insert(node_id, node_info);
+               }
+
+               let mut last_rapid_gossip_sync_timestamp: Option<u32> = None;
+               read_tlv_fields!(reader, {
+                       (1, last_rapid_gossip_sync_timestamp, option),
+               });
+
+               Ok(NetworkGraph {
+                       genesis_hash,
+                       channels: RwLock::new(channels),
+                       nodes: RwLock::new(nodes),
+                       last_rapid_gossip_sync_timestamp: Mutex::new(last_rapid_gossip_sync_timestamp),
+               })
+       }
+}
+
+impl fmt::Display for NetworkGraph {
+       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
+               writeln!(f, "Network map\n[Channels]")?;
+               for (key, val) in self.channels.read().unwrap().iter() {
+                       writeln!(f, " {}: {}", key, val)?;
+               }
+               writeln!(f, "[Nodes]")?;
+               for (&node_id, val) in self.nodes.read().unwrap().iter() {
+                       writeln!(f, " {}: {}", log_bytes!(node_id.as_slice()), val)?;
+               }
+               Ok(())
+       }
+}
+
+impl PartialEq for NetworkGraph {
+       fn eq(&self, other: &Self) -> bool {
+               self.genesis_hash == other.genesis_hash &&
+                       *self.channels.read().unwrap() == *other.channels.read().unwrap() &&
+                       *self.nodes.read().unwrap() == *other.nodes.read().unwrap()
+       }
+}
+
+impl NetworkGraph {
+       /// Creates a new, empty, network graph.
+       pub fn new(genesis_hash: BlockHash) -> NetworkGraph {
+               Self {
+                       genesis_hash,
+                       channels: RwLock::new(BTreeMap::new()),
+                       nodes: RwLock::new(BTreeMap::new()),
+                       last_rapid_gossip_sync_timestamp: Mutex::new(None),
+               }
+       }
+
+       /// Returns a read-only view of the network graph.
+       pub fn read_only(&'_ self) -> ReadOnlyNetworkGraph<'_> {
+               let channels = self.channels.read().unwrap();
+               let nodes = self.nodes.read().unwrap();
+               ReadOnlyNetworkGraph {
+                       channels,
+                       nodes,
+               }
+       }
+
+       /// The unix timestamp provided by the most recent rapid gossip sync.
+       /// It will be set by the rapid sync process after every sync completion.
+       pub fn get_last_rapid_gossip_sync_timestamp(&self) -> Option<u32> {
+               self.last_rapid_gossip_sync_timestamp.lock().unwrap().clone()
+       }
+
+       /// Update the unix timestamp provided by the most recent rapid gossip sync.
+       /// This should be done automatically by the rapid sync process after every sync completion.
+       pub fn set_last_rapid_gossip_sync_timestamp(&self, last_rapid_gossip_sync_timestamp: u32) {
+               self.last_rapid_gossip_sync_timestamp.lock().unwrap().replace(last_rapid_gossip_sync_timestamp);
+       }
+
+       /// Clears the `NodeAnnouncementInfo` field for all nodes in the `NetworkGraph` for testing
+       /// purposes.
+       #[cfg(test)]
+       pub fn clear_nodes_announcement_info(&self) {
+               for node in self.nodes.write().unwrap().iter_mut() {
+                       node.1.announcement_info = None;
+               }
+       }
+
+       /// For an already known node (from channel announcements), update its stored properties from a
+       /// given node announcement.
+       ///
+       /// You probably don't want to call this directly, instead relying on a P2PGossipSync's
+       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
+       /// routing messages from a source using a protocol other than the lightning P2P protocol.
+       pub fn update_node_from_announcement<T: secp256k1::Verification>(&self, msg: &msgs::NodeAnnouncement, secp_ctx: &Secp256k1<T>) -> Result<(), LightningError> {
+               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.contents.encode()[..])[..]);
+               secp_verify_sig!(secp_ctx, &msg_hash, &msg.signature, &msg.contents.node_id, "node_announcement");
+               self.update_node_from_announcement_intern(&msg.contents, Some(&msg))
+       }
+
+       /// For an already known node (from channel announcements), update its stored properties from a
+       /// given node announcement without verifying the associated signatures. Because we aren't
+       /// given the associated signatures here we cannot relay the node announcement to any of our
+       /// peers.
+       pub fn update_node_from_unsigned_announcement(&self, msg: &msgs::UnsignedNodeAnnouncement) -> Result<(), LightningError> {
+               self.update_node_from_announcement_intern(msg, None)
+       }
+
+       fn update_node_from_announcement_intern(&self, msg: &msgs::UnsignedNodeAnnouncement, full_msg: Option<&msgs::NodeAnnouncement>) -> Result<(), LightningError> {
+               match self.nodes.write().unwrap().get_mut(&NodeId::from_pubkey(&msg.node_id)) {
+                       None => Err(LightningError{err: "No existing channels for node_announcement".to_owned(), action: ErrorAction::IgnoreError}),
+                       Some(node) => {
+                               if let Some(node_info) = node.announcement_info.as_ref() {
+                                       // The timestamp field is somewhat of a misnomer - the BOLTs use it to order
+                                       // updates to ensure you always have the latest one, only vaguely suggesting
+                                       // that it be at least the current time.
+                                       if node_info.last_update  > msg.timestamp {
+                                               return Err(LightningError{err: "Update older than last processed update".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
+                                       } else if node_info.last_update  == msg.timestamp {
+                                               return Err(LightningError{err: "Update had the same timestamp as last processed update".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
+                                       }
+                               }
+
+                               let should_relay =
+                                       msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
+                                       msg.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
+                                       msg.excess_data.len() + msg.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY;
+                               node.announcement_info = Some(NodeAnnouncementInfo {
+                                       features: msg.features.clone(),
+                                       last_update: msg.timestamp,
+                                       rgb: msg.rgb,
+                                       alias: msg.alias,
+                                       addresses: msg.addresses.clone(),
+                                       announcement_message: if should_relay { full_msg.cloned() } else { None },
+                               });
+
+                               Ok(())
+                       }
+               }
+       }
+
+       /// Store or update channel info from a channel announcement.
+       ///
+       /// You probably don't want to call this directly, instead relying on a P2PGossipSync's
+       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
+       /// routing messages from a source using a protocol other than the lightning P2P protocol.
+       ///
+       /// If a `chain::Access` object is provided via `chain_access`, it will be called to verify
+       /// the corresponding UTXO exists on chain and is correctly-formatted.
+       pub fn update_channel_from_announcement<T: secp256k1::Verification, C: Deref>(
+               &self, msg: &msgs::ChannelAnnouncement, chain_access: &Option<C>, secp_ctx: &Secp256k1<T>
+       ) -> Result<(), LightningError>
+       where
+               C::Target: chain::Access,
+       {
+               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.contents.encode()[..])[..]);
+               secp_verify_sig!(secp_ctx, &msg_hash, &msg.node_signature_1, &msg.contents.node_id_1, "channel_announcement");
+               secp_verify_sig!(secp_ctx, &msg_hash, &msg.node_signature_2, &msg.contents.node_id_2, "channel_announcement");
+               secp_verify_sig!(secp_ctx, &msg_hash, &msg.bitcoin_signature_1, &msg.contents.bitcoin_key_1, "channel_announcement");
+               secp_verify_sig!(secp_ctx, &msg_hash, &msg.bitcoin_signature_2, &msg.contents.bitcoin_key_2, "channel_announcement");
+               self.update_channel_from_unsigned_announcement_intern(&msg.contents, Some(msg), chain_access)
+       }
+
+       /// Store or update channel info from a channel announcement without verifying the associated
+       /// signatures. Because we aren't given the associated signatures here we cannot relay the
+       /// channel announcement to any of our peers.
+       ///
+       /// If a `chain::Access` object is provided via `chain_access`, it will be called to verify
+       /// the corresponding UTXO exists on chain and is correctly-formatted.
+       pub fn update_channel_from_unsigned_announcement<C: Deref>(
+               &self, msg: &msgs::UnsignedChannelAnnouncement, chain_access: &Option<C>
+       ) -> Result<(), LightningError>
+       where
+               C::Target: chain::Access,
+       {
+               self.update_channel_from_unsigned_announcement_intern(msg, None, chain_access)
+       }
+
+       /// Update channel from partial announcement data received via rapid gossip sync
+       ///
+       /// `timestamp: u64`: Timestamp emulating the backdated original announcement receipt (by the
+       /// rapid gossip sync server)
+       ///
+       /// All other parameters as used in [`msgs::UnsignedChannelAnnouncement`] fields.
+       pub fn add_channel_from_partial_announcement(&self, short_channel_id: u64, timestamp: u64, features: ChannelFeatures, node_id_1: PublicKey, node_id_2: PublicKey) -> Result<(), LightningError> {
+               if node_id_1 == node_id_2 {
+                       return Err(LightningError{err: "Channel announcement node had a channel with itself".to_owned(), action: ErrorAction::IgnoreError});
+               };
+
+               let node_1 = NodeId::from_pubkey(&node_id_1);
+               let node_2 = NodeId::from_pubkey(&node_id_2);
+               let channel_info = ChannelInfo {
+                       features,
+                       node_one: node_1.clone(),
+                       one_to_two: None,
+                       node_two: node_2.clone(),
+                       two_to_one: None,
+                       capacity_sats: None,
+                       announcement_message: None,
+                       announcement_received_time: timestamp,
+               };
+
+               self.add_channel_between_nodes(short_channel_id, channel_info, None)
+       }
+
+       fn add_channel_between_nodes(&self, short_channel_id: u64, channel_info: ChannelInfo, utxo_value: Option<u64>) -> Result<(), LightningError> {
+               let mut channels = self.channels.write().unwrap();
+               let mut nodes = self.nodes.write().unwrap();
+
+               let node_id_a = channel_info.node_one.clone();
+               let node_id_b = channel_info.node_two.clone();
+
+               match channels.entry(short_channel_id) {
+                       BtreeEntry::Occupied(mut entry) => {
+                               //TODO: because asking the blockchain if short_channel_id is valid is only optional
+                               //in the blockchain API, we need to handle it smartly here, though it's unclear
+                               //exactly how...
+                               if utxo_value.is_some() {
+                                       // Either our UTXO provider is busted, there was a reorg, or the UTXO provider
+                                       // only sometimes returns results. In any case remove the previous entry. Note
+                                       // that the spec expects us to "blacklist" the node_ids involved, but we can't
+                                       // do that because
+                                       // a) we don't *require* a UTXO provider that always returns results.
+                                       // b) we don't track UTXOs of channels we know about and remove them if they
+                                       //    get reorg'd out.
+                                       // c) it's unclear how to do so without exposing ourselves to massive DoS risk.
+                                       Self::remove_channel_in_nodes(&mut nodes, &entry.get(), short_channel_id);
+                                       *entry.get_mut() = channel_info;
+                               } else {
+                                       return Err(LightningError{err: "Already have knowledge of channel".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
+                               }
+                       },
+                       BtreeEntry::Vacant(entry) => {
+                               entry.insert(channel_info);
+                       }
+               };
+
+               for current_node_id in [node_id_a, node_id_b].iter() {
+                       match nodes.entry(current_node_id.clone()) {
+                               BtreeEntry::Occupied(node_entry) => {
+                                       node_entry.into_mut().channels.push(short_channel_id);
+                               },
+                               BtreeEntry::Vacant(node_entry) => {
+                                       node_entry.insert(NodeInfo {
+                                               channels: vec!(short_channel_id),
+                                               lowest_inbound_channel_fees: None,
+                                               announcement_info: None,
+                                       });
+                               }
+                       };
+               };
+
+               Ok(())
+       }
+
+       fn update_channel_from_unsigned_announcement_intern<C: Deref>(
+               &self, msg: &msgs::UnsignedChannelAnnouncement, full_msg: Option<&msgs::ChannelAnnouncement>, chain_access: &Option<C>
+       ) -> Result<(), LightningError>
+       where
+               C::Target: chain::Access,
+       {
+               if msg.node_id_1 == msg.node_id_2 || msg.bitcoin_key_1 == msg.bitcoin_key_2 {
+                       return Err(LightningError{err: "Channel announcement node had a channel with itself".to_owned(), action: ErrorAction::IgnoreError});
+               }
+
+               let utxo_value = match &chain_access {
+                       &None => {
+                               // Tentatively accept, potentially exposing us to DoS attacks
+                               None
+                       },
+                       &Some(ref chain_access) => {
+                               match chain_access.get_utxo(&msg.chain_hash, msg.short_channel_id) {
+                                       Ok(TxOut { value, script_pubkey }) => {
+                                               let expected_script = Builder::new().push_opcode(opcodes::all::OP_PUSHNUM_2)
+                                                                                   .push_slice(&msg.bitcoin_key_1.serialize())
+                                                                                   .push_slice(&msg.bitcoin_key_2.serialize())
+                                                                                   .push_opcode(opcodes::all::OP_PUSHNUM_2)
+                                                                                   .push_opcode(opcodes::all::OP_CHECKMULTISIG).into_script().to_v0_p2wsh();
+                                               if script_pubkey != expected_script {
+                                                       return Err(LightningError{err: format!("Channel announcement key ({}) didn't match on-chain script ({})", script_pubkey.to_hex(), expected_script.to_hex()), action: ErrorAction::IgnoreError});
+                                               }
+                                               //TODO: Check if value is worth storing, use it to inform routing, and compare it
+                                               //to the new HTLC max field in channel_update
+                                               Some(value)
+                                       },
+                                       Err(chain::AccessError::UnknownChain) => {
+                                               return Err(LightningError{err: format!("Channel announced on an unknown chain ({})", msg.chain_hash.encode().to_hex()), action: ErrorAction::IgnoreError});
+                                       },
+                                       Err(chain::AccessError::UnknownTx) => {
+                                               return Err(LightningError{err: "Channel announced without corresponding UTXO entry".to_owned(), action: ErrorAction::IgnoreError});
+                                       },
+                               }
+                       },
+               };
+
+               #[allow(unused_mut, unused_assignments)]
+               let mut announcement_received_time = 0;
+               #[cfg(feature = "std")]
+               {
+                       announcement_received_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+               }
+
+               let chan_info = ChannelInfo {
+                       features: msg.features.clone(),
+                       node_one: NodeId::from_pubkey(&msg.node_id_1),
+                       one_to_two: None,
+                       node_two: NodeId::from_pubkey(&msg.node_id_2),
+                       two_to_one: None,
+                       capacity_sats: utxo_value,
+                       announcement_message: if msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY
+                               { full_msg.cloned() } else { None },
+                       announcement_received_time,
+               };
+
+               self.add_channel_between_nodes(msg.short_channel_id, chan_info, utxo_value)
+       }
+
+       /// Marks a channel in the graph as failed if a corresponding HTLC fail was sent.
+       /// If permanent, removes a channel from the local storage.
+       /// May cause the removal of nodes too, if this was their last channel.
+       /// If not permanent, makes channels unavailable for routing.
+       pub fn channel_failed(&self, short_channel_id: u64, is_permanent: bool) {
+               let mut channels = self.channels.write().unwrap();
+               if is_permanent {
+                       if let Some(chan) = channels.remove(&short_channel_id) {
+                               let mut nodes = self.nodes.write().unwrap();
+                               Self::remove_channel_in_nodes(&mut nodes, &chan, short_channel_id);
+                       }
+               } else {
+                       if let Some(chan) = channels.get_mut(&short_channel_id) {
+                               if let Some(one_to_two) = chan.one_to_two.as_mut() {
+                                       one_to_two.enabled = false;
+                               }
+                               if let Some(two_to_one) = chan.two_to_one.as_mut() {
+                                       two_to_one.enabled = false;
+                               }
+                       }
+               }
+       }
+
+       /// Marks a node in the graph as failed.
+       pub fn node_failed(&self, _node_id: &PublicKey, is_permanent: bool) {
+               if is_permanent {
+                       // TODO: Wholly remove the node
+               } else {
+                       // TODO: downgrade the node
+               }
+       }
+
+       #[cfg(feature = "std")]
+       /// Removes information about channels that we haven't heard any updates about in some time.
+       /// This can be used regularly to prune the network graph of channels that likely no longer
+       /// exist.
+       ///
+       /// While there is no formal requirement that nodes regularly re-broadcast their channel
+       /// updates every two weeks, the non-normative section of BOLT 7 currently suggests that
+       /// pruning occur for updates which are at least two weeks old, which we implement here.
+       ///
+       /// Note that for users of the `lightning-background-processor` crate this method may be
+       /// automatically called regularly for you.
+       ///
+       /// This method is only available with the `std` feature. See
+       /// [`NetworkGraph::remove_stale_channels_with_time`] for `no-std` use.
+       pub fn remove_stale_channels(&self) {
+               let time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+               self.remove_stale_channels_with_time(time);
+       }
+
+       /// Removes information about channels that we haven't heard any updates about in some time.
+       /// This can be used regularly to prune the network graph of channels that likely no longer
+       /// exist.
+       ///
+       /// While there is no formal requirement that nodes regularly re-broadcast their channel
+       /// updates every two weeks, the non-normative section of BOLT 7 currently suggests that
+       /// pruning occur for updates which are at least two weeks old, which we implement here.
+       ///
+       /// This function takes the current unix time as an argument. For users with the `std` feature
+       /// enabled, [`NetworkGraph::remove_stale_channels`] may be preferable.
+       pub fn remove_stale_channels_with_time(&self, current_time_unix: u64) {
+               let mut channels = self.channels.write().unwrap();
+               // Time out if we haven't received an update in at least 14 days.
+               if current_time_unix > u32::max_value() as u64 { return; } // Remove by 2106
+               if current_time_unix < STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS { return; }
+               let min_time_unix: u32 = (current_time_unix - STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS) as u32;
+               // Sadly BTreeMap::retain was only stabilized in 1.53 so we can't switch to it for some
+               // time.
+               let mut scids_to_remove = Vec::new();
+               for (scid, info) in channels.iter_mut() {
+                       if info.one_to_two.is_some() && info.one_to_two.as_ref().unwrap().last_update < min_time_unix {
+                               info.one_to_two = None;
+                       }
+                       if info.two_to_one.is_some() && info.two_to_one.as_ref().unwrap().last_update < min_time_unix {
+                               info.two_to_one = None;
+                       }
+                       if info.one_to_two.is_none() && info.two_to_one.is_none() {
+                               // We check the announcement_received_time here to ensure we don't drop
+                               // announcements that we just received and are just waiting for our peer to send a
+                               // channel_update for.
+                               if info.announcement_received_time < min_time_unix as u64 {
+                                       scids_to_remove.push(*scid);
+                               }
+                       }
+               }
+               if !scids_to_remove.is_empty() {
+                       let mut nodes = self.nodes.write().unwrap();
+                       for scid in scids_to_remove {
+                               let info = channels.remove(&scid).expect("We just accessed this scid, it should be present");
+                               Self::remove_channel_in_nodes(&mut nodes, &info, scid);
+                       }
+               }
+       }
+
+       /// For an already known (from announcement) channel, update info about one of the directions
+       /// of the channel.
+       ///
+       /// You probably don't want to call this directly, instead relying on a P2PGossipSync's
+       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
+       /// routing messages from a source using a protocol other than the lightning P2P protocol.
+       ///
+       /// If built with `no-std`, any updates with a timestamp more than two weeks in the past or
+       /// materially in the future will be rejected.
+       pub fn update_channel<T: secp256k1::Verification>(&self, msg: &msgs::ChannelUpdate, secp_ctx: &Secp256k1<T>) -> Result<(), LightningError> {
+               self.update_channel_intern(&msg.contents, Some(&msg), Some((&msg.signature, secp_ctx)))
+       }
+
+       /// For an already known (from announcement) channel, update info about one of the directions
+       /// of the channel without verifying the associated signatures. Because we aren't given the
+       /// associated signatures here we cannot relay the channel update to any of our peers.
+       ///
+       /// If built with `no-std`, any updates with a timestamp more than two weeks in the past or
+       /// materially in the future will be rejected.
+       pub fn update_channel_unsigned(&self, msg: &msgs::UnsignedChannelUpdate) -> Result<(), LightningError> {
+               self.update_channel_intern(msg, None, None::<(&secp256k1::ecdsa::Signature, &Secp256k1<secp256k1::VerifyOnly>)>)
+       }
+
+       fn update_channel_intern<T: secp256k1::Verification>(&self, msg: &msgs::UnsignedChannelUpdate, full_msg: Option<&msgs::ChannelUpdate>, sig_info: Option<(&secp256k1::ecdsa::Signature, &Secp256k1<T>)>) -> Result<(), LightningError> {
+               let dest_node_id;
+               let chan_enabled = msg.flags & (1 << 1) != (1 << 1);
+               let chan_was_enabled;
+
+               #[cfg(all(feature = "std", not(test), not(feature = "_test_utils")))]
+               {
+                       // Note that many tests rely on being able to set arbitrarily old timestamps, thus we
+                       // disable this check during tests!
+                       let time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+                       if (msg.timestamp as u64) < time - STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS {
+                               return Err(LightningError{err: "channel_update is older than two weeks old".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
+                       }
+                       if msg.timestamp as u64 > time + 60 * 60 * 24 {
+                               return Err(LightningError{err: "channel_update has a timestamp more than a day in the future".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
+                       }
+               }
+
+               let mut channels = self.channels.write().unwrap();
+               match channels.get_mut(&msg.short_channel_id) {
+                       None => return Err(LightningError{err: "Couldn't find channel for update".to_owned(), action: ErrorAction::IgnoreError}),
+                       Some(channel) => {
+                               if let OptionalField::Present(htlc_maximum_msat) = msg.htlc_maximum_msat {
+                                       if htlc_maximum_msat > MAX_VALUE_MSAT {
+                                               return Err(LightningError{err: "htlc_maximum_msat is larger than maximum possible msats".to_owned(), action: ErrorAction::IgnoreError});
+                                       }
+
+                                       if let Some(capacity_sats) = channel.capacity_sats {
+                                               // It's possible channel capacity is available now, although it wasn't available at announcement (so the field is None).
+                                               // Don't query UTXO set here to reduce DoS risks.
+                                               if capacity_sats > MAX_VALUE_MSAT / 1000 || htlc_maximum_msat > capacity_sats * 1000 {
+                                                       return Err(LightningError{err: "htlc_maximum_msat is larger than channel capacity or capacity is bogus".to_owned(), action: ErrorAction::IgnoreError});
+                                               }
+                                       }
+                               }
+                               macro_rules! check_update_latest {
+                                       ($target: expr) => {
+                                               if let Some(existing_chan_info) = $target.as_ref() {
+                                                       // The timestamp field is somewhat of a misnomer - the BOLTs use it to
+                                                       // order updates to ensure you always have the latest one, only
+                                                       // suggesting  that it be at least the current time. For
+                                                       // channel_updates specifically, the BOLTs discuss the possibility of
+                                                       // pruning based on the timestamp field being more than two weeks old,
+                                                       // but only in the non-normative section.
+                                                       if existing_chan_info.last_update > msg.timestamp {
+                                                               return Err(LightningError{err: "Update older than last processed update".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
+                                                       } else if existing_chan_info.last_update == msg.timestamp {
+                                                               return Err(LightningError{err: "Update had same timestamp as last processed update".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
+                                                       }
+                                                       chan_was_enabled = existing_chan_info.enabled;
+                                               } else {
+                                                       chan_was_enabled = false;
+                                               }
+                                       }
+                               }
+
+                               macro_rules! get_new_channel_info {
+                                       () => { {
+                                               let last_update_message = if msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY
+                                                       { full_msg.cloned() } else { None };
+
+                                               let updated_channel_update_info = ChannelUpdateInfo {
+                                                       enabled: chan_enabled,
+                                                       last_update: msg.timestamp,
+                                                       cltv_expiry_delta: msg.cltv_expiry_delta,
+                                                       htlc_minimum_msat: msg.htlc_minimum_msat,
+                                                       htlc_maximum_msat: if let OptionalField::Present(max_value) = msg.htlc_maximum_msat { Some(max_value) } else { None },
+                                                       fees: RoutingFees {
+                                                               base_msat: msg.fee_base_msat,
+                                                               proportional_millionths: msg.fee_proportional_millionths,
+                                                       },
+                                                       last_update_message
+                                               };
+                                               Some(updated_channel_update_info)
+                                       } }
+                               }
+
+                               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.encode()[..])[..]);
+                               if msg.flags & 1 == 1 {
+                                       dest_node_id = channel.node_one.clone();
+                                       check_update_latest!(channel.two_to_one);
+                                       if let Some((sig, ctx)) = sig_info {
+                                               secp_verify_sig!(ctx, &msg_hash, &sig, &PublicKey::from_slice(channel.node_two.as_slice()).map_err(|_| LightningError{
+                                                       err: "Couldn't parse source node pubkey".to_owned(),
+                                                       action: ErrorAction::IgnoreAndLog(Level::Debug)
+                                               })?, "channel_update");
+                                       }
+                                       channel.two_to_one = get_new_channel_info!();
+                               } else {
+                                       dest_node_id = channel.node_two.clone();
+                                       check_update_latest!(channel.one_to_two);
+                                       if let Some((sig, ctx)) = sig_info {
+                                               secp_verify_sig!(ctx, &msg_hash, &sig, &PublicKey::from_slice(channel.node_one.as_slice()).map_err(|_| LightningError{
+                                                       err: "Couldn't parse destination node pubkey".to_owned(),
+                                                       action: ErrorAction::IgnoreAndLog(Level::Debug)
+                                               })?, "channel_update");
+                                       }
+                                       channel.one_to_two = get_new_channel_info!();
+                               }
+                       }
+               }
+
+               let mut nodes = self.nodes.write().unwrap();
+               if chan_enabled {
+                       let node = nodes.get_mut(&dest_node_id).unwrap();
+                       let mut base_msat = msg.fee_base_msat;
+                       let mut proportional_millionths = msg.fee_proportional_millionths;
+                       if let Some(fees) = node.lowest_inbound_channel_fees {
+                               base_msat = cmp::min(base_msat, fees.base_msat);
+                               proportional_millionths = cmp::min(proportional_millionths, fees.proportional_millionths);
+                       }
+                       node.lowest_inbound_channel_fees = Some(RoutingFees {
+                               base_msat,
+                               proportional_millionths
+                       });
+               } else if chan_was_enabled {
+                       let node = nodes.get_mut(&dest_node_id).unwrap();
+                       let mut lowest_inbound_channel_fees = None;
+
+                       for chan_id in node.channels.iter() {
+                               let chan = channels.get(chan_id).unwrap();
+                               let chan_info_opt;
+                               if chan.node_one == dest_node_id {
+                                       chan_info_opt = chan.two_to_one.as_ref();
+                               } else {
+                                       chan_info_opt = chan.one_to_two.as_ref();
+                               }
+                               if let Some(chan_info) = chan_info_opt {
+                                       if chan_info.enabled {
+                                               let fees = lowest_inbound_channel_fees.get_or_insert(RoutingFees {
+                                                       base_msat: u32::max_value(), proportional_millionths: u32::max_value() });
+                                               fees.base_msat = cmp::min(fees.base_msat, chan_info.fees.base_msat);
+                                               fees.proportional_millionths = cmp::min(fees.proportional_millionths, chan_info.fees.proportional_millionths);
+                                       }
+                               }
+                       }
+
+                       node.lowest_inbound_channel_fees = lowest_inbound_channel_fees;
+               }
+
+               Ok(())
+       }
+
+       fn remove_channel_in_nodes(nodes: &mut BTreeMap<NodeId, NodeInfo>, chan: &ChannelInfo, short_channel_id: u64) {
+               macro_rules! remove_from_node {
+                       ($node_id: expr) => {
+                               if let BtreeEntry::Occupied(mut entry) = nodes.entry($node_id) {
+                                       entry.get_mut().channels.retain(|chan_id| {
+                                               short_channel_id != *chan_id
+                                       });
+                                       if entry.get().channels.is_empty() {
+                                               entry.remove_entry();
+                                       }
+                               } else {
+                                       panic!("Had channel that pointed to unknown node (ie inconsistent network map)!");
+                               }
+                       }
+               }
+
+               remove_from_node!(chan.node_one);
+               remove_from_node!(chan.node_two);
+       }
+}
+
+impl ReadOnlyNetworkGraph<'_> {
+       /// Returns all known valid channels' short ids along with announced channel info.
+       ///
+       /// (C-not exported) because we have no mapping for `BTreeMap`s
+       pub fn channels(&self) -> &BTreeMap<u64, ChannelInfo> {
+               &*self.channels
+       }
+
+       /// Returns all known nodes' public keys along with announced node info.
+       ///
+       /// (C-not exported) because we have no mapping for `BTreeMap`s
+       pub fn nodes(&self) -> &BTreeMap<NodeId, NodeInfo> {
+               &*self.nodes
+       }
+
+       /// Get network addresses by node id.
+       /// Returns None if the requested node is completely unknown,
+       /// or if node announcement for the node was never received.
+       pub fn get_addresses(&self, pubkey: &PublicKey) -> Option<Vec<NetAddress>> {
+               if let Some(node) = self.nodes.get(&NodeId::from_pubkey(&pubkey)) {
+                       if let Some(node_info) = node.announcement_info.as_ref() {
+                               return Some(node_info.addresses.clone())
+                       }
+               }
+               None
+       }
+}
+
+#[cfg(test)]
+mod tests {
+       use chain;
+       use ln::PaymentHash;
+       use ln::features::{ChannelFeatures, InitFeatures, NodeFeatures};
+       use routing::gossip::{P2PGossipSync, NetworkGraph, NetworkUpdate, MAX_EXCESS_BYTES_FOR_RELAY};
+       use ln::msgs::{Init, OptionalField, RoutingMessageHandler, UnsignedNodeAnnouncement, NodeAnnouncement,
+               UnsignedChannelAnnouncement, ChannelAnnouncement, UnsignedChannelUpdate, ChannelUpdate,
+               ReplyChannelRange, QueryChannelRange, QueryShortChannelIds, MAX_VALUE_MSAT};
+       use util::test_utils;
+       use util::logger::Logger;
+       use util::ser::{Readable, Writeable};
+       use util::events::{Event, EventHandler, MessageSendEvent, MessageSendEventsProvider};
+       use util::scid_utils::scid_from_parts;
+
+       use super::STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS;
+
+       use bitcoin::hashes::sha256d::Hash as Sha256dHash;
+       use bitcoin::hashes::Hash;
+       use bitcoin::network::constants::Network;
+       use bitcoin::blockdata::constants::genesis_block;
+       use bitcoin::blockdata::script::{Builder, Script};
+       use bitcoin::blockdata::transaction::TxOut;
+       use bitcoin::blockdata::opcodes;
+
+       use hex;
+
+       use bitcoin::secp256k1::{PublicKey, SecretKey};
+       use bitcoin::secp256k1::{All, Secp256k1};
+
+       use io;
+       use bitcoin::secp256k1;
+       use prelude::*;
+       use sync::Arc;
+
+       fn create_network_graph() -> NetworkGraph {
+               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
+               NetworkGraph::new(genesis_hash)
+       }
+
+       fn create_gossip_sync(network_graph: &NetworkGraph) -> (
+               Secp256k1<All>, P2PGossipSync<&NetworkGraph, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>
+       ) {
+               let secp_ctx = Secp256k1::new();
+               let logger = Arc::new(test_utils::TestLogger::new());
+               let gossip_sync = P2PGossipSync::new(network_graph, None, Arc::clone(&logger));
+               (secp_ctx, gossip_sync)
+       }
+
+       #[test]
+       fn request_full_sync_finite_times() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+               let node_id = PublicKey::from_secret_key(&secp_ctx, &SecretKey::from_slice(&hex::decode("0202020202020202020202020202020202020202020202020202020202020202").unwrap()[..]).unwrap());
+
+               assert!(gossip_sync.should_request_full_sync(&node_id));
+               assert!(gossip_sync.should_request_full_sync(&node_id));
+               assert!(gossip_sync.should_request_full_sync(&node_id));
+               assert!(gossip_sync.should_request_full_sync(&node_id));
+               assert!(gossip_sync.should_request_full_sync(&node_id));
+               assert!(!gossip_sync.should_request_full_sync(&node_id));
+       }
+
+       fn get_signed_node_announcement<F: Fn(&mut UnsignedNodeAnnouncement)>(f: F, node_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> NodeAnnouncement {
+               let node_id = PublicKey::from_secret_key(&secp_ctx, node_key);
+               let mut unsigned_announcement = UnsignedNodeAnnouncement {
+                       features: NodeFeatures::known(),
+                       timestamp: 100,
+                       node_id: node_id,
+                       rgb: [0; 3],
+                       alias: [0; 32],
+                       addresses: Vec::new(),
+                       excess_address_data: Vec::new(),
+                       excess_data: Vec::new(),
+               };
+               f(&mut unsigned_announcement);
+               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_announcement.encode()[..])[..]);
+               NodeAnnouncement {
+                       signature: secp_ctx.sign_ecdsa(&msghash, node_key),
+                       contents: unsigned_announcement
+               }
+       }
+
+       fn get_signed_channel_announcement<F: Fn(&mut UnsignedChannelAnnouncement)>(f: F, node_1_key: &SecretKey, node_2_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> ChannelAnnouncement {
+               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_1_key);
+               let node_id_2 = PublicKey::from_secret_key(&secp_ctx, node_2_key);
+               let node_1_btckey = &SecretKey::from_slice(&[40; 32]).unwrap();
+               let node_2_btckey = &SecretKey::from_slice(&[39; 32]).unwrap();
+
+               let mut unsigned_announcement = UnsignedChannelAnnouncement {
+                       features: ChannelFeatures::known(),
+                       chain_hash: genesis_block(Network::Testnet).header.block_hash(),
+                       short_channel_id: 0,
+                       node_id_1,
+                       node_id_2,
+                       bitcoin_key_1: PublicKey::from_secret_key(&secp_ctx, node_1_btckey),
+                       bitcoin_key_2: PublicKey::from_secret_key(&secp_ctx, node_2_btckey),
+                       excess_data: Vec::new(),
+               };
+               f(&mut unsigned_announcement);
+               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_announcement.encode()[..])[..]);
+               ChannelAnnouncement {
+                       node_signature_1: secp_ctx.sign_ecdsa(&msghash, node_1_key),
+                       node_signature_2: secp_ctx.sign_ecdsa(&msghash, node_2_key),
+                       bitcoin_signature_1: secp_ctx.sign_ecdsa(&msghash, node_1_btckey),
+                       bitcoin_signature_2: secp_ctx.sign_ecdsa(&msghash, node_2_btckey),
+                       contents: unsigned_announcement,
+               }
+       }
+
+       fn get_channel_script(secp_ctx: &Secp256k1<secp256k1::All>) -> Script {
+               let node_1_btckey = &SecretKey::from_slice(&[40; 32]).unwrap();
+               let node_2_btckey = &SecretKey::from_slice(&[39; 32]).unwrap();
+               Builder::new().push_opcode(opcodes::all::OP_PUSHNUM_2)
+                             .push_slice(&PublicKey::from_secret_key(&secp_ctx, node_1_btckey).serialize())
+                             .push_slice(&PublicKey::from_secret_key(&secp_ctx, node_2_btckey).serialize())
+                             .push_opcode(opcodes::all::OP_PUSHNUM_2)
+                             .push_opcode(opcodes::all::OP_CHECKMULTISIG).into_script()
+                             .to_v0_p2wsh()
+       }
+
+       fn get_signed_channel_update<F: Fn(&mut UnsignedChannelUpdate)>(f: F, node_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> ChannelUpdate {
+               let mut unsigned_channel_update = UnsignedChannelUpdate {
+                       chain_hash: genesis_block(Network::Testnet).header.block_hash(),
+                       short_channel_id: 0,
+                       timestamp: 100,
+                       flags: 0,
+                       cltv_expiry_delta: 144,
+                       htlc_minimum_msat: 1_000_000,
+                       htlc_maximum_msat: OptionalField::Absent,
+                       fee_base_msat: 10_000,
+                       fee_proportional_millionths: 20,
+                       excess_data: Vec::new()
+               };
+               f(&mut unsigned_channel_update);
+               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_channel_update.encode()[..])[..]);
+               ChannelUpdate {
+                       signature: secp_ctx.sign_ecdsa(&msghash, node_key),
+                       contents: unsigned_channel_update
+               }
+       }
+
+       #[test]
+       fn handling_node_announcements() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+               let zero_hash = Sha256dHash::hash(&[0; 32]);
+
+               let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_node_announcement(&valid_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!("No existing channels for node_announcement", e.err)
+               };
+
+               {
+                       // Announce a channel to add a corresponding node.
+                       let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+                       match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                               Ok(res) => assert!(res),
+                               _ => panic!()
+                       };
+               }
+
+               match gossip_sync.handle_node_announcement(&valid_announcement) {
+                       Ok(res) => assert!(res),
+                       Err(_) => panic!()
+               };
+
+               let fake_msghash = hash_to_message!(&zero_hash);
+               match gossip_sync.handle_node_announcement(
+                       &NodeAnnouncement {
+                               signature: secp_ctx.sign_ecdsa(&fake_msghash, node_1_privkey),
+                               contents: valid_announcement.contents.clone()
+               }) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Invalid signature on node_announcement message")
+               };
+
+               let announcement_with_data = get_signed_node_announcement(|unsigned_announcement| {
+                       unsigned_announcement.timestamp += 1000;
+                       unsigned_announcement.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
+               }, node_1_privkey, &secp_ctx);
+               // Return false because contains excess data.
+               match gossip_sync.handle_node_announcement(&announcement_with_data) {
+                       Ok(res) => assert!(!res),
+                       Err(_) => panic!()
+               };
+
+               // Even though previous announcement was not relayed further, we still accepted it,
+               // so we now won't accept announcements before the previous one.
+               let outdated_announcement = get_signed_node_announcement(|unsigned_announcement| {
+                       unsigned_announcement.timestamp += 1000 - 10;
+               }, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_node_announcement(&outdated_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Update older than last processed update")
+               };
+       }
+
+       #[test]
+       fn handling_channel_announcements() {
+               let secp_ctx = Secp256k1::new();
+               let logger: Arc<Logger> = Arc::new(test_utils::TestLogger::new());
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               let good_script = get_channel_script(&secp_ctx);
+               let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+
+               // Test if the UTXO lookups were not supported
+               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
+               let mut gossip_sync = P2PGossipSync::new(&network_graph, None, Arc::clone(&logger));
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(res) => assert!(res),
+                       _ => panic!()
+               };
+
+               {
+                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
+                               None => panic!(),
+                               Some(_) => ()
+                       };
+               }
+
+               // If we receive announcement for the same channel (with UTXO lookups disabled),
+               // drop new one on the floor, since we can't see any changes.
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Already have knowledge of channel")
+               };
+
+               // Test if an associated transaction were not on-chain (or not confirmed).
+               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
+               *chain_source.utxo_ret.lock().unwrap() = Err(chain::AccessError::UnknownTx);
+               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
+               gossip_sync = P2PGossipSync::new(&network_graph, Some(chain_source.clone()), Arc::clone(&logger));
+
+               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
+                       unsigned_announcement.short_channel_id += 1;
+               }, node_1_privkey, node_2_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Channel announced without corresponding UTXO entry")
+               };
+
+               // Now test if the transaction is found in the UTXO set and the script is correct.
+               *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: 0, script_pubkey: good_script.clone() });
+               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
+                       unsigned_announcement.short_channel_id += 2;
+               }, node_1_privkey, node_2_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(res) => assert!(res),
+                       _ => panic!()
+               };
+
+               {
+                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
+                               None => panic!(),
+                               Some(_) => ()
+                       };
+               }
+
+               // If we receive announcement for the same channel (but TX is not confirmed),
+               // drop new one on the floor, since we can't see any changes.
+               *chain_source.utxo_ret.lock().unwrap() = Err(chain::AccessError::UnknownTx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Channel announced without corresponding UTXO entry")
+               };
+
+               // But if it is confirmed, replace the channel
+               *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: 0, script_pubkey: good_script });
+               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
+                       unsigned_announcement.features = ChannelFeatures::empty();
+                       unsigned_announcement.short_channel_id += 2;
+               }, node_1_privkey, node_2_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(res) => assert!(res),
+                       _ => panic!()
+               };
+               {
+                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
+                               Some(channel_entry) => {
+                                       assert_eq!(channel_entry.features, ChannelFeatures::empty());
+                               },
+                               _ => panic!()
+                       };
+               }
+
+               // Don't relay valid channels with excess data
+               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
+                       unsigned_announcement.short_channel_id += 3;
+                       unsigned_announcement.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
+               }, node_1_privkey, node_2_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(res) => assert!(!res),
+                       _ => panic!()
+               };
+
+               let mut invalid_sig_announcement = valid_announcement.clone();
+               invalid_sig_announcement.contents.excess_data = Vec::new();
+               match gossip_sync.handle_channel_announcement(&invalid_sig_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Invalid signature on channel_announcement message")
+               };
+
+               let channel_to_itself_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&channel_to_itself_announcement) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Channel announcement node had a channel with itself")
+               };
+       }
+
+       #[test]
+       fn handling_channel_update() {
+               let secp_ctx = Secp256k1::new();
+               let logger: Arc<Logger> = Arc::new(test_utils::TestLogger::new());
+               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
+               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
+               let gossip_sync = P2PGossipSync::new(&network_graph, Some(chain_source.clone()), Arc::clone(&logger));
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               let amount_sats = 1000_000;
+               let short_channel_id;
+
+               {
+                       // Announce a channel we will update
+                       let good_script = get_channel_script(&secp_ctx);
+                       *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: amount_sats, script_pubkey: good_script.clone() });
+
+                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
+                       match gossip_sync.handle_channel_announcement(&valid_channel_announcement) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+
+               }
+
+               let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(res) => assert!(res),
+                       _ => panic!()
+               };
+
+               {
+                       match network_graph.read_only().channels().get(&short_channel_id) {
+                               None => panic!(),
+                               Some(channel_info) => {
+                                       assert_eq!(channel_info.one_to_two.as_ref().unwrap().cltv_expiry_delta, 144);
+                                       assert!(channel_info.two_to_one.is_none());
+                               }
+                       };
+               }
+
+               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.timestamp += 100;
+                       unsigned_channel_update.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
+               }, node_1_privkey, &secp_ctx);
+               // Return false because contains excess data
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(res) => assert!(!res),
+                       _ => panic!()
+               };
+
+               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.timestamp += 110;
+                       unsigned_channel_update.short_channel_id += 1;
+               }, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Couldn't find channel for update")
+               };
+
+               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.htlc_maximum_msat = OptionalField::Present(MAX_VALUE_MSAT + 1);
+                       unsigned_channel_update.timestamp += 110;
+               }, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "htlc_maximum_msat is larger than maximum possible msats")
+               };
+
+               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.htlc_maximum_msat = OptionalField::Present(amount_sats * 1000 + 1);
+                       unsigned_channel_update.timestamp += 110;
+               }, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "htlc_maximum_msat is larger than channel capacity or capacity is bogus")
+               };
+
+               // Even though previous update was not relayed further, we still accepted it,
+               // so we now won't accept update before the previous one.
+               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.timestamp += 100;
+               }, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Update had same timestamp as last processed update")
+               };
+
+               let mut invalid_sig_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                       unsigned_channel_update.timestamp += 500;
+               }, node_1_privkey, &secp_ctx);
+               let zero_hash = Sha256dHash::hash(&[0; 32]);
+               let fake_msghash = hash_to_message!(&zero_hash);
+               invalid_sig_channel_update.signature = secp_ctx.sign_ecdsa(&fake_msghash, node_1_privkey);
+               match gossip_sync.handle_channel_update(&invalid_sig_channel_update) {
+                       Ok(_) => panic!(),
+                       Err(e) => assert_eq!(e.err, "Invalid signature on channel_update message")
+               };
+       }
+
+       #[test]
+       fn handling_network_update() {
+               let logger = test_utils::TestLogger::new();
+               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
+               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
+               let network_graph = NetworkGraph::new(genesis_hash);
+               let gossip_sync = P2PGossipSync::new(&network_graph, Some(chain_source.clone()), &logger);
+               let secp_ctx = Secp256k1::new();
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               {
+                       // There is no nodes in the table at the beginning.
+                       assert_eq!(network_graph.read_only().nodes().len(), 0);
+               }
+
+               let short_channel_id;
+               {
+                       // Announce a channel we will update
+                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
+                       let chain_source: Option<&test_utils::TestChainSource> = None;
+                       assert!(network_graph.update_channel_from_announcement(&valid_channel_announcement, &chain_source, &secp_ctx).is_ok());
+                       assert!(network_graph.read_only().channels().get(&short_channel_id).is_some());
+
+                       let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
+                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_none());
+
+                       gossip_sync.handle_event(&Event::PaymentPathFailed {
+                               payment_id: None,
+                               payment_hash: PaymentHash([0; 32]),
+                               rejected_by_dest: false,
+                               all_paths_failed: true,
+                               path: vec![],
+                               network_update: Some(NetworkUpdate::ChannelUpdateMessage {
+                                       msg: valid_channel_update,
+                               }),
+                               short_channel_id: None,
+                               retry: None,
+                               error_code: None,
+                               error_data: None,
+                       });
+
+                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_some());
+               }
+
+               // Non-permanent closing just disables a channel
+               {
+                       match network_graph.read_only().channels().get(&short_channel_id) {
+                               None => panic!(),
+                               Some(channel_info) => {
+                                       assert!(channel_info.one_to_two.as_ref().unwrap().enabled);
+                               }
+                       };
+
+                       gossip_sync.handle_event(&Event::PaymentPathFailed {
+                               payment_id: None,
+                               payment_hash: PaymentHash([0; 32]),
+                               rejected_by_dest: false,
+                               all_paths_failed: true,
+                               path: vec![],
+                               network_update: Some(NetworkUpdate::ChannelFailure {
+                                       short_channel_id,
+                                       is_permanent: false,
+                               }),
+                               short_channel_id: None,
+                               retry: None,
+                               error_code: None,
+                               error_data: None,
+                       });
+
+                       match network_graph.read_only().channels().get(&short_channel_id) {
+                               None => panic!(),
+                               Some(channel_info) => {
+                                       assert!(!channel_info.one_to_two.as_ref().unwrap().enabled);
+                               }
+                       };
+               }
+
+               // Permanent closing deletes a channel
+               gossip_sync.handle_event(&Event::PaymentPathFailed {
+                       payment_id: None,
+                       payment_hash: PaymentHash([0; 32]),
+                       rejected_by_dest: false,
+                       all_paths_failed: true,
+                       path: vec![],
+                       network_update: Some(NetworkUpdate::ChannelFailure {
+                               short_channel_id,
+                               is_permanent: true,
+                       }),
+                       short_channel_id: None,
+                       retry: None,
+                       error_code: None,
+                       error_data: None,
+               });
+
+               assert_eq!(network_graph.read_only().channels().len(), 0);
+               // Nodes are also deleted because there are no associated channels anymore
+               assert_eq!(network_graph.read_only().nodes().len(), 0);
+               // TODO: Test NetworkUpdate::NodeFailure, which is not implemented yet.
+       }
+
+       #[test]
+       fn test_channel_timeouts() {
+               // Test the removal of channels with `remove_stale_channels`.
+               let logger = test_utils::TestLogger::new();
+               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
+               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
+               let network_graph = NetworkGraph::new(genesis_hash);
+               let gossip_sync = P2PGossipSync::new(&network_graph, Some(chain_source.clone()), &logger);
+               let secp_ctx = Secp256k1::new();
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+               let short_channel_id = valid_channel_announcement.contents.short_channel_id;
+               let chain_source: Option<&test_utils::TestChainSource> = None;
+               assert!(network_graph.update_channel_from_announcement(&valid_channel_announcement, &chain_source, &secp_ctx).is_ok());
+               assert!(network_graph.read_only().channels().get(&short_channel_id).is_some());
+
+               let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
+               assert!(gossip_sync.handle_channel_update(&valid_channel_update).is_ok());
+               assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_some());
+
+               network_graph.remove_stale_channels_with_time(100 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
+               assert_eq!(network_graph.read_only().channels().len(), 1);
+               assert_eq!(network_graph.read_only().nodes().len(), 2);
+
+               network_graph.remove_stale_channels_with_time(101 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
+               #[cfg(feature = "std")]
+               {
+                       // In std mode, a further check is performed before fully removing the channel -
+                       // the channel_announcement must have been received at least two weeks ago. We
+                       // fudge that here by indicating the time has jumped two weeks. Note that the
+                       // directional channel information will have been removed already..
+                       assert_eq!(network_graph.read_only().channels().len(), 1);
+                       assert_eq!(network_graph.read_only().nodes().len(), 2);
+                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_none());
+
+                       use std::time::{SystemTime, UNIX_EPOCH};
+                       let announcement_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+                       network_graph.remove_stale_channels_with_time(announcement_time + 1 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
+               }
+
+               assert_eq!(network_graph.read_only().channels().len(), 0);
+               assert_eq!(network_graph.read_only().nodes().len(), 0);
+       }
+
+       #[test]
+       fn getting_next_channel_announcements() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               // Channels were not announced yet.
+               let channels_with_announcements = gossip_sync.get_next_channel_announcements(0, 1);
+               assert_eq!(channels_with_announcements.len(), 0);
+
+               let short_channel_id;
+               {
+                       // Announce a channel we will update
+                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
+                       match gossip_sync.handle_channel_announcement(&valid_channel_announcement) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+               }
+
+               // Contains initial channel announcement now.
+               let channels_with_announcements = gossip_sync.get_next_channel_announcements(short_channel_id, 1);
+               assert_eq!(channels_with_announcements.len(), 1);
+               if let Some(channel_announcements) = channels_with_announcements.first() {
+                       let &(_, ref update_1, ref update_2) = channel_announcements;
+                       assert_eq!(update_1, &None);
+                       assert_eq!(update_2, &None);
+               } else {
+                       panic!();
+               }
+
+
+               {
+                       // Valid channel update
+                       let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                               unsigned_channel_update.timestamp = 101;
+                       }, node_1_privkey, &secp_ctx);
+                       match gossip_sync.handle_channel_update(&valid_channel_update) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+               }
+
+               // Now contains an initial announcement and an update.
+               let channels_with_announcements = gossip_sync.get_next_channel_announcements(short_channel_id, 1);
+               assert_eq!(channels_with_announcements.len(), 1);
+               if let Some(channel_announcements) = channels_with_announcements.first() {
+                       let &(_, ref update_1, ref update_2) = channel_announcements;
+                       assert_ne!(update_1, &None);
+                       assert_eq!(update_2, &None);
+               } else {
+                       panic!();
+               }
+
+               {
+                       // Channel update with excess data.
+                       let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
+                               unsigned_channel_update.timestamp = 102;
+                               unsigned_channel_update.excess_data = [1; MAX_EXCESS_BYTES_FOR_RELAY + 1].to_vec();
+                       }, node_1_privkey, &secp_ctx);
+                       match gossip_sync.handle_channel_update(&valid_channel_update) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+               }
+
+               // Test that announcements with excess data won't be returned
+               let channels_with_announcements = gossip_sync.get_next_channel_announcements(short_channel_id, 1);
+               assert_eq!(channels_with_announcements.len(), 1);
+               if let Some(channel_announcements) = channels_with_announcements.first() {
+                       let &(_, ref update_1, ref update_2) = channel_announcements;
+                       assert_eq!(update_1, &None);
+                       assert_eq!(update_2, &None);
+               } else {
+                       panic!();
+               }
+
+               // Further starting point have no channels after it
+               let channels_with_announcements = gossip_sync.get_next_channel_announcements(short_channel_id + 1000, 1);
+               assert_eq!(channels_with_announcements.len(), 0);
+       }
+
+       #[test]
+       fn getting_next_node_announcements() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_1_privkey);
+
+               // No nodes yet.
+               let next_announcements = gossip_sync.get_next_node_announcements(None, 10);
+               assert_eq!(next_announcements.len(), 0);
+
+               {
+                       // Announce a channel to add 2 nodes
+                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+                       match gossip_sync.handle_channel_announcement(&valid_channel_announcement) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+               }
+
+
+               // Nodes were never announced
+               let next_announcements = gossip_sync.get_next_node_announcements(None, 3);
+               assert_eq!(next_announcements.len(), 0);
+
+               {
+                       let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
+                       match gossip_sync.handle_node_announcement(&valid_announcement) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+
+                       let valid_announcement = get_signed_node_announcement(|_| {}, node_2_privkey, &secp_ctx);
+                       match gossip_sync.handle_node_announcement(&valid_announcement) {
+                               Ok(_) => (),
+                               Err(_) => panic!()
+                       };
+               }
+
+               let next_announcements = gossip_sync.get_next_node_announcements(None, 3);
+               assert_eq!(next_announcements.len(), 2);
+
+               // Skip the first node.
+               let next_announcements = gossip_sync.get_next_node_announcements(Some(&node_id_1), 2);
+               assert_eq!(next_announcements.len(), 1);
+
+               {
+                       // Later announcement which should not be relayed (excess data) prevent us from sharing a node
+                       let valid_announcement = get_signed_node_announcement(|unsigned_announcement| {
+                               unsigned_announcement.timestamp += 10;
+                               unsigned_announcement.excess_data = [1; MAX_EXCESS_BYTES_FOR_RELAY + 1].to_vec();
+                       }, node_2_privkey, &secp_ctx);
+                       match gossip_sync.handle_node_announcement(&valid_announcement) {
+                               Ok(res) => assert!(!res),
+                               Err(_) => panic!()
+                       };
+               }
+
+               let next_announcements = gossip_sync.get_next_node_announcements(Some(&node_id_1), 2);
+               assert_eq!(next_announcements.len(), 0);
+       }
+
+       #[test]
+       fn network_graph_serialization() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+
+               // Announce a channel to add a corresponding node.
+               let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                       Ok(res) => assert!(res),
+                       _ => panic!()
+               };
+
+               let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
+               match gossip_sync.handle_node_announcement(&valid_announcement) {
+                       Ok(_) => (),
+                       Err(_) => panic!()
+               };
+
+               let mut w = test_utils::TestVecWriter(Vec::new());
+               assert!(!network_graph.read_only().nodes().is_empty());
+               assert!(!network_graph.read_only().channels().is_empty());
+               network_graph.write(&mut w).unwrap();
+               assert!(<NetworkGraph>::read(&mut io::Cursor::new(&w.0)).unwrap() == network_graph);
+       }
+
+       #[test]
+       fn network_graph_tlv_serialization() {
+               let network_graph = create_network_graph();
+               network_graph.set_last_rapid_gossip_sync_timestamp(42);
+
+               let mut w = test_utils::TestVecWriter(Vec::new());
+               network_graph.write(&mut w).unwrap();
+               let reassembled_network_graph: NetworkGraph = Readable::read(&mut io::Cursor::new(&w.0)).unwrap();
+               assert!(reassembled_network_graph == network_graph);
+               assert_eq!(reassembled_network_graph.get_last_rapid_gossip_sync_timestamp().unwrap(), 42);
+       }
+
+       #[test]
+       #[cfg(feature = "std")]
+       fn calling_sync_routing_table() {
+               use std::time::{SystemTime, UNIX_EPOCH};
+
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+               let node_privkey_1 = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_privkey_1);
+
+               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
+
+               // It should ignore if gossip_queries feature is not enabled
+               {
+                       let init_msg = Init { features: InitFeatures::known().clear_gossip_queries(), remote_network_address: None };
+                       gossip_sync.peer_connected(&node_id_1, &init_msg);
+                       let events = gossip_sync.get_and_clear_pending_msg_events();
+                       assert_eq!(events.len(), 0);
+               }
+
+               // It should send a gossip_timestamp_filter with the correct information
+               {
+                       let init_msg = Init { features: InitFeatures::known(), remote_network_address: None };
+                       gossip_sync.peer_connected(&node_id_1, &init_msg);
+                       let events = gossip_sync.get_and_clear_pending_msg_events();
+                       assert_eq!(events.len(), 1);
+                       match &events[0] {
+                               MessageSendEvent::SendGossipTimestampFilter{ node_id, msg } => {
+                                       assert_eq!(node_id, &node_id_1);
+                                       assert_eq!(msg.chain_hash, chain_hash);
+                                       let expected_timestamp = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
+                                       assert!((msg.first_timestamp as u64) >= expected_timestamp - 60*60*24*7*2);
+                                       assert!((msg.first_timestamp as u64) < expected_timestamp - 60*60*24*7*2 + 10);
+                                       assert_eq!(msg.timestamp_range, u32::max_value());
+                               },
+                               _ => panic!("Expected MessageSendEvent::SendChannelRangeQuery")
+                       };
+               }
+       }
+
+       #[test]
+       fn handling_query_channel_range() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+
+               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
+               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
+               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+               let node_id_2 = PublicKey::from_secret_key(&secp_ctx, node_2_privkey);
+
+               let mut scids: Vec<u64> = vec![
+                       scid_from_parts(0xfffffe, 0xffffff, 0xffff).unwrap(), // max
+                       scid_from_parts(0xffffff, 0xffffff, 0xffff).unwrap(), // never
+               ];
+
+               // used for testing multipart reply across blocks
+               for block in 100000..=108001 {
+                       scids.push(scid_from_parts(block, 0, 0).unwrap());
+               }
+
+               // used for testing resumption on same block
+               scids.push(scid_from_parts(108001, 1, 0).unwrap());
+
+               for scid in scids {
+                       let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
+                               unsigned_announcement.short_channel_id = scid;
+                       }, node_1_privkey, node_2_privkey, &secp_ctx);
+                       match gossip_sync.handle_channel_announcement(&valid_announcement) {
+                               Ok(_) => (),
+                               _ => panic!()
+                       };
+               }
+
+               // Error when number_of_blocks=0
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0,
+                               number_of_blocks: 0,
+                       },
+                       false,
+                       vec![ReplyChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0,
+                               number_of_blocks: 0,
+                               sync_complete: true,
+                               short_channel_ids: vec![]
+                       }]
+               );
+
+               // Error when wrong chain
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: genesis_block(Network::Bitcoin).header.block_hash(),
+                               first_blocknum: 0,
+                               number_of_blocks: 0xffff_ffff,
+                       },
+                       false,
+                       vec![ReplyChannelRange {
+                               chain_hash: genesis_block(Network::Bitcoin).header.block_hash(),
+                               first_blocknum: 0,
+                               number_of_blocks: 0xffff_ffff,
+                               sync_complete: true,
+                               short_channel_ids: vec![],
+                       }]
+               );
+
+               // Error when first_blocknum > 0xffffff
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0x01000000,
+                               number_of_blocks: 0xffff_ffff,
+                       },
+                       false,
+                       vec![ReplyChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0x01000000,
+                               number_of_blocks: 0xffff_ffff,
+                               sync_complete: true,
+                               short_channel_ids: vec![]
+                       }]
+               );
+
+               // Empty reply when max valid SCID block num
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0xffffff,
+                               number_of_blocks: 1,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 0xffffff,
+                                       number_of_blocks: 1,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![]
+                               },
+                       ]
+               );
+
+               // No results in valid query range
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 1000,
+                               number_of_blocks: 1000,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 1000,
+                                       number_of_blocks: 1000,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![],
+                               }
+                       ]
+               );
+
+               // Overflow first_blocknum + number_of_blocks
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 0xfe0000,
+                               number_of_blocks: 0xffffffff,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 0xfe0000,
+                                       number_of_blocks: 0xffffffff - 0xfe0000,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![
+                                               0xfffffe_ffffff_ffff, // max
+                                       ]
+                               }
+                       ]
+               );
+
+               // Single block exactly full
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 100000,
+                               number_of_blocks: 8000,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 100000,
+                                       number_of_blocks: 8000,
+                                       sync_complete: true,
+                                       short_channel_ids: (100000..=107999)
+                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
+                                               .collect(),
+                               },
+                       ]
+               );
+
+               // Multiple split on new block
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 100000,
+                               number_of_blocks: 8001,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 100000,
+                                       number_of_blocks: 7999,
+                                       sync_complete: false,
+                                       short_channel_ids: (100000..=107999)
+                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
+                                               .collect(),
+                               },
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 107999,
+                                       number_of_blocks: 2,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![
+                                               scid_from_parts(108000, 0, 0).unwrap(),
+                                       ],
+                               }
+                       ]
+               );
+
+               // Multiple split on same block
+               do_handling_query_channel_range(
+                       &gossip_sync,
+                       &node_id_2,
+                       QueryChannelRange {
+                               chain_hash: chain_hash.clone(),
+                               first_blocknum: 100002,
+                               number_of_blocks: 8000,
+                       },
+                       true,
+                       vec![
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 100002,
+                                       number_of_blocks: 7999,
+                                       sync_complete: false,
+                                       short_channel_ids: (100002..=108001)
+                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
+                                               .collect(),
+                               },
+                               ReplyChannelRange {
+                                       chain_hash: chain_hash.clone(),
+                                       first_blocknum: 108001,
+                                       number_of_blocks: 1,
+                                       sync_complete: true,
+                                       short_channel_ids: vec![
+                                               scid_from_parts(108001, 1, 0).unwrap(),
+                                       ],
+                               }
+                       ]
+               );
+       }
+
+       fn do_handling_query_channel_range(
+               gossip_sync: &P2PGossipSync<&NetworkGraph, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
+               test_node_id: &PublicKey,
+               msg: QueryChannelRange,
+               expected_ok: bool,
+               expected_replies: Vec<ReplyChannelRange>
+       ) {
+               let mut max_firstblocknum = msg.first_blocknum.saturating_sub(1);
+               let mut c_lightning_0_9_prev_end_blocknum = max_firstblocknum;
+               let query_end_blocknum = msg.end_blocknum();
+               let result = gossip_sync.handle_query_channel_range(test_node_id, msg);
+
+               if expected_ok {
+                       assert!(result.is_ok());
+               } else {
+                       assert!(result.is_err());
+               }
+
+               let events = gossip_sync.get_and_clear_pending_msg_events();
+               assert_eq!(events.len(), expected_replies.len());
+
+               for i in 0..events.len() {
+                       let expected_reply = &expected_replies[i];
+                       match &events[i] {
+                               MessageSendEvent::SendReplyChannelRange { node_id, msg } => {
+                                       assert_eq!(node_id, test_node_id);
+                                       assert_eq!(msg.chain_hash, expected_reply.chain_hash);
+                                       assert_eq!(msg.first_blocknum, expected_reply.first_blocknum);
+                                       assert_eq!(msg.number_of_blocks, expected_reply.number_of_blocks);
+                                       assert_eq!(msg.sync_complete, expected_reply.sync_complete);
+                                       assert_eq!(msg.short_channel_ids, expected_reply.short_channel_ids);
+
+                                       // Enforce exactly the sequencing requirements present on c-lightning v0.9.3
+                                       assert!(msg.first_blocknum == c_lightning_0_9_prev_end_blocknum || msg.first_blocknum == c_lightning_0_9_prev_end_blocknum.saturating_add(1));
+                                       assert!(msg.first_blocknum >= max_firstblocknum);
+                                       max_firstblocknum = msg.first_blocknum;
+                                       c_lightning_0_9_prev_end_blocknum = msg.first_blocknum.saturating_add(msg.number_of_blocks);
+
+                                       // Check that the last block count is >= the query's end_blocknum
+                                       if i == events.len() - 1 {
+                                               assert!(msg.first_blocknum.saturating_add(msg.number_of_blocks) >= query_end_blocknum);
+                                       }
+                               },
+                               _ => panic!("expected MessageSendEvent::SendReplyChannelRange"),
+                       }
+               }
+       }
+
+       #[test]
+       fn handling_query_short_channel_ids() {
+               let network_graph = create_network_graph();
+               let (secp_ctx, gossip_sync) = create_gossip_sync(&network_graph);
+               let node_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
+               let node_id = PublicKey::from_secret_key(&secp_ctx, node_privkey);
+
+               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
+
+               let result = gossip_sync.handle_query_short_channel_ids(&node_id, QueryShortChannelIds {
+                       chain_hash,
+                       short_channel_ids: vec![0x0003e8_000000_0000],
+               });
+               assert!(result.is_err());
+       }
+}
+
+#[cfg(all(test, feature = "_bench_unstable"))]
+mod benches {
+       use super::*;
+
+       use test::Bencher;
+       use std::io::Read;
+
+       #[bench]
+       fn read_network_graph(bench: &mut Bencher) {
+               let mut d = ::routing::router::test_utils::get_route_file().unwrap();
+               let mut v = Vec::new();
+               d.read_to_end(&mut v).unwrap();
+               bench.iter(|| {
+                       let _ = NetworkGraph::read(&mut std::io::Cursor::new(&v)).unwrap();
+               });
+       }
+
+       #[bench]
+       fn write_network_graph(bench: &mut Bencher) {
+               let mut d = ::routing::router::test_utils::get_route_file().unwrap();
+               let net_graph = NetworkGraph::read(&mut d).unwrap();
+               bench.iter(|| {
+                       let _ = net_graph.encode();
+               });
+       }
+}
index a3ab6c0c1b30ff04e50ab5892b91ef358a1980f2..c7babbe3027babb77b878f82768963d7a6319d92 100644 (file)
@@ -9,6 +9,6 @@
 
 //! Structs and impls for receiving messages about the network and storing the topology live here.
 
-pub mod network_graph;
+pub mod gossip;
 pub mod router;
 pub mod scoring;
diff --git a/lightning/src/routing/network_graph.rs b/lightning/src/routing/network_graph.rs
deleted file mode 100644 (file)
index 18b846c..0000000
+++ /dev/null
@@ -1,2771 +0,0 @@
-// This file is Copyright its original authors, visible in version control
-// history.
-//
-// This file is licensed under the Apache License, Version 2.0 <LICENSE-APACHE
-// or http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
-// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your option.
-// You may not use this file except in accordance with one or both of these
-// licenses.
-
-//! The top-level network map tracking logic lives here.
-
-use bitcoin::secp256k1::constants::PUBLIC_KEY_SIZE;
-use bitcoin::secp256k1::PublicKey;
-use bitcoin::secp256k1::Secp256k1;
-use bitcoin::secp256k1;
-
-use bitcoin::hashes::sha256d::Hash as Sha256dHash;
-use bitcoin::hashes::Hash;
-use bitcoin::blockdata::script::Builder;
-use bitcoin::blockdata::transaction::TxOut;
-use bitcoin::blockdata::opcodes;
-use bitcoin::hash_types::BlockHash;
-
-use chain;
-use chain::Access;
-use ln::features::{ChannelFeatures, NodeFeatures};
-use ln::msgs::{DecodeError, ErrorAction, Init, LightningError, RoutingMessageHandler, NetAddress, MAX_VALUE_MSAT};
-use ln::msgs::{ChannelAnnouncement, ChannelUpdate, NodeAnnouncement, OptionalField, GossipTimestampFilter};
-use ln::msgs::{QueryChannelRange, ReplyChannelRange, QueryShortChannelIds, ReplyShortChannelIdsEnd};
-use ln::msgs;
-use util::ser::{Writeable, Readable, Writer};
-use util::logger::{Logger, Level};
-use util::events::{Event, EventHandler, MessageSendEvent, MessageSendEventsProvider};
-use util::scid_utils::{block_from_scid, scid_from_parts, MAX_SCID_BLOCK};
-
-use io;
-use prelude::*;
-use alloc::collections::{BTreeMap, btree_map::Entry as BtreeEntry};
-use core::{cmp, fmt};
-use sync::{RwLock, RwLockReadGuard};
-use core::sync::atomic::{AtomicUsize, Ordering};
-use sync::Mutex;
-use core::ops::Deref;
-use bitcoin::hashes::hex::ToHex;
-
-#[cfg(feature = "std")]
-use std::time::{SystemTime, UNIX_EPOCH};
-
-/// We remove stale channel directional info two weeks after the last update, per BOLT 7's
-/// suggestion.
-const STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS: u64 = 60 * 60 * 24 * 14;
-
-/// The maximum number of extra bytes which we do not understand in a gossip message before we will
-/// refuse to relay the message.
-const MAX_EXCESS_BYTES_FOR_RELAY: usize = 1024;
-
-/// Maximum number of short_channel_ids that will be encoded in one gossip reply message.
-/// This value ensures a reply fits within the 65k payload limit and is consistent with other implementations.
-const MAX_SCIDS_PER_REPLY: usize = 8000;
-
-/// Represents the compressed public key of a node
-#[derive(Clone, Copy)]
-pub struct NodeId([u8; PUBLIC_KEY_SIZE]);
-
-impl NodeId {
-       /// Create a new NodeId from a public key
-       pub fn from_pubkey(pubkey: &PublicKey) -> Self {
-               NodeId(pubkey.serialize())
-       }
-
-       /// Get the public key slice from this NodeId
-       pub fn as_slice(&self) -> &[u8] {
-               &self.0
-       }
-}
-
-impl fmt::Debug for NodeId {
-       fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
-               write!(f, "NodeId({})", log_bytes!(self.0))
-       }
-}
-
-impl core::hash::Hash for NodeId {
-       fn hash<H: core::hash::Hasher>(&self, hasher: &mut H) {
-               self.0.hash(hasher);
-       }
-}
-
-impl Eq for NodeId {}
-
-impl PartialEq for NodeId {
-       fn eq(&self, other: &Self) -> bool {
-               self.0[..] == other.0[..]
-       }
-}
-
-impl cmp::PartialOrd for NodeId {
-       fn partial_cmp(&self, other: &Self) -> Option<cmp::Ordering> {
-               Some(self.cmp(other))
-       }
-}
-
-impl Ord for NodeId {
-       fn cmp(&self, other: &Self) -> cmp::Ordering {
-               self.0[..].cmp(&other.0[..])
-       }
-}
-
-impl Writeable for NodeId {
-       fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
-               writer.write_all(&self.0)?;
-               Ok(())
-       }
-}
-
-impl Readable for NodeId {
-       fn read<R: io::Read>(reader: &mut R) -> Result<Self, DecodeError> {
-               let mut buf = [0; PUBLIC_KEY_SIZE];
-               reader.read_exact(&mut buf)?;
-               Ok(Self(buf))
-       }
-}
-
-/// Represents the network as nodes and channels between them
-pub struct NetworkGraph {
-       last_rapid_gossip_sync_timestamp: Mutex<Option<u32>>,
-       genesis_hash: BlockHash,
-       // Lock order: channels -> nodes
-       channels: RwLock<BTreeMap<u64, ChannelInfo>>,
-       nodes: RwLock<BTreeMap<NodeId, NodeInfo>>,
-}
-
-impl Clone for NetworkGraph {
-       fn clone(&self) -> Self {
-               let channels = self.channels.read().unwrap();
-               let nodes = self.nodes.read().unwrap();
-               let last_rapid_gossip_sync_timestamp = self.get_last_rapid_gossip_sync_timestamp();
-               Self {
-                       genesis_hash: self.genesis_hash.clone(),
-                       channels: RwLock::new(channels.clone()),
-                       nodes: RwLock::new(nodes.clone()),
-                       last_rapid_gossip_sync_timestamp: Mutex::new(last_rapid_gossip_sync_timestamp)
-               }
-       }
-}
-
-/// A read-only view of [`NetworkGraph`].
-pub struct ReadOnlyNetworkGraph<'a> {
-       channels: RwLockReadGuard<'a, BTreeMap<u64, ChannelInfo>>,
-       nodes: RwLockReadGuard<'a, BTreeMap<NodeId, NodeInfo>>,
-}
-
-/// Update to the [`NetworkGraph`] based on payment failure information conveyed via the Onion
-/// return packet by a node along the route. See [BOLT #4] for details.
-///
-/// [BOLT #4]: https://github.com/lightning/bolts/blob/master/04-onion-routing.md
-#[derive(Clone, Debug, PartialEq)]
-pub enum NetworkUpdate {
-       /// An error indicating a `channel_update` messages should be applied via
-       /// [`NetworkGraph::update_channel`].
-       ChannelUpdateMessage {
-               /// The update to apply via [`NetworkGraph::update_channel`].
-               msg: ChannelUpdate,
-       },
-       /// An error indicating only that a channel has been closed, which should be applied via
-       /// [`NetworkGraph::close_channel_from_update`].
-       ChannelClosed {
-               /// The short channel id of the closed channel.
-               short_channel_id: u64,
-               /// Whether the channel should be permanently removed or temporarily disabled until a new
-               /// `channel_update` message is received.
-               is_permanent: bool,
-       },
-       /// An error indicating only that a node has failed, which should be applied via
-       /// [`NetworkGraph::fail_node`].
-       NodeFailure {
-               /// The node id of the failed node.
-               node_id: PublicKey,
-               /// Whether the node should be permanently removed from consideration or can be restored
-               /// when a new `channel_update` message is received.
-               is_permanent: bool,
-       }
-}
-
-impl_writeable_tlv_based_enum_upgradable!(NetworkUpdate,
-       (0, ChannelUpdateMessage) => {
-               (0, msg, required),
-       },
-       (2, ChannelClosed) => {
-               (0, short_channel_id, required),
-               (2, is_permanent, required),
-       },
-       (4, NodeFailure) => {
-               (0, node_id, required),
-               (2, is_permanent, required),
-       },
-);
-
-impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> EventHandler for NetGraphMsgHandler<G, C, L>
-where C::Target: chain::Access, L::Target: Logger {
-       fn handle_event(&self, event: &Event) {
-               if let Event::PaymentPathFailed { payment_hash: _, rejected_by_dest: _, network_update, .. } = event {
-                       if let Some(network_update) = network_update {
-                               self.handle_network_update(network_update);
-                       }
-               }
-       }
-}
-
-/// Receives and validates network updates from peers,
-/// stores authentic and relevant data as a network graph.
-/// This network graph is then used for routing payments.
-/// Provides interface to help with initial routing sync by
-/// serving historical announcements.
-///
-/// Serves as an [`EventHandler`] for applying updates from [`Event::PaymentPathFailed`] to the
-/// [`NetworkGraph`].
-pub struct NetGraphMsgHandler<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref>
-where C::Target: chain::Access, L::Target: Logger
-{
-       secp_ctx: Secp256k1<secp256k1::VerifyOnly>,
-       network_graph: G,
-       chain_access: Option<C>,
-       full_syncs_requested: AtomicUsize,
-       pending_events: Mutex<Vec<MessageSendEvent>>,
-       logger: L,
-}
-
-impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> NetGraphMsgHandler<G, C, L>
-where C::Target: chain::Access, L::Target: Logger
-{
-       /// Creates a new tracker of the actual state of the network of channels and nodes,
-       /// assuming an existing Network Graph.
-       /// Chain monitor is used to make sure announced channels exist on-chain,
-       /// channel data is correct, and that the announcement is signed with
-       /// channel owners' keys.
-       pub fn new(network_graph: G, chain_access: Option<C>, logger: L) -> Self {
-               NetGraphMsgHandler {
-                       secp_ctx: Secp256k1::verification_only(),
-                       network_graph,
-                       full_syncs_requested: AtomicUsize::new(0),
-                       chain_access,
-                       pending_events: Mutex::new(vec![]),
-                       logger,
-               }
-       }
-
-       /// Adds a provider used to check new announcements. Does not affect
-       /// existing announcements unless they are updated.
-       /// Add, update or remove the provider would replace the current one.
-       pub fn add_chain_access(&mut self, chain_access: Option<C>) {
-               self.chain_access = chain_access;
-       }
-
-       /// Gets a reference to the underlying [`NetworkGraph`] which was provided in
-       /// [`NetGraphMsgHandler::new`].
-       ///
-       /// (C-not exported) as bindings don't support a reference-to-a-reference yet
-       pub fn network_graph(&self) -> &G {
-               &self.network_graph
-       }
-
-       /// Returns true when a full routing table sync should be performed with a peer.
-       fn should_request_full_sync(&self, _node_id: &PublicKey) -> bool {
-               //TODO: Determine whether to request a full sync based on the network map.
-               const FULL_SYNCS_TO_REQUEST: usize = 5;
-               if self.full_syncs_requested.load(Ordering::Acquire) < FULL_SYNCS_TO_REQUEST {
-                       self.full_syncs_requested.fetch_add(1, Ordering::AcqRel);
-                       true
-               } else {
-                       false
-               }
-       }
-
-       /// Applies changes to the [`NetworkGraph`] from the given update.
-       fn handle_network_update(&self, update: &NetworkUpdate) {
-               match *update {
-                       NetworkUpdate::ChannelUpdateMessage { ref msg } => {
-                               let short_channel_id = msg.contents.short_channel_id;
-                               let is_enabled = msg.contents.flags & (1 << 1) != (1 << 1);
-                               let status = if is_enabled { "enabled" } else { "disabled" };
-                               log_debug!(self.logger, "Updating channel with channel_update from a payment failure. Channel {} is {}.", short_channel_id, status);
-                               let _ = self.network_graph.update_channel(msg, &self.secp_ctx);
-                       },
-                       NetworkUpdate::ChannelClosed { short_channel_id, is_permanent } => {
-                               let action = if is_permanent { "Removing" } else { "Disabling" };
-                               log_debug!(self.logger, "{} channel graph entry for {} due to a payment failure.", action, short_channel_id);
-                               self.network_graph.close_channel_from_update(short_channel_id, is_permanent);
-                       },
-                       NetworkUpdate::NodeFailure { ref node_id, is_permanent } => {
-                               let action = if is_permanent { "Removing" } else { "Disabling" };
-                               log_debug!(self.logger, "{} node graph entry for {} due to a payment failure.", action, node_id);
-                               self.network_graph.fail_node(node_id, is_permanent);
-                       },
-               }
-       }
-}
-
-macro_rules! secp_verify_sig {
-       ( $secp_ctx: expr, $msg: expr, $sig: expr, $pubkey: expr, $msg_type: expr ) => {
-               match $secp_ctx.verify_ecdsa($msg, $sig, $pubkey) {
-                       Ok(_) => {},
-                       Err(_) => {
-                               return Err(LightningError {
-                                       err: format!("Invalid signature on {} message", $msg_type),
-                                       action: ErrorAction::SendWarningMessage {
-                                               msg: msgs::WarningMessage {
-                                                       channel_id: [0; 32],
-                                                       data: format!("Invalid signature on {} message", $msg_type),
-                                               },
-                                               log_level: Level::Trace,
-                                       },
-                               });
-                       },
-               }
-       };
-}
-
-impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> RoutingMessageHandler for NetGraphMsgHandler<G, C, L>
-where C::Target: chain::Access, L::Target: Logger
-{
-       fn handle_node_announcement(&self, msg: &msgs::NodeAnnouncement) -> Result<bool, LightningError> {
-               self.network_graph.update_node_from_announcement(msg, &self.secp_ctx)?;
-               Ok(msg.contents.excess_data.len() <=  MAX_EXCESS_BYTES_FOR_RELAY &&
-                  msg.contents.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
-                  msg.contents.excess_data.len() + msg.contents.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
-       }
-
-       fn handle_channel_announcement(&self, msg: &msgs::ChannelAnnouncement) -> Result<bool, LightningError> {
-               self.network_graph.update_channel_from_announcement(msg, &self.chain_access, &self.secp_ctx)?;
-               log_gossip!(self.logger, "Added channel_announcement for {}{}", msg.contents.short_channel_id, if !msg.contents.excess_data.is_empty() { " with excess uninterpreted data!" } else { "" });
-               Ok(msg.contents.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
-       }
-
-       fn handle_channel_update(&self, msg: &msgs::ChannelUpdate) -> Result<bool, LightningError> {
-               self.network_graph.update_channel(msg, &self.secp_ctx)?;
-               Ok(msg.contents.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY)
-       }
-
-       fn get_next_channel_announcements(&self, starting_point: u64, batch_amount: u8) -> Vec<(ChannelAnnouncement, Option<ChannelUpdate>, Option<ChannelUpdate>)> {
-               let mut result = Vec::with_capacity(batch_amount as usize);
-               let channels = self.network_graph.channels.read().unwrap();
-               let mut iter = channels.range(starting_point..);
-               while result.len() < batch_amount as usize {
-                       if let Some((_, ref chan)) = iter.next() {
-                               if chan.announcement_message.is_some() {
-                                       let chan_announcement = chan.announcement_message.clone().unwrap();
-                                       let mut one_to_two_announcement: Option<msgs::ChannelUpdate> = None;
-                                       let mut two_to_one_announcement: Option<msgs::ChannelUpdate> = None;
-                                       if let Some(one_to_two) = chan.one_to_two.as_ref() {
-                                               one_to_two_announcement = one_to_two.last_update_message.clone();
-                                       }
-                                       if let Some(two_to_one) = chan.two_to_one.as_ref() {
-                                               two_to_one_announcement = two_to_one.last_update_message.clone();
-                                       }
-                                       result.push((chan_announcement, one_to_two_announcement, two_to_one_announcement));
-                               } else {
-                                       // TODO: We may end up sending un-announced channel_updates if we are sending
-                                       // initial sync data while receiving announce/updates for this channel.
-                               }
-                       } else {
-                               return result;
-                       }
-               }
-               result
-       }
-
-       fn get_next_node_announcements(&self, starting_point: Option<&PublicKey>, batch_amount: u8) -> Vec<NodeAnnouncement> {
-               let mut result = Vec::with_capacity(batch_amount as usize);
-               let nodes = self.network_graph.nodes.read().unwrap();
-               let mut iter = if let Some(pubkey) = starting_point {
-                               let mut iter = nodes.range(NodeId::from_pubkey(pubkey)..);
-                               iter.next();
-                               iter
-                       } else {
-                               nodes.range::<NodeId, _>(..)
-                       };
-               while result.len() < batch_amount as usize {
-                       if let Some((_, ref node)) = iter.next() {
-                               if let Some(node_info) = node.announcement_info.as_ref() {
-                                       if node_info.announcement_message.is_some() {
-                                               result.push(node_info.announcement_message.clone().unwrap());
-                                       }
-                               }
-                       } else {
-                               return result;
-                       }
-               }
-               result
-       }
-
-       /// Initiates a stateless sync of routing gossip information with a peer
-       /// using gossip_queries. The default strategy used by this implementation
-       /// is to sync the full block range with several peers.
-       ///
-       /// We should expect one or more reply_channel_range messages in response
-       /// to our query_channel_range. Each reply will enqueue a query_scid message
-       /// to request gossip messages for each channel. The sync is considered complete
-       /// when the final reply_scids_end message is received, though we are not
-       /// tracking this directly.
-       fn peer_connected(&self, their_node_id: &PublicKey, init_msg: &Init) {
-               // We will only perform a sync with peers that support gossip_queries.
-               if !init_msg.features.supports_gossip_queries() {
-                       return ();
-               }
-
-               // The lightning network's gossip sync system is completely broken in numerous ways.
-               //
-               // Given no broadly-available set-reconciliation protocol, the only reasonable approach is
-               // to do a full sync from the first few peers we connect to, and then receive gossip
-               // updates from all our peers normally.
-               //
-               // Originally, we could simply tell a peer to dump us the entire gossip table on startup,
-               // wasting lots of bandwidth but ensuring we have the full network graph. After the initial
-               // dump peers would always send gossip and we'd stay up-to-date with whatever our peer has
-               // seen.
-               //
-               // In order to reduce the bandwidth waste, "gossip queries" were introduced, allowing you
-               // to ask for the SCIDs of all channels in your peer's routing graph, and then only request
-               // channel data which you are missing. Except there was no way at all to identify which
-               // `channel_update`s you were missing, so you still had to request everything, just in a
-               // very complicated way with some queries instead of just getting the dump.
-               //
-               // Later, an option was added to fetch the latest timestamps of the `channel_update`s to
-               // make efficient sync possible, however it has yet to be implemented in lnd, which makes
-               // relying on it useless.
-               //
-               // After gossip queries were introduced, support for receiving a full gossip table dump on
-               // connection was removed from several nodes, making it impossible to get a full sync
-               // without using the "gossip queries" messages.
-               //
-               // Once you opt into "gossip queries" the only way to receive any gossip updates that a
-               // peer receives after you connect, you must send a `gossip_timestamp_filter` message. This
-               // message, as the name implies, tells the peer to not forward any gossip messages with a
-               // timestamp older than a given value (not the time the peer received the filter, but the
-               // timestamp in the update message, which is often hours behind when the peer received the
-               // message).
-               //
-               // Obnoxiously, `gossip_timestamp_filter` isn't *just* a filter, but its also a request for
-               // your peer to send you the full routing graph (subject to the filter). Thus, in order to
-               // tell a peer to send you any updates as it sees them, you have to also ask for the full
-               // routing graph to be synced. If you set a timestamp filter near the current time, peers
-               // will simply not forward any new updates they see to you which were generated some time
-               // ago (which is not uncommon). If you instead set a timestamp filter near 0 (or two weeks
-               // ago), you will always get the full routing graph from all your peers.
-               //
-               // Most lightning nodes today opt to simply turn off receiving gossip data which only
-               // propagated some time after it was generated, and, worse, often disable gossiping with
-               // several peers after their first connection. The second behavior can cause gossip to not
-               // propagate fully if there are cuts in the gossiping subgraph.
-               //
-               // In an attempt to cut a middle ground between always fetching the full graph from all of
-               // our peers and never receiving gossip from peers at all, we send all of our peers a
-               // `gossip_timestamp_filter`, with the filter time set either two weeks ago or an hour ago.
-               //
-               // For no-std builds, we bury our head in the sand and do a full sync on each connection.
-               let should_request_full_sync = self.should_request_full_sync(&their_node_id);
-               #[allow(unused_mut, unused_assignments)]
-               let mut gossip_start_time = 0;
-               #[cfg(feature = "std")]
-               {
-                       gossip_start_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-                       if should_request_full_sync {
-                               gossip_start_time -= 60 * 60 * 24 * 7 * 2; // 2 weeks ago
-                       } else {
-                               gossip_start_time -= 60 * 60; // an hour ago
-                       }
-               }
-
-               let mut pending_events = self.pending_events.lock().unwrap();
-               pending_events.push(MessageSendEvent::SendGossipTimestampFilter {
-                       node_id: their_node_id.clone(),
-                       msg: GossipTimestampFilter {
-                               chain_hash: self.network_graph.genesis_hash,
-                               first_timestamp: gossip_start_time as u32, // 2106 issue!
-                               timestamp_range: u32::max_value(),
-                       },
-               });
-       }
-
-       fn handle_reply_channel_range(&self, _their_node_id: &PublicKey, _msg: ReplyChannelRange) -> Result<(), LightningError> {
-               // We don't make queries, so should never receive replies. If, in the future, the set
-               // reconciliation extensions to gossip queries become broadly supported, we should revert
-               // this code to its state pre-0.0.106.
-               Ok(())
-       }
-
-       fn handle_reply_short_channel_ids_end(&self, _their_node_id: &PublicKey, _msg: ReplyShortChannelIdsEnd) -> Result<(), LightningError> {
-               // We don't make queries, so should never receive replies. If, in the future, the set
-               // reconciliation extensions to gossip queries become broadly supported, we should revert
-               // this code to its state pre-0.0.106.
-               Ok(())
-       }
-
-       /// Processes a query from a peer by finding announced/public channels whose funding UTXOs
-       /// are in the specified block range. Due to message size limits, large range
-       /// queries may result in several reply messages. This implementation enqueues
-       /// all reply messages into pending events. Each message will allocate just under 65KiB. A full
-       /// sync of the public routing table with 128k channels will generated 16 messages and allocate ~1MB.
-       /// Logic can be changed to reduce allocation if/when a full sync of the routing table impacts
-       /// memory constrained systems.
-       fn handle_query_channel_range(&self, their_node_id: &PublicKey, msg: QueryChannelRange) -> Result<(), LightningError> {
-               log_debug!(self.logger, "Handling query_channel_range peer={}, first_blocknum={}, number_of_blocks={}", log_pubkey!(their_node_id), msg.first_blocknum, msg.number_of_blocks);
-
-               let inclusive_start_scid = scid_from_parts(msg.first_blocknum as u64, 0, 0);
-
-               // We might receive valid queries with end_blocknum that would overflow SCID conversion.
-               // If so, we manually cap the ending block to avoid this overflow.
-               let exclusive_end_scid = scid_from_parts(cmp::min(msg.end_blocknum() as u64, MAX_SCID_BLOCK), 0, 0);
-
-               // Per spec, we must reply to a query. Send an empty message when things are invalid.
-               if msg.chain_hash != self.network_graph.genesis_hash || inclusive_start_scid.is_err() || exclusive_end_scid.is_err() || msg.number_of_blocks == 0 {
-                       let mut pending_events = self.pending_events.lock().unwrap();
-                       pending_events.push(MessageSendEvent::SendReplyChannelRange {
-                               node_id: their_node_id.clone(),
-                               msg: ReplyChannelRange {
-                                       chain_hash: msg.chain_hash.clone(),
-                                       first_blocknum: msg.first_blocknum,
-                                       number_of_blocks: msg.number_of_blocks,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![],
-                               }
-                       });
-                       return Err(LightningError {
-                               err: String::from("query_channel_range could not be processed"),
-                               action: ErrorAction::IgnoreError,
-                       });
-               }
-
-               // Creates channel batches. We are not checking if the channel is routable
-               // (has at least one update). A peer may still want to know the channel
-               // exists even if its not yet routable.
-               let mut batches: Vec<Vec<u64>> = vec![Vec::with_capacity(MAX_SCIDS_PER_REPLY)];
-               let channels = self.network_graph.channels.read().unwrap();
-               for (_, ref chan) in channels.range(inclusive_start_scid.unwrap()..exclusive_end_scid.unwrap()) {
-                       if let Some(chan_announcement) = &chan.announcement_message {
-                               // Construct a new batch if last one is full
-                               if batches.last().unwrap().len() == batches.last().unwrap().capacity() {
-                                       batches.push(Vec::with_capacity(MAX_SCIDS_PER_REPLY));
-                               }
-
-                               let batch = batches.last_mut().unwrap();
-                               batch.push(chan_announcement.contents.short_channel_id);
-                       }
-               }
-               drop(channels);
-
-               let mut pending_events = self.pending_events.lock().unwrap();
-               let batch_count = batches.len();
-               let mut prev_batch_endblock = msg.first_blocknum;
-               for (batch_index, batch) in batches.into_iter().enumerate() {
-                       // Per spec, the initial `first_blocknum` needs to be <= the query's `first_blocknum`
-                       // and subsequent `first_blocknum`s must be >= the prior reply's `first_blocknum`.
-                       //
-                       // Additionally, c-lightning versions < 0.10 require that the `first_blocknum` of each
-                       // reply is >= the previous reply's `first_blocknum` and either exactly the previous
-                       // reply's `first_blocknum + number_of_blocks` or exactly one greater. This is a
-                       // significant diversion from the requirements set by the spec, and, in case of blocks
-                       // with no channel opens (e.g. empty blocks), requires that we use the previous value
-                       // and *not* derive the first_blocknum from the actual first block of the reply.
-                       let first_blocknum = prev_batch_endblock;
-
-                       // Each message carries the number of blocks (from the `first_blocknum`) its contents
-                       // fit in. Though there is no requirement that we use exactly the number of blocks its
-                       // contents are from, except for the bogus requirements c-lightning enforces, above.
-                       //
-                       // Per spec, the last end block (ie `first_blocknum + number_of_blocks`) needs to be
-                       // >= the query's end block. Thus, for the last reply, we calculate the difference
-                       // between the query's end block and the start of the reply.
-                       //
-                       // Overflow safe since end_blocknum=msg.first_block_num+msg.number_of_blocks and
-                       // first_blocknum will be either msg.first_blocknum or a higher block height.
-                       let (sync_complete, number_of_blocks) = if batch_index == batch_count-1 {
-                               (true, msg.end_blocknum() - first_blocknum)
-                       }
-                       // Prior replies should use the number of blocks that fit into the reply. Overflow
-                       // safe since first_blocknum is always <= last SCID's block.
-                       else {
-                               (false, block_from_scid(batch.last().unwrap()) - first_blocknum)
-                       };
-
-                       prev_batch_endblock = first_blocknum + number_of_blocks;
-
-                       pending_events.push(MessageSendEvent::SendReplyChannelRange {
-                               node_id: their_node_id.clone(),
-                               msg: ReplyChannelRange {
-                                       chain_hash: msg.chain_hash.clone(),
-                                       first_blocknum,
-                                       number_of_blocks,
-                                       sync_complete,
-                                       short_channel_ids: batch,
-                               }
-                       });
-               }
-
-               Ok(())
-       }
-
-       fn handle_query_short_channel_ids(&self, _their_node_id: &PublicKey, _msg: QueryShortChannelIds) -> Result<(), LightningError> {
-               // TODO
-               Err(LightningError {
-                       err: String::from("Not implemented"),
-                       action: ErrorAction::IgnoreError,
-               })
-       }
-}
-
-impl<G: Deref<Target=NetworkGraph>, C: Deref, L: Deref> MessageSendEventsProvider for NetGraphMsgHandler<G, C, L>
-where
-       C::Target: chain::Access,
-       L::Target: Logger,
-{
-       fn get_and_clear_pending_msg_events(&self) -> Vec<MessageSendEvent> {
-               let mut ret = Vec::new();
-               let mut pending_events = self.pending_events.lock().unwrap();
-               core::mem::swap(&mut ret, &mut pending_events);
-               ret
-       }
-}
-
-#[derive(Clone, Debug, PartialEq)]
-/// Details about one direction of a channel as received within a [`ChannelUpdate`].
-pub struct ChannelUpdateInfo {
-       /// When the last update to the channel direction was issued.
-       /// Value is opaque, as set in the announcement.
-       pub last_update: u32,
-       /// Whether the channel can be currently used for payments (in this one direction).
-       pub enabled: bool,
-       /// The difference in CLTV values that you must have when routing through this channel.
-       pub cltv_expiry_delta: u16,
-       /// The minimum value, which must be relayed to the next hop via the channel
-       pub htlc_minimum_msat: u64,
-       /// The maximum value which may be relayed to the next hop via the channel.
-       pub htlc_maximum_msat: Option<u64>,
-       /// Fees charged when the channel is used for routing
-       pub fees: RoutingFees,
-       /// Most recent update for the channel received from the network
-       /// Mostly redundant with the data we store in fields explicitly.
-       /// Everything else is useful only for sending out for initial routing sync.
-       /// Not stored if contains excess data to prevent DoS.
-       pub last_update_message: Option<ChannelUpdate>,
-}
-
-impl fmt::Display for ChannelUpdateInfo {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               write!(f, "last_update {}, enabled {}, cltv_expiry_delta {}, htlc_minimum_msat {}, fees {:?}", self.last_update, self.enabled, self.cltv_expiry_delta, self.htlc_minimum_msat, self.fees)?;
-               Ok(())
-       }
-}
-
-impl_writeable_tlv_based!(ChannelUpdateInfo, {
-       (0, last_update, required),
-       (2, enabled, required),
-       (4, cltv_expiry_delta, required),
-       (6, htlc_minimum_msat, required),
-       (8, htlc_maximum_msat, required),
-       (10, fees, required),
-       (12, last_update_message, required),
-});
-
-#[derive(Clone, Debug, PartialEq)]
-/// Details about a channel (both directions).
-/// Received within a channel announcement.
-pub struct ChannelInfo {
-       /// Protocol features of a channel communicated during its announcement
-       pub features: ChannelFeatures,
-       /// Source node of the first direction of a channel
-       pub node_one: NodeId,
-       /// Details about the first direction of a channel
-       pub one_to_two: Option<ChannelUpdateInfo>,
-       /// Source node of the second direction of a channel
-       pub node_two: NodeId,
-       /// Details about the second direction of a channel
-       pub two_to_one: Option<ChannelUpdateInfo>,
-       /// The channel capacity as seen on-chain, if chain lookup is available.
-       pub capacity_sats: Option<u64>,
-       /// An initial announcement of the channel
-       /// Mostly redundant with the data we store in fields explicitly.
-       /// Everything else is useful only for sending out for initial routing sync.
-       /// Not stored if contains excess data to prevent DoS.
-       pub announcement_message: Option<ChannelAnnouncement>,
-       /// The timestamp when we received the announcement, if we are running with feature = "std"
-       /// (which we can probably assume we are - no-std environments probably won't have a full
-       /// network graph in memory!).
-       announcement_received_time: u64,
-}
-
-impl ChannelInfo {
-       /// Returns a [`DirectedChannelInfo`] for the channel directed to the given `target` from a
-       /// returned `source`, or `None` if `target` is not one of the channel's counterparties.
-       pub fn as_directed_to(&self, target: &NodeId) -> Option<(DirectedChannelInfo, &NodeId)> {
-               let (direction, source) = {
-                       if target == &self.node_one {
-                               (self.two_to_one.as_ref(), &self.node_two)
-                       } else if target == &self.node_two {
-                               (self.one_to_two.as_ref(), &self.node_one)
-                       } else {
-                               return None;
-                       }
-               };
-               Some((DirectedChannelInfo::new(self, direction), source))
-       }
-
-       /// Returns a [`DirectedChannelInfo`] for the channel directed from the given `source` to a
-       /// returned `target`, or `None` if `source` is not one of the channel's counterparties.
-       pub fn as_directed_from(&self, source: &NodeId) -> Option<(DirectedChannelInfo, &NodeId)> {
-               let (direction, target) = {
-                       if source == &self.node_one {
-                               (self.one_to_two.as_ref(), &self.node_two)
-                       } else if source == &self.node_two {
-                               (self.two_to_one.as_ref(), &self.node_one)
-                       } else {
-                               return None;
-                       }
-               };
-               Some((DirectedChannelInfo::new(self, direction), target))
-       }
-
-       /// Returns a [`ChannelUpdateInfo`] based on the direction implied by the channel_flag.
-       pub fn get_directional_info(&self, channel_flags: u8) -> Option<&ChannelUpdateInfo> {
-               let direction = channel_flags & 1u8;
-               if direction == 0 {
-                       self.one_to_two.as_ref()
-               } else {
-                       self.two_to_one.as_ref()
-               }
-       }
-}
-
-impl fmt::Display for ChannelInfo {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               write!(f, "features: {}, node_one: {}, one_to_two: {:?}, node_two: {}, two_to_one: {:?}",
-                  log_bytes!(self.features.encode()), log_bytes!(self.node_one.as_slice()), self.one_to_two, log_bytes!(self.node_two.as_slice()), self.two_to_one)?;
-               Ok(())
-       }
-}
-
-impl_writeable_tlv_based!(ChannelInfo, {
-       (0, features, required),
-       (1, announcement_received_time, (default_value, 0)),
-       (2, node_one, required),
-       (4, one_to_two, required),
-       (6, node_two, required),
-       (8, two_to_one, required),
-       (10, capacity_sats, required),
-       (12, announcement_message, required),
-});
-
-/// A wrapper around [`ChannelInfo`] representing information about the channel as directed from a
-/// source node to a target node.
-#[derive(Clone)]
-pub struct DirectedChannelInfo<'a> {
-       channel: &'a ChannelInfo,
-       direction: Option<&'a ChannelUpdateInfo>,
-       htlc_maximum_msat: u64,
-       effective_capacity: EffectiveCapacity,
-}
-
-impl<'a> DirectedChannelInfo<'a> {
-       #[inline]
-       fn new(channel: &'a ChannelInfo, direction: Option<&'a ChannelUpdateInfo>) -> Self {
-               let htlc_maximum_msat = direction.and_then(|direction| direction.htlc_maximum_msat);
-               let capacity_msat = channel.capacity_sats.map(|capacity_sats| capacity_sats * 1000);
-
-               let (htlc_maximum_msat, effective_capacity) = match (htlc_maximum_msat, capacity_msat) {
-                       (Some(amount_msat), Some(capacity_msat)) => {
-                               let htlc_maximum_msat = cmp::min(amount_msat, capacity_msat);
-                               (htlc_maximum_msat, EffectiveCapacity::Total { capacity_msat })
-                       },
-                       (Some(amount_msat), None) => {
-                               (amount_msat, EffectiveCapacity::MaximumHTLC { amount_msat })
-                       },
-                       (None, Some(capacity_msat)) => {
-                               (capacity_msat, EffectiveCapacity::Total { capacity_msat })
-                       },
-                       (None, None) => (EffectiveCapacity::Unknown.as_msat(), EffectiveCapacity::Unknown),
-               };
-
-               Self {
-                       channel, direction, htlc_maximum_msat, effective_capacity
-               }
-       }
-
-       /// Returns information for the channel.
-       pub fn channel(&self) -> &'a ChannelInfo { self.channel }
-
-       /// Returns information for the direction.
-       pub fn direction(&self) -> Option<&'a ChannelUpdateInfo> { self.direction }
-
-       /// Returns the maximum HTLC amount allowed over the channel in the direction.
-       pub fn htlc_maximum_msat(&self) -> u64 {
-               self.htlc_maximum_msat
-       }
-
-       /// Returns the [`EffectiveCapacity`] of the channel in the direction.
-       ///
-       /// This is either the total capacity from the funding transaction, if known, or the
-       /// `htlc_maximum_msat` for the direction as advertised by the gossip network, if known,
-       /// otherwise.
-       pub fn effective_capacity(&self) -> EffectiveCapacity {
-               self.effective_capacity
-       }
-
-       /// Returns `Some` if [`ChannelUpdateInfo`] is available in the direction.
-       pub(super) fn with_update(self) -> Option<DirectedChannelInfoWithUpdate<'a>> {
-               match self.direction {
-                       Some(_) => Some(DirectedChannelInfoWithUpdate { inner: self }),
-                       None => None,
-               }
-       }
-}
-
-impl<'a> fmt::Debug for DirectedChannelInfo<'a> {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               f.debug_struct("DirectedChannelInfo")
-                       .field("channel", &self.channel)
-                       .finish()
-       }
-}
-
-/// A [`DirectedChannelInfo`] with [`ChannelUpdateInfo`] available in its direction.
-#[derive(Clone)]
-pub(super) struct DirectedChannelInfoWithUpdate<'a> {
-       inner: DirectedChannelInfo<'a>,
-}
-
-impl<'a> DirectedChannelInfoWithUpdate<'a> {
-       /// Returns information for the channel.
-       #[inline]
-       pub(super) fn channel(&self) -> &'a ChannelInfo { &self.inner.channel }
-
-       /// Returns information for the direction.
-       #[inline]
-       pub(super) fn direction(&self) -> &'a ChannelUpdateInfo { self.inner.direction.unwrap() }
-
-       /// Returns the [`EffectiveCapacity`] of the channel in the direction.
-       #[inline]
-       pub(super) fn effective_capacity(&self) -> EffectiveCapacity { self.inner.effective_capacity() }
-
-       /// Returns the maximum HTLC amount allowed over the channel in the direction.
-       #[inline]
-       pub(super) fn htlc_maximum_msat(&self) -> u64 { self.inner.htlc_maximum_msat() }
-}
-
-impl<'a> fmt::Debug for DirectedChannelInfoWithUpdate<'a> {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               self.inner.fmt(f)
-       }
-}
-
-/// The effective capacity of a channel for routing purposes.
-///
-/// While this may be smaller than the actual channel capacity, amounts greater than
-/// [`Self::as_msat`] should not be routed through the channel.
-#[derive(Clone, Copy)]
-pub enum EffectiveCapacity {
-       /// The available liquidity in the channel known from being a channel counterparty, and thus a
-       /// direct hop.
-       ExactLiquidity {
-               /// Either the inbound or outbound liquidity depending on the direction, denominated in
-               /// millisatoshi.
-               liquidity_msat: u64,
-       },
-       /// The maximum HTLC amount in one direction as advertised on the gossip network.
-       MaximumHTLC {
-               /// The maximum HTLC amount denominated in millisatoshi.
-               amount_msat: u64,
-       },
-       /// The total capacity of the channel as determined by the funding transaction.
-       Total {
-               /// The funding amount denominated in millisatoshi.
-               capacity_msat: u64,
-       },
-       /// A capacity sufficient to route any payment, typically used for private channels provided by
-       /// an invoice.
-       Infinite,
-       /// A capacity that is unknown possibly because either the chain state is unavailable to know
-       /// the total capacity or the `htlc_maximum_msat` was not advertised on the gossip network.
-       Unknown,
-}
-
-/// The presumed channel capacity denominated in millisatoshi for [`EffectiveCapacity::Unknown`] to
-/// use when making routing decisions.
-pub const UNKNOWN_CHANNEL_CAPACITY_MSAT: u64 = 250_000 * 1000;
-
-impl EffectiveCapacity {
-       /// Returns the effective capacity denominated in millisatoshi.
-       pub fn as_msat(&self) -> u64 {
-               match self {
-                       EffectiveCapacity::ExactLiquidity { liquidity_msat } => *liquidity_msat,
-                       EffectiveCapacity::MaximumHTLC { amount_msat } => *amount_msat,
-                       EffectiveCapacity::Total { capacity_msat } => *capacity_msat,
-                       EffectiveCapacity::Infinite => u64::max_value(),
-                       EffectiveCapacity::Unknown => UNKNOWN_CHANNEL_CAPACITY_MSAT,
-               }
-       }
-}
-
-/// Fees for routing via a given channel or a node
-#[derive(Eq, PartialEq, Copy, Clone, Debug, Hash)]
-pub struct RoutingFees {
-       /// Flat routing fee in satoshis
-       pub base_msat: u32,
-       /// Liquidity-based routing fee in millionths of a routed amount.
-       /// In other words, 10000 is 1%.
-       pub proportional_millionths: u32,
-}
-
-impl_writeable_tlv_based!(RoutingFees, {
-       (0, base_msat, required),
-       (2, proportional_millionths, required)
-});
-
-#[derive(Clone, Debug, PartialEq)]
-/// Information received in the latest node_announcement from this node.
-pub struct NodeAnnouncementInfo {
-       /// Protocol features the node announced support for
-       pub features: NodeFeatures,
-       /// When the last known update to the node state was issued.
-       /// Value is opaque, as set in the announcement.
-       pub last_update: u32,
-       /// Color assigned to the node
-       pub rgb: [u8; 3],
-       /// Moniker assigned to the node.
-       /// May be invalid or malicious (eg control chars),
-       /// should not be exposed to the user.
-       pub alias: [u8; 32],
-       /// Internet-level addresses via which one can connect to the node
-       pub addresses: Vec<NetAddress>,
-       /// An initial announcement of the node
-       /// Mostly redundant with the data we store in fields explicitly.
-       /// Everything else is useful only for sending out for initial routing sync.
-       /// Not stored if contains excess data to prevent DoS.
-       pub announcement_message: Option<NodeAnnouncement>
-}
-
-impl_writeable_tlv_based!(NodeAnnouncementInfo, {
-       (0, features, required),
-       (2, last_update, required),
-       (4, rgb, required),
-       (6, alias, required),
-       (8, announcement_message, option),
-       (10, addresses, vec_type),
-});
-
-#[derive(Clone, Debug, PartialEq)]
-/// Details about a node in the network, known from the network announcement.
-pub struct NodeInfo {
-       /// All valid channels a node has announced
-       pub channels: Vec<u64>,
-       /// Lowest fees enabling routing via any of the enabled, known channels to a node.
-       /// The two fields (flat and proportional fee) are independent,
-       /// meaning they don't have to refer to the same channel.
-       pub lowest_inbound_channel_fees: Option<RoutingFees>,
-       /// More information about a node from node_announcement.
-       /// Optional because we store a Node entry after learning about it from
-       /// a channel announcement, but before receiving a node announcement.
-       pub announcement_info: Option<NodeAnnouncementInfo>
-}
-
-impl fmt::Display for NodeInfo {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               write!(f, "lowest_inbound_channel_fees: {:?}, channels: {:?}, announcement_info: {:?}",
-                  self.lowest_inbound_channel_fees, &self.channels[..], self.announcement_info)?;
-               Ok(())
-       }
-}
-
-impl_writeable_tlv_based!(NodeInfo, {
-       (0, lowest_inbound_channel_fees, option),
-       (2, announcement_info, option),
-       (4, channels, vec_type),
-});
-
-const SERIALIZATION_VERSION: u8 = 1;
-const MIN_SERIALIZATION_VERSION: u8 = 1;
-
-impl Writeable for NetworkGraph {
-       fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
-               write_ver_prefix!(writer, SERIALIZATION_VERSION, MIN_SERIALIZATION_VERSION);
-
-               self.genesis_hash.write(writer)?;
-               let channels = self.channels.read().unwrap();
-               (channels.len() as u64).write(writer)?;
-               for (ref chan_id, ref chan_info) in channels.iter() {
-                       (*chan_id).write(writer)?;
-                       chan_info.write(writer)?;
-               }
-               let nodes = self.nodes.read().unwrap();
-               (nodes.len() as u64).write(writer)?;
-               for (ref node_id, ref node_info) in nodes.iter() {
-                       node_id.write(writer)?;
-                       node_info.write(writer)?;
-               }
-
-               let last_rapid_gossip_sync_timestamp = self.get_last_rapid_gossip_sync_timestamp();
-               write_tlv_fields!(writer, {
-                       (1, last_rapid_gossip_sync_timestamp, option),
-               });
-               Ok(())
-       }
-}
-
-impl Readable for NetworkGraph {
-       fn read<R: io::Read>(reader: &mut R) -> Result<NetworkGraph, DecodeError> {
-               let _ver = read_ver_prefix!(reader, SERIALIZATION_VERSION);
-
-               let genesis_hash: BlockHash = Readable::read(reader)?;
-               let channels_count: u64 = Readable::read(reader)?;
-               let mut channels = BTreeMap::new();
-               for _ in 0..channels_count {
-                       let chan_id: u64 = Readable::read(reader)?;
-                       let chan_info = Readable::read(reader)?;
-                       channels.insert(chan_id, chan_info);
-               }
-               let nodes_count: u64 = Readable::read(reader)?;
-               let mut nodes = BTreeMap::new();
-               for _ in 0..nodes_count {
-                       let node_id = Readable::read(reader)?;
-                       let node_info = Readable::read(reader)?;
-                       nodes.insert(node_id, node_info);
-               }
-
-               let mut last_rapid_gossip_sync_timestamp: Option<u32> = None;
-               read_tlv_fields!(reader, {
-                       (1, last_rapid_gossip_sync_timestamp, option),
-               });
-
-               Ok(NetworkGraph {
-                       genesis_hash,
-                       channels: RwLock::new(channels),
-                       nodes: RwLock::new(nodes),
-                       last_rapid_gossip_sync_timestamp: Mutex::new(last_rapid_gossip_sync_timestamp),
-               })
-       }
-}
-
-impl fmt::Display for NetworkGraph {
-       fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
-               writeln!(f, "Network map\n[Channels]")?;
-               for (key, val) in self.channels.read().unwrap().iter() {
-                       writeln!(f, " {}: {}", key, val)?;
-               }
-               writeln!(f, "[Nodes]")?;
-               for (&node_id, val) in self.nodes.read().unwrap().iter() {
-                       writeln!(f, " {}: {}", log_bytes!(node_id.as_slice()), val)?;
-               }
-               Ok(())
-       }
-}
-
-impl PartialEq for NetworkGraph {
-       fn eq(&self, other: &Self) -> bool {
-               self.genesis_hash == other.genesis_hash &&
-                       *self.channels.read().unwrap() == *other.channels.read().unwrap() &&
-                       *self.nodes.read().unwrap() == *other.nodes.read().unwrap()
-       }
-}
-
-impl NetworkGraph {
-       /// Creates a new, empty, network graph.
-       pub fn new(genesis_hash: BlockHash) -> NetworkGraph {
-               Self {
-                       genesis_hash,
-                       channels: RwLock::new(BTreeMap::new()),
-                       nodes: RwLock::new(BTreeMap::new()),
-                       last_rapid_gossip_sync_timestamp: Mutex::new(None),
-               }
-       }
-
-       /// Returns a read-only view of the network graph.
-       pub fn read_only(&'_ self) -> ReadOnlyNetworkGraph<'_> {
-               let channels = self.channels.read().unwrap();
-               let nodes = self.nodes.read().unwrap();
-               ReadOnlyNetworkGraph {
-                       channels,
-                       nodes,
-               }
-       }
-
-       /// The unix timestamp provided by the most recent rapid gossip sync.
-       /// It will be set by the rapid sync process after every sync completion.
-       pub fn get_last_rapid_gossip_sync_timestamp(&self) -> Option<u32> {
-               self.last_rapid_gossip_sync_timestamp.lock().unwrap().clone()
-       }
-
-       /// Update the unix timestamp provided by the most recent rapid gossip sync.
-       /// This should be done automatically by the rapid sync process after every sync completion.
-       pub fn set_last_rapid_gossip_sync_timestamp(&self, last_rapid_gossip_sync_timestamp: u32) {
-               self.last_rapid_gossip_sync_timestamp.lock().unwrap().replace(last_rapid_gossip_sync_timestamp);
-       }
-
-       /// Clears the `NodeAnnouncementInfo` field for all nodes in the `NetworkGraph` for testing
-       /// purposes.
-       #[cfg(test)]
-       pub fn clear_nodes_announcement_info(&self) {
-               for node in self.nodes.write().unwrap().iter_mut() {
-                       node.1.announcement_info = None;
-               }
-       }
-
-       /// For an already known node (from channel announcements), update its stored properties from a
-       /// given node announcement.
-       ///
-       /// You probably don't want to call this directly, instead relying on a NetGraphMsgHandler's
-       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
-       /// routing messages from a source using a protocol other than the lightning P2P protocol.
-       pub fn update_node_from_announcement<T: secp256k1::Verification>(&self, msg: &msgs::NodeAnnouncement, secp_ctx: &Secp256k1<T>) -> Result<(), LightningError> {
-               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.contents.encode()[..])[..]);
-               secp_verify_sig!(secp_ctx, &msg_hash, &msg.signature, &msg.contents.node_id, "node_announcement");
-               self.update_node_from_announcement_intern(&msg.contents, Some(&msg))
-       }
-
-       /// For an already known node (from channel announcements), update its stored properties from a
-       /// given node announcement without verifying the associated signatures. Because we aren't
-       /// given the associated signatures here we cannot relay the node announcement to any of our
-       /// peers.
-       pub fn update_node_from_unsigned_announcement(&self, msg: &msgs::UnsignedNodeAnnouncement) -> Result<(), LightningError> {
-               self.update_node_from_announcement_intern(msg, None)
-       }
-
-       fn update_node_from_announcement_intern(&self, msg: &msgs::UnsignedNodeAnnouncement, full_msg: Option<&msgs::NodeAnnouncement>) -> Result<(), LightningError> {
-               match self.nodes.write().unwrap().get_mut(&NodeId::from_pubkey(&msg.node_id)) {
-                       None => Err(LightningError{err: "No existing channels for node_announcement".to_owned(), action: ErrorAction::IgnoreError}),
-                       Some(node) => {
-                               if let Some(node_info) = node.announcement_info.as_ref() {
-                                       // The timestamp field is somewhat of a misnomer - the BOLTs use it to order
-                                       // updates to ensure you always have the latest one, only vaguely suggesting
-                                       // that it be at least the current time.
-                                       if node_info.last_update  > msg.timestamp {
-                                               return Err(LightningError{err: "Update older than last processed update".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
-                                       } else if node_info.last_update  == msg.timestamp {
-                                               return Err(LightningError{err: "Update had the same timestamp as last processed update".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
-                                       }
-                               }
-
-                               let should_relay =
-                                       msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
-                                       msg.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY &&
-                                       msg.excess_data.len() + msg.excess_address_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY;
-                               node.announcement_info = Some(NodeAnnouncementInfo {
-                                       features: msg.features.clone(),
-                                       last_update: msg.timestamp,
-                                       rgb: msg.rgb,
-                                       alias: msg.alias,
-                                       addresses: msg.addresses.clone(),
-                                       announcement_message: if should_relay { full_msg.cloned() } else { None },
-                               });
-
-                               Ok(())
-                       }
-               }
-       }
-
-       /// Store or update channel info from a channel announcement.
-       ///
-       /// You probably don't want to call this directly, instead relying on a NetGraphMsgHandler's
-       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
-       /// routing messages from a source using a protocol other than the lightning P2P protocol.
-       ///
-       /// If a `chain::Access` object is provided via `chain_access`, it will be called to verify
-       /// the corresponding UTXO exists on chain and is correctly-formatted.
-       pub fn update_channel_from_announcement<T: secp256k1::Verification, C: Deref>(
-               &self, msg: &msgs::ChannelAnnouncement, chain_access: &Option<C>, secp_ctx: &Secp256k1<T>
-       ) -> Result<(), LightningError>
-       where
-               C::Target: chain::Access,
-       {
-               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.contents.encode()[..])[..]);
-               secp_verify_sig!(secp_ctx, &msg_hash, &msg.node_signature_1, &msg.contents.node_id_1, "channel_announcement");
-               secp_verify_sig!(secp_ctx, &msg_hash, &msg.node_signature_2, &msg.contents.node_id_2, "channel_announcement");
-               secp_verify_sig!(secp_ctx, &msg_hash, &msg.bitcoin_signature_1, &msg.contents.bitcoin_key_1, "channel_announcement");
-               secp_verify_sig!(secp_ctx, &msg_hash, &msg.bitcoin_signature_2, &msg.contents.bitcoin_key_2, "channel_announcement");
-               self.update_channel_from_unsigned_announcement_intern(&msg.contents, Some(msg), chain_access)
-       }
-
-       /// Store or update channel info from a channel announcement without verifying the associated
-       /// signatures. Because we aren't given the associated signatures here we cannot relay the
-       /// channel announcement to any of our peers.
-       ///
-       /// If a `chain::Access` object is provided via `chain_access`, it will be called to verify
-       /// the corresponding UTXO exists on chain and is correctly-formatted.
-       pub fn update_channel_from_unsigned_announcement<C: Deref>(
-               &self, msg: &msgs::UnsignedChannelAnnouncement, chain_access: &Option<C>
-       ) -> Result<(), LightningError>
-       where
-               C::Target: chain::Access,
-       {
-               self.update_channel_from_unsigned_announcement_intern(msg, None, chain_access)
-       }
-
-       /// Update channel from partial announcement data received via rapid gossip sync
-       ///
-       /// `timestamp: u64`: Timestamp emulating the backdated original announcement receipt (by the
-       /// rapid gossip sync server)
-       ///
-       /// All other parameters as used in [`msgs::UnsignedChannelAnnouncement`] fields.
-       pub fn add_channel_from_partial_announcement(&self, short_channel_id: u64, timestamp: u64, features: ChannelFeatures, node_id_1: PublicKey, node_id_2: PublicKey) -> Result<(), LightningError> {
-               if node_id_1 == node_id_2 {
-                       return Err(LightningError{err: "Channel announcement node had a channel with itself".to_owned(), action: ErrorAction::IgnoreError});
-               };
-
-               let node_1 = NodeId::from_pubkey(&node_id_1);
-               let node_2 = NodeId::from_pubkey(&node_id_2);
-               let channel_info = ChannelInfo {
-                       features,
-                       node_one: node_1.clone(),
-                       one_to_two: None,
-                       node_two: node_2.clone(),
-                       two_to_one: None,
-                       capacity_sats: None,
-                       announcement_message: None,
-                       announcement_received_time: timestamp,
-               };
-
-               self.add_channel_between_nodes(short_channel_id, channel_info, None)
-       }
-
-       fn add_channel_between_nodes(&self, short_channel_id: u64, channel_info: ChannelInfo, utxo_value: Option<u64>) -> Result<(), LightningError> {
-               let mut channels = self.channels.write().unwrap();
-               let mut nodes = self.nodes.write().unwrap();
-
-               let node_id_a = channel_info.node_one.clone();
-               let node_id_b = channel_info.node_two.clone();
-
-               match channels.entry(short_channel_id) {
-                       BtreeEntry::Occupied(mut entry) => {
-                               //TODO: because asking the blockchain if short_channel_id is valid is only optional
-                               //in the blockchain API, we need to handle it smartly here, though it's unclear
-                               //exactly how...
-                               if utxo_value.is_some() {
-                                       // Either our UTXO provider is busted, there was a reorg, or the UTXO provider
-                                       // only sometimes returns results. In any case remove the previous entry. Note
-                                       // that the spec expects us to "blacklist" the node_ids involved, but we can't
-                                       // do that because
-                                       // a) we don't *require* a UTXO provider that always returns results.
-                                       // b) we don't track UTXOs of channels we know about and remove them if they
-                                       //    get reorg'd out.
-                                       // c) it's unclear how to do so without exposing ourselves to massive DoS risk.
-                                       Self::remove_channel_in_nodes(&mut nodes, &entry.get(), short_channel_id);
-                                       *entry.get_mut() = channel_info;
-                               } else {
-                                       return Err(LightningError{err: "Already have knowledge of channel".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
-                               }
-                       },
-                       BtreeEntry::Vacant(entry) => {
-                               entry.insert(channel_info);
-                       }
-               };
-
-               for current_node_id in [node_id_a, node_id_b].iter() {
-                       match nodes.entry(current_node_id.clone()) {
-                               BtreeEntry::Occupied(node_entry) => {
-                                       node_entry.into_mut().channels.push(short_channel_id);
-                               },
-                               BtreeEntry::Vacant(node_entry) => {
-                                       node_entry.insert(NodeInfo {
-                                               channels: vec!(short_channel_id),
-                                               lowest_inbound_channel_fees: None,
-                                               announcement_info: None,
-                                       });
-                               }
-                       };
-               };
-
-               Ok(())
-       }
-
-       fn update_channel_from_unsigned_announcement_intern<C: Deref>(
-               &self, msg: &msgs::UnsignedChannelAnnouncement, full_msg: Option<&msgs::ChannelAnnouncement>, chain_access: &Option<C>
-       ) -> Result<(), LightningError>
-       where
-               C::Target: chain::Access,
-       {
-               if msg.node_id_1 == msg.node_id_2 || msg.bitcoin_key_1 == msg.bitcoin_key_2 {
-                       return Err(LightningError{err: "Channel announcement node had a channel with itself".to_owned(), action: ErrorAction::IgnoreError});
-               }
-
-               let utxo_value = match &chain_access {
-                       &None => {
-                               // Tentatively accept, potentially exposing us to DoS attacks
-                               None
-                       },
-                       &Some(ref chain_access) => {
-                               match chain_access.get_utxo(&msg.chain_hash, msg.short_channel_id) {
-                                       Ok(TxOut { value, script_pubkey }) => {
-                                               let expected_script = Builder::new().push_opcode(opcodes::all::OP_PUSHNUM_2)
-                                                                                   .push_slice(&msg.bitcoin_key_1.serialize())
-                                                                                   .push_slice(&msg.bitcoin_key_2.serialize())
-                                                                                   .push_opcode(opcodes::all::OP_PUSHNUM_2)
-                                                                                   .push_opcode(opcodes::all::OP_CHECKMULTISIG).into_script().to_v0_p2wsh();
-                                               if script_pubkey != expected_script {
-                                                       return Err(LightningError{err: format!("Channel announcement key ({}) didn't match on-chain script ({})", script_pubkey.to_hex(), expected_script.to_hex()), action: ErrorAction::IgnoreError});
-                                               }
-                                               //TODO: Check if value is worth storing, use it to inform routing, and compare it
-                                               //to the new HTLC max field in channel_update
-                                               Some(value)
-                                       },
-                                       Err(chain::AccessError::UnknownChain) => {
-                                               return Err(LightningError{err: format!("Channel announced on an unknown chain ({})", msg.chain_hash.encode().to_hex()), action: ErrorAction::IgnoreError});
-                                       },
-                                       Err(chain::AccessError::UnknownTx) => {
-                                               return Err(LightningError{err: "Channel announced without corresponding UTXO entry".to_owned(), action: ErrorAction::IgnoreError});
-                                       },
-                               }
-                       },
-               };
-
-               #[allow(unused_mut, unused_assignments)]
-               let mut announcement_received_time = 0;
-               #[cfg(feature = "std")]
-               {
-                       announcement_received_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-               }
-
-               let chan_info = ChannelInfo {
-                       features: msg.features.clone(),
-                       node_one: NodeId::from_pubkey(&msg.node_id_1),
-                       one_to_two: None,
-                       node_two: NodeId::from_pubkey(&msg.node_id_2),
-                       two_to_one: None,
-                       capacity_sats: utxo_value,
-                       announcement_message: if msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY
-                               { full_msg.cloned() } else { None },
-                       announcement_received_time,
-               };
-
-               self.add_channel_between_nodes(msg.short_channel_id, chan_info, utxo_value)
-       }
-
-       /// Close a channel if a corresponding HTLC fail was sent.
-       /// If permanent, removes a channel from the local storage.
-       /// May cause the removal of nodes too, if this was their last channel.
-       /// If not permanent, makes channels unavailable for routing.
-       pub fn close_channel_from_update(&self, short_channel_id: u64, is_permanent: bool) {
-               let mut channels = self.channels.write().unwrap();
-               if is_permanent {
-                       if let Some(chan) = channels.remove(&short_channel_id) {
-                               let mut nodes = self.nodes.write().unwrap();
-                               Self::remove_channel_in_nodes(&mut nodes, &chan, short_channel_id);
-                       }
-               } else {
-                       if let Some(chan) = channels.get_mut(&short_channel_id) {
-                               if let Some(one_to_two) = chan.one_to_two.as_mut() {
-                                       one_to_two.enabled = false;
-                               }
-                               if let Some(two_to_one) = chan.two_to_one.as_mut() {
-                                       two_to_one.enabled = false;
-                               }
-                       }
-               }
-       }
-
-       /// Marks a node in the graph as failed.
-       pub fn fail_node(&self, _node_id: &PublicKey, is_permanent: bool) {
-               if is_permanent {
-                       // TODO: Wholly remove the node
-               } else {
-                       // TODO: downgrade the node
-               }
-       }
-
-       #[cfg(feature = "std")]
-       /// Removes information about channels that we haven't heard any updates about in some time.
-       /// This can be used regularly to prune the network graph of channels that likely no longer
-       /// exist.
-       ///
-       /// While there is no formal requirement that nodes regularly re-broadcast their channel
-       /// updates every two weeks, the non-normative section of BOLT 7 currently suggests that
-       /// pruning occur for updates which are at least two weeks old, which we implement here.
-       ///
-       /// Note that for users of the `lightning-background-processor` crate this method may be
-       /// automatically called regularly for you.
-       ///
-       /// This method is only available with the `std` feature. See
-       /// [`NetworkGraph::remove_stale_channels_with_time`] for `no-std` use.
-       pub fn remove_stale_channels(&self) {
-               let time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-               self.remove_stale_channels_with_time(time);
-       }
-
-       /// Removes information about channels that we haven't heard any updates about in some time.
-       /// This can be used regularly to prune the network graph of channels that likely no longer
-       /// exist.
-       ///
-       /// While there is no formal requirement that nodes regularly re-broadcast their channel
-       /// updates every two weeks, the non-normative section of BOLT 7 currently suggests that
-       /// pruning occur for updates which are at least two weeks old, which we implement here.
-       ///
-       /// This function takes the current unix time as an argument. For users with the `std` feature
-       /// enabled, [`NetworkGraph::remove_stale_channels`] may be preferable.
-       pub fn remove_stale_channels_with_time(&self, current_time_unix: u64) {
-               let mut channels = self.channels.write().unwrap();
-               // Time out if we haven't received an update in at least 14 days.
-               if current_time_unix > u32::max_value() as u64 { return; } // Remove by 2106
-               if current_time_unix < STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS { return; }
-               let min_time_unix: u32 = (current_time_unix - STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS) as u32;
-               // Sadly BTreeMap::retain was only stabilized in 1.53 so we can't switch to it for some
-               // time.
-               let mut scids_to_remove = Vec::new();
-               for (scid, info) in channels.iter_mut() {
-                       if info.one_to_two.is_some() && info.one_to_two.as_ref().unwrap().last_update < min_time_unix {
-                               info.one_to_two = None;
-                       }
-                       if info.two_to_one.is_some() && info.two_to_one.as_ref().unwrap().last_update < min_time_unix {
-                               info.two_to_one = None;
-                       }
-                       if info.one_to_two.is_none() && info.two_to_one.is_none() {
-                               // We check the announcement_received_time here to ensure we don't drop
-                               // announcements that we just received and are just waiting for our peer to send a
-                               // channel_update for.
-                               if info.announcement_received_time < min_time_unix as u64 {
-                                       scids_to_remove.push(*scid);
-                               }
-                       }
-               }
-               if !scids_to_remove.is_empty() {
-                       let mut nodes = self.nodes.write().unwrap();
-                       for scid in scids_to_remove {
-                               let info = channels.remove(&scid).expect("We just accessed this scid, it should be present");
-                               Self::remove_channel_in_nodes(&mut nodes, &info, scid);
-                       }
-               }
-       }
-
-       /// For an already known (from announcement) channel, update info about one of the directions
-       /// of the channel.
-       ///
-       /// You probably don't want to call this directly, instead relying on a NetGraphMsgHandler's
-       /// RoutingMessageHandler implementation to call it indirectly. This may be useful to accept
-       /// routing messages from a source using a protocol other than the lightning P2P protocol.
-       ///
-       /// If built with `no-std`, any updates with a timestamp more than two weeks in the past or
-       /// materially in the future will be rejected.
-       pub fn update_channel<T: secp256k1::Verification>(&self, msg: &msgs::ChannelUpdate, secp_ctx: &Secp256k1<T>) -> Result<(), LightningError> {
-               self.update_channel_intern(&msg.contents, Some(&msg), Some((&msg.signature, secp_ctx)))
-       }
-
-       /// For an already known (from announcement) channel, update info about one of the directions
-       /// of the channel without verifying the associated signatures. Because we aren't given the
-       /// associated signatures here we cannot relay the channel update to any of our peers.
-       ///
-       /// If built with `no-std`, any updates with a timestamp more than two weeks in the past or
-       /// materially in the future will be rejected.
-       pub fn update_channel_unsigned(&self, msg: &msgs::UnsignedChannelUpdate) -> Result<(), LightningError> {
-               self.update_channel_intern(msg, None, None::<(&secp256k1::ecdsa::Signature, &Secp256k1<secp256k1::VerifyOnly>)>)
-       }
-
-       fn update_channel_intern<T: secp256k1::Verification>(&self, msg: &msgs::UnsignedChannelUpdate, full_msg: Option<&msgs::ChannelUpdate>, sig_info: Option<(&secp256k1::ecdsa::Signature, &Secp256k1<T>)>) -> Result<(), LightningError> {
-               let dest_node_id;
-               let chan_enabled = msg.flags & (1 << 1) != (1 << 1);
-               let chan_was_enabled;
-
-               #[cfg(all(feature = "std", not(test), not(feature = "_test_utils")))]
-               {
-                       // Note that many tests rely on being able to set arbitrarily old timestamps, thus we
-                       // disable this check during tests!
-                       let time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-                       if (msg.timestamp as u64) < time - STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS {
-                               return Err(LightningError{err: "channel_update is older than two weeks old".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
-                       }
-                       if msg.timestamp as u64 > time + 60 * 60 * 24 {
-                               return Err(LightningError{err: "channel_update has a timestamp more than a day in the future".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
-                       }
-               }
-
-               let mut channels = self.channels.write().unwrap();
-               match channels.get_mut(&msg.short_channel_id) {
-                       None => return Err(LightningError{err: "Couldn't find channel for update".to_owned(), action: ErrorAction::IgnoreError}),
-                       Some(channel) => {
-                               if let OptionalField::Present(htlc_maximum_msat) = msg.htlc_maximum_msat {
-                                       if htlc_maximum_msat > MAX_VALUE_MSAT {
-                                               return Err(LightningError{err: "htlc_maximum_msat is larger than maximum possible msats".to_owned(), action: ErrorAction::IgnoreError});
-                                       }
-
-                                       if let Some(capacity_sats) = channel.capacity_sats {
-                                               // It's possible channel capacity is available now, although it wasn't available at announcement (so the field is None).
-                                               // Don't query UTXO set here to reduce DoS risks.
-                                               if capacity_sats > MAX_VALUE_MSAT / 1000 || htlc_maximum_msat > capacity_sats * 1000 {
-                                                       return Err(LightningError{err: "htlc_maximum_msat is larger than channel capacity or capacity is bogus".to_owned(), action: ErrorAction::IgnoreError});
-                                               }
-                                       }
-                               }
-                               macro_rules! check_update_latest {
-                                       ($target: expr) => {
-                                               if let Some(existing_chan_info) = $target.as_ref() {
-                                                       // The timestamp field is somewhat of a misnomer - the BOLTs use it to
-                                                       // order updates to ensure you always have the latest one, only
-                                                       // suggesting  that it be at least the current time. For
-                                                       // channel_updates specifically, the BOLTs discuss the possibility of
-                                                       // pruning based on the timestamp field being more than two weeks old,
-                                                       // but only in the non-normative section.
-                                                       if existing_chan_info.last_update > msg.timestamp {
-                                                               return Err(LightningError{err: "Update older than last processed update".to_owned(), action: ErrorAction::IgnoreAndLog(Level::Gossip)});
-                                                       } else if existing_chan_info.last_update == msg.timestamp {
-                                                               return Err(LightningError{err: "Update had same timestamp as last processed update".to_owned(), action: ErrorAction::IgnoreDuplicateGossip});
-                                                       }
-                                                       chan_was_enabled = existing_chan_info.enabled;
-                                               } else {
-                                                       chan_was_enabled = false;
-                                               }
-                                       }
-                               }
-
-                               macro_rules! get_new_channel_info {
-                                       () => { {
-                                               let last_update_message = if msg.excess_data.len() <= MAX_EXCESS_BYTES_FOR_RELAY
-                                                       { full_msg.cloned() } else { None };
-
-                                               let updated_channel_update_info = ChannelUpdateInfo {
-                                                       enabled: chan_enabled,
-                                                       last_update: msg.timestamp,
-                                                       cltv_expiry_delta: msg.cltv_expiry_delta,
-                                                       htlc_minimum_msat: msg.htlc_minimum_msat,
-                                                       htlc_maximum_msat: if let OptionalField::Present(max_value) = msg.htlc_maximum_msat { Some(max_value) } else { None },
-                                                       fees: RoutingFees {
-                                                               base_msat: msg.fee_base_msat,
-                                                               proportional_millionths: msg.fee_proportional_millionths,
-                                                       },
-                                                       last_update_message
-                                               };
-                                               Some(updated_channel_update_info)
-                                       } }
-                               }
-
-                               let msg_hash = hash_to_message!(&Sha256dHash::hash(&msg.encode()[..])[..]);
-                               if msg.flags & 1 == 1 {
-                                       dest_node_id = channel.node_one.clone();
-                                       check_update_latest!(channel.two_to_one);
-                                       if let Some((sig, ctx)) = sig_info {
-                                               secp_verify_sig!(ctx, &msg_hash, &sig, &PublicKey::from_slice(channel.node_two.as_slice()).map_err(|_| LightningError{
-                                                       err: "Couldn't parse source node pubkey".to_owned(),
-                                                       action: ErrorAction::IgnoreAndLog(Level::Debug)
-                                               })?, "channel_update");
-                                       }
-                                       channel.two_to_one = get_new_channel_info!();
-                               } else {
-                                       dest_node_id = channel.node_two.clone();
-                                       check_update_latest!(channel.one_to_two);
-                                       if let Some((sig, ctx)) = sig_info {
-                                               secp_verify_sig!(ctx, &msg_hash, &sig, &PublicKey::from_slice(channel.node_one.as_slice()).map_err(|_| LightningError{
-                                                       err: "Couldn't parse destination node pubkey".to_owned(),
-                                                       action: ErrorAction::IgnoreAndLog(Level::Debug)
-                                               })?, "channel_update");
-                                       }
-                                       channel.one_to_two = get_new_channel_info!();
-                               }
-                       }
-               }
-
-               let mut nodes = self.nodes.write().unwrap();
-               if chan_enabled {
-                       let node = nodes.get_mut(&dest_node_id).unwrap();
-                       let mut base_msat = msg.fee_base_msat;
-                       let mut proportional_millionths = msg.fee_proportional_millionths;
-                       if let Some(fees) = node.lowest_inbound_channel_fees {
-                               base_msat = cmp::min(base_msat, fees.base_msat);
-                               proportional_millionths = cmp::min(proportional_millionths, fees.proportional_millionths);
-                       }
-                       node.lowest_inbound_channel_fees = Some(RoutingFees {
-                               base_msat,
-                               proportional_millionths
-                       });
-               } else if chan_was_enabled {
-                       let node = nodes.get_mut(&dest_node_id).unwrap();
-                       let mut lowest_inbound_channel_fees = None;
-
-                       for chan_id in node.channels.iter() {
-                               let chan = channels.get(chan_id).unwrap();
-                               let chan_info_opt;
-                               if chan.node_one == dest_node_id {
-                                       chan_info_opt = chan.two_to_one.as_ref();
-                               } else {
-                                       chan_info_opt = chan.one_to_two.as_ref();
-                               }
-                               if let Some(chan_info) = chan_info_opt {
-                                       if chan_info.enabled {
-                                               let fees = lowest_inbound_channel_fees.get_or_insert(RoutingFees {
-                                                       base_msat: u32::max_value(), proportional_millionths: u32::max_value() });
-                                               fees.base_msat = cmp::min(fees.base_msat, chan_info.fees.base_msat);
-                                               fees.proportional_millionths = cmp::min(fees.proportional_millionths, chan_info.fees.proportional_millionths);
-                                       }
-                               }
-                       }
-
-                       node.lowest_inbound_channel_fees = lowest_inbound_channel_fees;
-               }
-
-               Ok(())
-       }
-
-       fn remove_channel_in_nodes(nodes: &mut BTreeMap<NodeId, NodeInfo>, chan: &ChannelInfo, short_channel_id: u64) {
-               macro_rules! remove_from_node {
-                       ($node_id: expr) => {
-                               if let BtreeEntry::Occupied(mut entry) = nodes.entry($node_id) {
-                                       entry.get_mut().channels.retain(|chan_id| {
-                                               short_channel_id != *chan_id
-                                       });
-                                       if entry.get().channels.is_empty() {
-                                               entry.remove_entry();
-                                       }
-                               } else {
-                                       panic!("Had channel that pointed to unknown node (ie inconsistent network map)!");
-                               }
-                       }
-               }
-
-               remove_from_node!(chan.node_one);
-               remove_from_node!(chan.node_two);
-       }
-}
-
-impl ReadOnlyNetworkGraph<'_> {
-       /// Returns all known valid channels' short ids along with announced channel info.
-       ///
-       /// (C-not exported) because we have no mapping for `BTreeMap`s
-       pub fn channels(&self) -> &BTreeMap<u64, ChannelInfo> {
-               &*self.channels
-       }
-
-       /// Returns all known nodes' public keys along with announced node info.
-       ///
-       /// (C-not exported) because we have no mapping for `BTreeMap`s
-       pub fn nodes(&self) -> &BTreeMap<NodeId, NodeInfo> {
-               &*self.nodes
-       }
-
-       /// Get network addresses by node id.
-       /// Returns None if the requested node is completely unknown,
-       /// or if node announcement for the node was never received.
-       pub fn get_addresses(&self, pubkey: &PublicKey) -> Option<Vec<NetAddress>> {
-               if let Some(node) = self.nodes.get(&NodeId::from_pubkey(&pubkey)) {
-                       if let Some(node_info) = node.announcement_info.as_ref() {
-                               return Some(node_info.addresses.clone())
-                       }
-               }
-               None
-       }
-}
-
-#[cfg(test)]
-mod tests {
-       use chain;
-       use ln::PaymentHash;
-       use ln::features::{ChannelFeatures, InitFeatures, NodeFeatures};
-       use routing::network_graph::{NetGraphMsgHandler, NetworkGraph, NetworkUpdate, MAX_EXCESS_BYTES_FOR_RELAY};
-       use ln::msgs::{Init, OptionalField, RoutingMessageHandler, UnsignedNodeAnnouncement, NodeAnnouncement,
-               UnsignedChannelAnnouncement, ChannelAnnouncement, UnsignedChannelUpdate, ChannelUpdate,
-               ReplyChannelRange, QueryChannelRange, QueryShortChannelIds, MAX_VALUE_MSAT};
-       use util::test_utils;
-       use util::logger::Logger;
-       use util::ser::{Readable, Writeable};
-       use util::events::{Event, EventHandler, MessageSendEvent, MessageSendEventsProvider};
-       use util::scid_utils::scid_from_parts;
-
-       use super::STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS;
-
-       use bitcoin::hashes::sha256d::Hash as Sha256dHash;
-       use bitcoin::hashes::Hash;
-       use bitcoin::network::constants::Network;
-       use bitcoin::blockdata::constants::genesis_block;
-       use bitcoin::blockdata::script::{Builder, Script};
-       use bitcoin::blockdata::transaction::TxOut;
-       use bitcoin::blockdata::opcodes;
-
-       use hex;
-
-       use bitcoin::secp256k1::{PublicKey, SecretKey};
-       use bitcoin::secp256k1::{All, Secp256k1};
-
-       use io;
-       use bitcoin::secp256k1;
-       use prelude::*;
-       use sync::Arc;
-
-       fn create_network_graph() -> NetworkGraph {
-               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
-               NetworkGraph::new(genesis_hash)
-       }
-
-       fn create_net_graph_msg_handler(network_graph: &NetworkGraph) -> (
-               Secp256k1<All>, NetGraphMsgHandler<&NetworkGraph, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>
-       ) {
-               let secp_ctx = Secp256k1::new();
-               let logger = Arc::new(test_utils::TestLogger::new());
-               let net_graph_msg_handler = NetGraphMsgHandler::new(network_graph, None, Arc::clone(&logger));
-               (secp_ctx, net_graph_msg_handler)
-       }
-
-       #[test]
-       fn request_full_sync_finite_times() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-               let node_id = PublicKey::from_secret_key(&secp_ctx, &SecretKey::from_slice(&hex::decode("0202020202020202020202020202020202020202020202020202020202020202").unwrap()[..]).unwrap());
-
-               assert!(net_graph_msg_handler.should_request_full_sync(&node_id));
-               assert!(net_graph_msg_handler.should_request_full_sync(&node_id));
-               assert!(net_graph_msg_handler.should_request_full_sync(&node_id));
-               assert!(net_graph_msg_handler.should_request_full_sync(&node_id));
-               assert!(net_graph_msg_handler.should_request_full_sync(&node_id));
-               assert!(!net_graph_msg_handler.should_request_full_sync(&node_id));
-       }
-
-       fn get_signed_node_announcement<F: Fn(&mut UnsignedNodeAnnouncement)>(f: F, node_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> NodeAnnouncement {
-               let node_id = PublicKey::from_secret_key(&secp_ctx, node_key);
-               let mut unsigned_announcement = UnsignedNodeAnnouncement {
-                       features: NodeFeatures::known(),
-                       timestamp: 100,
-                       node_id: node_id,
-                       rgb: [0; 3],
-                       alias: [0; 32],
-                       addresses: Vec::new(),
-                       excess_address_data: Vec::new(),
-                       excess_data: Vec::new(),
-               };
-               f(&mut unsigned_announcement);
-               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_announcement.encode()[..])[..]);
-               NodeAnnouncement {
-                       signature: secp_ctx.sign_ecdsa(&msghash, node_key),
-                       contents: unsigned_announcement
-               }
-       }
-
-       fn get_signed_channel_announcement<F: Fn(&mut UnsignedChannelAnnouncement)>(f: F, node_1_key: &SecretKey, node_2_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> ChannelAnnouncement {
-               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_1_key);
-               let node_id_2 = PublicKey::from_secret_key(&secp_ctx, node_2_key);
-               let node_1_btckey = &SecretKey::from_slice(&[40; 32]).unwrap();
-               let node_2_btckey = &SecretKey::from_slice(&[39; 32]).unwrap();
-
-               let mut unsigned_announcement = UnsignedChannelAnnouncement {
-                       features: ChannelFeatures::known(),
-                       chain_hash: genesis_block(Network::Testnet).header.block_hash(),
-                       short_channel_id: 0,
-                       node_id_1,
-                       node_id_2,
-                       bitcoin_key_1: PublicKey::from_secret_key(&secp_ctx, node_1_btckey),
-                       bitcoin_key_2: PublicKey::from_secret_key(&secp_ctx, node_2_btckey),
-                       excess_data: Vec::new(),
-               };
-               f(&mut unsigned_announcement);
-               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_announcement.encode()[..])[..]);
-               ChannelAnnouncement {
-                       node_signature_1: secp_ctx.sign_ecdsa(&msghash, node_1_key),
-                       node_signature_2: secp_ctx.sign_ecdsa(&msghash, node_2_key),
-                       bitcoin_signature_1: secp_ctx.sign_ecdsa(&msghash, node_1_btckey),
-                       bitcoin_signature_2: secp_ctx.sign_ecdsa(&msghash, node_2_btckey),
-                       contents: unsigned_announcement,
-               }
-       }
-
-       fn get_channel_script(secp_ctx: &Secp256k1<secp256k1::All>) -> Script {
-               let node_1_btckey = &SecretKey::from_slice(&[40; 32]).unwrap();
-               let node_2_btckey = &SecretKey::from_slice(&[39; 32]).unwrap();
-               Builder::new().push_opcode(opcodes::all::OP_PUSHNUM_2)
-                             .push_slice(&PublicKey::from_secret_key(&secp_ctx, node_1_btckey).serialize())
-                             .push_slice(&PublicKey::from_secret_key(&secp_ctx, node_2_btckey).serialize())
-                             .push_opcode(opcodes::all::OP_PUSHNUM_2)
-                             .push_opcode(opcodes::all::OP_CHECKMULTISIG).into_script()
-                             .to_v0_p2wsh()
-       }
-
-       fn get_signed_channel_update<F: Fn(&mut UnsignedChannelUpdate)>(f: F, node_key: &SecretKey, secp_ctx: &Secp256k1<secp256k1::All>) -> ChannelUpdate {
-               let mut unsigned_channel_update = UnsignedChannelUpdate {
-                       chain_hash: genesis_block(Network::Testnet).header.block_hash(),
-                       short_channel_id: 0,
-                       timestamp: 100,
-                       flags: 0,
-                       cltv_expiry_delta: 144,
-                       htlc_minimum_msat: 1_000_000,
-                       htlc_maximum_msat: OptionalField::Absent,
-                       fee_base_msat: 10_000,
-                       fee_proportional_millionths: 20,
-                       excess_data: Vec::new()
-               };
-               f(&mut unsigned_channel_update);
-               let msghash = hash_to_message!(&Sha256dHash::hash(&unsigned_channel_update.encode()[..])[..]);
-               ChannelUpdate {
-                       signature: secp_ctx.sign_ecdsa(&msghash, node_key),
-                       contents: unsigned_channel_update
-               }
-       }
-
-       #[test]
-       fn handling_node_announcements() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-               let zero_hash = Sha256dHash::hash(&[0; 32]);
-
-               let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!("No existing channels for node_announcement", e.err)
-               };
-
-               {
-                       // Announce a channel to add a corresponding node.
-                       let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                               Ok(res) => assert!(res),
-                               _ => panic!()
-                       };
-               }
-
-               match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                       Ok(res) => assert!(res),
-                       Err(_) => panic!()
-               };
-
-               let fake_msghash = hash_to_message!(&zero_hash);
-               match net_graph_msg_handler.handle_node_announcement(
-                       &NodeAnnouncement {
-                               signature: secp_ctx.sign_ecdsa(&fake_msghash, node_1_privkey),
-                               contents: valid_announcement.contents.clone()
-               }) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Invalid signature on node_announcement message")
-               };
-
-               let announcement_with_data = get_signed_node_announcement(|unsigned_announcement| {
-                       unsigned_announcement.timestamp += 1000;
-                       unsigned_announcement.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
-               }, node_1_privkey, &secp_ctx);
-               // Return false because contains excess data.
-               match net_graph_msg_handler.handle_node_announcement(&announcement_with_data) {
-                       Ok(res) => assert!(!res),
-                       Err(_) => panic!()
-               };
-
-               // Even though previous announcement was not relayed further, we still accepted it,
-               // so we now won't accept announcements before the previous one.
-               let outdated_announcement = get_signed_node_announcement(|unsigned_announcement| {
-                       unsigned_announcement.timestamp += 1000 - 10;
-               }, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_node_announcement(&outdated_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Update older than last processed update")
-               };
-       }
-
-       #[test]
-       fn handling_channel_announcements() {
-               let secp_ctx = Secp256k1::new();
-               let logger: Arc<Logger> = Arc::new(test_utils::TestLogger::new());
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               let good_script = get_channel_script(&secp_ctx);
-               let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-
-               // Test if the UTXO lookups were not supported
-               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
-               let mut net_graph_msg_handler = NetGraphMsgHandler::new(&network_graph, None, Arc::clone(&logger));
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(res) => assert!(res),
-                       _ => panic!()
-               };
-
-               {
-                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
-                               None => panic!(),
-                               Some(_) => ()
-                       };
-               }
-
-               // If we receive announcement for the same channel (with UTXO lookups disabled),
-               // drop new one on the floor, since we can't see any changes.
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Already have knowledge of channel")
-               };
-
-               // Test if an associated transaction were not on-chain (or not confirmed).
-               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
-               *chain_source.utxo_ret.lock().unwrap() = Err(chain::AccessError::UnknownTx);
-               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
-               net_graph_msg_handler = NetGraphMsgHandler::new(&network_graph, Some(chain_source.clone()), Arc::clone(&logger));
-
-               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
-                       unsigned_announcement.short_channel_id += 1;
-               }, node_1_privkey, node_2_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Channel announced without corresponding UTXO entry")
-               };
-
-               // Now test if the transaction is found in the UTXO set and the script is correct.
-               *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: 0, script_pubkey: good_script.clone() });
-               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
-                       unsigned_announcement.short_channel_id += 2;
-               }, node_1_privkey, node_2_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(res) => assert!(res),
-                       _ => panic!()
-               };
-
-               {
-                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
-                               None => panic!(),
-                               Some(_) => ()
-                       };
-               }
-
-               // If we receive announcement for the same channel (but TX is not confirmed),
-               // drop new one on the floor, since we can't see any changes.
-               *chain_source.utxo_ret.lock().unwrap() = Err(chain::AccessError::UnknownTx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Channel announced without corresponding UTXO entry")
-               };
-
-               // But if it is confirmed, replace the channel
-               *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: 0, script_pubkey: good_script });
-               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
-                       unsigned_announcement.features = ChannelFeatures::empty();
-                       unsigned_announcement.short_channel_id += 2;
-               }, node_1_privkey, node_2_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(res) => assert!(res),
-                       _ => panic!()
-               };
-               {
-                       match network_graph.read_only().channels().get(&valid_announcement.contents.short_channel_id) {
-                               Some(channel_entry) => {
-                                       assert_eq!(channel_entry.features, ChannelFeatures::empty());
-                               },
-                               _ => panic!()
-                       };
-               }
-
-               // Don't relay valid channels with excess data
-               let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
-                       unsigned_announcement.short_channel_id += 3;
-                       unsigned_announcement.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
-               }, node_1_privkey, node_2_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(res) => assert!(!res),
-                       _ => panic!()
-               };
-
-               let mut invalid_sig_announcement = valid_announcement.clone();
-               invalid_sig_announcement.contents.excess_data = Vec::new();
-               match net_graph_msg_handler.handle_channel_announcement(&invalid_sig_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Invalid signature on channel_announcement message")
-               };
-
-               let channel_to_itself_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&channel_to_itself_announcement) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Channel announcement node had a channel with itself")
-               };
-       }
-
-       #[test]
-       fn handling_channel_update() {
-               let secp_ctx = Secp256k1::new();
-               let logger: Arc<Logger> = Arc::new(test_utils::TestLogger::new());
-               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
-               let network_graph = NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash());
-               let net_graph_msg_handler = NetGraphMsgHandler::new(&network_graph, Some(chain_source.clone()), Arc::clone(&logger));
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               let amount_sats = 1000_000;
-               let short_channel_id;
-
-               {
-                       // Announce a channel we will update
-                       let good_script = get_channel_script(&secp_ctx);
-                       *chain_source.utxo_ret.lock().unwrap() = Ok(TxOut { value: amount_sats, script_pubkey: good_script.clone() });
-
-                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
-                       match net_graph_msg_handler.handle_channel_announcement(&valid_channel_announcement) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-
-               }
-
-               let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(res) => assert!(res),
-                       _ => panic!()
-               };
-
-               {
-                       match network_graph.read_only().channels().get(&short_channel_id) {
-                               None => panic!(),
-                               Some(channel_info) => {
-                                       assert_eq!(channel_info.one_to_two.as_ref().unwrap().cltv_expiry_delta, 144);
-                                       assert!(channel_info.two_to_one.is_none());
-                               }
-                       };
-               }
-
-               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.timestamp += 100;
-                       unsigned_channel_update.excess_data.resize(MAX_EXCESS_BYTES_FOR_RELAY + 1, 0);
-               }, node_1_privkey, &secp_ctx);
-               // Return false because contains excess data
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(res) => assert!(!res),
-                       _ => panic!()
-               };
-
-               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.timestamp += 110;
-                       unsigned_channel_update.short_channel_id += 1;
-               }, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Couldn't find channel for update")
-               };
-
-               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.htlc_maximum_msat = OptionalField::Present(MAX_VALUE_MSAT + 1);
-                       unsigned_channel_update.timestamp += 110;
-               }, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "htlc_maximum_msat is larger than maximum possible msats")
-               };
-
-               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.htlc_maximum_msat = OptionalField::Present(amount_sats * 1000 + 1);
-                       unsigned_channel_update.timestamp += 110;
-               }, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "htlc_maximum_msat is larger than channel capacity or capacity is bogus")
-               };
-
-               // Even though previous update was not relayed further, we still accepted it,
-               // so we now won't accept update before the previous one.
-               let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.timestamp += 100;
-               }, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Update had same timestamp as last processed update")
-               };
-
-               let mut invalid_sig_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                       unsigned_channel_update.timestamp += 500;
-               }, node_1_privkey, &secp_ctx);
-               let zero_hash = Sha256dHash::hash(&[0; 32]);
-               let fake_msghash = hash_to_message!(&zero_hash);
-               invalid_sig_channel_update.signature = secp_ctx.sign_ecdsa(&fake_msghash, node_1_privkey);
-               match net_graph_msg_handler.handle_channel_update(&invalid_sig_channel_update) {
-                       Ok(_) => panic!(),
-                       Err(e) => assert_eq!(e.err, "Invalid signature on channel_update message")
-               };
-       }
-
-       #[test]
-       fn handling_network_update() {
-               let logger = test_utils::TestLogger::new();
-               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
-               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
-               let network_graph = NetworkGraph::new(genesis_hash);
-               let net_graph_msg_handler = NetGraphMsgHandler::new(&network_graph, Some(chain_source.clone()), &logger);
-               let secp_ctx = Secp256k1::new();
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               {
-                       // There is no nodes in the table at the beginning.
-                       assert_eq!(network_graph.read_only().nodes().len(), 0);
-               }
-
-               let short_channel_id;
-               {
-                       // Announce a channel we will update
-                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
-                       let chain_source: Option<&test_utils::TestChainSource> = None;
-                       assert!(network_graph.update_channel_from_announcement(&valid_channel_announcement, &chain_source, &secp_ctx).is_ok());
-                       assert!(network_graph.read_only().channels().get(&short_channel_id).is_some());
-
-                       let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
-                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_none());
-
-                       net_graph_msg_handler.handle_event(&Event::PaymentPathFailed {
-                               payment_id: None,
-                               payment_hash: PaymentHash([0; 32]),
-                               rejected_by_dest: false,
-                               all_paths_failed: true,
-                               path: vec![],
-                               network_update: Some(NetworkUpdate::ChannelUpdateMessage {
-                                       msg: valid_channel_update,
-                               }),
-                               short_channel_id: None,
-                               retry: None,
-                               error_code: None,
-                               error_data: None,
-                       });
-
-                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_some());
-               }
-
-               // Non-permanent closing just disables a channel
-               {
-                       match network_graph.read_only().channels().get(&short_channel_id) {
-                               None => panic!(),
-                               Some(channel_info) => {
-                                       assert!(channel_info.one_to_two.as_ref().unwrap().enabled);
-                               }
-                       };
-
-                       net_graph_msg_handler.handle_event(&Event::PaymentPathFailed {
-                               payment_id: None,
-                               payment_hash: PaymentHash([0; 32]),
-                               rejected_by_dest: false,
-                               all_paths_failed: true,
-                               path: vec![],
-                               network_update: Some(NetworkUpdate::ChannelClosed {
-                                       short_channel_id,
-                                       is_permanent: false,
-                               }),
-                               short_channel_id: None,
-                               retry: None,
-                               error_code: None,
-                               error_data: None,
-                       });
-
-                       match network_graph.read_only().channels().get(&short_channel_id) {
-                               None => panic!(),
-                               Some(channel_info) => {
-                                       assert!(!channel_info.one_to_two.as_ref().unwrap().enabled);
-                               }
-                       };
-               }
-
-               // Permanent closing deletes a channel
-               net_graph_msg_handler.handle_event(&Event::PaymentPathFailed {
-                       payment_id: None,
-                       payment_hash: PaymentHash([0; 32]),
-                       rejected_by_dest: false,
-                       all_paths_failed: true,
-                       path: vec![],
-                       network_update: Some(NetworkUpdate::ChannelClosed {
-                               short_channel_id,
-                               is_permanent: true,
-                       }),
-                       short_channel_id: None,
-                       retry: None,
-                       error_code: None,
-                       error_data: None,
-               });
-
-               assert_eq!(network_graph.read_only().channels().len(), 0);
-               // Nodes are also deleted because there are no associated channels anymore
-               assert_eq!(network_graph.read_only().nodes().len(), 0);
-               // TODO: Test NetworkUpdate::NodeFailure, which is not implemented yet.
-       }
-
-       #[test]
-       fn test_channel_timeouts() {
-               // Test the removal of channels with `remove_stale_channels`.
-               let logger = test_utils::TestLogger::new();
-               let chain_source = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
-               let genesis_hash = genesis_block(Network::Testnet).header.block_hash();
-               let network_graph = NetworkGraph::new(genesis_hash);
-               let net_graph_msg_handler = NetGraphMsgHandler::new(&network_graph, Some(chain_source.clone()), &logger);
-               let secp_ctx = Secp256k1::new();
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-               let short_channel_id = valid_channel_announcement.contents.short_channel_id;
-               let chain_source: Option<&test_utils::TestChainSource> = None;
-               assert!(network_graph.update_channel_from_announcement(&valid_channel_announcement, &chain_source, &secp_ctx).is_ok());
-               assert!(network_graph.read_only().channels().get(&short_channel_id).is_some());
-
-               let valid_channel_update = get_signed_channel_update(|_| {}, node_1_privkey, &secp_ctx);
-               assert!(net_graph_msg_handler.handle_channel_update(&valid_channel_update).is_ok());
-               assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_some());
-
-               network_graph.remove_stale_channels_with_time(100 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
-               assert_eq!(network_graph.read_only().channels().len(), 1);
-               assert_eq!(network_graph.read_only().nodes().len(), 2);
-
-               network_graph.remove_stale_channels_with_time(101 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
-               #[cfg(feature = "std")]
-               {
-                       // In std mode, a further check is performed before fully removing the channel -
-                       // the channel_announcement must have been received at least two weeks ago. We
-                       // fudge that here by indicating the time has jumped two weeks. Note that the
-                       // directional channel information will have been removed already..
-                       assert_eq!(network_graph.read_only().channels().len(), 1);
-                       assert_eq!(network_graph.read_only().nodes().len(), 2);
-                       assert!(network_graph.read_only().channels().get(&short_channel_id).unwrap().one_to_two.is_none());
-
-                       use std::time::{SystemTime, UNIX_EPOCH};
-                       let announcement_time = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-                       network_graph.remove_stale_channels_with_time(announcement_time + 1 + STALE_CHANNEL_UPDATE_AGE_LIMIT_SECS);
-               }
-
-               assert_eq!(network_graph.read_only().channels().len(), 0);
-               assert_eq!(network_graph.read_only().nodes().len(), 0);
-       }
-
-       #[test]
-       fn getting_next_channel_announcements() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               // Channels were not announced yet.
-               let channels_with_announcements = net_graph_msg_handler.get_next_channel_announcements(0, 1);
-               assert_eq!(channels_with_announcements.len(), 0);
-
-               let short_channel_id;
-               {
-                       // Announce a channel we will update
-                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-                       short_channel_id = valid_channel_announcement.contents.short_channel_id;
-                       match net_graph_msg_handler.handle_channel_announcement(&valid_channel_announcement) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-               }
-
-               // Contains initial channel announcement now.
-               let channels_with_announcements = net_graph_msg_handler.get_next_channel_announcements(short_channel_id, 1);
-               assert_eq!(channels_with_announcements.len(), 1);
-               if let Some(channel_announcements) = channels_with_announcements.first() {
-                       let &(_, ref update_1, ref update_2) = channel_announcements;
-                       assert_eq!(update_1, &None);
-                       assert_eq!(update_2, &None);
-               } else {
-                       panic!();
-               }
-
-
-               {
-                       // Valid channel update
-                       let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                               unsigned_channel_update.timestamp = 101;
-                       }, node_1_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-               }
-
-               // Now contains an initial announcement and an update.
-               let channels_with_announcements = net_graph_msg_handler.get_next_channel_announcements(short_channel_id, 1);
-               assert_eq!(channels_with_announcements.len(), 1);
-               if let Some(channel_announcements) = channels_with_announcements.first() {
-                       let &(_, ref update_1, ref update_2) = channel_announcements;
-                       assert_ne!(update_1, &None);
-                       assert_eq!(update_2, &None);
-               } else {
-                       panic!();
-               }
-
-               {
-                       // Channel update with excess data.
-                       let valid_channel_update = get_signed_channel_update(|unsigned_channel_update| {
-                               unsigned_channel_update.timestamp = 102;
-                               unsigned_channel_update.excess_data = [1; MAX_EXCESS_BYTES_FOR_RELAY + 1].to_vec();
-                       }, node_1_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-               }
-
-               // Test that announcements with excess data won't be returned
-               let channels_with_announcements = net_graph_msg_handler.get_next_channel_announcements(short_channel_id, 1);
-               assert_eq!(channels_with_announcements.len(), 1);
-               if let Some(channel_announcements) = channels_with_announcements.first() {
-                       let &(_, ref update_1, ref update_2) = channel_announcements;
-                       assert_eq!(update_1, &None);
-                       assert_eq!(update_2, &None);
-               } else {
-                       panic!();
-               }
-
-               // Further starting point have no channels after it
-               let channels_with_announcements = net_graph_msg_handler.get_next_channel_announcements(short_channel_id + 1000, 1);
-               assert_eq!(channels_with_announcements.len(), 0);
-       }
-
-       #[test]
-       fn getting_next_node_announcements() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_1_privkey);
-
-               // No nodes yet.
-               let next_announcements = net_graph_msg_handler.get_next_node_announcements(None, 10);
-               assert_eq!(next_announcements.len(), 0);
-
-               {
-                       // Announce a channel to add 2 nodes
-                       let valid_channel_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_channel_announcement(&valid_channel_announcement) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-               }
-
-
-               // Nodes were never announced
-               let next_announcements = net_graph_msg_handler.get_next_node_announcements(None, 3);
-               assert_eq!(next_announcements.len(), 0);
-
-               {
-                       let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-
-                       let valid_announcement = get_signed_node_announcement(|_| {}, node_2_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                               Ok(_) => (),
-                               Err(_) => panic!()
-                       };
-               }
-
-               let next_announcements = net_graph_msg_handler.get_next_node_announcements(None, 3);
-               assert_eq!(next_announcements.len(), 2);
-
-               // Skip the first node.
-               let next_announcements = net_graph_msg_handler.get_next_node_announcements(Some(&node_id_1), 2);
-               assert_eq!(next_announcements.len(), 1);
-
-               {
-                       // Later announcement which should not be relayed (excess data) prevent us from sharing a node
-                       let valid_announcement = get_signed_node_announcement(|unsigned_announcement| {
-                               unsigned_announcement.timestamp += 10;
-                               unsigned_announcement.excess_data = [1; MAX_EXCESS_BYTES_FOR_RELAY + 1].to_vec();
-                       }, node_2_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                               Ok(res) => assert!(!res),
-                               Err(_) => panic!()
-                       };
-               }
-
-               let next_announcements = net_graph_msg_handler.get_next_node_announcements(Some(&node_id_1), 2);
-               assert_eq!(next_announcements.len(), 0);
-       }
-
-       #[test]
-       fn network_graph_serialization() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-
-               // Announce a channel to add a corresponding node.
-               let valid_announcement = get_signed_channel_announcement(|_| {}, node_1_privkey, node_2_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                       Ok(res) => assert!(res),
-                       _ => panic!()
-               };
-
-               let valid_announcement = get_signed_node_announcement(|_| {}, node_1_privkey, &secp_ctx);
-               match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
-                       Ok(_) => (),
-                       Err(_) => panic!()
-               };
-
-               let mut w = test_utils::TestVecWriter(Vec::new());
-               assert!(!network_graph.read_only().nodes().is_empty());
-               assert!(!network_graph.read_only().channels().is_empty());
-               network_graph.write(&mut w).unwrap();
-               assert!(<NetworkGraph>::read(&mut io::Cursor::new(&w.0)).unwrap() == network_graph);
-       }
-
-       #[test]
-       fn network_graph_tlv_serialization() {
-               let mut network_graph = create_network_graph();
-               network_graph.set_last_rapid_gossip_sync_timestamp(42);
-
-               let mut w = test_utils::TestVecWriter(Vec::new());
-               network_graph.write(&mut w).unwrap();
-               let reassembled_network_graph: NetworkGraph = Readable::read(&mut io::Cursor::new(&w.0)).unwrap();
-               assert!(reassembled_network_graph == network_graph);
-               assert_eq!(reassembled_network_graph.get_last_rapid_gossip_sync_timestamp().unwrap(), 42);
-       }
-
-       #[test]
-       #[cfg(feature = "std")]
-       fn calling_sync_routing_table() {
-               use std::time::{SystemTime, UNIX_EPOCH};
-
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-               let node_privkey_1 = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_privkey_1);
-
-               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
-
-               // It should ignore if gossip_queries feature is not enabled
-               {
-                       let init_msg = Init { features: InitFeatures::known().clear_gossip_queries(), remote_network_address: None };
-                       net_graph_msg_handler.peer_connected(&node_id_1, &init_msg);
-                       let events = net_graph_msg_handler.get_and_clear_pending_msg_events();
-                       assert_eq!(events.len(), 0);
-               }
-
-               // It should send a gossip_timestamp_filter with the correct information
-               {
-                       let init_msg = Init { features: InitFeatures::known(), remote_network_address: None };
-                       net_graph_msg_handler.peer_connected(&node_id_1, &init_msg);
-                       let events = net_graph_msg_handler.get_and_clear_pending_msg_events();
-                       assert_eq!(events.len(), 1);
-                       match &events[0] {
-                               MessageSendEvent::SendGossipTimestampFilter{ node_id, msg } => {
-                                       assert_eq!(node_id, &node_id_1);
-                                       assert_eq!(msg.chain_hash, chain_hash);
-                                       let expected_timestamp = SystemTime::now().duration_since(UNIX_EPOCH).expect("Time must be > 1970").as_secs();
-                                       assert!((msg.first_timestamp as u64) >= expected_timestamp - 60*60*24*7*2);
-                                       assert!((msg.first_timestamp as u64) < expected_timestamp - 60*60*24*7*2 + 10);
-                                       assert_eq!(msg.timestamp_range, u32::max_value());
-                               },
-                               _ => panic!("Expected MessageSendEvent::SendChannelRangeQuery")
-                       };
-               }
-       }
-
-       #[test]
-       fn handling_query_channel_range() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-
-               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
-               let node_1_privkey = &SecretKey::from_slice(&[42; 32]).unwrap();
-               let node_2_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-               let node_id_2 = PublicKey::from_secret_key(&secp_ctx, node_2_privkey);
-
-               let mut scids: Vec<u64> = vec![
-                       scid_from_parts(0xfffffe, 0xffffff, 0xffff).unwrap(), // max
-                       scid_from_parts(0xffffff, 0xffffff, 0xffff).unwrap(), // never
-               ];
-
-               // used for testing multipart reply across blocks
-               for block in 100000..=108001 {
-                       scids.push(scid_from_parts(block, 0, 0).unwrap());
-               }
-
-               // used for testing resumption on same block
-               scids.push(scid_from_parts(108001, 1, 0).unwrap());
-
-               for scid in scids {
-                       let valid_announcement = get_signed_channel_announcement(|unsigned_announcement| {
-                               unsigned_announcement.short_channel_id = scid;
-                       }, node_1_privkey, node_2_privkey, &secp_ctx);
-                       match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
-                               Ok(_) => (),
-                               _ => panic!()
-                       };
-               }
-
-               // Error when number_of_blocks=0
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0,
-                               number_of_blocks: 0,
-                       },
-                       false,
-                       vec![ReplyChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0,
-                               number_of_blocks: 0,
-                               sync_complete: true,
-                               short_channel_ids: vec![]
-                       }]
-               );
-
-               // Error when wrong chain
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: genesis_block(Network::Bitcoin).header.block_hash(),
-                               first_blocknum: 0,
-                               number_of_blocks: 0xffff_ffff,
-                       },
-                       false,
-                       vec![ReplyChannelRange {
-                               chain_hash: genesis_block(Network::Bitcoin).header.block_hash(),
-                               first_blocknum: 0,
-                               number_of_blocks: 0xffff_ffff,
-                               sync_complete: true,
-                               short_channel_ids: vec![],
-                       }]
-               );
-
-               // Error when first_blocknum > 0xffffff
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0x01000000,
-                               number_of_blocks: 0xffff_ffff,
-                       },
-                       false,
-                       vec![ReplyChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0x01000000,
-                               number_of_blocks: 0xffff_ffff,
-                               sync_complete: true,
-                               short_channel_ids: vec![]
-                       }]
-               );
-
-               // Empty reply when max valid SCID block num
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0xffffff,
-                               number_of_blocks: 1,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 0xffffff,
-                                       number_of_blocks: 1,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![]
-                               },
-                       ]
-               );
-
-               // No results in valid query range
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 1000,
-                               number_of_blocks: 1000,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 1000,
-                                       number_of_blocks: 1000,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![],
-                               }
-                       ]
-               );
-
-               // Overflow first_blocknum + number_of_blocks
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 0xfe0000,
-                               number_of_blocks: 0xffffffff,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 0xfe0000,
-                                       number_of_blocks: 0xffffffff - 0xfe0000,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![
-                                               0xfffffe_ffffff_ffff, // max
-                                       ]
-                               }
-                       ]
-               );
-
-               // Single block exactly full
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 100000,
-                               number_of_blocks: 8000,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 100000,
-                                       number_of_blocks: 8000,
-                                       sync_complete: true,
-                                       short_channel_ids: (100000..=107999)
-                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
-                                               .collect(),
-                               },
-                       ]
-               );
-
-               // Multiple split on new block
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 100000,
-                               number_of_blocks: 8001,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 100000,
-                                       number_of_blocks: 7999,
-                                       sync_complete: false,
-                                       short_channel_ids: (100000..=107999)
-                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
-                                               .collect(),
-                               },
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 107999,
-                                       number_of_blocks: 2,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![
-                                               scid_from_parts(108000, 0, 0).unwrap(),
-                                       ],
-                               }
-                       ]
-               );
-
-               // Multiple split on same block
-               do_handling_query_channel_range(
-                       &net_graph_msg_handler,
-                       &node_id_2,
-                       QueryChannelRange {
-                               chain_hash: chain_hash.clone(),
-                               first_blocknum: 100002,
-                               number_of_blocks: 8000,
-                       },
-                       true,
-                       vec![
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 100002,
-                                       number_of_blocks: 7999,
-                                       sync_complete: false,
-                                       short_channel_ids: (100002..=108001)
-                                               .map(|block| scid_from_parts(block, 0, 0).unwrap())
-                                               .collect(),
-                               },
-                               ReplyChannelRange {
-                                       chain_hash: chain_hash.clone(),
-                                       first_blocknum: 108001,
-                                       number_of_blocks: 1,
-                                       sync_complete: true,
-                                       short_channel_ids: vec![
-                                               scid_from_parts(108001, 1, 0).unwrap(),
-                                       ],
-                               }
-                       ]
-               );
-       }
-
-       fn do_handling_query_channel_range(
-               net_graph_msg_handler: &NetGraphMsgHandler<&NetworkGraph, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
-               test_node_id: &PublicKey,
-               msg: QueryChannelRange,
-               expected_ok: bool,
-               expected_replies: Vec<ReplyChannelRange>
-       ) {
-               let mut max_firstblocknum = msg.first_blocknum.saturating_sub(1);
-               let mut c_lightning_0_9_prev_end_blocknum = max_firstblocknum;
-               let query_end_blocknum = msg.end_blocknum();
-               let result = net_graph_msg_handler.handle_query_channel_range(test_node_id, msg);
-
-               if expected_ok {
-                       assert!(result.is_ok());
-               } else {
-                       assert!(result.is_err());
-               }
-
-               let events = net_graph_msg_handler.get_and_clear_pending_msg_events();
-               assert_eq!(events.len(), expected_replies.len());
-
-               for i in 0..events.len() {
-                       let expected_reply = &expected_replies[i];
-                       match &events[i] {
-                               MessageSendEvent::SendReplyChannelRange { node_id, msg } => {
-                                       assert_eq!(node_id, test_node_id);
-                                       assert_eq!(msg.chain_hash, expected_reply.chain_hash);
-                                       assert_eq!(msg.first_blocknum, expected_reply.first_blocknum);
-                                       assert_eq!(msg.number_of_blocks, expected_reply.number_of_blocks);
-                                       assert_eq!(msg.sync_complete, expected_reply.sync_complete);
-                                       assert_eq!(msg.short_channel_ids, expected_reply.short_channel_ids);
-
-                                       // Enforce exactly the sequencing requirements present on c-lightning v0.9.3
-                                       assert!(msg.first_blocknum == c_lightning_0_9_prev_end_blocknum || msg.first_blocknum == c_lightning_0_9_prev_end_blocknum.saturating_add(1));
-                                       assert!(msg.first_blocknum >= max_firstblocknum);
-                                       max_firstblocknum = msg.first_blocknum;
-                                       c_lightning_0_9_prev_end_blocknum = msg.first_blocknum.saturating_add(msg.number_of_blocks);
-
-                                       // Check that the last block count is >= the query's end_blocknum
-                                       if i == events.len() - 1 {
-                                               assert!(msg.first_blocknum.saturating_add(msg.number_of_blocks) >= query_end_blocknum);
-                                       }
-                               },
-                               _ => panic!("expected MessageSendEvent::SendReplyChannelRange"),
-                       }
-               }
-       }
-
-       #[test]
-       fn handling_query_short_channel_ids() {
-               let network_graph = create_network_graph();
-               let (secp_ctx, net_graph_msg_handler) = create_net_graph_msg_handler(&network_graph);
-               let node_privkey = &SecretKey::from_slice(&[41; 32]).unwrap();
-               let node_id = PublicKey::from_secret_key(&secp_ctx, node_privkey);
-
-               let chain_hash = genesis_block(Network::Testnet).header.block_hash();
-
-               let result = net_graph_msg_handler.handle_query_short_channel_ids(&node_id, QueryShortChannelIds {
-                       chain_hash,
-                       short_channel_ids: vec![0x0003e8_000000_0000],
-               });
-               assert!(result.is_err());
-       }
-}
-
-#[cfg(all(test, feature = "_bench_unstable"))]
-mod benches {
-       use super::*;
-
-       use test::Bencher;
-       use std::io::Read;
-
-       #[bench]
-       fn read_network_graph(bench: &mut Bencher) {
-               let mut d = ::routing::router::test_utils::get_route_file().unwrap();
-               let mut v = Vec::new();
-               d.read_to_end(&mut v).unwrap();
-               bench.iter(|| {
-                       let _ = NetworkGraph::read(&mut std::io::Cursor::new(&v)).unwrap();
-               });
-       }
-
-       #[bench]
-       fn write_network_graph(bench: &mut Bencher) {
-               let mut d = ::routing::router::test_utils::get_route_file().unwrap();
-               let net_graph = NetworkGraph::read(&mut d).unwrap();
-               bench.iter(|| {
-                       let _ = net_graph.encode();
-               });
-       }
-}
index 5d4d4ac123818ca0cda313a4e768a4b479f8ff3d..301ae76ddbb315e4aa02cec072e394f0bf87924f 100644 (file)
@@ -9,7 +9,7 @@
 
 //! The top-level routing/network map tracking logic lives here.
 //!
-//! You probably want to create a NetGraphMsgHandler and use that as your RoutingMessageHandler and then
+//! You probably want to create a P2PGossipSync and use that as your RoutingMessageHandler and then
 //! interrogate it to get routes for your own payments.
 
 use bitcoin::secp256k1::PublicKey;
@@ -17,8 +17,8 @@ use bitcoin::secp256k1::PublicKey;
 use ln::channelmanager::ChannelDetails;
 use ln::features::{ChannelFeatures, InvoiceFeatures, NodeFeatures};
 use ln::msgs::{DecodeError, ErrorAction, LightningError, MAX_VALUE_MSAT};
+use routing::gossip::{DirectedChannelInfoWithUpdate, EffectiveCapacity, NetworkGraph, ReadOnlyNetworkGraph, NodeId, RoutingFees};
 use routing::scoring::{ChannelUsage, Score};
-use routing::network_graph::{DirectedChannelInfoWithUpdate, EffectiveCapacity, NetworkGraph, ReadOnlyNetworkGraph, NodeId, RoutingFees};
 use util::ser::{Writeable, Readable, Writer};
 use util::logger::{Level, Logger};
 use util::chacha20::ChaCha20;
@@ -1858,7 +1858,7 @@ fn build_route_from_hops_internal<L: Deref>(
 
 #[cfg(test)]
 mod tests {
-       use routing::network_graph::{NetworkGraph, NetGraphMsgHandler, NodeId};
+       use routing::gossip::{NetworkGraph, P2PGossipSync, NodeId};
        use routing::router::{get_route, build_route_from_hops_internal, add_random_cltv_offset, default_node_features,
                PaymentParameters, Route, RouteHint, RouteHintHop, RouteHop, RoutingFees,
                DEFAULT_MAX_TOTAL_CLTV_EXPIRY_DELTA, MAX_PATH_LENGTH_ESTIMATE};
@@ -1926,7 +1926,7 @@ mod tests {
 
        // Using the same keys for LN and BTC ids
        fn add_channel(
-               net_graph_msg_handler: &NetGraphMsgHandler<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
+               gossip_sync: &P2PGossipSync<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
                secp_ctx: &Secp256k1<All>, node_1_privkey: &SecretKey, node_2_privkey: &SecretKey, features: ChannelFeatures, short_channel_id: u64
        ) {
                let node_id_1 = PublicKey::from_secret_key(&secp_ctx, node_1_privkey);
@@ -1951,14 +1951,14 @@ mod tests {
                        bitcoin_signature_2: secp_ctx.sign_ecdsa(&msghash, node_2_privkey),
                        contents: unsigned_announcement.clone(),
                };
-               match net_graph_msg_handler.handle_channel_announcement(&valid_announcement) {
+               match gossip_sync.handle_channel_announcement(&valid_announcement) {
                        Ok(res) => assert!(res),
                        _ => panic!()
                };
        }
 
        fn update_channel(
-               net_graph_msg_handler: &NetGraphMsgHandler<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
+               gossip_sync: &P2PGossipSync<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
                secp_ctx: &Secp256k1<All>, node_privkey: &SecretKey, update: UnsignedChannelUpdate
        ) {
                let msghash = hash_to_message!(&Sha256dHash::hash(&update.encode()[..])[..]);
@@ -1967,14 +1967,14 @@ mod tests {
                        contents: update.clone()
                };
 
-               match net_graph_msg_handler.handle_channel_update(&valid_channel_update) {
+               match gossip_sync.handle_channel_update(&valid_channel_update) {
                        Ok(res) => assert!(res),
                        Err(_) => panic!()
                };
        }
 
        fn add_or_update_node(
-               net_graph_msg_handler: &NetGraphMsgHandler<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
+               gossip_sync: &P2PGossipSync<Arc<NetworkGraph>, Arc<test_utils::TestChainSource>, Arc<test_utils::TestLogger>>,
                secp_ctx: &Secp256k1<All>, node_privkey: &SecretKey, features: NodeFeatures, timestamp: u32
        ) {
                let node_id = PublicKey::from_secret_key(&secp_ctx, node_privkey);
@@ -1994,7 +1994,7 @@ mod tests {
                        contents: unsigned_announcement.clone()
                };
 
-               match net_graph_msg_handler.handle_node_announcement(&valid_announcement) {
+               match gossip_sync.handle_node_announcement(&valid_announcement) {
                        Ok(_) => (),
                        Err(_) => panic!()
                };
@@ -2029,7 +2029,7 @@ mod tests {
        }
 
        fn build_line_graph() -> (
-               Secp256k1<All>, sync::Arc<NetworkGraph>, NetGraphMsgHandler<sync::Arc<NetworkGraph>,
+               Secp256k1<All>, sync::Arc<NetworkGraph>, P2PGossipSync<sync::Arc<NetworkGraph>,
                sync::Arc<test_utils::TestChainSource>, sync::Arc<crate::util::test_utils::TestLogger>>,
                sync::Arc<test_utils::TestChainSource>, sync::Arc<test_utils::TestLogger>,
        ) {
@@ -2037,7 +2037,7 @@ mod tests {
                let logger = Arc::new(test_utils::TestLogger::new());
                let chain_monitor = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
                let network_graph = Arc::new(NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash()));
-               let net_graph_msg_handler = NetGraphMsgHandler::new(Arc::clone(&network_graph), None, Arc::clone(&logger));
+               let gossip_sync = P2PGossipSync::new(Arc::clone(&network_graph), None, Arc::clone(&logger));
 
                // Build network from our_id to node 19:
                // our_id -1(1)2- node0 -1(2)2- node1 - ... - node19
@@ -2046,9 +2046,9 @@ mod tests {
                for (idx, (cur_privkey, next_privkey)) in core::iter::once(&our_privkey)
                        .chain(privkeys.iter()).zip(privkeys.iter()).enumerate() {
                        let cur_short_channel_id = (idx as u64) + 1;
-                       add_channel(&net_graph_msg_handler, &secp_ctx, &cur_privkey, &next_privkey,
+                       add_channel(&gossip_sync, &secp_ctx, &cur_privkey, &next_privkey,
                                ChannelFeatures::from_le_bytes(id_to_feature_flags(1)), cur_short_channel_id);
-                       update_channel(&net_graph_msg_handler, &secp_ctx, &cur_privkey, UnsignedChannelUpdate {
+                       update_channel(&gossip_sync, &secp_ctx, &cur_privkey, UnsignedChannelUpdate {
                                chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                                short_channel_id: cur_short_channel_id,
                                timestamp: idx as u32,
@@ -2060,7 +2060,7 @@ mod tests {
                                fee_proportional_millionths: 0,
                                excess_data: Vec::new()
                        });
-                       update_channel(&net_graph_msg_handler, &secp_ctx, &next_privkey, UnsignedChannelUpdate {
+                       update_channel(&gossip_sync, &secp_ctx, &next_privkey, UnsignedChannelUpdate {
                                chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                                short_channel_id: cur_short_channel_id,
                                timestamp: (idx as u32)+1,
@@ -2072,17 +2072,17 @@ mod tests {
                                fee_proportional_millionths: 0,
                                excess_data: Vec::new()
                        });
-                       add_or_update_node(&net_graph_msg_handler, &secp_ctx, next_privkey,
+                       add_or_update_node(&gossip_sync, &secp_ctx, next_privkey,
                                NodeFeatures::from_le_bytes(id_to_feature_flags(1)), 0);
                }
 
-               (secp_ctx, network_graph, net_graph_msg_handler, chain_monitor, logger)
+               (secp_ctx, network_graph, gossip_sync, chain_monitor, logger)
        }
 
        fn build_graph() -> (
                Secp256k1<All>,
                sync::Arc<NetworkGraph>,
-               NetGraphMsgHandler<sync::Arc<NetworkGraph>, sync::Arc<test_utils::TestChainSource>, sync::Arc<crate::util::test_utils::TestLogger>>,
+               P2PGossipSync<sync::Arc<NetworkGraph>, sync::Arc<test_utils::TestChainSource>, sync::Arc<crate::util::test_utils::TestLogger>>,
                sync::Arc<test_utils::TestChainSource>,
                sync::Arc<test_utils::TestLogger>,
        ) {
@@ -2090,7 +2090,7 @@ mod tests {
                let logger = Arc::new(test_utils::TestLogger::new());
                let chain_monitor = Arc::new(test_utils::TestChainSource::new(Network::Testnet));
                let network_graph = Arc::new(NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash()));
-               let net_graph_msg_handler = NetGraphMsgHandler::new(Arc::clone(&network_graph), None, Arc::clone(&logger));
+               let gossip_sync = P2PGossipSync::new(Arc::clone(&network_graph), None, Arc::clone(&logger));
                // Build network from our_id to node6:
                //
                //        -1(1)2-  node0  -1(3)2-
@@ -2152,8 +2152,8 @@ mod tests {
 
                let (our_privkey, _, privkeys, _) = get_nodes(&secp_ctx);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, &privkeys[0], ChannelFeatures::from_le_bytes(id_to_feature_flags(1)), 1);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &our_privkey, &privkeys[0], ChannelFeatures::from_le_bytes(id_to_feature_flags(1)), 1);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 1,
@@ -2166,10 +2166,10 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[0], NodeFeatures::from_le_bytes(id_to_feature_flags(1)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[0], NodeFeatures::from_le_bytes(id_to_feature_flags(1)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, &privkeys[1], ChannelFeatures::from_le_bytes(id_to_feature_flags(2)), 2);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &our_privkey, &privkeys[1], ChannelFeatures::from_le_bytes(id_to_feature_flags(2)), 2);
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 1,
@@ -2181,7 +2181,7 @@ mod tests {
                        fee_proportional_millionths: u32::max_value(),
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 1,
@@ -2194,10 +2194,10 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[1], NodeFeatures::from_le_bytes(id_to_feature_flags(2)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[1], NodeFeatures::from_le_bytes(id_to_feature_flags(2)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, &privkeys[7], ChannelFeatures::from_le_bytes(id_to_feature_flags(12)), 12);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &our_privkey, &privkeys[7], ChannelFeatures::from_le_bytes(id_to_feature_flags(12)), 12);
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 1,
@@ -2209,7 +2209,7 @@ mod tests {
                        fee_proportional_millionths: u32::max_value(),
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 1,
@@ -2222,10 +2222,10 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[7], NodeFeatures::from_le_bytes(id_to_feature_flags(8)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[7], NodeFeatures::from_le_bytes(id_to_feature_flags(8)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 3);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[0], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 3);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 1,
@@ -2237,7 +2237,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 1,
@@ -2250,8 +2250,8 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(4)), 4);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[1], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(4)), 4);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 1,
@@ -2263,7 +2263,7 @@ mod tests {
                        fee_proportional_millionths: 1000000,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 1,
@@ -2276,8 +2276,8 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(13)), 13);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[7], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(13)), 13);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 1,
@@ -2289,7 +2289,7 @@ mod tests {
                        fee_proportional_millionths: 2000000,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 1,
@@ -2302,10 +2302,10 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[2], NodeFeatures::from_le_bytes(id_to_feature_flags(3)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[2], NodeFeatures::from_le_bytes(id_to_feature_flags(3)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(6)), 6);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(6)), 6);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 1,
@@ -2317,7 +2317,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 1,
@@ -2330,8 +2330,8 @@ mod tests {
                        excess_data: Vec::new(),
                });
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(11)), 11);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[4], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(11)), 11);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 1,
@@ -2343,7 +2343,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[3], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[3], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 1,
@@ -2356,12 +2356,12 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[4], NodeFeatures::from_le_bytes(id_to_feature_flags(5)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[4], NodeFeatures::from_le_bytes(id_to_feature_flags(5)), 0);
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[3], NodeFeatures::from_le_bytes(id_to_feature_flags(4)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[3], NodeFeatures::from_le_bytes(id_to_feature_flags(4)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[5], ChannelFeatures::from_le_bytes(id_to_feature_flags(7)), 7);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[5], ChannelFeatures::from_le_bytes(id_to_feature_flags(7)), 7);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 1,
@@ -2373,7 +2373,7 @@ mod tests {
                        fee_proportional_millionths: 1000000,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[5], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[5], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 1,
@@ -2386,9 +2386,9 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[5], NodeFeatures::from_le_bytes(id_to_feature_flags(6)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[5], NodeFeatures::from_le_bytes(id_to_feature_flags(6)), 0);
 
-               (secp_ctx, network_graph, net_graph_msg_handler, chain_monitor, logger)
+               (secp_ctx, network_graph, gossip_sync, chain_monitor, logger)
        }
 
        #[test]
@@ -2448,7 +2448,7 @@ mod tests {
 
        #[test]
        fn htlc_minimum_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let payment_params = PaymentParameters::from_node_id(nodes[2]);
                let scorer = test_utils::TestScorer::with_penalty(0);
@@ -2458,7 +2458,7 @@ mod tests {
                // Simple route to 2 via 1
 
                // Disable other paths
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -2470,7 +2470,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -2482,7 +2482,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -2494,7 +2494,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -2506,7 +2506,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -2521,7 +2521,7 @@ mod tests {
 
                // Check against amount_to_transfer_over_msat.
                // Set minimal HTLC of 200_000_000 msat.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 3,
@@ -2536,7 +2536,7 @@ mod tests {
 
                // Second hop only allows to forward 199_999_999 at most, thus not allowing the first hop to
                // be used.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 3,
@@ -2555,7 +2555,7 @@ mod tests {
                } else { panic!(); }
 
                // Lift the restriction on the first hop.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 4,
@@ -2575,7 +2575,7 @@ mod tests {
 
        #[test]
        fn htlc_minimum_overpay_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let payment_params = PaymentParameters::from_node_id(nodes[2]).with_features(InvoiceFeatures::known());
                let scorer = test_utils::TestScorer::with_penalty(0);
@@ -2585,7 +2585,7 @@ mod tests {
                // A route to node#2 via two paths.
                // One path allows transferring 35-40 sats, another one also allows 35-40 sats.
                // Thus, they can't send 60 without overpaying.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -2597,7 +2597,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 3,
@@ -2611,7 +2611,7 @@ mod tests {
                });
 
                // Make 0 fee.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -2623,7 +2623,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
@@ -2637,7 +2637,7 @@ mod tests {
                });
 
                // Disable other paths
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 3,
@@ -2658,7 +2658,7 @@ mod tests {
 
                // Now, test that if there are 2 paths, a "cheaper" by fee path wouldn't be prioritized
                // while taking even more fee to match htlc_minimum_msat.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 4,
@@ -2670,7 +2670,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 3,
@@ -2682,7 +2682,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 4,
@@ -2713,7 +2713,7 @@ mod tests {
 
        #[test]
        fn disable_channels_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let payment_params = PaymentParameters::from_node_id(nodes[2]);
                let scorer = test_utils::TestScorer::with_penalty(0);
@@ -2721,7 +2721,7 @@ mod tests {
                let random_seed_bytes = keys_manager.get_secure_random_bytes();
 
                // // Disable channels 4 and 12 by flags=2
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
@@ -2733,7 +2733,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -2773,7 +2773,7 @@ mod tests {
 
        #[test]
        fn disable_node_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (_, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let payment_params = PaymentParameters::from_node_id(nodes[2]);
                let scorer = test_utils::TestScorer::with_penalty(0);
@@ -2783,9 +2783,9 @@ mod tests {
                // Disable nodes 1, 2, and 8 by requiring unknown feature bits
                let mut unknown_features = NodeFeatures::known();
                unknown_features.set_unknown_feature_required();
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[0], unknown_features.clone(), 1);
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[1], unknown_features.clone(), 1);
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[7], unknown_features.clone(), 1);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[0], unknown_features.clone(), 1);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[1], unknown_features.clone(), 1);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[7], unknown_features.clone(), 1);
 
                // If all nodes require some features we don't understand, route should fail
                if let Err(LightningError{err, action: ErrorAction::IgnoreError}) = get_route(&our_id, &payment_params, &network_graph.read_only(), None, 100, 42, Arc::clone(&logger), &scorer, &random_seed_bytes) {
@@ -3128,7 +3128,7 @@ mod tests {
 
        #[test]
        fn multi_hint_last_hops_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (_, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let last_hops = multi_hop_last_hops_hint([nodes[2], nodes[3]]);
                let payment_params = PaymentParameters::from_node_id(nodes[6]).with_route_hints(last_hops.clone());
@@ -3139,7 +3139,7 @@ mod tests {
                // Test shows that multiple hop hints are considered.
 
                // Disabling channels 6 & 7 by flags=2
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -3151,7 +3151,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -3198,7 +3198,7 @@ mod tests {
 
        #[test]
        fn private_multi_hint_last_hops_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (_, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
 
                let non_announced_privkey = SecretKey::from_slice(&hex::decode(format!("{:02x}", 0xf0).repeat(32)).unwrap()[..]).unwrap();
@@ -3211,7 +3211,7 @@ mod tests {
                // Test shows that multiple hop hints are considered.
 
                // Disabling channels 6 & 7 by flags=2
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -3223,7 +3223,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -3541,7 +3541,7 @@ mod tests {
        fn available_amount_while_routing_test() {
                // Tests whether we choose the correct available channel amount while routing.
 
-               let (secp_ctx, network_graph, mut net_graph_msg_handler, chain_monitor, logger) = build_graph();
+               let (secp_ctx, network_graph, mut gossip_sync, chain_monitor, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -3552,7 +3552,7 @@ mod tests {
                // our node to node2 via node0: channels {1, 3}.
 
                // First disable all other paths.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -3564,7 +3564,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -3579,7 +3579,7 @@ mod tests {
 
                // Make the first channel (#1) very permissive,
                // and we will be testing all limits on the second channel.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -3594,7 +3594,7 @@ mod tests {
 
                // First, let's see if routing works if we have absolutely no idea about the available amount.
                // In this case, it should be set to 250_000 sats.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -3627,7 +3627,7 @@ mod tests {
 
                // Check that setting next_outbound_htlc_limit_msat in first_hops limits the channels.
                // Disable channel #1 and use another first hop.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 3,
@@ -3662,7 +3662,7 @@ mod tests {
                }
 
                // Enable channel #1 back.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 4,
@@ -3677,7 +3677,7 @@ mod tests {
 
 
                // Now let's see if routing works if we know only htlc_maximum_msat.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 3,
@@ -3712,7 +3712,7 @@ mod tests {
 
                // We can't change UTXO capacity on the fly, so we'll disable
                // the existing channel and add another one with the capacity we need.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 4,
@@ -3732,10 +3732,10 @@ mod tests {
                .push_opcode(opcodes::all::OP_CHECKMULTISIG).into_script().to_v0_p2wsh();
 
                *chain_monitor.utxo_ret.lock().unwrap() = Ok(TxOut { value: 15, script_pubkey: good_script.clone() });
-               net_graph_msg_handler.add_chain_access(Some(chain_monitor));
+               gossip_sync.add_chain_access(Some(chain_monitor));
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 333);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[0], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 333);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 333,
                        timestamp: 1,
@@ -3747,7 +3747,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 333,
                        timestamp: 1,
@@ -3779,7 +3779,7 @@ mod tests {
                }
 
                // Now let's see if routing chooses htlc_maximum_msat over UTXO capacity.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 333,
                        timestamp: 6,
@@ -3815,7 +3815,7 @@ mod tests {
        fn available_liquidity_last_hop_test() {
                // Check that available liquidity properly limits the path even when only
                // one of the latter hops is limited.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -3827,7 +3827,7 @@ mod tests {
                // Total capacity: 50 sats.
 
                // Disable other potential paths.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -3839,7 +3839,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -3854,7 +3854,7 @@ mod tests {
 
                // Limit capacities
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -3866,7 +3866,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -3879,7 +3879,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -3891,7 +3891,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 2,
@@ -3940,7 +3940,7 @@ mod tests {
 
        #[test]
        fn ignore_fee_first_hop_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -3948,7 +3948,7 @@ mod tests {
                let payment_params = PaymentParameters::from_node_id(nodes[2]);
 
                // Path via node0 is channels {1, 3}. Limit them to 100 and 50 sats (total limit 50).
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -3960,7 +3960,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -3988,7 +3988,7 @@ mod tests {
 
        #[test]
        fn simple_mpp_route_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4004,7 +4004,7 @@ mod tests {
                // Their aggregate capacity will be 50 + 60 + 180 = 290 sats.
 
                // Path via node0 is channels {1, 3}. Limit them to 100 and 50 sats (total limit 50).
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4016,7 +4016,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -4031,7 +4031,7 @@ mod tests {
 
                // Path via node7 is channels {12, 13}. Limit them to 60 and 60 sats
                // (total limit 60).
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -4043,7 +4043,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4058,7 +4058,7 @@ mod tests {
 
                // Path via node1 is channels {2, 4}. Limit them to 200 and 180 sats
                // (total capacity 180 sats).
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4070,7 +4070,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
@@ -4121,7 +4121,7 @@ mod tests {
 
        #[test]
        fn long_mpp_route_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4136,7 +4136,7 @@ mod tests {
                // are used twice will have 200 sats capacity.
 
                // Disable other potential paths.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4148,7 +4148,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -4162,7 +4162,7 @@ mod tests {
                });
 
                // Path via {node0, node2} is channels {1, 3, 5}.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4174,7 +4174,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -4188,8 +4188,8 @@ mod tests {
                });
 
                // Capacity of 200 sats because this channel will be used by 3rd path as well.
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 5,
                        timestamp: 2,
@@ -4205,7 +4205,7 @@ mod tests {
                // Path via {node7, node2, node4} is channels {12, 13, 6, 11}.
                // Add 100 sats to the capacities of {12, 13}, because these channels
                // are also used for 3rd path. 100 sats for the rest. Total capacity: 100 sats.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -4217,7 +4217,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4230,7 +4230,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -4242,7 +4242,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 2,
@@ -4285,7 +4285,7 @@ mod tests {
 
        #[test]
        fn mpp_cheaper_route_test() {
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4304,7 +4304,7 @@ mod tests {
                // are used twice will have 200 sats capacity.
 
                // Disable other potential paths.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4316,7 +4316,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -4330,7 +4330,7 @@ mod tests {
                });
 
                // Path via {node0, node2} is channels {1, 3, 5}.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4342,7 +4342,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -4356,8 +4356,8 @@ mod tests {
                });
 
                // Capacity of 200 sats because this channel will be used by 3rd path as well.
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 5,
                        timestamp: 2,
@@ -4373,7 +4373,7 @@ mod tests {
                // Path via {node7, node2, node4} is channels {12, 13, 6, 11}.
                // Add 100 sats to the capacities of {12, 13}, because these channels
                // are also used for 3rd path. 100 sats for the rest. Total capacity: 100 sats.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -4385,7 +4385,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4398,7 +4398,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -4410,7 +4410,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 2,
@@ -4454,7 +4454,7 @@ mod tests {
                // This test makes sure that MPP algorithm properly takes into account
                // fees charged on the channels, by making the fees impactful:
                // if the fee is not properly accounted for, the behavior is different.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4471,7 +4471,7 @@ mod tests {
                // It's fine to ignore this concern for now.
 
                // Disable other potential paths.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4484,7 +4484,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 7,
                        timestamp: 2,
@@ -4498,7 +4498,7 @@ mod tests {
                });
 
                // Path via {node0, node2} is channels {1, 3, 5}.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4510,7 +4510,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -4523,8 +4523,8 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 5,
                        timestamp: 2,
@@ -4547,7 +4547,7 @@ mod tests {
                // - channel 12 capacity is 250 sats
                // - fee for channel 6 is 150 sats
                // Let's test this by enforcing these 2 conditions and removing other limits.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -4559,7 +4559,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4572,7 +4572,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 2,
@@ -4584,7 +4584,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 11,
                        timestamp: 2,
@@ -4635,7 +4635,7 @@ mod tests {
                // to only have the remaining to-collect amount in available liquidity.
                //
                // This bug appeared in production in some specific channel configurations.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4655,7 +4655,7 @@ mod tests {
                // would first use the no-fee route and then fail to find a path along the second route as
                // we think we can only send up to 1 additional sat over the last-hop but refuse to as its
                // under 5% of our payment amount.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4667,7 +4667,7 @@ mod tests {
                        fee_proportional_millionths: u32::max_value(),
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4679,7 +4679,7 @@ mod tests {
                        fee_proportional_millionths: u32::max_value(),
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
@@ -4691,7 +4691,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4725,7 +4725,7 @@ mod tests {
        fn drop_lowest_channel_mpp_route_test() {
                // This test checks that low-capacity channel is dropped when after
                // path finding we realize that we found more capacity than we need.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -4742,7 +4742,7 @@ mod tests {
                // Their aggregate capacity will be 50 + 60 + 20 = 130 sats.
 
                // Path via node0 is channels {1, 3}. Limit them to 100 and 50 sats (total limit 50);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 2,
@@ -4754,7 +4754,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[0], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 2,
@@ -4768,7 +4768,7 @@ mod tests {
                });
 
                // Path via node7 is channels {12, 13}. Limit them to 60 and 60 sats (total limit 60);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -4780,7 +4780,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[7], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 13,
                        timestamp: 2,
@@ -4794,7 +4794,7 @@ mod tests {
                });
 
                // Path via node1 is channels {2, 4}. Limit them to 20 and 20 sats (total capacity 20 sats).
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -4806,7 +4806,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
@@ -4884,15 +4884,15 @@ mod tests {
                let secp_ctx = Secp256k1::new();
                let logger = Arc::new(test_utils::TestLogger::new());
                let network = Arc::new(NetworkGraph::new(genesis_block(Network::Testnet).header.block_hash()));
-               let net_graph_msg_handler = NetGraphMsgHandler::new(Arc::clone(&network), None, Arc::clone(&logger));
+               let gossip_sync = P2PGossipSync::new(Arc::clone(&network), None, Arc::clone(&logger));
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
                let random_seed_bytes = keys_manager.get_secure_random_bytes();
                let payment_params = PaymentParameters::from_node_id(nodes[6]);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, &privkeys[1], ChannelFeatures::from_le_bytes(id_to_feature_flags(6)), 6);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &our_privkey, &privkeys[1], ChannelFeatures::from_le_bytes(id_to_feature_flags(6)), 6);
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 6,
                        timestamp: 1,
@@ -4904,10 +4904,10 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[1], NodeFeatures::from_le_bytes(id_to_feature_flags(1)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[1], NodeFeatures::from_le_bytes(id_to_feature_flags(1)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[1], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(5)), 5);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 5,
                        timestamp: 1,
@@ -4919,10 +4919,10 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[4], NodeFeatures::from_le_bytes(id_to_feature_flags(4)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[4], NodeFeatures::from_le_bytes(id_to_feature_flags(4)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(4)), 4);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[4], &privkeys[3], ChannelFeatures::from_le_bytes(id_to_feature_flags(4)), 4);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 1,
@@ -4934,10 +4934,10 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[3], NodeFeatures::from_le_bytes(id_to_feature_flags(3)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[3], NodeFeatures::from_le_bytes(id_to_feature_flags(3)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[3], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 3);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[3], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[3], &privkeys[2], ChannelFeatures::from_le_bytes(id_to_feature_flags(3)), 3);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[3], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 3,
                        timestamp: 1,
@@ -4949,10 +4949,10 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[2], NodeFeatures::from_le_bytes(id_to_feature_flags(2)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[2], NodeFeatures::from_le_bytes(id_to_feature_flags(2)), 0);
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(2)), 2);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[2], &privkeys[4], ChannelFeatures::from_le_bytes(id_to_feature_flags(2)), 2);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[2], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 1,
@@ -4965,8 +4965,8 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               add_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], &privkeys[6], ChannelFeatures::from_le_bytes(id_to_feature_flags(1)), 1);
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
+               add_channel(&gossip_sync, &secp_ctx, &privkeys[4], &privkeys[6], ChannelFeatures::from_le_bytes(id_to_feature_flags(1)), 1);
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[4], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 1,
                        timestamp: 1,
@@ -4978,7 +4978,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               add_or_update_node(&net_graph_msg_handler, &secp_ctx, &privkeys[6], NodeFeatures::from_le_bytes(id_to_feature_flags(6)), 0);
+               add_or_update_node(&gossip_sync, &secp_ctx, &privkeys[6], NodeFeatures::from_le_bytes(id_to_feature_flags(6)), 0);
 
                {
                        // Now ensure the route flows simply over nodes 1 and 4 to 6.
@@ -5015,7 +5015,7 @@ mod tests {
                // Test that if, while walking the graph, we find a hop that has exactly enough liquidity
                // for us, including later hop fees, we take it. In the first version of our MPP algorithm
                // we calculated fees on a higher value, resulting in us ignoring such paths.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, _, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -5024,7 +5024,7 @@ mod tests {
 
                // We modify the graph to set the htlc_maximum of channel 2 to below the value we wish to
                // send.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -5037,7 +5037,7 @@ mod tests {
                        excess_data: Vec::new()
                });
 
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 12,
                        timestamp: 2,
@@ -5079,7 +5079,7 @@ mod tests {
                // htlc_maximum_msat, we don't end up undershooting a later htlc_minimum_msat. In the
                // initial version of MPP we'd accept such routes but reject them while recalculating fees,
                // resulting in us thinking there is no possible path, even if other paths exist.
-               let (secp_ctx, network_graph, net_graph_msg_handler, _, logger) = build_graph();
+               let (secp_ctx, network_graph, gossip_sync, _, logger) = build_graph();
                let (our_privkey, our_id, privkeys, nodes) = get_nodes(&secp_ctx);
                let scorer = test_utils::TestScorer::with_penalty(0);
                let keys_manager = test_utils::TestKeysInterface::new(&[0u8; 32], Network::Testnet);
@@ -5089,7 +5089,7 @@ mod tests {
                // We modify the graph to set the htlc_minimum of channel 2 and 4 as needed - channel 2
                // gets an htlc_maximum_msat of 80_000 and channel 4 an htlc_minimum_msat of 90_000. We
                // then try to send 90_000.
-               update_channel(&net_graph_msg_handler, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &our_privkey, UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 2,
                        timestamp: 2,
@@ -5101,7 +5101,7 @@ mod tests {
                        fee_proportional_millionths: 0,
                        excess_data: Vec::new()
                });
-               update_channel(&net_graph_msg_handler, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
+               update_channel(&gossip_sync, &secp_ctx, &privkeys[1], UnsignedChannelUpdate {
                        chain_hash: genesis_block(Network::Testnet).header.block_hash(),
                        short_channel_id: 4,
                        timestamp: 2,
index 339979b1b020f708e3f1dc55597275f36254250d..54400624dce63a30ec0e506bfc1c87800a6bac4a 100644 (file)
@@ -17,7 +17,7 @@
 //! ```
 //! # extern crate bitcoin;
 //! #
-//! # use lightning::routing::network_graph::NetworkGraph;
+//! # use lightning::routing::gossip::NetworkGraph;
 //! # use lightning::routing::router::{RouteParameters, find_route};
 //! # use lightning::routing::scoring::{ProbabilisticScorer, ProbabilisticScoringParameters};
 //! # use lightning::chain::keysinterface::{KeysManager, KeysInterface};
@@ -55,7 +55,7 @@
 //! [`find_route`]: crate::routing::router::find_route
 
 use ln::msgs::DecodeError;
-use routing::network_graph::{EffectiveCapacity, NetworkGraph, NodeId};
+use routing::gossip::{EffectiveCapacity, NetworkGraph, NodeId};
 use routing::router::RouteHop;
 use util::ser::{Readable, ReadableArgs, Writeable, Writer};
 use util::logger::Logger;
@@ -1120,9 +1120,9 @@ mod tests {
 
        use ln::features::{ChannelFeatures, NodeFeatures};
        use ln::msgs::{ChannelAnnouncement, ChannelUpdate, OptionalField, UnsignedChannelAnnouncement, UnsignedChannelUpdate};
-       use routing::scoring::{ChannelUsage, Score};
-       use routing::network_graph::{EffectiveCapacity, NetworkGraph, NodeId};
+       use routing::gossip::{EffectiveCapacity, NetworkGraph, NodeId};
        use routing::router::RouteHop;
+       use routing::scoring::{ChannelUsage, Score};
        use util::ser::{ReadableArgs, Writeable};
        use util::test_utils::TestLogger;
 
index 342dbb25545059660a10de797b952c229cf81d3e..a2794c329a26d827c214846d7cd571274e481b7e 100644 (file)
@@ -21,7 +21,7 @@ use ln::features::ChannelTypeFeatures;
 use ln::msgs;
 use ln::msgs::DecodeError;
 use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
-use routing::network_graph::NetworkUpdate;
+use routing::gossip::NetworkUpdate;
 use util::ser::{BigSize, FixedLengthReader, Writeable, Writer, MaybeReadable, Readable, VecReadWrapper, VecWriteWrapper};
 use routing::router::{RouteHop, RouteParameters};
 
@@ -337,10 +337,10 @@ pub enum Event {
                /// payment route.
                ///
                /// Should be applied to the [`NetworkGraph`] so that routing decisions can take into
-               /// account the update. [`NetGraphMsgHandler`] is capable of doing this.
+               /// account the update. [`P2PGossipSync`] is capable of doing this.
                ///
-               /// [`NetworkGraph`]: crate::routing::network_graph::NetworkGraph
-               /// [`NetGraphMsgHandler`]: crate::routing::network_graph::NetGraphMsgHandler
+               /// [`NetworkGraph`]: crate::routing::gossip::NetworkGraph
+               /// [`P2PGossipSync`]: crate::routing::gossip::P2PGossipSync
                network_update: Option<NetworkUpdate>,
                /// For both single-path and multi-path payments, this is set if all paths of the payment have
                /// failed. This will be set to false if (1) this is an MPP payment and (2) other parts of the
index 5c124c21afdb25984d9d43362726f858fb292c8b..117eff491b343d4ae6cdfd0d0af7cd7a1df5ae91 100644 (file)
@@ -13,7 +13,7 @@ use bitcoin::hashes::hex::ToHex;
 use io::{self};
 use routing::scoring::WriteableScore;
 
-use crate::{chain::{keysinterface::{Sign, KeysInterface}, self, transaction::{OutPoint}, chaininterface::{BroadcasterInterface, FeeEstimator}, chainmonitor::{Persist, MonitorUpdateId}, channelmonitor::{ChannelMonitor, ChannelMonitorUpdate}}, ln::channelmanager::ChannelManager, routing::network_graph::NetworkGraph};
+use crate::{chain::{keysinterface::{Sign, KeysInterface}, self, transaction::{OutPoint}, chaininterface::{BroadcasterInterface, FeeEstimator}, chainmonitor::{Persist, MonitorUpdateId}, channelmonitor::{ChannelMonitor, ChannelMonitorUpdate}}, ln::channelmanager::ChannelManager, routing::gossip::NetworkGraph};
 use super::{logger::Logger, ser::Writeable};
 
 /// Trait for a key-value store for persisting some writeable object at some key