impl more message encodings
[rust-lightning] / src / ln / channelmanager.rs
index 50aac54c3a0010a27e5e2be465fe99f35a6677ed..dff6cb19e953ecd6abf93d64c9116b2ebefb2ea3 100644 (file)
@@ -32,29 +32,57 @@ use std::collections::hash_map;
 use std::{ptr, mem};
 use std::time::{Instant,Duration};
 
-/// Stores the info we will need to send when we want to forward an HTLC onwards
-pub struct PendingForwardHTLCInfo {
-       onion_packet: Option<msgs::OnionPacket>,
-       payment_hash: [u8; 32],
-       short_channel_id: u64,
-       prev_short_channel_id: u64,
-       amt_to_forward: u64,
-       outgoing_cltv_value: u32,
-}
+mod channel_held_info {
+       use ln::msgs;
 
-#[cfg(feature = "fuzztarget")]
-impl PendingForwardHTLCInfo {
-       pub fn dummy() -> Self {
-               Self {
-                       onion_packet: None,
-                       payment_hash: [0; 32],
-                       short_channel_id: 0,
-                       prev_short_channel_id: 0,
-                       amt_to_forward: 0,
-                       outgoing_cltv_value: 0,
+       /// Stores the info we will need to send when we want to forward an HTLC onwards
+       pub struct PendingForwardHTLCInfo {
+               pub(super) onion_packet: Option<msgs::OnionPacket>,
+               pub(super) payment_hash: [u8; 32],
+               pub(super) short_channel_id: u64,
+               pub(super) prev_short_channel_id: u64,
+               pub(super) amt_to_forward: u64,
+               pub(super) outgoing_cltv_value: u32,
+       }
+
+       #[cfg(feature = "fuzztarget")]
+       impl PendingForwardHTLCInfo {
+               pub fn dummy() -> Self {
+                       Self {
+                               onion_packet: None,
+                               payment_hash: [0; 32],
+                               short_channel_id: 0,
+                               prev_short_channel_id: 0,
+                               amt_to_forward: 0,
+                               outgoing_cltv_value: 0,
+                       }
+               }
+       }
+
+       #[derive(Clone)] // See Channel::revoke_and_ack for why, tl;dr: Rust bug
+       pub enum HTLCFailReason {
+               ErrorPacket {
+                       err: msgs::OnionErrorPacket,
+               },
+               Reason {
+                       failure_code: u16,
+                       data: Vec<u8>,
+               }
+       }
+
+       #[cfg(feature = "fuzztarget")]
+       impl HTLCFailReason {
+               pub fn dummy() -> Self {
+                       HTLCFailReason::Reason {
+                               failure_code: 0, data: Vec::new(),
+                       }
                }
        }
 }
+#[cfg(feature = "fuzztarget")]
+pub use self::channel_held_info::*;
+#[cfg(not(feature = "fuzztarget"))]
+pub(crate) use self::channel_held_info::*;
 
 enum PendingOutboundHTLC {
        IntermediaryHopData {
@@ -72,16 +100,6 @@ enum PendingOutboundHTLC {
        }
 }
 
-enum HTLCFailReason<'a> {
-       ErrorPacket {
-               err: &'a msgs::OnionErrorPacket,
-       },
-       Reason {
-               failure_code: u16,
-               data: &'a[u8],
-       }
-}
-
 /// We hold back HTLCs we intend to relay for a random interval in the range (this, 5*this). This
 /// provides some limited amount of privacy. Ideally this would range from somewhere like 1 second
 /// to 30 seconds, but people expect lightning to be, you know, kinda fast, sadly. We could
@@ -257,7 +275,7 @@ impl ChannelManager {
                };
                for payment_hash in res.1 {
                        // unknown_next_peer...I dunno who that is anymore....
-                       self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 10, data: &[0; 0] });
+                       self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 10, data: Vec::new() });
                }
                Ok(res.0)
        }
@@ -706,8 +724,8 @@ impl ChannelManager {
 
                for failed_forward in failed_forwards.drain(..) {
                        match failed_forward.2 {
-                               None => self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &failed_forward.0, HTLCFailReason::Reason { failure_code: failed_forward.1, data: &[0;0] }),
-                               Some(chan_update) => self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &failed_forward.0, HTLCFailReason::Reason { failure_code: failed_forward.1, data: &chan_update.encode()[..] }),
+                               None => self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &failed_forward.0, HTLCFailReason::Reason { failure_code: failed_forward.1, data: Vec::new() }),
+                               Some(chan_update) => self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &failed_forward.0, HTLCFailReason::Reason { failure_code: failed_forward.1, data: chan_update.encode() }),
                        };
                }
 
@@ -722,7 +740,7 @@ impl ChannelManager {
 
        /// Indicates that the preimage for payment_hash is unknown after a PaymentReceived event.
        pub fn fail_htlc_backwards(&self, payment_hash: &[u8; 32]) -> bool {
-               self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 15, data: &[0;0] })
+               self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 15, data: Vec::new() })
        }
 
        fn fail_htlc_backwards_internal(&self, mut channel_state: MutexGuard<ChannelHolder>, payment_hash: &[u8; 32], onion_error: HTLCFailReason) -> bool {
@@ -754,7 +772,7 @@ impl ChannelManager {
                        PendingOutboundHTLC::IntermediaryHopData { source_short_channel_id, incoming_packet_shared_secret } => {
                                let err_packet = match onion_error {
                                        HTLCFailReason::Reason { failure_code, data } => {
-                                               let packet = ChannelManager::build_failure_packet(&incoming_packet_shared_secret, failure_code, data).encode();
+                                               let packet = ChannelManager::build_failure_packet(&incoming_packet_shared_secret, failure_code, &data[..]).encode();
                                                ChannelManager::encrypt_failure_packet(&incoming_packet_shared_secret, &packet)
                                        },
                                        HTLCFailReason::ErrorPacket { err } => {
@@ -762,14 +780,14 @@ impl ChannelManager {
                                        }
                                };
 
-                               let (node_id, fail_msg) = {
+                               let (node_id, fail_msgs) = {
                                        let chan_id = match channel_state.short_to_id.get(&source_short_channel_id) {
                                                Some(chan_id) => chan_id.clone(),
                                                None => return false
                                        };
 
                                        let chan = channel_state.by_id.get_mut(&chan_id).unwrap();
-                                       match chan.get_update_fail_htlc(payment_hash, err_packet) {
+                                       match chan.get_update_fail_htlc_and_commit(payment_hash, err_packet) {
                                                Ok(msg) => (chan.get_their_node_id(), msg),
                                                Err(_e) => {
                                                        //TODO: Do something with e?
@@ -778,12 +796,18 @@ impl ChannelManager {
                                        }
                                };
 
-                               mem::drop(channel_state);
-                               let mut pending_events = self.pending_events.lock().unwrap();
-                               pending_events.push(events::Event::SendFailHTLC {
-                                       node_id,
-                                       msg: fail_msg
-                               });
+                               match fail_msgs {
+                                       Some(msgs) => {
+                                               mem::drop(channel_state);
+                                               let mut pending_events = self.pending_events.lock().unwrap();
+                                               pending_events.push(events::Event::SendFailHTLC {
+                                                       node_id,
+                                                       msg: msgs.0,
+                                                       commitment_msg: msgs.1,
+                                               });
+                                       },
+                                       None => {},
+                               }
 
                                true
                        },
@@ -793,6 +817,7 @@ impl ChannelManager {
        /// Provides a payment preimage in response to a PaymentReceived event, returning true and
        /// generating message events for the net layer to claim the payment, if possible. Thus, you
        /// should probably kick the net layer to go send messages if this returns true!
+       /// May panic if called except in response to a PaymentReceived event.
        pub fn claim_funds(&self, payment_preimage: [u8; 32]) -> bool {
                self.claim_funds_internal(payment_preimage, true)
        }
@@ -838,14 +863,14 @@ impl ChannelManager {
                                false
                        },
                        PendingOutboundHTLC::IntermediaryHopData { source_short_channel_id, .. } => {
-                               let (node_id, fulfill_msg, monitor) = {
+                               let (node_id, fulfill_msgs, monitor) = {
                                        let chan_id = match channel_state.short_to_id.get(&source_short_channel_id) {
                                                Some(chan_id) => chan_id.clone(),
                                                None => return false
                                        };
 
                                        let chan = channel_state.by_id.get_mut(&chan_id).unwrap();
-                                       match chan.get_update_fulfill_htlc(payment_preimage) {
+                                       match chan.get_update_fulfill_htlc_and_commit(payment_preimage) {
                                                Ok(msg) => (chan.get_their_node_id(), msg, if from_user { Some(chan.channel_monitor()) } else { None }),
                                                Err(_e) => {
                                                        //TODO: Do something with e?
@@ -854,13 +879,17 @@ impl ChannelManager {
                                        }
                                };
 
-                               {
-                                       mem::drop(channel_state);
-                                       let mut pending_events = self.pending_events.lock().unwrap();
-                                       pending_events.push(events::Event::SendFulfillHTLC {
-                                               node_id: node_id,
-                                               msg: fulfill_msg
-                                       });
+                               mem::drop(channel_state);
+                               match fulfill_msgs {
+                                       Some(msgs) => {
+                                               let mut pending_events = self.pending_events.lock().unwrap();
+                                               pending_events.push(events::Event::SendFulfillHTLC {
+                                                       node_id: node_id,
+                                                       msg: msgs.0,
+                                                       commitment_msg: msgs.1,
+                                               });
+                                       },
+                                       None => {},
                                }
 
                                //TODO: It may not be possible to handle add_update_monitor fails gracefully, maybe
@@ -1066,7 +1095,7 @@ impl ChannelMessageHandler for ChannelManager {
                };
                for payment_hash in res.2 {
                        // unknown_next_peer...I dunno who that is anymore....
-                       self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 10, data: &[0; 0] });
+                       self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &payment_hash, HTLCFailReason::Reason { failure_code: 0x4000 | 10, data: Vec::new() });
                }
                Ok((res.0, res.1))
        }
@@ -1328,6 +1357,11 @@ impl ChannelMessageHandler for ChannelManager {
        }
 
        fn handle_update_fulfill_htlc(&self, their_node_id: &PublicKey, msg: &msgs::UpdateFulfillHTLC) -> Result<(), HandleError> {
+               //TODO: Delay the claimed_funds relaying just like we do outbound relay!
+               // Claim funds first, cause we don't really care if the channel we received the message on
+               // is broken, we may have enough info to get our own money!
+               self.claim_funds_internal(msg.payment_preimage.clone(), false);
+
                let monitor = {
                        let mut channel_state = self.channel_state.lock().unwrap();
                        match channel_state.by_id.get_mut(&msg.channel_id) {
@@ -1341,48 +1375,40 @@ impl ChannelMessageHandler for ChannelManager {
                                None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
                        }
                };
-               //TODO: Delay the claimed_funds relaying just like we do outbound relay!
-               self.claim_funds_internal(msg.payment_preimage.clone(), false);
                self.monitor.add_update_monitor(monitor.get_funding_txo().unwrap(), monitor)?;
                Ok(())
        }
 
        fn handle_update_fail_htlc(&self, their_node_id: &PublicKey, msg: &msgs::UpdateFailHTLC) -> Result<(), HandleError> {
                let mut channel_state = self.channel_state.lock().unwrap();
-               let payment_hash = match channel_state.by_id.get_mut(&msg.channel_id) {
+               match channel_state.by_id.get_mut(&msg.channel_id) {
                        Some(chan) => {
                                if chan.get_their_node_id() != *their_node_id {
                                        return Err(HandleError{err: "Got a message for a channel from the wrong node!", msg: None})
                                }
-                               chan.update_fail_htlc(&msg)?
+                               chan.update_fail_htlc(&msg, HTLCFailReason::ErrorPacket { err: msg.reason.clone() })
                        },
                        None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
-               };
-               self.fail_htlc_backwards_internal(channel_state, &payment_hash, HTLCFailReason::ErrorPacket { err: &msg.reason });
-               Ok(())
+               }
        }
 
        fn handle_update_fail_malformed_htlc(&self, their_node_id: &PublicKey, msg: &msgs::UpdateFailMalformedHTLC) -> Result<(), HandleError> {
                let mut channel_state = self.channel_state.lock().unwrap();
-               let payment_hash = match channel_state.by_id.get_mut(&msg.channel_id) {
+               match channel_state.by_id.get_mut(&msg.channel_id) {
                        Some(chan) => {
                                if chan.get_their_node_id() != *their_node_id {
                                        return Err(HandleError{err: "Got a message for a channel from the wrong node!", msg: None})
                                }
-                               chan.update_fail_malformed_htlc(&msg)?
+                               chan.update_fail_malformed_htlc(&msg, HTLCFailReason::Reason { failure_code: msg.failure_code, data: Vec::new() })
                        },
                        None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
-               };
-               self.fail_htlc_backwards_internal(channel_state, &payment_hash, HTLCFailReason::Reason { failure_code: msg.failure_code, data: &[0;0] });
-               Ok(())
+               }
        }
 
-       fn handle_commitment_signed(&self, their_node_id: &PublicKey, msg: &msgs::CommitmentSigned) -> Result<msgs::RevokeAndACK, HandleError> {
-               let mut forward_event = None;
+       fn handle_commitment_signed(&self, their_node_id: &PublicKey, msg: &msgs::CommitmentSigned) -> Result<(msgs::RevokeAndACK, Option<msgs::CommitmentSigned>), HandleError> {
                let (res, monitor) = {
                        let mut channel_state = self.channel_state.lock().unwrap();
-
-                       let ((res, mut forwarding_infos), monitor) = match channel_state.by_id.get_mut(&msg.channel_id) {
+                       match channel_state.by_id.get_mut(&msg.channel_id) {
                                Some(chan) => {
                                        if chan.get_their_node_id() != *their_node_id {
                                                return Err(HandleError{err: "Got a message for a channel from the wrong node!", msg: None})
@@ -1390,13 +1416,40 @@ impl ChannelMessageHandler for ChannelManager {
                                        (chan.commitment_signed(&msg)?, chan.channel_monitor())
                                },
                                None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
-                       };
+                       }
+               };
+               //TODO: Only if we store HTLC sigs
+               self.monitor.add_update_monitor(monitor.get_funding_txo().unwrap(), monitor)?;
+
+               Ok(res)
+       }
 
+       fn handle_revoke_and_ack(&self, their_node_id: &PublicKey, msg: &msgs::RevokeAndACK) -> Result<Option<msgs::CommitmentUpdate>, HandleError> {
+               let ((res, mut pending_forwards, mut pending_failures), monitor) = {
+                       let mut channel_state = self.channel_state.lock().unwrap();
+                       match channel_state.by_id.get_mut(&msg.channel_id) {
+                               Some(chan) => {
+                                       if chan.get_their_node_id() != *their_node_id {
+                                               return Err(HandleError{err: "Got a message for a channel from the wrong node!", msg: None})
+                                       }
+                                       (chan.revoke_and_ack(&msg)?, chan.channel_monitor())
+                               },
+                               None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
+                       }
+               };
+               self.monitor.add_update_monitor(monitor.get_funding_txo().unwrap(), monitor)?;
+               for failure in pending_failures.drain(..) {
+                       self.fail_htlc_backwards_internal(self.channel_state.lock().unwrap(), &failure.0, failure.1);
+               }
+
+               let mut forward_event = None;
+               if !pending_forwards.is_empty() {
+                       let mut channel_state = self.channel_state.lock().unwrap();
                        if channel_state.forward_htlcs.is_empty() {
                                forward_event = Some(Instant::now() + Duration::from_millis(((rng::rand_f32() * 4.0 + 1.0) * MIN_HTLC_RELAY_HOLDING_CELL_MILLIS as f32) as u64));
                                channel_state.next_forward = forward_event.unwrap();
                        }
-                       for forward_info in forwarding_infos.drain(..) {
+                       for forward_info in pending_forwards.drain(..) {
                                match channel_state.forward_htlcs.entry(forward_info.short_channel_id) {
                                        hash_map::Entry::Occupied(mut entry) => {
                                                entry.get_mut().push(forward_info);
@@ -1406,12 +1459,7 @@ impl ChannelMessageHandler for ChannelManager {
                                        }
                                }
                        }
-
-                       (res, monitor)
-               };
-               //TODO: Only if we store HTLC sigs
-               self.monitor.add_update_monitor(monitor.get_funding_txo().unwrap(), monitor)?;
-
+               }
                match forward_event {
                        Some(time) => {
                                let mut pending_events = self.pending_events.lock().unwrap();
@@ -1425,23 +1473,6 @@ impl ChannelMessageHandler for ChannelManager {
                Ok(res)
        }
 
-       fn handle_revoke_and_ack(&self, their_node_id: &PublicKey, msg: &msgs::RevokeAndACK) -> Result<Option<(Vec<msgs::UpdateAddHTLC>, msgs::CommitmentSigned)>, HandleError> {
-               let (res, monitor) = {
-                       let mut channel_state = self.channel_state.lock().unwrap();
-                       match channel_state.by_id.get_mut(&msg.channel_id) {
-                               Some(chan) => {
-                                       if chan.get_their_node_id() != *their_node_id {
-                                               return Err(HandleError{err: "Got a message for a channel from the wrong node!", msg: None})
-                                       }
-                                       (chan.revoke_and_ack(&msg)?, chan.channel_monitor())
-                               },
-                               None => return Err(HandleError{err: "Failed to find corresponding channel", msg: None})
-                       }
-               };
-               self.monitor.add_update_monitor(monitor.get_funding_txo().unwrap(), monitor)?;
-               Ok(res)
-       }
-
        fn handle_update_fee(&self, their_node_id: &PublicKey, msg: &msgs::UpdateFee) -> Result<(), HandleError> {
                let mut channel_state = self.channel_state.lock().unwrap();
                match channel_state.by_id.get_mut(&msg.channel_id) {
@@ -1552,8 +1583,9 @@ mod tests {
 
        use rand::{thread_rng,Rng};
 
-       use std::sync::{Arc, Mutex};
+       use std::collections::HashMap;
        use std::default::Default;
+       use std::sync::{Arc, Mutex};
        use std::time::Instant;
 
        fn build_test_onion_keys() -> Vec<OnionKeys> {
@@ -1919,9 +1951,17 @@ mod tests {
                                assert_eq!(added_monitors.len(), 1);
                                added_monitors.clear();
                        }
-                       assert!(prev_node.0.handle_revoke_and_ack(&node.get_our_node_id(), &revoke_and_ack).unwrap().is_none());
+                       assert!(prev_node.0.handle_revoke_and_ack(&node.get_our_node_id(), &revoke_and_ack.0).unwrap().is_none());
+                       let prev_revoke_and_ack = prev_node.0.handle_commitment_signed(&node.get_our_node_id(), &revoke_and_ack.1.unwrap()).unwrap();
                        {
                                let mut added_monitors = prev_node.1.added_monitors.lock().unwrap();
+                               assert_eq!(added_monitors.len(), 2);
+                               added_monitors.clear();
+                       }
+                       assert!(node.handle_revoke_and_ack(&prev_node.0.get_our_node_id(), &prev_revoke_and_ack.0).unwrap().is_none());
+                       assert!(prev_revoke_and_ack.1.is_none());
+                       {
+                               let mut added_monitors = monitor.added_monitors.lock().unwrap();
                                assert_eq!(added_monitors.len(), 1);
                                added_monitors.clear();
                        }
@@ -1967,42 +2007,58 @@ mod tests {
                        added_monitors.clear();
                }
 
-               let mut expected_next_node = expected_route.last().unwrap().0.get_our_node_id();
-               let mut prev_node = expected_route.last().unwrap().0;
-               let mut next_msg = None;
-               for &(node, monitor) in expected_route.iter().rev() {
-                       assert_eq!(expected_next_node, node.get_our_node_id());
-                       match next_msg {
-                               Some(msg) => {
-                                       node.handle_update_fulfill_htlc(&prev_node.get_our_node_id(), &msg).unwrap();
+               let mut next_msgs: Option<(msgs::UpdateFulfillHTLC, msgs::CommitmentSigned)> = None;
+               macro_rules! update_fulfill_dance {
+                       ($node: expr, $monitor: expr, $prev_node: expr, $prev_monitor: expr) => {
+                               {
+                                       $node.handle_update_fulfill_htlc(&$prev_node.get_our_node_id(), &next_msgs.as_ref().unwrap().0).unwrap();
+                                       let revoke_and_commit = $node.handle_commitment_signed(&$prev_node.get_our_node_id(), &next_msgs.as_ref().unwrap().1).unwrap();
+                                       {
+                                               let mut added_monitors = $monitor.added_monitors.lock().unwrap();
+                                               assert_eq!(added_monitors.len(), 2);
+                                               added_monitors.clear();
+                                       }
+                                       assert!($prev_node.handle_revoke_and_ack(&$node.get_our_node_id(), &revoke_and_commit.0).unwrap().is_none());
+                                       let revoke_and_ack = $prev_node.handle_commitment_signed(&$node.get_our_node_id(), &revoke_and_commit.1.unwrap()).unwrap();
+                                       assert!(revoke_and_ack.1.is_none());
                                        {
-                                               let mut added_monitors = monitor.added_monitors.lock().unwrap();
+                                               let mut added_monitors = $prev_monitor.added_monitors.lock().unwrap();
+                                               assert_eq!(added_monitors.len(), 2);
+                                               added_monitors.clear();
+                                       }
+                                       assert!($node.handle_revoke_and_ack(&$prev_node.get_our_node_id(), &revoke_and_ack.0).unwrap().is_none());
+                                       {
+                                               let mut added_monitors = $monitor.added_monitors.lock().unwrap();
                                                assert_eq!(added_monitors.len(), 1);
                                                added_monitors.clear();
                                        }
-                               }, None => {}
+                               }
+                       }
+               }
+
+               let mut expected_next_node = expected_route.last().unwrap().0.get_our_node_id();
+               let mut prev_node = (expected_route.last().unwrap().0, expected_route.last().unwrap().1);
+               for &(node, monitor) in expected_route.iter().rev() {
+                       assert_eq!(expected_next_node, node.get_our_node_id());
+                       if next_msgs.is_some() {
+                               update_fulfill_dance!(node, monitor, prev_node.0, prev_node.1);
                        }
 
                        let events = node.get_and_clear_pending_events();
                        assert_eq!(events.len(), 1);
                        match events[0] {
-                               Event::SendFulfillHTLC { ref node_id, ref msg } => {
+                               Event::SendFulfillHTLC { ref node_id, ref msg, ref commitment_msg } => {
                                        expected_next_node = node_id.clone();
-                                       next_msg = Some(msg.clone());
+                                       next_msgs = Some((msg.clone(), commitment_msg.clone()));
                                },
                                _ => panic!("Unexpected event"),
                        };
 
-                       prev_node = node;
+                       prev_node = (node, monitor);
                }
 
                assert_eq!(expected_next_node, origin_node.get_our_node_id());
-               origin_node.handle_update_fulfill_htlc(&expected_route.first().unwrap().0.get_our_node_id(), &next_msg.unwrap()).unwrap();
-               {
-                       let mut added_monitors = origin_monitor.added_monitors.lock().unwrap();
-                       assert_eq!(added_monitors.len(), 1);
-                       added_monitors.clear();
-               }
+               update_fulfill_dance!(origin_node, origin_monitor, expected_route.first().unwrap().0, expected_route.first().unwrap().1);
 
                let events = origin_node.get_and_clear_pending_events();
                assert_eq!(events.len(), 1);
@@ -2060,32 +2116,58 @@ mod tests {
 
                assert!(expected_route.last().unwrap().0.fail_htlc_backwards(&our_payment_hash));
 
+               let mut next_msgs: Option<(msgs::UpdateFailHTLC, msgs::CommitmentSigned)> = None;
+               macro_rules! update_fail_dance {
+                       ($node: expr, $monitor: expr, $prev_node: expr, $prev_monitor: expr) => {
+                               {
+                                       $node.handle_update_fail_htlc(&$prev_node.get_our_node_id(), &next_msgs.as_ref().unwrap().0).unwrap();
+                                       let revoke_and_commit = $node.handle_commitment_signed(&$prev_node.get_our_node_id(), &next_msgs.as_ref().unwrap().1).unwrap();
+                                       {
+                                               let mut added_monitors = $monitor.added_monitors.lock().unwrap();
+                                               assert_eq!(added_monitors.len(), 1);
+                                               added_monitors.clear();
+                                       }
+                                       assert!($prev_node.handle_revoke_and_ack(&$node.get_our_node_id(), &revoke_and_commit.0).unwrap().is_none());
+                                       let revoke_and_ack = $prev_node.handle_commitment_signed(&$node.get_our_node_id(), &revoke_and_commit.1.unwrap()).unwrap();
+                                       assert!(revoke_and_ack.1.is_none());
+                                       {
+                                               let mut added_monitors = $prev_monitor.added_monitors.lock().unwrap();
+                                               assert_eq!(added_monitors.len(), 2);
+                                               added_monitors.clear();
+                                       }
+                                       assert!($node.handle_revoke_and_ack(&$prev_node.get_our_node_id(), &revoke_and_ack.0).unwrap().is_none());
+                                       {
+                                               let mut added_monitors = $monitor.added_monitors.lock().unwrap();
+                                               assert_eq!(added_monitors.len(), 1);
+                                               added_monitors.clear();
+                                       }
+                               }
+                       }
+               }
+
                let mut expected_next_node = expected_route.last().unwrap().0.get_our_node_id();
-               let mut prev_node = expected_route.last().unwrap().0;
-               let mut next_msg = None;
-               for &(node, _) in expected_route.iter().rev() {
+               let mut prev_node = (expected_route.last().unwrap().0, expected_route.last().unwrap().1);
+               for &(node, monitor) in expected_route.iter().rev() {
                        assert_eq!(expected_next_node, node.get_our_node_id());
-                       match next_msg {
-                               Some(msg) => {
-                                       node.handle_update_fail_htlc(&prev_node.get_our_node_id(), &msg).unwrap();
-                               }, None => {}
+                       if next_msgs.is_some() {
+                               update_fail_dance!(node, monitor, prev_node.0, prev_node.1);
                        }
 
                        let events = node.get_and_clear_pending_events();
                        assert_eq!(events.len(), 1);
                        match events[0] {
-                               Event::SendFailHTLC { ref node_id, ref msg } => {
+                               Event::SendFailHTLC { ref node_id, ref msg, ref commitment_msg } => {
                                        expected_next_node = node_id.clone();
-                                       next_msg = Some(msg.clone());
+                                       next_msgs = Some((msg.clone(), commitment_msg.clone()));
                                },
                                _ => panic!("Unexpected event"),
                        };
 
-                       prev_node = node;
+                       prev_node = (node, monitor);
                }
 
                assert_eq!(expected_next_node, origin_node.get_our_node_id());
-               origin_node.handle_update_fail_htlc(&expected_route.first().unwrap().0.get_our_node_id(), &next_msg.unwrap()).unwrap();
+               update_fail_dance!(origin_node, origin_monitor, expected_route.first().unwrap().0, expected_route.first().unwrap().1);
 
                let events = origin_node.get_and_clear_pending_events();
                assert_eq!(events.len(), 1);