use bitcoin::blockdata::script::Script;
use bitcoin::hash_types::{Txid, BlockHash};
-use ln::features::{ChannelFeatures, InitFeatures, NodeFeatures};
+use ln::features::{ChannelFeatures, ChannelTypeFeatures, InitFeatures, NodeFeatures};
-use std::{cmp, fmt};
-use std::fmt::Debug;
-use std::io::Read;
+use prelude::*;
+use core::{cmp, fmt};
+use core::fmt::Debug;
+use io::{self, Read};
+use io_extras::read_to_end;
use util::events::MessageSendEventsProvider;
+use util::logger;
use util::ser::{Readable, Writeable, Writer, FixedLengthReader, HighZeroBytesDroppedVarInt};
-use ln::channelmanager::{PaymentPreimage, PaymentHash, PaymentSecret};
+use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
/// 21 million * 10^8 * 1000
pub(crate) const MAX_VALUE_MSAT: u64 = 21_000_000_0000_0000_000;
/// An error in decoding a message or struct.
-#[derive(Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub enum DecodeError {
/// A version byte specified something we don't know how to handle.
/// Includes unknown realm byte in an OnionHopData packet
BadLengthDescriptor,
/// Error from std::io
Io(/// (C-not exported) as ErrorKind doesn't have a reasonable mapping
- ::std::io::ErrorKind),
+ io::ErrorKind),
+ /// The message included zlib-compressed values, which we don't support.
+ UnsupportedCompression,
}
/// An init message to be sent or received from a peer
pub channel_flags: u8,
/// Optionally, a request to pre-set the to-sender output's scriptPubkey for when we collaboratively close
pub shutdown_scriptpubkey: OptionalField<Script>,
+ /// The channel type that this channel will represent. If none is set, we derive the channel
+ /// type from the intersection of our feature bits with our counterparty's feature bits from
+ /// the Init message.
+ pub channel_type: Option<ChannelTypeFeatures>,
}
/// An accept_channel message to be sent or received from a peer
pub funding_txid: Txid,
/// The specific output index funding this channel
pub funding_output_index: u16,
- /// The signature of the channel initiator (funder) on the funding transaction
+ /// The signature of the channel initiator (funder) on the initial commitment transaction
pub signature: Signature,
}
pub struct FundingSigned {
/// The channel ID
pub channel_id: [u8; 32],
- /// The signature of the channel acceptor (fundee) on the funding transaction
+ /// The signature of the channel acceptor (fundee) on the initial commitment transaction
pub signature: Signature,
}
pub scriptpubkey: Script,
}
+/// The minimum and maximum fees which the sender is willing to place on the closing transaction.
+/// This is provided in [`ClosingSigned`] by both sides to indicate the fee range they are willing
+/// to use.
+#[derive(Clone, Debug, PartialEq)]
+pub struct ClosingSignedFeeRange {
+ /// The minimum absolute fee, in satoshis, which the sender is willing to place on the closing
+ /// transaction.
+ pub min_fee_satoshis: u64,
+ /// The maximum absolute fee, in satoshis, which the sender is willing to place on the closing
+ /// transaction.
+ pub max_fee_satoshis: u64,
+}
+
/// A closing_signed message to be sent or received from a peer
#[derive(Clone, Debug, PartialEq)]
pub struct ClosingSigned {
pub fee_satoshis: u64,
/// A signature on the closing transaction
pub signature: Signature,
+ /// The minimum and maximum fees which the sender is willing to accept, provided only by new
+ /// nodes.
+ pub fee_range: Option<ClosingSignedFeeRange>,
}
/// An update_add_htlc message to be sent or received from a peer
},
}
impl NetAddress {
+ /// Gets the ID of this address type. Addresses in node_announcement messages should be sorted
+ /// by this.
+ pub(crate) fn get_id(&self) -> u8 {
+ match self {
+ &NetAddress::IPv4 {..} => { 1 },
+ &NetAddress::IPv6 {..} => { 2 },
+ &NetAddress::OnionV2 {..} => { 3 },
+ &NetAddress::OnionV3 {..} => { 4 },
+ }
+ }
+
/// Strict byte-length of address descriptor, 1-byte type not recorded
fn len(&self) -> u16 {
match self {
}
impl Writeable for NetAddress {
- fn write<W: Writer>(&self, writer: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
match self {
&NetAddress::IPv4 { ref addr, ref port } => {
1u8.write(writer)?;
}
}
+impl Readable for NetAddress {
+ fn read<R: Read>(reader: &mut R) -> Result<NetAddress, DecodeError> {
+ match Readable::read(reader) {
+ Ok(Ok(res)) => Ok(res),
+ Ok(Err(_)) => Err(DecodeError::UnknownVersion),
+ Err(e) => Err(e),
+ }
+ }
+}
+
+
/// The unsigned part of a node_announcement
#[derive(Clone, Debug, PartialEq)]
pub struct UnsignedNodeAnnouncement {
pub timestamp: u32,
/// Channel flags
pub flags: u8,
- /// The number of blocks to subtract from incoming HTLC cltv_expiry values
+ /// The number of blocks such that if:
+ /// `incoming_htlc.cltv_expiry < outgoing_htlc.cltv_expiry + cltv_expiry_delta`
+ /// then we need to fail the HTLC backwards. When forwarding an HTLC, cltv_expiry_delta determines
+ /// the outgoing HTLC's minimum cltv_expiry value -- so, if an incoming HTLC comes in with a
+ /// cltv_expiry of 100000, and the node we're forwarding to has a cltv_expiry_delta value of 10,
+ /// then we'll check that the outgoing HTLC's cltv_expiry value is at least 100010 before
+ /// forwarding. Note that the HTLC sender is the one who originally sets this value when
+ /// constructing the route.
pub cltv_expiry_delta: u16,
/// The minimum HTLC size incoming to sender, in milli-satoshi
pub htlc_minimum_msat: u64,
msg: Option<ErrorMessage>
},
/// The peer did something harmless that we weren't able to process, just log and ignore
+ // New code should *not* use this. New code must use IgnoreAndLog, below!
IgnoreError,
+ /// The peer did something harmless that we weren't able to meaningfully process.
+ /// If the error is logged, log it at the given level.
+ IgnoreAndLog(logger::Level),
/// The peer did something incorrect. Tell them.
SendErrorMessage {
/// The message to send.
pub commitment_signed: CommitmentSigned,
}
-/// The information we received from a peer along the route of a payment we originated. This is
-/// returned by ChannelMessageHandler::handle_update_fail_htlc to be passed into
-/// RoutingMessageHandler::handle_htlc_fail_channel_update to update our network map.
-#[derive(Clone, Debug, PartialEq)]
-pub enum HTLCFailChannelUpdate {
- /// We received an error which included a full ChannelUpdate message.
- ChannelUpdateMessage {
- /// The unwrapped message we received
- msg: ChannelUpdate,
- },
- /// We received an error which indicated only that a channel has been closed
- ChannelClosed {
- /// The short_channel_id which has now closed.
- short_channel_id: u64,
- /// when this true, this channel should be permanently removed from the
- /// consideration. Otherwise, this channel can be restored as new channel_update is received
- is_permanent: bool,
- },
- /// We received an error which indicated only that a node has failed
- NodeFailure {
- /// The node_id that has failed.
- node_id: PublicKey,
- /// when this true, node should be permanently removed from the
- /// consideration. Otherwise, the channels connected to this node can be
- /// restored as new channel_update is received
- is_permanent: bool,
- }
-}
-
/// Messages could have optional fields to use with extended features
/// As we wish to serialize these differently from Option<T>s (Options get a tag byte, but
/// OptionalFeild simply gets Present if there are enough bytes to read into it), we have a
///
/// Messages MAY be called in parallel when they originate from different their_node_ids, however
/// they MUST NOT be called in parallel when the two calls have the same their_node_id.
-pub trait ChannelMessageHandler : MessageSendEventsProvider + Send + Sync {
+pub trait ChannelMessageHandler : MessageSendEventsProvider {
//Channel init:
/// Handle an incoming open_channel message from the given peer.
fn handle_open_channel(&self, their_node_id: &PublicKey, their_features: InitFeatures, msg: &OpenChannel);
// Channl close:
/// Handle an incoming shutdown message from the given peer.
- fn handle_shutdown(&self, their_node_id: &PublicKey, msg: &Shutdown);
+ fn handle_shutdown(&self, their_node_id: &PublicKey, their_features: &InitFeatures, msg: &Shutdown);
/// Handle an incoming closing_signed message from the given peer.
fn handle_closing_signed(&self, their_node_id: &PublicKey, msg: &ClosingSigned);
/// Handle an incoming channel_reestablish message from the given peer.
fn handle_channel_reestablish(&self, their_node_id: &PublicKey, msg: &ChannelReestablish);
+ /// Handle an incoming channel update from the given peer.
+ fn handle_channel_update(&self, their_node_id: &PublicKey, msg: &ChannelUpdate);
+
// Error:
/// Handle an incoming error message from the given peer.
fn handle_error(&self, their_node_id: &PublicKey, msg: &ErrorMessage);
/// For `gossip_queries` messages there are potential DoS vectors when handling
/// inbound queries. Implementors using an on-disk network graph should be aware of
/// repeated disk I/O for queries accessing different parts of the network graph.
-pub trait RoutingMessageHandler : Send + Sync + MessageSendEventsProvider {
+pub trait RoutingMessageHandler : MessageSendEventsProvider {
/// Handle an incoming node_announcement message, returning true if it should be forwarded on,
/// false or returning an Err otherwise.
fn handle_node_announcement(&self, msg: &NodeAnnouncement) -> Result<bool, LightningError>;
/// Handle an incoming channel_update message, returning true if it should be forwarded on,
/// false or returning an Err otherwise.
fn handle_channel_update(&self, msg: &ChannelUpdate) -> Result<bool, LightningError>;
- /// Handle some updates to the route graph that we learned due to an outbound failed payment.
- fn handle_htlc_fail_channel_update(&self, update: &HTLCFailChannelUpdate);
/// Gets a subset of the channel announcements and updates required to dump our routing table
/// to a remote node, starting at the short_channel_id indicated by starting_point and
/// including the batch_amount entries immediately higher in numerical value than starting_point.
}
mod fuzzy_internal_msgs {
- use ln::channelmanager::PaymentSecret;
+ use prelude::*;
+ use ln::{PaymentPreimage, PaymentSecret};
// These types aren't intended to be pub, but are exposed for direct fuzzing (as we deserialize
// them from untrusted input):
},
FinalNode {
payment_data: Option<FinalOnionHopData>,
+ keysend_preimage: Option<PaymentPreimage>,
},
}
DecodeError::ShortRead => f.write_str("Packet extended beyond the provided bytes"),
DecodeError::BadLengthDescriptor => f.write_str("A length descriptor in the packet didn't describe the later data correctly"),
DecodeError::Io(ref e) => e.fmt(f),
+ DecodeError::UnsupportedCompression => f.write_str("We don't support receiving messages with zlib-compressed fields"),
}
}
}
-impl From<::std::io::Error> for DecodeError {
- fn from(e: ::std::io::Error) -> Self {
- if e.kind() == ::std::io::ErrorKind::UnexpectedEof {
+impl From<io::Error> for DecodeError {
+ fn from(e: io::Error) -> Self {
+ if e.kind() == io::ErrorKind::UnexpectedEof {
DecodeError::ShortRead
} else {
DecodeError::Io(e.kind())
}
impl Writeable for OptionalField<Script> {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
match *self {
OptionalField::Present(ref script) => {
// Note that Writeable for script includes the 16-bit length tag for us
}
impl Writeable for OptionalField<u64> {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
match *self {
OptionalField::Present(ref value) => {
value.write(w)?;
}
-impl_writeable_len_match!(AcceptChannel, {
- {AcceptChannel{ shutdown_scriptpubkey: OptionalField::Present(ref script), .. }, 270 + 2 + script.len()},
- {_, 270}
- }, {
+impl_writeable_msg!(AcceptChannel, {
temporary_channel_id,
dust_limit_satoshis,
max_htlc_value_in_flight_msat,
htlc_basepoint,
first_per_commitment_point,
shutdown_scriptpubkey
-});
+}, {});
-impl_writeable!(AnnouncementSignatures, 32+8+64*2, {
+impl_writeable_msg!(AnnouncementSignatures, {
channel_id,
short_channel_id,
node_signature,
bitcoin_signature
-});
+}, {});
impl Writeable for ChannelReestablish {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(if let OptionalField::Present(..) = self.data_loss_protect { 32+2*8+33+32 } else { 32+2*8 });
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.channel_id.write(w)?;
self.next_local_commitment_number.write(w)?;
self.next_remote_commitment_number.write(w)?;
}
}
-impl_writeable!(ClosingSigned, 32+8+64, {
- channel_id,
- fee_satoshis,
- signature
+impl_writeable_msg!(ClosingSigned,
+ { channel_id, fee_satoshis, signature },
+ { (1, fee_range, option) }
+);
+
+impl_writeable!(ClosingSignedFeeRange, {
+ min_fee_satoshis,
+ max_fee_satoshis
});
-impl_writeable_len_match!(CommitmentSigned, {
- { CommitmentSigned { ref htlc_signatures, .. }, 32+64+2+htlc_signatures.len()*64 }
- }, {
+impl_writeable_msg!(CommitmentSigned, {
channel_id,
signature,
htlc_signatures
-});
+}, {});
-impl_writeable_len_match!(DecodedOnionErrorPacket, {
- { DecodedOnionErrorPacket { ref failuremsg, ref pad, .. }, 32 + 4 + failuremsg.len() + pad.len() }
- }, {
+impl_writeable!(DecodedOnionErrorPacket, {
hmac,
failuremsg,
pad
});
-impl_writeable!(FundingCreated, 32+32+2+64, {
+impl_writeable_msg!(FundingCreated, {
temporary_channel_id,
funding_txid,
funding_output_index,
signature
-});
+}, {});
-impl_writeable!(FundingSigned, 32+64, {
+impl_writeable_msg!(FundingSigned, {
channel_id,
signature
-});
+}, {});
-impl_writeable!(FundingLocked, 32+33, {
+impl_writeable_msg!(FundingLocked, {
channel_id,
- next_per_commitment_point
-});
+ next_per_commitment_point,
+}, {});
impl Writeable for Init {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
// global_features gets the bottom 13 bits of our features, and local_features gets all of
// our relevant feature bits. This keeps us compatible with old nodes.
self.features.write_up_to_13(w)?;
}
}
-impl_writeable_len_match!(OpenChannel, {
- { OpenChannel { shutdown_scriptpubkey: OptionalField::Present(ref script), .. }, 319 + 2 + script.len() },
- { _, 319 }
- }, {
+impl_writeable_msg!(OpenChannel, {
chain_hash,
temporary_channel_id,
funding_satoshis,
first_per_commitment_point,
channel_flags,
shutdown_scriptpubkey
+}, {
+ (1, channel_type, option),
});
-impl_writeable!(RevokeAndACK, 32+32+33, {
+impl_writeable_msg!(RevokeAndACK, {
channel_id,
per_commitment_secret,
next_per_commitment_point
-});
+}, {});
-impl_writeable_len_match!(Shutdown, {
- { Shutdown { ref scriptpubkey, .. }, 32 + 2 + scriptpubkey.len() }
- }, {
+impl_writeable_msg!(Shutdown, {
channel_id,
scriptpubkey
-});
+}, {});
-impl_writeable_len_match!(UpdateFailHTLC, {
- { UpdateFailHTLC { ref reason, .. }, 32 + 10 + reason.data.len() }
- }, {
+impl_writeable_msg!(UpdateFailHTLC, {
channel_id,
htlc_id,
reason
-});
+}, {});
-impl_writeable!(UpdateFailMalformedHTLC, 32+8+32+2, {
+impl_writeable_msg!(UpdateFailMalformedHTLC, {
channel_id,
htlc_id,
sha256_of_onion,
failure_code
-});
+}, {});
-impl_writeable!(UpdateFee, 32+4, {
+impl_writeable_msg!(UpdateFee, {
channel_id,
feerate_per_kw
-});
+}, {});
-impl_writeable!(UpdateFulfillHTLC, 32+8+32, {
+impl_writeable_msg!(UpdateFulfillHTLC, {
channel_id,
htlc_id,
payment_preimage
-});
+}, {});
-impl_writeable_len_match!(OnionErrorPacket, {
- { OnionErrorPacket { ref data, .. }, 2 + data.len() }
- }, {
+// Note that this is written as a part of ChannelManager objects, and thus cannot change its
+// serialization format in a way which assumes we know the total serialized length/message end
+// position.
+impl_writeable!(OnionErrorPacket, {
data
});
+// Note that this is written as a part of ChannelManager objects, and thus cannot change its
+// serialization format in a way which assumes we know the total serialized length/message end
+// position.
impl Writeable for OnionPacket {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(1 + 33 + 20*65 + 32);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.version.write(w)?;
match self.public_key {
Ok(pubkey) => pubkey.write(w)?,
}
}
-impl_writeable!(UpdateAddHTLC, 32+8+8+32+4+1366, {
+impl_writeable_msg!(UpdateAddHTLC, {
channel_id,
htlc_id,
amount_msat,
payment_hash,
cltv_expiry,
onion_routing_packet
-});
+}, {});
impl Writeable for FinalOnionHopData {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 8 - (self.total_msat.leading_zeros()/8) as usize);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.payment_secret.0.write(w)?;
HighZeroBytesDroppedVarInt(self.total_msat).write(w)
}
}
impl Writeable for OnionHopData {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(33);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
// Note that this should never be reachable if Rust-Lightning generated the message, as we
// check values are sane long before we get here, though its possible in the future
// user-generated messages may hit this.
},
OnionHopDataFormat::NonFinalNode { short_channel_id } => {
encode_varint_length_prefixed_tlv!(w, {
- (2, HighZeroBytesDroppedVarInt(self.amt_to_forward)),
- (4, HighZeroBytesDroppedVarInt(self.outgoing_cltv_value)),
- (6, short_channel_id)
+ (2, HighZeroBytesDroppedVarInt(self.amt_to_forward), required),
+ (4, HighZeroBytesDroppedVarInt(self.outgoing_cltv_value), required),
+ (6, short_channel_id, required)
});
},
- OnionHopDataFormat::FinalNode { payment_data: Some(ref final_data) } => {
- if final_data.total_msat > MAX_VALUE_MSAT { panic!("We should never be sending infinite/overflow onion payments"); }
- encode_varint_length_prefixed_tlv!(w, {
- (2, HighZeroBytesDroppedVarInt(self.amt_to_forward)),
- (4, HighZeroBytesDroppedVarInt(self.outgoing_cltv_value)),
- (8, final_data)
- });
- },
- OnionHopDataFormat::FinalNode { payment_data: None } => {
+ OnionHopDataFormat::FinalNode { ref payment_data, ref keysend_preimage } => {
+ if let Some(final_data) = payment_data {
+ if final_data.total_msat > MAX_VALUE_MSAT { panic!("We should never be sending infinite/overflow onion payments"); }
+ }
encode_varint_length_prefixed_tlv!(w, {
- (2, HighZeroBytesDroppedVarInt(self.amt_to_forward)),
- (4, HighZeroBytesDroppedVarInt(self.outgoing_cltv_value))
+ (2, HighZeroBytesDroppedVarInt(self.amt_to_forward), required),
+ (4, HighZeroBytesDroppedVarInt(self.outgoing_cltv_value), required),
+ (8, payment_data, option),
+ (5482373484, keysend_preimage, option)
});
},
}
let mut cltv_value = HighZeroBytesDroppedVarInt(0u32);
let mut short_id: Option<u64> = None;
let mut payment_data: Option<FinalOnionHopData> = None;
- decode_tlv!(&mut rd, {
- (2, amt),
- (4, cltv_value)
- }, {
- (6, short_id),
- (8, payment_data)
+ let mut keysend_preimage: Option<PaymentPreimage> = None;
+ // The TLV type is chosen to be compatible with lnd and c-lightning.
+ decode_tlv_stream!(&mut rd, {
+ (2, amt, required),
+ (4, cltv_value, required),
+ (6, short_id, option),
+ (8, payment_data, option),
+ (5482373484, keysend_preimage, option)
});
rd.eat_remaining().map_err(|_| DecodeError::ShortRead)?;
let format = if let Some(short_channel_id) = short_id {
}
}
OnionHopDataFormat::FinalNode {
- payment_data
+ payment_data,
+ keysend_preimage,
}
};
(format, amt.0, cltv_value.0)
}
impl Writeable for Ping {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(self.byteslen as usize + 4);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.ponglen.write(w)?;
vec![0u8; self.byteslen as usize].write(w)?; // size-unchecked write
Ok(())
}
impl Writeable for Pong {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(self.byteslen as usize + 2);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
vec![0u8; self.byteslen as usize].write(w)?; // size-unchecked write
Ok(())
}
}
impl Writeable for UnsignedChannelAnnouncement {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(2 + 2*32 + 4*33 + self.features.byte_count() + self.excess_data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.features.write(w)?;
self.chain_hash.write(w)?;
self.short_channel_id.write(w)?;
node_id_2: Readable::read(r)?,
bitcoin_key_1: Readable::read(r)?,
bitcoin_key_2: Readable::read(r)?,
- excess_data: {
- let mut excess_data = vec![];
- r.read_to_end(&mut excess_data)?;
- excess_data
- },
+ excess_data: read_to_end(r)?,
})
}
}
-impl_writeable_len_match!(ChannelAnnouncement, {
- { ChannelAnnouncement { contents: UnsignedChannelAnnouncement {ref features, ref excess_data, ..}, .. },
- 2 + 2*32 + 4*33 + features.byte_count() + excess_data.len() + 4*64 }
- }, {
+impl_writeable!(ChannelAnnouncement, {
node_signature_1,
node_signature_2,
bitcoin_signature_1,
});
impl Writeable for UnsignedChannelUpdate {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- let mut size = 64 + self.excess_data.len();
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
let mut message_flags: u8 = 0;
if let OptionalField::Present(_) = self.htlc_maximum_msat {
- size += 8;
message_flags = 1;
}
- w.size_hint(size);
self.chain_hash.write(w)?;
self.short_channel_id.write(w)?;
self.timestamp.write(w)?;
fee_base_msat: Readable::read(r)?,
fee_proportional_millionths: Readable::read(r)?,
htlc_maximum_msat: if has_htlc_maximum_msat { Readable::read(r)? } else { OptionalField::Absent },
- excess_data: {
- let mut excess_data = vec![];
- r.read_to_end(&mut excess_data)?;
- excess_data
- },
+ excess_data: read_to_end(r)?,
})
}
}
-impl_writeable_len_match!(ChannelUpdate, {
- { ChannelUpdate { contents: UnsignedChannelUpdate {ref excess_data, ..}, .. },
- 64 + excess_data.len() + 64 }
- }, {
+impl_writeable!(ChannelUpdate, {
signature,
contents
});
impl Writeable for ErrorMessage {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 2 + self.data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.channel_id.write(w)?;
(self.data.len() as u16).write(w)?;
w.write_all(self.data.as_bytes())?;
channel_id: Readable::read(r)?,
data: {
let mut sz: usize = <u16 as Readable>::read(r)? as usize;
- let mut data = vec![];
- let data_len = r.read_to_end(&mut data)?;
- sz = cmp::min(data_len, sz);
+ let data = read_to_end(r)?;
+ sz = cmp::min(data.len(), sz);
match String::from_utf8(data[..sz as usize].to_vec()) {
Ok(s) => s,
Err(_) => return Err(DecodeError::InvalidValue),
}
impl Writeable for UnsignedNodeAnnouncement {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(64 + 76 + self.features.byte_count() + self.addresses.len()*38 + self.excess_address_data.len() + self.excess_data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.features.write(w)?;
self.timestamp.write(w)?;
self.node_id.write(w)?;
}
Vec::new()
};
- r.read_to_end(&mut excess_data)?;
+ excess_data.extend(read_to_end(r)?.iter());
Ok(UnsignedNodeAnnouncement {
features,
timestamp,
}
}
-impl_writeable_len_match!(NodeAnnouncement, {
- { NodeAnnouncement { contents: UnsignedNodeAnnouncement { ref features, ref addresses, ref excess_address_data, ref excess_data, ..}, .. },
- 64 + 76 + features.byte_count() + addresses.len()*(NetAddress::MAX_LEN as usize + 1) + excess_address_data.len() + excess_data.len() }
- }, {
+impl_writeable!(NodeAnnouncement, {
signature,
contents
});
fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
let chain_hash: BlockHash = Readable::read(r)?;
- // We expect the encoding_len to always includes the 1-byte
- // encoding_type and that short_channel_ids are 8-bytes each
let encoding_len: u16 = Readable::read(r)?;
- if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
- return Err(DecodeError::InvalidValue);
- }
+ let encoding_type: u8 = Readable::read(r)?;
// Must be encoding_type=0 uncompressed serialization. We do not
// support encoding_type=1 zlib serialization.
- let encoding_type: u8 = Readable::read(r)?;
if encoding_type != EncodingType::Uncompressed as u8 {
+ return Err(DecodeError::UnsupportedCompression);
+ }
+
+ // We expect the encoding_len to always includes the 1-byte
+ // encoding_type and that short_channel_ids are 8-bytes each
+ if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
return Err(DecodeError::InvalidValue);
}
}
impl Writeable for QueryShortChannelIds {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
// Calculated from 1-byte encoding_type plus 8-bytes per short_channel_id
let encoding_len: u16 = 1 + self.short_channel_ids.len() as u16 * 8;
- w.size_hint(32 + 2 + encoding_len as usize);
self.chain_hash.write(w)?;
encoding_len.write(w)?;
}
}
-impl Readable for ReplyShortChannelIdsEnd {
- fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
- let chain_hash: BlockHash = Readable::read(r)?;
- let full_information: bool = Readable::read(r)?;
- Ok(ReplyShortChannelIdsEnd {
- chain_hash,
- full_information,
- })
- }
-}
-
-impl Writeable for ReplyShortChannelIdsEnd {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 1);
- self.chain_hash.write(w)?;
- self.full_information.write(w)?;
- Ok(())
- }
-}
-
-impl Readable for QueryChannelRange {
- fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
- let chain_hash: BlockHash = Readable::read(r)?;
- let first_blocknum: u32 = Readable::read(r)?;
- let number_of_blocks: u32 = Readable::read(r)?;
- Ok(QueryChannelRange {
- chain_hash,
- first_blocknum,
- number_of_blocks
- })
+impl_writeable_msg!(ReplyShortChannelIdsEnd, {
+ chain_hash,
+ full_information,
+}, {});
+
+impl QueryChannelRange {
+ /**
+ * Calculates the overflow safe ending block height for the query.
+ * Overflow returns `0xffffffff`, otherwise returns `first_blocknum + number_of_blocks`
+ */
+ pub fn end_blocknum(&self) -> u32 {
+ match self.first_blocknum.checked_add(self.number_of_blocks) {
+ Some(block) => block,
+ None => u32::max_value(),
+ }
}
}
-impl Writeable for QueryChannelRange {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 4 + 4);
- self.chain_hash.write(w)?;
- self.first_blocknum.write(w)?;
- self.number_of_blocks.write(w)?;
- Ok(())
- }
-}
+impl_writeable_msg!(QueryChannelRange, {
+ chain_hash,
+ first_blocknum,
+ number_of_blocks
+}, {});
impl Readable for ReplyChannelRange {
fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
let number_of_blocks: u32 = Readable::read(r)?;
let sync_complete: bool = Readable::read(r)?;
- // We expect the encoding_len to always includes the 1-byte
- // encoding_type and that short_channel_ids are 8-bytes each
let encoding_len: u16 = Readable::read(r)?;
- if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
- return Err(DecodeError::InvalidValue);
- }
+ let encoding_type: u8 = Readable::read(r)?;
// Must be encoding_type=0 uncompressed serialization. We do not
// support encoding_type=1 zlib serialization.
- let encoding_type: u8 = Readable::read(r)?;
if encoding_type != EncodingType::Uncompressed as u8 {
+ return Err(DecodeError::UnsupportedCompression);
+ }
+
+ // We expect the encoding_len to always includes the 1-byte
+ // encoding_type and that short_channel_ids are 8-bytes each
+ if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
return Err(DecodeError::InvalidValue);
}
}
impl Writeable for ReplyChannelRange {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
let encoding_len: u16 = 1 + self.short_channel_ids.len() as u16 * 8;
- w.size_hint(32 + 4 + 4 + 1 + 2 + encoding_len as usize);
self.chain_hash.write(w)?;
self.first_blocknum.write(w)?;
self.number_of_blocks.write(w)?;
}
}
-impl Readable for GossipTimestampFilter {
- fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
- let chain_hash: BlockHash = Readable::read(r)?;
- let first_timestamp: u32 = Readable::read(r)?;
- let timestamp_range: u32 = Readable::read(r)?;
- Ok(GossipTimestampFilter {
- chain_hash,
- first_timestamp,
- timestamp_range,
- })
- }
-}
-
-impl Writeable for GossipTimestampFilter {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 4 + 4);
- self.chain_hash.write(w)?;
- self.first_timestamp.write(w)?;
- self.timestamp_range.write(w)?;
- Ok(())
- }
-}
-
+impl_writeable_msg!(GossipTimestampFilter, {
+ chain_hash,
+ first_timestamp,
+ timestamp_range,
+}, {});
#[cfg(test)]
mod tests {
use hex;
+ use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
+ use ln::features::{ChannelFeatures, ChannelTypeFeatures, InitFeatures, NodeFeatures};
use ln::msgs;
- use ln::msgs::{ChannelFeatures, FinalOnionHopData, InitFeatures, NodeFeatures, OptionalField, OnionErrorPacket, OnionHopDataFormat};
- use ln::channelmanager::{PaymentPreimage, PaymentHash, PaymentSecret};
+ use ln::msgs::{FinalOnionHopData, OptionalField, OnionErrorPacket, OnionHopDataFormat};
use util::ser::{Writeable, Readable};
use bitcoin::hashes::hex::FromHex;
use bitcoin::secp256k1::key::{PublicKey,SecretKey};
use bitcoin::secp256k1::{Secp256k1, Message};
- use std::io::Cursor;
+ use io::Cursor;
+ use prelude::*;
#[test]
fn encoding_channel_reestablish_no_secret() {
do_encoding_channel_update(true, true, true, true);
}
- fn do_encoding_open_channel(random_bit: bool, shutdown: bool) {
+ fn do_encoding_open_channel(random_bit: bool, shutdown: bool, incl_chan_type: bool) {
let secp_ctx = Secp256k1::new();
let (_, pubkey_1) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
let (_, pubkey_2) = get_keys_from!("0202020202020202020202020202020202020202020202020202020202020202", secp_ctx);
htlc_basepoint: pubkey_5,
first_per_commitment_point: pubkey_6,
channel_flags: if random_bit { 1 << 5 } else { 0 },
- shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent }
+ shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent },
+ channel_type: if incl_chan_type { Some(ChannelTypeFeatures::empty()) } else { None },
};
let encoded_value = open_channel.encode();
let mut target_value = Vec::new();
if shutdown {
target_value.append(&mut hex::decode("001976a91479b000887626b294a914501a4cd226b58b23598388ac").unwrap());
}
+ if incl_chan_type {
+ target_value.append(&mut hex::decode("0100").unwrap());
+ }
assert_eq!(encoded_value, target_value);
}
#[test]
fn encoding_open_channel() {
- do_encoding_open_channel(false, false);
- do_encoding_open_channel(true, false);
- do_encoding_open_channel(false, true);
- do_encoding_open_channel(true, true);
+ do_encoding_open_channel(false, false, false);
+ do_encoding_open_channel(false, false, true);
+ do_encoding_open_channel(false, true, false);
+ do_encoding_open_channel(false, true, true);
+ do_encoding_open_channel(true, false, false);
+ do_encoding_open_channel(true, false, true);
+ do_encoding_open_channel(true, true, false);
+ do_encoding_open_channel(true, true, true);
}
fn do_encoding_accept_channel(shutdown: bool) {
channel_id: [2; 32],
fee_satoshis: 2316138423780173,
signature: sig_1,
+ fee_range: None,
};
let encoded_value = closing_signed.encode();
let target_value = hex::decode("020202020202020202020202020202020202020202020202020202020202020200083a840000034dd977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073a").unwrap();
assert_eq!(encoded_value, target_value);
+ assert_eq!(msgs::ClosingSigned::read(&mut Cursor::new(&target_value)).unwrap(), closing_signed);
+
+ let closing_signed_with_range = msgs::ClosingSigned {
+ channel_id: [2; 32],
+ fee_satoshis: 2316138423780173,
+ signature: sig_1,
+ fee_range: Some(msgs::ClosingSignedFeeRange {
+ min_fee_satoshis: 0xdeadbeef,
+ max_fee_satoshis: 0x1badcafe01234567,
+ }),
+ };
+ let encoded_value_with_range = closing_signed_with_range.encode();
+ let target_value_with_range = hex::decode("020202020202020202020202020202020202020202020202020202020202020200083a840000034dd977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073a011000000000deadbeef1badcafe01234567").unwrap();
+ assert_eq!(encoded_value_with_range, target_value_with_range);
+ assert_eq!(msgs::ClosingSigned::read(&mut Cursor::new(&target_value_with_range)).unwrap(),
+ closing_signed_with_range);
}
#[test]
let mut msg = msgs::OnionHopData {
format: OnionHopDataFormat::FinalNode {
payment_data: None,
+ keysend_preimage: None,
},
amt_to_forward: 0x0badf00d01020304,
outgoing_cltv_value: 0xffffffff,
let target_value = hex::decode("1002080badf00d010203040404ffffffff").unwrap();
assert_eq!(encoded_value, target_value);
msg = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
- if let OnionHopDataFormat::FinalNode { payment_data: None } = msg.format { } else { panic!(); }
+ if let OnionHopDataFormat::FinalNode { payment_data: None, .. } = msg.format { } else { panic!(); }
assert_eq!(msg.amt_to_forward, 0x0badf00d01020304);
assert_eq!(msg.outgoing_cltv_value, 0xffffffff);
}
payment_secret: expected_payment_secret,
total_msat: 0x1badca1f
}),
+ keysend_preimage: None,
},
amt_to_forward: 0x0badf00d01020304,
outgoing_cltv_value: 0xffffffff,
payment_data: Some(FinalOnionHopData {
payment_secret,
total_msat: 0x1badca1f
- })
+ }),
+ keysend_preimage: None,
} = msg.format {
assert_eq!(payment_secret, expected_payment_secret);
} else { panic!(); }
assert_eq!(msg.outgoing_cltv_value, 0xffffffff);
}
+ #[test]
+ fn query_channel_range_end_blocknum() {
+ let tests: Vec<(u32, u32, u32)> = vec![
+ (10000, 1500, 11500),
+ (0, 0xffffffff, 0xffffffff),
+ (1, 0xffffffff, 0xffffffff),
+ ];
+
+ for (first_blocknum, number_of_blocks, expected) in tests.into_iter() {
+ let sut = msgs::QueryChannelRange {
+ chain_hash: BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap(),
+ first_blocknum,
+ number_of_blocks,
+ };
+ assert_eq!(sut.end_blocknum(), expected);
+ }
+ }
+
#[test]
fn encoding_query_channel_range() {
let mut query_channel_range = msgs::QueryChannelRange {