+// This file is Copyright its original authors, visible in version control
+// history.
+//
+// This file is licensed under the Apache License, Version 2.0 <LICENSE-APACHE
+// or http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
+// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your option.
+// You may not use this file except in accordance with one or both of these
+// licenses.
+
//! Wire messages, traits representing wire message handlers, and a few error types live here.
//!
//! For a normal node you probably don't need to use anything here, however, if you wish to split a
//! raw socket events into your non-internet-facing system and then send routing events back to
//! track the network on the less-secure system.
-use secp256k1::key::PublicKey;
-use secp256k1::Signature;
-use secp256k1;
-use bitcoin_hashes::sha256d::Hash as Sha256dHash;
+use bitcoin::secp256k1::PublicKey;
+use bitcoin::secp256k1::ecdsa::Signature;
+use bitcoin::secp256k1;
use bitcoin::blockdata::script::Script;
+use bitcoin::hash_types::{Txid, BlockHash};
+
+use ln::features::{ChannelFeatures, ChannelTypeFeatures, InitFeatures, NodeFeatures};
+use ln::onion_utils;
+use onion_message;
+
+use prelude::*;
+use core::fmt;
+use core::fmt::Debug;
+use io::{self, Read};
+use io_extras::read_to_end;
-use std::error::Error;
-use std::{cmp, fmt};
-use std::io::Read;
-use std::result::Result;
+use util::events::{MessageSendEventsProvider, OnionMessageProvider};
+use util::logger;
+use util::ser::{BigSize, LengthReadable, Readable, ReadableArgs, Writeable, Writer, FixedLengthReader, HighZeroBytesDroppedBigSize, Hostname};
-use util::events;
-use util::ser::{Readable, Writeable, Writer};
+use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
-use ln::channelmanager::{PaymentPreimage, PaymentHash};
+/// 21 million * 10^8 * 1000
+pub(crate) const MAX_VALUE_MSAT: u64 = 21_000_000_0000_0000_000;
/// An error in decoding a message or struct.
-#[derive(Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub enum DecodeError {
/// A version byte specified something we don't know how to handle.
/// Includes unknown realm byte in an OnionHopData packet
UnknownVersion,
- /// Unknown feature mandating we fail to parse message
+ /// Unknown feature mandating we fail to parse message (eg TLV with an even, unknown type)
UnknownRequiredFeature,
/// Value was invalid, eg a byte which was supposed to be a bool was something other than a 0
- /// or 1, a public key/private key/signature was invalid, text wasn't UTF-8, etc
+ /// or 1, a public key/private key/signature was invalid, text wasn't UTF-8, TLV was
+ /// syntactically incorrect, etc
InvalidValue,
/// Buffer too short
ShortRead,
- /// node_announcement included more than one address of a given type!
- ExtraAddressesPerType,
/// A length descriptor in the packet didn't describe the later data correctly
BadLengthDescriptor,
/// Error from std::io
- Io(::std::io::Error),
-}
-
-/// Tracks localfeatures which are only in init messages
-#[derive(Clone, PartialEq)]
-pub struct LocalFeatures {
- flags: Vec<u8>,
-}
-
-impl LocalFeatures {
- /// Create a blank LocalFeatures flags (visibility extended for fuzz tests)
- #[cfg(not(feature = "fuzztarget"))]
- pub(crate) fn new() -> LocalFeatures {
- LocalFeatures {
- flags: vec![2 | 1 << 5],
- }
- }
- #[cfg(feature = "fuzztarget")]
- pub fn new() -> LocalFeatures {
- LocalFeatures {
- flags: vec![2 | 1 << 5],
- }
- }
-
- pub(crate) fn supports_data_loss_protect(&self) -> bool {
- self.flags.len() > 0 && (self.flags[0] & 3) != 0
- }
- pub(crate) fn initial_routing_sync(&self) -> bool {
- self.flags.len() > 0 && (self.flags[0] & (1 << 3)) != 0
- }
- pub(crate) fn set_initial_routing_sync(&mut self) {
- if self.flags.len() == 0 {
- self.flags.resize(1, 1 << 3);
- } else {
- self.flags[0] |= 1 << 3;
- }
- }
-
- pub(crate) fn supports_upfront_shutdown_script(&self) -> bool {
- self.flags.len() > 0 && (self.flags[0] & (3 << 4)) != 0
- }
- #[cfg(test)]
- pub(crate) fn unset_upfront_shutdown_script(&mut self) {
- self.flags[0] ^= 1 << 5;
- }
-
- pub(crate) fn requires_unknown_bits(&self) -> bool {
- self.flags.iter().enumerate().any(|(idx, &byte)| {
- ( idx != 0 && (byte & 0x55) != 0 ) || ( idx == 0 && (byte & 0x14) != 0 )
- })
- }
-
- pub(crate) fn supports_unknown_bits(&self) -> bool {
- self.flags.iter().enumerate().any(|(idx, &byte)| {
- ( idx != 0 && byte != 0 ) || ( idx == 0 && (byte & 0xc4) != 0 )
- })
- }
-}
-
-/// Tracks globalfeatures which are in init messages and routing announcements
-#[derive(Clone, PartialEq, Debug)]
-pub struct GlobalFeatures {
- #[cfg(not(test))]
- flags: Vec<u8>,
- // Used to test encoding of diverse msgs
- #[cfg(test)]
- pub flags: Vec<u8>
-}
-
-impl GlobalFeatures {
- pub(crate) fn new() -> GlobalFeatures {
- GlobalFeatures {
- flags: Vec::new(),
- }
- }
-
- pub(crate) fn requires_unknown_bits(&self) -> bool {
- for &byte in self.flags.iter() {
- if (byte & 0x55) != 0 {
- return true;
- }
- }
- return false;
- }
-
- pub(crate) fn supports_unknown_bits(&self) -> bool {
- for &byte in self.flags.iter() {
- if byte != 0 {
- return true;
- }
- }
- return false;
- }
+ Io(/// (C-not exported) as ErrorKind doesn't have a reasonable mapping
+ io::ErrorKind),
+ /// The message included zlib-compressed values, which we don't support.
+ UnsupportedCompression,
}
/// An init message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
pub struct Init {
- pub(crate) global_features: GlobalFeatures,
- pub(crate) local_features: LocalFeatures,
+ /// The relevant features which the sender supports
+ pub features: InitFeatures,
+ /// The receipient's network address. This adds the option to report a remote IP address
+ /// back to a connecting peer using the init message. A node can decide to use that information
+ /// to discover a potential update to its public IPv4 address (NAT) and use
+ /// that for a node_announcement update message containing the new address.
+ pub remote_network_address: Option<NetAddress>,
}
/// An error message to be sent or received from a peer
-#[derive(Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct ErrorMessage {
- pub(crate) channel_id: [u8; 32],
- pub(crate) data: String,
+ /// The channel ID involved in the error.
+ ///
+ /// All-0s indicates a general error unrelated to a specific channel, after which all channels
+ /// with the sending peer should be closed.
+ pub channel_id: [u8; 32],
+ /// A possibly human-readable error description.
+ /// The string should be sanitized before it is used (e.g. emitted to logs or printed to
+ /// stdout). Otherwise, a well crafted error message may trigger a security vulnerability in
+ /// the terminal emulator or the logging subsystem.
+ pub data: String,
+}
+
+/// A warning message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
+pub struct WarningMessage {
+ /// The channel ID involved in the warning.
+ ///
+ /// All-0s indicates a warning unrelated to a specific channel.
+ pub channel_id: [u8; 32],
+ /// A possibly human-readable warning description.
+ /// The string should be sanitized before it is used (e.g. emitted to logs or printed to
+ /// stdout). Otherwise, a well crafted error message may trigger a security vulnerability in
+ /// the terminal emulator or the logging subsystem.
+ pub data: String,
}
/// A ping message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
pub struct Ping {
- pub(crate) ponglen: u16,
- pub(crate) byteslen: u16,
+ /// The desired response length
+ pub ponglen: u16,
+ /// The ping packet size.
+ /// This field is not sent on the wire. byteslen zeros are sent.
+ pub byteslen: u16,
}
/// A pong message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
pub struct Pong {
- pub(crate) byteslen: u16,
+ /// The pong packet size.
+ /// This field is not sent on the wire. byteslen zeros are sent.
+ pub byteslen: u16,
}
/// An open_channel message to be sent or received from a peer
-#[derive(Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct OpenChannel {
- pub(crate) chain_hash: Sha256dHash,
- pub(crate) temporary_channel_id: [u8; 32],
- pub(crate) funding_satoshis: u64,
- pub(crate) push_msat: u64,
- pub(crate) dust_limit_satoshis: u64,
- pub(crate) max_htlc_value_in_flight_msat: u64,
- pub(crate) channel_reserve_satoshis: u64,
- pub(crate) htlc_minimum_msat: u64,
- pub(crate) feerate_per_kw: u32,
- pub(crate) to_self_delay: u16,
- pub(crate) max_accepted_htlcs: u16,
- pub(crate) funding_pubkey: PublicKey,
- pub(crate) revocation_basepoint: PublicKey,
- pub(crate) payment_basepoint: PublicKey,
- pub(crate) delayed_payment_basepoint: PublicKey,
- pub(crate) htlc_basepoint: PublicKey,
- pub(crate) first_per_commitment_point: PublicKey,
- pub(crate) channel_flags: u8,
- pub(crate) shutdown_scriptpubkey: OptionalField<Script>,
+ /// The genesis hash of the blockchain where the channel is to be opened
+ pub chain_hash: BlockHash,
+ /// A temporary channel ID, until the funding outpoint is announced
+ pub temporary_channel_id: [u8; 32],
+ /// The channel value
+ pub funding_satoshis: u64,
+ /// The amount to push to the counterparty as part of the open, in milli-satoshi
+ pub push_msat: u64,
+ /// The threshold below which outputs on transactions broadcast by sender will be omitted
+ pub dust_limit_satoshis: u64,
+ /// The maximum inbound HTLC value in flight towards sender, in milli-satoshi
+ pub max_htlc_value_in_flight_msat: u64,
+ /// The minimum value unencumbered by HTLCs for the counterparty to keep in the channel
+ pub channel_reserve_satoshis: u64,
+ /// The minimum HTLC size incoming to sender, in milli-satoshi
+ pub htlc_minimum_msat: u64,
+ /// The feerate per 1000-weight of sender generated transactions, until updated by update_fee
+ pub feerate_per_kw: u32,
+ /// The number of blocks which the counterparty will have to wait to claim on-chain funds if they broadcast a commitment transaction
+ pub to_self_delay: u16,
+ /// The maximum number of inbound HTLCs towards sender
+ pub max_accepted_htlcs: u16,
+ /// The sender's key controlling the funding transaction
+ pub funding_pubkey: PublicKey,
+ /// Used to derive a revocation key for transactions broadcast by counterparty
+ pub revocation_basepoint: PublicKey,
+ /// A payment key to sender for transactions broadcast by counterparty
+ pub payment_point: PublicKey,
+ /// Used to derive a payment key to sender for transactions broadcast by sender
+ pub delayed_payment_basepoint: PublicKey,
+ /// Used to derive an HTLC payment key to sender
+ pub htlc_basepoint: PublicKey,
+ /// The first to-be-broadcast-by-sender transaction's per commitment point
+ pub first_per_commitment_point: PublicKey,
+ /// Channel flags
+ pub channel_flags: u8,
+ /// Optionally, a request to pre-set the to-sender output's scriptPubkey for when we collaboratively close
+ pub shutdown_scriptpubkey: OptionalField<Script>,
+ /// The channel type that this channel will represent. If none is set, we derive the channel
+ /// type from the intersection of our feature bits with our counterparty's feature bits from
+ /// the Init message.
+ pub channel_type: Option<ChannelTypeFeatures>,
}
/// An accept_channel message to be sent or received from a peer
-#[derive(Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct AcceptChannel {
- pub(crate) temporary_channel_id: [u8; 32],
- pub(crate) dust_limit_satoshis: u64,
- pub(crate) max_htlc_value_in_flight_msat: u64,
- pub(crate) channel_reserve_satoshis: u64,
- pub(crate) htlc_minimum_msat: u64,
- pub(crate) minimum_depth: u32,
- pub(crate) to_self_delay: u16,
- pub(crate) max_accepted_htlcs: u16,
- pub(crate) funding_pubkey: PublicKey,
- pub(crate) revocation_basepoint: PublicKey,
- pub(crate) payment_basepoint: PublicKey,
- pub(crate) delayed_payment_basepoint: PublicKey,
- pub(crate) htlc_basepoint: PublicKey,
- pub(crate) first_per_commitment_point: PublicKey,
- pub(crate) shutdown_scriptpubkey: OptionalField<Script>
+ /// A temporary channel ID, until the funding outpoint is announced
+ pub temporary_channel_id: [u8; 32],
+ /// The threshold below which outputs on transactions broadcast by sender will be omitted
+ pub dust_limit_satoshis: u64,
+ /// The maximum inbound HTLC value in flight towards sender, in milli-satoshi
+ pub max_htlc_value_in_flight_msat: u64,
+ /// The minimum value unencumbered by HTLCs for the counterparty to keep in the channel
+ pub channel_reserve_satoshis: u64,
+ /// The minimum HTLC size incoming to sender, in milli-satoshi
+ pub htlc_minimum_msat: u64,
+ /// Minimum depth of the funding transaction before the channel is considered open
+ pub minimum_depth: u32,
+ /// The number of blocks which the counterparty will have to wait to claim on-chain funds if they broadcast a commitment transaction
+ pub to_self_delay: u16,
+ /// The maximum number of inbound HTLCs towards sender
+ pub max_accepted_htlcs: u16,
+ /// The sender's key controlling the funding transaction
+ pub funding_pubkey: PublicKey,
+ /// Used to derive a revocation key for transactions broadcast by counterparty
+ pub revocation_basepoint: PublicKey,
+ /// A payment key to sender for transactions broadcast by counterparty
+ pub payment_point: PublicKey,
+ /// Used to derive a payment key to sender for transactions broadcast by sender
+ pub delayed_payment_basepoint: PublicKey,
+ /// Used to derive an HTLC payment key to sender for transactions broadcast by counterparty
+ pub htlc_basepoint: PublicKey,
+ /// The first to-be-broadcast-by-sender transaction's per commitment point
+ pub first_per_commitment_point: PublicKey,
+ /// Optionally, a request to pre-set the to-sender output's scriptPubkey for when we collaboratively close
+ pub shutdown_scriptpubkey: OptionalField<Script>,
+ /// The channel type that this channel will represent. If none is set, we derive the channel
+ /// type from the intersection of our feature bits with our counterparty's feature bits from
+ /// the Init message.
+ ///
+ /// This is required to match the equivalent field in [`OpenChannel::channel_type`].
+ pub channel_type: Option<ChannelTypeFeatures>,
}
/// A funding_created message to be sent or received from a peer
-#[derive(Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct FundingCreated {
- pub(crate) temporary_channel_id: [u8; 32],
- pub(crate) funding_txid: Sha256dHash,
- pub(crate) funding_output_index: u16,
- pub(crate) signature: Signature,
+ /// A temporary channel ID, until the funding is established
+ pub temporary_channel_id: [u8; 32],
+ /// The funding transaction ID
+ pub funding_txid: Txid,
+ /// The specific output index funding this channel
+ pub funding_output_index: u16,
+ /// The signature of the channel initiator (funder) on the initial commitment transaction
+ pub signature: Signature,
}
/// A funding_signed message to be sent or received from a peer
-#[derive(Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct FundingSigned {
- pub(crate) channel_id: [u8; 32],
- pub(crate) signature: Signature,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The signature of the channel acceptor (fundee) on the initial commitment transaction
+ pub signature: Signature,
}
-/// A funding_locked message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
-pub struct FundingLocked {
- pub(crate) channel_id: [u8; 32],
- pub(crate) next_per_commitment_point: PublicKey,
+/// A channel_ready message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
+pub struct ChannelReady {
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The per-commitment point of the second commitment transaction
+ pub next_per_commitment_point: PublicKey,
+ /// If set, provides a short_channel_id alias for this channel. The sender will accept payments
+ /// to be forwarded over this SCID and forward them to this messages' recipient.
+ pub short_channel_id_alias: Option<u64>,
}
/// A shutdown message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct Shutdown {
- pub(crate) channel_id: [u8; 32],
- pub(crate) scriptpubkey: Script,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The destination of this peer's funds on closing.
+ /// Must be in one of these forms: p2pkh, p2sh, p2wpkh, p2wsh.
+ pub scriptpubkey: Script,
+}
+
+/// The minimum and maximum fees which the sender is willing to place on the closing transaction.
+/// This is provided in [`ClosingSigned`] by both sides to indicate the fee range they are willing
+/// to use.
+#[derive(Clone, Debug, PartialEq)]
+pub struct ClosingSignedFeeRange {
+ /// The minimum absolute fee, in satoshis, which the sender is willing to place on the closing
+ /// transaction.
+ pub min_fee_satoshis: u64,
+ /// The maximum absolute fee, in satoshis, which the sender is willing to place on the closing
+ /// transaction.
+ pub max_fee_satoshis: u64,
}
/// A closing_signed message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct ClosingSigned {
- pub(crate) channel_id: [u8; 32],
- pub(crate) fee_satoshis: u64,
- pub(crate) signature: Signature,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The proposed total fee for the closing transaction
+ pub fee_satoshis: u64,
+ /// A signature on the closing transaction
+ pub signature: Signature,
+ /// The minimum and maximum fees which the sender is willing to accept, provided only by new
+ /// nodes.
+ pub fee_range: Option<ClosingSignedFeeRange>,
}
/// An update_add_htlc message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UpdateAddHTLC {
- pub(crate) channel_id: [u8; 32],
- pub(crate) htlc_id: u64,
- pub(crate) amount_msat: u64,
- pub(crate) payment_hash: PaymentHash,
- pub(crate) cltv_expiry: u32,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The HTLC ID
+ pub htlc_id: u64,
+ /// The HTLC value in milli-satoshi
+ pub amount_msat: u64,
+ /// The payment hash, the pre-image of which controls HTLC redemption
+ pub payment_hash: PaymentHash,
+ /// The expiry height of the HTLC
+ pub cltv_expiry: u32,
pub(crate) onion_routing_packet: OnionPacket,
}
+ /// An onion message to be sent or received from a peer
+#[derive(Clone, Debug, PartialEq)]
+pub struct OnionMessage {
+ /// Used in decrypting the onion packet's payload.
+ pub blinding_point: PublicKey,
+ pub(crate) onion_routing_packet: onion_message::Packet,
+}
+
/// An update_fulfill_htlc message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UpdateFulfillHTLC {
- pub(crate) channel_id: [u8; 32],
- pub(crate) htlc_id: u64,
- pub(crate) payment_preimage: PaymentPreimage,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The HTLC ID
+ pub htlc_id: u64,
+ /// The pre-image of the payment hash, allowing HTLC redemption
+ pub payment_preimage: PaymentPreimage,
}
/// An update_fail_htlc message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UpdateFailHTLC {
- pub(crate) channel_id: [u8; 32],
- pub(crate) htlc_id: u64,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The HTLC ID
+ pub htlc_id: u64,
pub(crate) reason: OnionErrorPacket,
}
/// An update_fail_malformed_htlc message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UpdateFailMalformedHTLC {
- pub(crate) channel_id: [u8; 32],
- pub(crate) htlc_id: u64,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The HTLC ID
+ pub htlc_id: u64,
pub(crate) sha256_of_onion: [u8; 32],
- pub(crate) failure_code: u16,
+ /// The failure code
+ pub failure_code: u16,
}
/// A commitment_signed message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct CommitmentSigned {
- pub(crate) channel_id: [u8; 32],
- pub(crate) signature: Signature,
- pub(crate) htlc_signatures: Vec<Signature>,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// A signature on the commitment transaction
+ pub signature: Signature,
+ /// Signatures on the HTLC transactions
+ pub htlc_signatures: Vec<Signature>,
}
/// A revoke_and_ack message to be sent or received from a peer
-#[derive(Clone, PartialEq)]
+#[derive(Clone, Debug, PartialEq)]
pub struct RevokeAndACK {
- pub(crate) channel_id: [u8; 32],
- pub(crate) per_commitment_secret: [u8; 32],
- pub(crate) next_per_commitment_point: PublicKey,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The secret corresponding to the per-commitment point
+ pub per_commitment_secret: [u8; 32],
+ /// The next sender-broadcast commitment transaction's per-commitment point
+ pub next_per_commitment_point: PublicKey,
}
/// An update_fee message to be sent or received from a peer
-#[derive(PartialEq, Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UpdateFee {
- pub(crate) channel_id: [u8; 32],
- pub(crate) feerate_per_kw: u32,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// Fee rate per 1000-weight of the transaction
+ pub feerate_per_kw: u32,
}
-#[derive(PartialEq, Clone)]
-pub(crate) struct DataLossProtect {
- pub(crate) your_last_per_commitment_secret: [u8; 32],
- pub(crate) my_current_per_commitment_point: PublicKey,
+#[derive(Clone, Debug, PartialEq)]
+/// Proof that the sender knows the per-commitment secret of the previous commitment transaction.
+/// This is used to convince the recipient that the channel is at a certain commitment
+/// number even if they lost that data due to a local failure. Of course, the peer may lie
+/// and even later commitments may have been revoked.
+pub struct DataLossProtect {
+ /// Proof that the sender knows the per-commitment secret of a specific commitment transaction
+ /// belonging to the recipient
+ pub your_last_per_commitment_secret: [u8; 32],
+ /// The sender's per-commitment point for their current commitment transaction
+ pub my_current_per_commitment_point: PublicKey,
}
/// A channel_reestablish message to be sent or received from a peer
-#[derive(PartialEq, Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct ChannelReestablish {
- pub(crate) channel_id: [u8; 32],
- pub(crate) next_local_commitment_number: u64,
- pub(crate) next_remote_commitment_number: u64,
- pub(crate) data_loss_protect: OptionalField<DataLossProtect>,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The next commitment number for the sender
+ pub next_local_commitment_number: u64,
+ /// The next commitment number for the recipient
+ pub next_remote_commitment_number: u64,
+ /// Optionally, a field proving that next_remote_commitment_number-1 has been revoked
+ pub data_loss_protect: OptionalField<DataLossProtect>,
}
/// An announcement_signatures message to be sent or received from a peer
-#[derive(PartialEq, Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub struct AnnouncementSignatures {
- pub(crate) channel_id: [u8; 32],
- pub(crate) short_channel_id: u64,
- pub(crate) node_signature: Signature,
- pub(crate) bitcoin_signature: Signature,
+ /// The channel ID
+ pub channel_id: [u8; 32],
+ /// The short channel ID
+ pub short_channel_id: u64,
+ /// A signature by the node key
+ pub node_signature: Signature,
+ /// A signature by the funding key
+ pub bitcoin_signature: Signature,
}
/// An address which can be used to connect to a remote peer
-#[derive(Clone, PartialEq, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub enum NetAddress {
/// An IPv4 address/port on which the peer is listening.
IPv4 {
port: u16,
},
/// An old-style Tor onion address/port on which the peer is listening.
- OnionV2 {
- /// The bytes (usually encoded in base32 with ".onion" appended)
- addr: [u8; 10],
- /// The port on which the node is listening
- port: u16,
- },
+ ///
+ /// This field is deprecated and the Tor network generally no longer supports V2 Onion
+ /// addresses. Thus, the details are not parsed here.
+ OnionV2([u8; 12]),
/// A new-style Tor onion address/port on which the peer is listening.
/// To create the human-readable "hostname", concatenate ed25519_pubkey, checksum, and version,
/// wrap as base32 and append ".onion".
/// The port on which the node is listening
port: u16,
},
+ /// A hostname/port on which the peer is listening.
+ Hostname {
+ /// The hostname on which the node is listening.
+ hostname: Hostname,
+ /// The port on which the node is listening.
+ port: u16,
+ },
}
impl NetAddress {
- fn get_id(&self) -> u8 {
+ /// Gets the ID of this address type. Addresses in node_announcement messages should be sorted
+ /// by this.
+ pub(crate) fn get_id(&self) -> u8 {
match self {
&NetAddress::IPv4 {..} => { 1 },
&NetAddress::IPv6 {..} => { 2 },
- &NetAddress::OnionV2 {..} => { 3 },
+ &NetAddress::OnionV2(_) => { 3 },
&NetAddress::OnionV3 {..} => { 4 },
+ &NetAddress::Hostname {..} => { 5 },
}
}
match self {
&NetAddress::IPv4 { .. } => { 6 },
&NetAddress::IPv6 { .. } => { 18 },
- &NetAddress::OnionV2 { .. } => { 12 },
+ &NetAddress::OnionV2(_) => { 12 },
&NetAddress::OnionV3 { .. } => { 37 },
+ // Consists of 1-byte hostname length, hostname bytes, and 2-byte port.
+ &NetAddress::Hostname { ref hostname, .. } => { u16::from(hostname.len()) + 3 },
}
}
+
+ /// The maximum length of any address descriptor, not including the 1-byte type.
+ /// This maximum length is reached by a hostname address descriptor:
+ /// a hostname with a maximum length of 255, its 1-byte length and a 2-byte port.
+ pub(crate) const MAX_LEN: u16 = 258;
}
impl Writeable for NetAddress {
- fn write<W: Writer>(&self, writer: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, writer: &mut W) -> Result<(), io::Error> {
match self {
&NetAddress::IPv4 { ref addr, ref port } => {
1u8.write(writer)?;
addr.write(writer)?;
port.write(writer)?;
},
- &NetAddress::OnionV2 { ref addr, ref port } => {
+ &NetAddress::OnionV2(bytes) => {
3u8.write(writer)?;
- addr.write(writer)?;
- port.write(writer)?;
+ bytes.write(writer)?;
},
&NetAddress::OnionV3 { ref ed25519_pubkey, ref checksum, ref version, ref port } => {
4u8.write(writer)?;
checksum.write(writer)?;
version.write(writer)?;
port.write(writer)?;
- }
+ },
+ &NetAddress::Hostname { ref hostname, ref port } => {
+ 5u8.write(writer)?;
+ hostname.write(writer)?;
+ port.write(writer)?;
+ },
}
Ok(())
}
}
-impl<R: ::std::io::Read> Readable<R> for Result<NetAddress, u8> {
- fn read(reader: &mut R) -> Result<Result<NetAddress, u8>, DecodeError> {
- let byte = <u8 as Readable<R>>::read(reader)?;
+impl Readable for Result<NetAddress, u8> {
+ fn read<R: Read>(reader: &mut R) -> Result<Result<NetAddress, u8>, DecodeError> {
+ let byte = <u8 as Readable>::read(reader)?;
match byte {
1 => {
Ok(Ok(NetAddress::IPv4 {
port: Readable::read(reader)?,
}))
},
- 3 => {
- Ok(Ok(NetAddress::OnionV2 {
- addr: Readable::read(reader)?,
- port: Readable::read(reader)?,
- }))
- },
+ 3 => Ok(Ok(NetAddress::OnionV2(Readable::read(reader)?))),
4 => {
Ok(Ok(NetAddress::OnionV3 {
ed25519_pubkey: Readable::read(reader)?,
port: Readable::read(reader)?,
}))
},
+ 5 => {
+ Ok(Ok(NetAddress::Hostname {
+ hostname: Readable::read(reader)?,
+ port: Readable::read(reader)?,
+ }))
+ },
_ => return Ok(Err(byte)),
}
}
}
-// Only exposed as broadcast of node_announcement should be filtered by node_id
+impl Readable for NetAddress {
+ fn read<R: Read>(reader: &mut R) -> Result<NetAddress, DecodeError> {
+ match Readable::read(reader) {
+ Ok(Ok(res)) => Ok(res),
+ Ok(Err(_)) => Err(DecodeError::UnknownVersion),
+ Err(e) => Err(e),
+ }
+ }
+}
+
+
/// The unsigned part of a node_announcement
-#[derive(PartialEq, Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UnsignedNodeAnnouncement {
- pub(crate) features: GlobalFeatures,
- pub(crate) timestamp: u32,
+ /// The advertised features
+ pub features: NodeFeatures,
+ /// A strictly monotonic announcement counter, with gaps allowed
+ pub timestamp: u32,
/// The node_id this announcement originated from (don't rebroadcast the node_announcement back
/// to this node).
- pub node_id: PublicKey,
- pub(crate) rgb: [u8; 3],
- pub(crate) alias: [u8; 32],
- /// List of addresses on which this node is reachable. Note that you may only have up to one
- /// address of each type, if you have more, they may be silently discarded or we may panic!
- pub(crate) addresses: Vec<NetAddress>,
+ pub node_id: PublicKey,
+ /// An RGB color for UI purposes
+ pub rgb: [u8; 3],
+ /// An alias, for UI purposes. This should be sanitized before use. There is no guarantee
+ /// of uniqueness.
+ pub alias: [u8; 32],
+ /// List of addresses on which this node is reachable
+ pub addresses: Vec<NetAddress>,
pub(crate) excess_address_data: Vec<u8>,
pub(crate) excess_data: Vec<u8>,
}
-#[derive(PartialEq, Clone)]
+#[derive(Clone, Debug, PartialEq)]
/// A node_announcement message to be sent or received from a peer
pub struct NodeAnnouncement {
- pub(crate) signature: Signature,
- pub(crate) contents: UnsignedNodeAnnouncement,
+ /// The signature by the node key
+ pub signature: Signature,
+ /// The actual content of the announcement
+ pub contents: UnsignedNodeAnnouncement,
}
-// Only exposed as broadcast of channel_announcement should be filtered by node_id
/// The unsigned part of a channel_announcement
-#[derive(PartialEq, Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub struct UnsignedChannelAnnouncement {
- pub(crate) features: GlobalFeatures,
- pub(crate) chain_hash: Sha256dHash,
- pub(crate) short_channel_id: u64,
+ /// The advertised channel features
+ pub features: ChannelFeatures,
+ /// The genesis hash of the blockchain where the channel is to be opened
+ pub chain_hash: BlockHash,
+ /// The short channel ID
+ pub short_channel_id: u64,
/// One of the two node_ids which are endpoints of this channel
- pub node_id_1: PublicKey,
+ pub node_id_1: PublicKey,
/// The other of the two node_ids which are endpoints of this channel
- pub node_id_2: PublicKey,
- pub(crate) bitcoin_key_1: PublicKey,
- pub(crate) bitcoin_key_2: PublicKey,
+ pub node_id_2: PublicKey,
+ /// The funding key for the first node
+ pub bitcoin_key_1: PublicKey,
+ /// The funding key for the second node
+ pub bitcoin_key_2: PublicKey,
pub(crate) excess_data: Vec<u8>,
}
/// A channel_announcement message to be sent or received from a peer
-#[derive(PartialEq, Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub struct ChannelAnnouncement {
- pub(crate) node_signature_1: Signature,
- pub(crate) node_signature_2: Signature,
- pub(crate) bitcoin_signature_1: Signature,
- pub(crate) bitcoin_signature_2: Signature,
- pub(crate) contents: UnsignedChannelAnnouncement,
-}
-
-#[derive(PartialEq, Clone, Debug)]
-pub(crate) struct UnsignedChannelUpdate {
- pub(crate) chain_hash: Sha256dHash,
- pub(crate) short_channel_id: u64,
- pub(crate) timestamp: u32,
- pub(crate) flags: u16,
- pub(crate) cltv_expiry_delta: u16,
- pub(crate) htlc_minimum_msat: u64,
- pub(crate) fee_base_msat: u32,
- pub(crate) fee_proportional_millionths: u32,
- pub(crate) excess_data: Vec<u8>,
+ /// Authentication of the announcement by the first public node
+ pub node_signature_1: Signature,
+ /// Authentication of the announcement by the second public node
+ pub node_signature_2: Signature,
+ /// Proof of funding UTXO ownership by the first public node
+ pub bitcoin_signature_1: Signature,
+ /// Proof of funding UTXO ownership by the second public node
+ pub bitcoin_signature_2: Signature,
+ /// The actual announcement
+ pub contents: UnsignedChannelAnnouncement,
+}
+
+/// The unsigned part of a channel_update
+#[derive(Clone, Debug, PartialEq)]
+pub struct UnsignedChannelUpdate {
+ /// The genesis hash of the blockchain where the channel is to be opened
+ pub chain_hash: BlockHash,
+ /// The short channel ID
+ pub short_channel_id: u64,
+ /// A strictly monotonic announcement counter, with gaps allowed, specific to this channel
+ pub timestamp: u32,
+ /// Channel flags
+ pub flags: u8,
+ /// The number of blocks such that if:
+ /// `incoming_htlc.cltv_expiry < outgoing_htlc.cltv_expiry + cltv_expiry_delta`
+ /// then we need to fail the HTLC backwards. When forwarding an HTLC, cltv_expiry_delta determines
+ /// the outgoing HTLC's minimum cltv_expiry value -- so, if an incoming HTLC comes in with a
+ /// cltv_expiry of 100000, and the node we're forwarding to has a cltv_expiry_delta value of 10,
+ /// then we'll check that the outgoing HTLC's cltv_expiry value is at least 100010 before
+ /// forwarding. Note that the HTLC sender is the one who originally sets this value when
+ /// constructing the route.
+ pub cltv_expiry_delta: u16,
+ /// The minimum HTLC size incoming to sender, in milli-satoshi
+ pub htlc_minimum_msat: u64,
+ /// The maximum HTLC value incoming to sender, in milli-satoshi. Used to be optional.
+ pub htlc_maximum_msat: u64,
+ /// The base HTLC fee charged by sender, in milli-satoshi
+ pub fee_base_msat: u32,
+ /// The amount to fee multiplier, in micro-satoshi
+ pub fee_proportional_millionths: u32,
+ /// Excess data which was signed as a part of the message which we do not (yet) understand how
+ /// to decode. This is stored to ensure forward-compatibility as new fields are added to the
+ /// lightning gossip
+ pub excess_data: Vec<u8>,
}
/// A channel_update message to be sent or received from a peer
-#[derive(PartialEq, Clone, Debug)]
+#[derive(Clone, Debug, PartialEq)]
pub struct ChannelUpdate {
- pub(crate) signature: Signature,
- pub(crate) contents: UnsignedChannelUpdate,
+ /// A signature of the channel update
+ pub signature: Signature,
+ /// The actual channel update
+ pub contents: UnsignedChannelUpdate,
+}
+
+/// A query_channel_range message is used to query a peer for channel
+/// UTXOs in a range of blocks. The recipient of a query makes a best
+/// effort to reply to the query using one or more reply_channel_range
+/// messages.
+#[derive(Clone, Debug, PartialEq)]
+pub struct QueryChannelRange {
+ /// The genesis hash of the blockchain being queried
+ pub chain_hash: BlockHash,
+ /// The height of the first block for the channel UTXOs being queried
+ pub first_blocknum: u32,
+ /// The number of blocks to include in the query results
+ pub number_of_blocks: u32,
+}
+
+/// A reply_channel_range message is a reply to a query_channel_range
+/// message. Multiple reply_channel_range messages can be sent in reply
+/// to a single query_channel_range message. The query recipient makes a
+/// best effort to respond based on their local network view which may
+/// not be a perfect view of the network. The short_channel_ids in the
+/// reply are encoded. We only support encoding_type=0 uncompressed
+/// serialization and do not support encoding_type=1 zlib serialization.
+#[derive(Clone, Debug, PartialEq)]
+pub struct ReplyChannelRange {
+ /// The genesis hash of the blockchain being queried
+ pub chain_hash: BlockHash,
+ /// The height of the first block in the range of the reply
+ pub first_blocknum: u32,
+ /// The number of blocks included in the range of the reply
+ pub number_of_blocks: u32,
+ /// True when this is the final reply for a query
+ pub sync_complete: bool,
+ /// The short_channel_ids in the channel range
+ pub short_channel_ids: Vec<u64>,
+}
+
+/// A query_short_channel_ids message is used to query a peer for
+/// routing gossip messages related to one or more short_channel_ids.
+/// The query recipient will reply with the latest, if available,
+/// channel_announcement, channel_update and node_announcement messages
+/// it maintains for the requested short_channel_ids followed by a
+/// reply_short_channel_ids_end message. The short_channel_ids sent in
+/// this query are encoded. We only support encoding_type=0 uncompressed
+/// serialization and do not support encoding_type=1 zlib serialization.
+#[derive(Clone, Debug, PartialEq)]
+pub struct QueryShortChannelIds {
+ /// The genesis hash of the blockchain being queried
+ pub chain_hash: BlockHash,
+ /// The short_channel_ids that are being queried
+ pub short_channel_ids: Vec<u64>,
+}
+
+/// A reply_short_channel_ids_end message is sent as a reply to a
+/// query_short_channel_ids message. The query recipient makes a best
+/// effort to respond based on their local network view which may not be
+/// a perfect view of the network.
+#[derive(Clone, Debug, PartialEq)]
+pub struct ReplyShortChannelIdsEnd {
+ /// The genesis hash of the blockchain that was queried
+ pub chain_hash: BlockHash,
+ /// Indicates if the query recipient maintains up-to-date channel
+ /// information for the chain_hash
+ pub full_information: bool,
+}
+
+/// A gossip_timestamp_filter message is used by a node to request
+/// gossip relay for messages in the requested time range when the
+/// gossip_queries feature has been negotiated.
+#[derive(Clone, Debug, PartialEq)]
+pub struct GossipTimestampFilter {
+ /// The genesis hash of the blockchain for channel and node information
+ pub chain_hash: BlockHash,
+ /// The starting unix timestamp
+ pub first_timestamp: u32,
+ /// The range of information in seconds
+ pub timestamp_range: u32,
+}
+
+/// Encoding type for data compression of collections in gossip queries.
+/// We do not support encoding_type=1 zlib serialization defined in BOLT #7.
+enum EncodingType {
+ Uncompressed = 0x00,
}
/// Used to put an error message in a LightningError
-#[derive(Clone)]
+#[derive(Clone, Debug)]
pub enum ErrorAction {
/// The peer took some action which made us think they were useless. Disconnect them.
DisconnectPeer {
msg: Option<ErrorMessage>
},
/// The peer did something harmless that we weren't able to process, just log and ignore
+ // New code should *not* use this. New code must use IgnoreAndLog, below!
IgnoreError,
+ /// The peer did something harmless that we weren't able to meaningfully process.
+ /// If the error is logged, log it at the given level.
+ IgnoreAndLog(logger::Level),
+ /// The peer provided us with a gossip message which we'd already seen. In most cases this
+ /// should be ignored, but it may result in the message being forwarded if it is a duplicate of
+ /// our own channel announcements.
+ IgnoreDuplicateGossip,
/// The peer did something incorrect. Tell them.
SendErrorMessage {
/// The message to send.
- msg: ErrorMessage
+ msg: ErrorMessage,
+ },
+ /// The peer did something incorrect. Tell them without closing any channels.
+ SendWarningMessage {
+ /// The message to send.
+ msg: WarningMessage,
+ /// The peer may have done something harmless that we weren't able to meaningfully process,
+ /// though we should still tell them about it.
+ /// If this event is logged, log it at the given level.
+ log_level: logger::Level,
},
}
/// An Err type for failure to process messages.
+#[derive(Clone, Debug)]
pub struct LightningError {
/// A human-readable message describing the error
- pub err: &'static str,
+ pub err: String,
/// The action which should be taken against the offending peer.
pub action: ErrorAction,
}
/// Struct used to return values from revoke_and_ack messages, containing a bunch of commitment
/// transaction updates if they were pending.
-#[derive(PartialEq, Clone)]
+#[derive(Clone, Debug, PartialEq)]
pub struct CommitmentUpdate {
/// update_add_htlc messages which should be sent
pub update_add_htlcs: Vec<UpdateAddHTLC>,
pub commitment_signed: CommitmentSigned,
}
-/// The information we received from a peer along the route of a payment we originated. This is
-/// returned by ChannelMessageHandler::handle_update_fail_htlc to be passed into
-/// RoutingMessageHandler::handle_htlc_fail_channel_update to update our network map.
-#[derive(Clone)]
-pub enum HTLCFailChannelUpdate {
- /// We received an error which included a full ChannelUpdate message.
- ChannelUpdateMessage {
- /// The unwrapped message we received
- msg: ChannelUpdate,
- },
- /// We received an error which indicated only that a channel has been closed
- ChannelClosed {
- /// The short_channel_id which has now closed.
- short_channel_id: u64,
- /// when this true, this channel should be permanently removed from the
- /// consideration. Otherwise, this channel can be restored as new channel_update is received
- is_permanent: bool,
- },
- /// We received an error which indicated only that a node has failed
- NodeFailure {
- /// The node_id that has failed.
- node_id: PublicKey,
- /// when this true, node should be permanently removed from the
- /// consideration. Otherwise, the channels connected to this node can be
- /// restored as new channel_update is received
- is_permanent: bool,
- }
-}
-
/// Messages could have optional fields to use with extended features
/// As we wish to serialize these differently from Option<T>s (Options get a tag byte, but
/// OptionalFeild simply gets Present if there are enough bytes to read into it), we have a
/// separate enum type for them.
-#[derive(Clone, PartialEq)]
+/// (C-not exported) due to a free generic in T
+#[derive(Clone, Debug, PartialEq)]
pub enum OptionalField<T> {
/// Optional field is included in message
Present(T),
///
/// Messages MAY be called in parallel when they originate from different their_node_ids, however
/// they MUST NOT be called in parallel when the two calls have the same their_node_id.
-pub trait ChannelMessageHandler : events::MessageSendEventsProvider + Send + Sync {
+pub trait ChannelMessageHandler : MessageSendEventsProvider {
//Channel init:
/// Handle an incoming open_channel message from the given peer.
- fn handle_open_channel(&self, their_node_id: &PublicKey, their_local_features: LocalFeatures, msg: &OpenChannel) -> Result<(), LightningError>;
+ fn handle_open_channel(&self, their_node_id: &PublicKey, their_features: InitFeatures, msg: &OpenChannel);
/// Handle an incoming accept_channel message from the given peer.
- fn handle_accept_channel(&self, their_node_id: &PublicKey, their_local_features: LocalFeatures, msg: &AcceptChannel) -> Result<(), LightningError>;
+ fn handle_accept_channel(&self, their_node_id: &PublicKey, their_features: InitFeatures, msg: &AcceptChannel);
/// Handle an incoming funding_created message from the given peer.
- fn handle_funding_created(&self, their_node_id: &PublicKey, msg: &FundingCreated) -> Result<(), LightningError>;
+ fn handle_funding_created(&self, their_node_id: &PublicKey, msg: &FundingCreated);
/// Handle an incoming funding_signed message from the given peer.
- fn handle_funding_signed(&self, their_node_id: &PublicKey, msg: &FundingSigned) -> Result<(), LightningError>;
- /// Handle an incoming funding_locked message from the given peer.
- fn handle_funding_locked(&self, their_node_id: &PublicKey, msg: &FundingLocked) -> Result<(), LightningError>;
+ fn handle_funding_signed(&self, their_node_id: &PublicKey, msg: &FundingSigned);
+ /// Handle an incoming channel_ready message from the given peer.
+ fn handle_channel_ready(&self, their_node_id: &PublicKey, msg: &ChannelReady);
// Channl close:
/// Handle an incoming shutdown message from the given peer.
- fn handle_shutdown(&self, their_node_id: &PublicKey, msg: &Shutdown) -> Result<(), LightningError>;
+ fn handle_shutdown(&self, their_node_id: &PublicKey, their_features: &InitFeatures, msg: &Shutdown);
/// Handle an incoming closing_signed message from the given peer.
- fn handle_closing_signed(&self, their_node_id: &PublicKey, msg: &ClosingSigned) -> Result<(), LightningError>;
+ fn handle_closing_signed(&self, their_node_id: &PublicKey, msg: &ClosingSigned);
// HTLC handling:
/// Handle an incoming update_add_htlc message from the given peer.
- fn handle_update_add_htlc(&self, their_node_id: &PublicKey, msg: &UpdateAddHTLC) -> Result<(), LightningError>;
+ fn handle_update_add_htlc(&self, their_node_id: &PublicKey, msg: &UpdateAddHTLC);
/// Handle an incoming update_fulfill_htlc message from the given peer.
- fn handle_update_fulfill_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFulfillHTLC) -> Result<(), LightningError>;
+ fn handle_update_fulfill_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFulfillHTLC);
/// Handle an incoming update_fail_htlc message from the given peer.
- fn handle_update_fail_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFailHTLC) -> Result<(), LightningError>;
+ fn handle_update_fail_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFailHTLC);
/// Handle an incoming update_fail_malformed_htlc message from the given peer.
- fn handle_update_fail_malformed_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFailMalformedHTLC) -> Result<(), LightningError>;
+ fn handle_update_fail_malformed_htlc(&self, their_node_id: &PublicKey, msg: &UpdateFailMalformedHTLC);
/// Handle an incoming commitment_signed message from the given peer.
- fn handle_commitment_signed(&self, their_node_id: &PublicKey, msg: &CommitmentSigned) -> Result<(), LightningError>;
+ fn handle_commitment_signed(&self, their_node_id: &PublicKey, msg: &CommitmentSigned);
/// Handle an incoming revoke_and_ack message from the given peer.
- fn handle_revoke_and_ack(&self, their_node_id: &PublicKey, msg: &RevokeAndACK) -> Result<(), LightningError>;
+ fn handle_revoke_and_ack(&self, their_node_id: &PublicKey, msg: &RevokeAndACK);
/// Handle an incoming update_fee message from the given peer.
- fn handle_update_fee(&self, their_node_id: &PublicKey, msg: &UpdateFee) -> Result<(), LightningError>;
+ fn handle_update_fee(&self, their_node_id: &PublicKey, msg: &UpdateFee);
// Channel-to-announce:
/// Handle an incoming announcement_signatures message from the given peer.
- fn handle_announcement_signatures(&self, their_node_id: &PublicKey, msg: &AnnouncementSignatures) -> Result<(), LightningError>;
+ fn handle_announcement_signatures(&self, their_node_id: &PublicKey, msg: &AnnouncementSignatures);
// Connection loss/reestablish:
/// Indicates a connection to the peer failed/an existing connection was lost. If no connection
fn peer_disconnected(&self, their_node_id: &PublicKey, no_connection_possible: bool);
/// Handle a peer reconnecting, possibly generating channel_reestablish message(s).
- fn peer_connected(&self, their_node_id: &PublicKey);
+ fn peer_connected(&self, their_node_id: &PublicKey, msg: &Init);
/// Handle an incoming channel_reestablish message from the given peer.
- fn handle_channel_reestablish(&self, their_node_id: &PublicKey, msg: &ChannelReestablish) -> Result<(), LightningError>;
+ fn handle_channel_reestablish(&self, their_node_id: &PublicKey, msg: &ChannelReestablish);
+
+ /// Handle an incoming channel update from the given peer.
+ fn handle_channel_update(&self, their_node_id: &PublicKey, msg: &ChannelUpdate);
// Error:
/// Handle an incoming error message from the given peer.
}
/// A trait to describe an object which can receive routing messages.
-pub trait RoutingMessageHandler : Send + Sync {
+///
+/// # Implementor DoS Warnings
+///
+/// For `gossip_queries` messages there are potential DoS vectors when handling
+/// inbound queries. Implementors using an on-disk network graph should be aware of
+/// repeated disk I/O for queries accessing different parts of the network graph.
+pub trait RoutingMessageHandler : MessageSendEventsProvider {
/// Handle an incoming node_announcement message, returning true if it should be forwarded on,
/// false or returning an Err otherwise.
fn handle_node_announcement(&self, msg: &NodeAnnouncement) -> Result<bool, LightningError>;
/// Handle an incoming channel_update message, returning true if it should be forwarded on,
/// false or returning an Err otherwise.
fn handle_channel_update(&self, msg: &ChannelUpdate) -> Result<bool, LightningError>;
- /// Handle some updates to the route graph that we learned due to an outbound failed payment.
- fn handle_htlc_fail_channel_update(&self, update: &HTLCFailChannelUpdate);
- /// Gets a subset of the channel announcements and updates required to dump our routing table
- /// to a remote node, starting at the short_channel_id indicated by starting_point and
- /// including batch_amount entries.
- fn get_next_channel_announcements(&self, starting_point: u64, batch_amount: u8) -> Vec<(ChannelAnnouncement, ChannelUpdate, ChannelUpdate)>;
- /// Gets a subset of the node announcements required to dump our routing table to a remote node,
- /// starting at the node *after* the provided publickey and including batch_amount entries.
+ /// Gets channel announcements and updates required to dump our routing table to a remote node,
+ /// starting at the short_channel_id indicated by starting_point and including announcements
+ /// for a single channel.
+ fn get_next_channel_announcement(&self, starting_point: u64) -> Option<(ChannelAnnouncement, Option<ChannelUpdate>, Option<ChannelUpdate>)>;
+ /// Gets a node announcement required to dump our routing table to a remote node, starting at
+ /// the node *after* the provided pubkey and including up to one announcement immediately
+ /// higher (as defined by <PublicKey as Ord>::cmp) than starting_point.
/// If None is provided for starting_point, we start at the first node.
- fn get_next_node_announcements(&self, starting_point: Option<&PublicKey>, batch_amount: u8) -> Vec<NodeAnnouncement>;
-}
-
-pub(crate) struct OnionRealm0HopData {
- pub(crate) short_channel_id: u64,
- pub(crate) amt_to_forward: u64,
- pub(crate) outgoing_cltv_value: u32,
- // 12 bytes of 0-padding
+ fn get_next_node_announcement(&self, starting_point: Option<&PublicKey>) -> Option<NodeAnnouncement>;
+ /// Called when a connection is established with a peer. This can be used to
+ /// perform routing table synchronization using a strategy defined by the
+ /// implementor.
+ fn peer_connected(&self, their_node_id: &PublicKey, init: &Init);
+ /// Handles the reply of a query we initiated to learn about channels
+ /// for a given range of blocks. We can expect to receive one or more
+ /// replies to a single query.
+ fn handle_reply_channel_range(&self, their_node_id: &PublicKey, msg: ReplyChannelRange) -> Result<(), LightningError>;
+ /// Handles the reply of a query we initiated asking for routing gossip
+ /// messages for a list of channels. We should receive this message when
+ /// a node has completed its best effort to send us the pertaining routing
+ /// gossip messages.
+ fn handle_reply_short_channel_ids_end(&self, their_node_id: &PublicKey, msg: ReplyShortChannelIdsEnd) -> Result<(), LightningError>;
+ /// Handles when a peer asks us to send a list of short_channel_ids
+ /// for the requested range of blocks.
+ fn handle_query_channel_range(&self, their_node_id: &PublicKey, msg: QueryChannelRange) -> Result<(), LightningError>;
+ /// Handles when a peer asks us to send routing gossip messages for a
+ /// list of short_channel_ids.
+ fn handle_query_short_channel_ids(&self, their_node_id: &PublicKey, msg: QueryShortChannelIds) -> Result<(), LightningError>;
+}
+
+/// A trait to describe an object that can receive onion messages.
+pub trait OnionMessageHandler : OnionMessageProvider {
+ /// Handle an incoming onion_message message from the given peer.
+ fn handle_onion_message(&self, peer_node_id: &PublicKey, msg: &OnionMessage);
+ /// Called when a connection is established with a peer. Can be used to track which peers
+ /// advertise onion message support and are online.
+ fn peer_connected(&self, their_node_id: &PublicKey, init: &Init);
+ /// Indicates a connection to the peer failed/an existing connection was lost. Allows handlers to
+ /// drop and refuse to forward onion messages to this peer.
+ fn peer_disconnected(&self, their_node_id: &PublicKey, no_connection_possible: bool);
}
mod fuzzy_internal_msgs {
+ use prelude::*;
+ use ln::{PaymentPreimage, PaymentSecret};
+
// These types aren't intended to be pub, but are exposed for direct fuzzing (as we deserialize
// them from untrusted input):
+ #[derive(Clone)]
+ pub(crate) struct FinalOnionHopData {
+ pub(crate) payment_secret: PaymentSecret,
+ /// The total value, in msat, of the payment as received by the ultimate recipient.
+ /// Message serialization may panic if this value is more than 21 million Bitcoin.
+ pub(crate) total_msat: u64,
+ }
+
+ pub(crate) enum OnionHopDataFormat {
+ Legacy { // aka Realm-0
+ short_channel_id: u64,
+ },
+ NonFinalNode {
+ short_channel_id: u64,
+ },
+ FinalNode {
+ payment_data: Option<FinalOnionHopData>,
+ keysend_preimage: Option<PaymentPreimage>,
+ },
+ }
- use super::OnionRealm0HopData;
pub struct OnionHopData {
- pub(crate) realm: u8,
- pub(crate) data: OnionRealm0HopData,
- pub(crate) hmac: [u8; 32],
+ pub(crate) format: OnionHopDataFormat,
+ /// The value, in msat, of the payment after this hop's fee is deducted.
+ /// Message serialization may panic if this value is more than 21 million Bitcoin.
+ pub(crate) amt_to_forward: u64,
+ pub(crate) outgoing_cltv_value: u32,
+ // 12 bytes of 0-padding for Legacy format
}
pub struct DecodedOnionErrorPacket {
pub(crate) pad: Vec<u8>,
}
}
-#[cfg(feature = "fuzztarget")]
+#[cfg(fuzzing)]
pub use self::fuzzy_internal_msgs::*;
-#[cfg(not(feature = "fuzztarget"))]
+#[cfg(not(fuzzing))]
pub(crate) use self::fuzzy_internal_msgs::*;
#[derive(Clone)]
pub(crate) hmac: [u8; 32],
}
+impl onion_utils::Packet for OnionPacket {
+ type Data = onion_utils::FixedSizeOnionPacket;
+ fn new(pubkey: PublicKey, hop_data: onion_utils::FixedSizeOnionPacket, hmac: [u8; 32]) -> Self {
+ Self {
+ version: 0,
+ public_key: Ok(pubkey),
+ hop_data: hop_data.0,
+ hmac,
+ }
+ }
+}
+
impl PartialEq for OnionPacket {
fn eq(&self, other: &OnionPacket) -> bool {
for (i, j) in self.hop_data.iter().zip(other.hop_data.iter()) {
}
}
-#[derive(Clone, PartialEq)]
+impl fmt::Debug for OnionPacket {
+ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+ f.write_fmt(format_args!("OnionPacket version {} with hmac {:?}", self.version, &self.hmac[..]))
+ }
+}
+
+#[derive(Clone, Debug, PartialEq)]
pub(crate) struct OnionErrorPacket {
// This really should be a constant size slice, but the spec lets these things be up to 128KB?
// (TODO) We limit it in decode to much lower...
pub(crate) data: Vec<u8>,
}
-impl Error for DecodeError {
- fn description(&self) -> &str {
- match *self {
- DecodeError::UnknownVersion => "Unknown realm byte in Onion packet",
- DecodeError::UnknownRequiredFeature => "Unknown required feature preventing decode",
- DecodeError::InvalidValue => "Nonsense bytes didn't map to the type they were interpreted as",
- DecodeError::ShortRead => "Packet extended beyond the provided bytes",
- DecodeError::ExtraAddressesPerType => "More than one address of a single type",
- DecodeError::BadLengthDescriptor => "A length descriptor in the packet didn't describe the later data correctly",
- DecodeError::Io(ref e) => e.description(),
- }
- }
-}
impl fmt::Display for DecodeError {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
- f.write_str(self.description())
- }
-}
-
-impl fmt::Debug for LightningError {
- fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
- f.write_str(self.err)
+ match *self {
+ DecodeError::UnknownVersion => f.write_str("Unknown realm byte in Onion packet"),
+ DecodeError::UnknownRequiredFeature => f.write_str("Unknown required feature preventing decode"),
+ DecodeError::InvalidValue => f.write_str("Nonsense bytes didn't map to the type they were interpreted as"),
+ DecodeError::ShortRead => f.write_str("Packet extended beyond the provided bytes"),
+ DecodeError::BadLengthDescriptor => f.write_str("A length descriptor in the packet didn't describe the later data correctly"),
+ DecodeError::Io(ref e) => fmt::Debug::fmt(e, f),
+ DecodeError::UnsupportedCompression => f.write_str("We don't support receiving messages with zlib-compressed fields"),
+ }
}
}
-impl From<::std::io::Error> for DecodeError {
- fn from(e: ::std::io::Error) -> Self {
- if e.kind() == ::std::io::ErrorKind::UnexpectedEof {
+impl From<io::Error> for DecodeError {
+ fn from(e: io::Error) -> Self {
+ if e.kind() == io::ErrorKind::UnexpectedEof {
DecodeError::ShortRead
} else {
- DecodeError::Io(e)
+ DecodeError::Io(e.kind())
}
}
}
impl Writeable for OptionalField<Script> {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
match *self {
OptionalField::Present(ref script) => {
// Note that Writeable for script includes the 16-bit length tag for us
}
}
-impl<R: Read> Readable<R> for OptionalField<Script> {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
- match <u16 as Readable<R>>::read(r) {
+impl Readable for OptionalField<Script> {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ match <u16 as Readable>::read(r) {
Ok(len) => {
let mut buf = vec![0; len as usize];
r.read_exact(&mut buf)?;
}
}
-impl_writeable_len_match!(AcceptChannel, {
- {AcceptChannel{ shutdown_scriptpubkey: OptionalField::Present(ref script), .. }, 270 + 2 + script.len()},
- {_, 270}
- }, {
+impl Writeable for OptionalField<u64> {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ match *self {
+ OptionalField::Present(ref value) => {
+ value.write(w)?;
+ },
+ OptionalField::Absent => {}
+ }
+ Ok(())
+ }
+}
+
+impl Readable for OptionalField<u64> {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let value: u64 = Readable::read(r)?;
+ Ok(OptionalField::Present(value))
+ }
+}
+
+
+impl_writeable_msg!(AcceptChannel, {
temporary_channel_id,
dust_limit_satoshis,
max_htlc_value_in_flight_msat,
max_accepted_htlcs,
funding_pubkey,
revocation_basepoint,
- payment_basepoint,
+ payment_point,
delayed_payment_basepoint,
htlc_basepoint,
first_per_commitment_point,
shutdown_scriptpubkey
+}, {
+ (1, channel_type, option),
});
-impl_writeable!(AnnouncementSignatures, 32+8+64*2, {
+impl_writeable_msg!(AnnouncementSignatures, {
channel_id,
short_channel_id,
node_signature,
bitcoin_signature
-});
+}, {});
impl Writeable for ChannelReestablish {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(if let OptionalField::Present(..) = self.data_loss_protect { 32+2*8+33+32 } else { 32+2*8 });
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.channel_id.write(w)?;
self.next_local_commitment_number.write(w)?;
self.next_remote_commitment_number.write(w)?;
}
}
-impl<R: Read> Readable<R> for ChannelReestablish{
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for ChannelReestablish{
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Self {
channel_id: Readable::read(r)?,
next_local_commitment_number: Readable::read(r)?,
next_remote_commitment_number: Readable::read(r)?,
data_loss_protect: {
- match <[u8; 32] as Readable<R>>::read(r) {
+ match <[u8; 32] as Readable>::read(r) {
Ok(your_last_per_commitment_secret) =>
OptionalField::Present(DataLossProtect {
your_last_per_commitment_secret,
}
}
-impl_writeable!(ClosingSigned, 32+8+64, {
- channel_id,
- fee_satoshis,
- signature
+impl_writeable_msg!(ClosingSigned,
+ { channel_id, fee_satoshis, signature },
+ { (1, fee_range, option) }
+);
+
+impl_writeable!(ClosingSignedFeeRange, {
+ min_fee_satoshis,
+ max_fee_satoshis
});
-impl_writeable_len_match!(CommitmentSigned, {
- { CommitmentSigned { ref htlc_signatures, .. }, 32+64+2+htlc_signatures.len()*64 }
- }, {
+impl_writeable_msg!(CommitmentSigned, {
channel_id,
signature,
htlc_signatures
-});
+}, {});
-impl_writeable_len_match!(DecodedOnionErrorPacket, {
- { DecodedOnionErrorPacket { ref failuremsg, ref pad, .. }, 32 + 4 + failuremsg.len() + pad.len() }
- }, {
+impl_writeable!(DecodedOnionErrorPacket, {
hmac,
failuremsg,
pad
});
-impl_writeable!(FundingCreated, 32+32+2+64, {
+impl_writeable_msg!(FundingCreated, {
temporary_channel_id,
funding_txid,
funding_output_index,
signature
-});
+}, {});
-impl_writeable!(FundingSigned, 32+64, {
+impl_writeable_msg!(FundingSigned, {
channel_id,
signature
-});
+}, {});
-impl_writeable!(FundingLocked, 32+33, {
+impl_writeable_msg!(ChannelReady, {
channel_id,
- next_per_commitment_point
-});
-
-impl_writeable_len_match!(GlobalFeatures, {
- { GlobalFeatures { ref flags }, flags.len() + 2 }
- }, {
- flags
+ next_per_commitment_point,
+}, {
+ (1, short_channel_id_alias, option),
});
-impl_writeable_len_match!(LocalFeatures, {
- { LocalFeatures { ref flags }, flags.len() + 2 }
- }, {
- flags
-});
+impl Writeable for Init {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ // global_features gets the bottom 13 bits of our features, and local_features gets all of
+ // our relevant feature bits. This keeps us compatible with old nodes.
+ self.features.write_up_to_13(w)?;
+ self.features.write(w)?;
+ encode_tlv_stream!(w, {
+ (3, self.remote_network_address, option)
+ });
+ Ok(())
+ }
+}
-impl_writeable_len_match!(Init, {
- { Init { ref global_features, ref local_features }, global_features.flags.len() + local_features.flags.len() + 4 }
- }, {
- global_features,
- local_features
-});
+impl Readable for Init {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let global_features: InitFeatures = Readable::read(r)?;
+ let features: InitFeatures = Readable::read(r)?;
+ let mut remote_network_address: Option<NetAddress> = None;
+ decode_tlv_stream!(r, {
+ (3, remote_network_address, option)
+ });
+ Ok(Init {
+ features: features.or(global_features),
+ remote_network_address,
+ })
+ }
+}
-impl_writeable_len_match!(OpenChannel, {
- { OpenChannel { shutdown_scriptpubkey: OptionalField::Present(ref script), .. }, 319 + 2 + script.len() },
- { _, 319 }
- }, {
+impl_writeable_msg!(OpenChannel, {
chain_hash,
temporary_channel_id,
funding_satoshis,
max_accepted_htlcs,
funding_pubkey,
revocation_basepoint,
- payment_basepoint,
+ payment_point,
delayed_payment_basepoint,
htlc_basepoint,
first_per_commitment_point,
channel_flags,
shutdown_scriptpubkey
+}, {
+ (1, channel_type, option),
});
-impl_writeable!(RevokeAndACK, 32+32+33, {
+impl_writeable_msg!(RevokeAndACK, {
channel_id,
per_commitment_secret,
next_per_commitment_point
-});
+}, {});
-impl_writeable_len_match!(Shutdown, {
- { Shutdown { ref scriptpubkey, .. }, 32 + 2 + scriptpubkey.len() }
- }, {
+impl_writeable_msg!(Shutdown, {
channel_id,
scriptpubkey
-});
+}, {});
-impl_writeable_len_match!(UpdateFailHTLC, {
- { UpdateFailHTLC { ref reason, .. }, 32 + 10 + reason.data.len() }
- }, {
+impl_writeable_msg!(UpdateFailHTLC, {
channel_id,
htlc_id,
reason
-});
+}, {});
-impl_writeable!(UpdateFailMalformedHTLC, 32+8+32+2, {
+impl_writeable_msg!(UpdateFailMalformedHTLC, {
channel_id,
htlc_id,
sha256_of_onion,
failure_code
-});
+}, {});
-impl_writeable!(UpdateFee, 32+4, {
+impl_writeable_msg!(UpdateFee, {
channel_id,
feerate_per_kw
-});
+}, {});
-impl_writeable!(UpdateFulfillHTLC, 32+8+32, {
+impl_writeable_msg!(UpdateFulfillHTLC, {
channel_id,
htlc_id,
payment_preimage
-});
+}, {});
-impl_writeable_len_match!(OnionErrorPacket, {
- { OnionErrorPacket { ref data, .. }, 2 + data.len() }
- }, {
+// Note that this is written as a part of ChannelManager objects, and thus cannot change its
+// serialization format in a way which assumes we know the total serialized length/message end
+// position.
+impl_writeable!(OnionErrorPacket, {
data
});
+// Note that this is written as a part of ChannelManager objects, and thus cannot change its
+// serialization format in a way which assumes we know the total serialized length/message end
+// position.
impl Writeable for OnionPacket {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(1 + 33 + 20*65 + 32);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.version.write(w)?;
match self.public_key {
Ok(pubkey) => pubkey.write(w)?,
}
}
-impl<R: Read> Readable<R> for OnionPacket {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for OnionPacket {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(OnionPacket {
version: Readable::read(r)?,
public_key: {
}
}
-impl_writeable!(UpdateAddHTLC, 32+8+8+32+4+1366, {
+impl_writeable_msg!(UpdateAddHTLC, {
channel_id,
htlc_id,
amount_msat,
payment_hash,
cltv_expiry,
onion_routing_packet
-});
+}, {});
+
+impl Readable for OnionMessage {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let blinding_point: PublicKey = Readable::read(r)?;
+ let len: u16 = Readable::read(r)?;
+ let mut packet_reader = FixedLengthReader::new(r, len as u64);
+ let onion_routing_packet: onion_message::Packet = <onion_message::Packet as LengthReadable>::read(&mut packet_reader)?;
+ Ok(Self {
+ blinding_point,
+ onion_routing_packet,
+ })
+ }
+}
-impl Writeable for OnionRealm0HopData {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32);
- self.short_channel_id.write(w)?;
- self.amt_to_forward.write(w)?;
- self.outgoing_cltv_value.write(w)?;
- w.write_all(&[0;12])?;
+impl Writeable for OnionMessage {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ self.blinding_point.write(w)?;
+ let onion_packet_len = self.onion_routing_packet.serialized_length();
+ (onion_packet_len as u16).write(w)?;
+ self.onion_routing_packet.write(w)?;
Ok(())
}
}
-impl<R: Read> Readable<R> for OnionRealm0HopData {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
- Ok(OnionRealm0HopData {
- short_channel_id: Readable::read(r)?,
- amt_to_forward: Readable::read(r)?,
- outgoing_cltv_value: {
- let v: u32 = Readable::read(r)?;
- r.read_exact(&mut [0; 12])?;
- v
- }
- })
+impl Writeable for FinalOnionHopData {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ self.payment_secret.0.write(w)?;
+ HighZeroBytesDroppedBigSize(self.total_msat).write(w)
+ }
+}
+
+impl Readable for FinalOnionHopData {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let secret: [u8; 32] = Readable::read(r)?;
+ let amt: HighZeroBytesDroppedBigSize<u64> = Readable::read(r)?;
+ Ok(Self { payment_secret: PaymentSecret(secret), total_msat: amt.0 })
}
}
impl Writeable for OnionHopData {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(65);
- self.realm.write(w)?;
- self.data.write(w)?;
- self.hmac.write(w)?;
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ match self.format {
+ OnionHopDataFormat::Legacy { short_channel_id } => {
+ 0u8.write(w)?;
+ short_channel_id.write(w)?;
+ self.amt_to_forward.write(w)?;
+ self.outgoing_cltv_value.write(w)?;
+ w.write_all(&[0;12])?;
+ },
+ OnionHopDataFormat::NonFinalNode { short_channel_id } => {
+ encode_varint_length_prefixed_tlv!(w, {
+ (2, HighZeroBytesDroppedBigSize(self.amt_to_forward), required),
+ (4, HighZeroBytesDroppedBigSize(self.outgoing_cltv_value), required),
+ (6, short_channel_id, required)
+ });
+ },
+ OnionHopDataFormat::FinalNode { ref payment_data, ref keysend_preimage } => {
+ encode_varint_length_prefixed_tlv!(w, {
+ (2, HighZeroBytesDroppedBigSize(self.amt_to_forward), required),
+ (4, HighZeroBytesDroppedBigSize(self.outgoing_cltv_value), required),
+ (8, payment_data, option),
+ (5482373484, keysend_preimage, option)
+ });
+ },
+ }
Ok(())
}
}
-impl<R: Read> Readable<R> for OnionHopData {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
- Ok(OnionHopData {
- realm: {
- let r: u8 = Readable::read(r)?;
- if r != 0 {
- return Err(DecodeError::UnknownVersion);
+impl Readable for OnionHopData {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let b: BigSize = Readable::read(r)?;
+ const LEGACY_ONION_HOP_FLAG: u64 = 0;
+ let (format, amt, cltv_value) = if b.0 != LEGACY_ONION_HOP_FLAG {
+ let mut rd = FixedLengthReader::new(r, b.0);
+ let mut amt = HighZeroBytesDroppedBigSize(0u64);
+ let mut cltv_value = HighZeroBytesDroppedBigSize(0u32);
+ let mut short_id: Option<u64> = None;
+ let mut payment_data: Option<FinalOnionHopData> = None;
+ let mut keysend_preimage: Option<PaymentPreimage> = None;
+ decode_tlv_stream!(&mut rd, {
+ (2, amt, required),
+ (4, cltv_value, required),
+ (6, short_id, option),
+ (8, payment_data, option),
+ // See https://github.com/lightning/blips/blob/master/blip-0003.md
+ (5482373484, keysend_preimage, option)
+ });
+ rd.eat_remaining().map_err(|_| DecodeError::ShortRead)?;
+ let format = if let Some(short_channel_id) = short_id {
+ if payment_data.is_some() { return Err(DecodeError::InvalidValue); }
+ OnionHopDataFormat::NonFinalNode {
+ short_channel_id,
}
- r
- },
- data: Readable::read(r)?,
- hmac: Readable::read(r)?,
+ } else {
+ if let &Some(ref data) = &payment_data {
+ if data.total_msat > MAX_VALUE_MSAT {
+ return Err(DecodeError::InvalidValue);
+ }
+ }
+ OnionHopDataFormat::FinalNode {
+ payment_data,
+ keysend_preimage,
+ }
+ };
+ (format, amt.0, cltv_value.0)
+ } else {
+ let format = OnionHopDataFormat::Legacy {
+ short_channel_id: Readable::read(r)?,
+ };
+ let amt: u64 = Readable::read(r)?;
+ let cltv_value: u32 = Readable::read(r)?;
+ r.read_exact(&mut [0; 12])?;
+ (format, amt, cltv_value)
+ };
+
+ if amt > MAX_VALUE_MSAT {
+ return Err(DecodeError::InvalidValue);
+ }
+ Ok(OnionHopData {
+ format,
+ amt_to_forward: amt,
+ outgoing_cltv_value: cltv_value,
})
}
}
+// ReadableArgs because we need onion_utils::decode_next_hop to accommodate payment packets and
+// onion message packets.
+impl ReadableArgs<()> for OnionHopData {
+ fn read<R: Read>(r: &mut R, _arg: ()) -> Result<Self, DecodeError> {
+ <Self as Readable>::read(r)
+ }
+}
+
impl Writeable for Ping {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(self.byteslen as usize + 4);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.ponglen.write(w)?;
vec![0u8; self.byteslen as usize].write(w)?; // size-unchecked write
Ok(())
}
}
-impl<R: Read> Readable<R> for Ping {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for Ping {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Ping {
ponglen: Readable::read(r)?,
byteslen: {
}
impl Writeable for Pong {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(self.byteslen as usize + 2);
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
vec![0u8; self.byteslen as usize].write(w)?; // size-unchecked write
Ok(())
}
}
-impl<R: Read> Readable<R> for Pong {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for Pong {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Pong {
byteslen: {
let byteslen = Readable::read(r)?;
}
impl Writeable for UnsignedChannelAnnouncement {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(2 + 2*32 + 4*33 + self.features.flags.len() + self.excess_data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.features.write(w)?;
self.chain_hash.write(w)?;
self.short_channel_id.write(w)?;
}
}
-impl<R: Read> Readable<R> for UnsignedChannelAnnouncement {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for UnsignedChannelAnnouncement {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Self {
- features: {
- let f: GlobalFeatures = Readable::read(r)?;
- if f.requires_unknown_bits() {
- return Err(DecodeError::UnknownRequiredFeature);
- }
- f
- },
+ features: Readable::read(r)?,
chain_hash: Readable::read(r)?,
short_channel_id: Readable::read(r)?,
node_id_1: Readable::read(r)?,
node_id_2: Readable::read(r)?,
bitcoin_key_1: Readable::read(r)?,
bitcoin_key_2: Readable::read(r)?,
- excess_data: {
- let mut excess_data = vec![];
- r.read_to_end(&mut excess_data)?;
- excess_data
- },
+ excess_data: read_to_end(r)?,
})
}
}
-impl_writeable_len_match!(ChannelAnnouncement, {
- { ChannelAnnouncement { contents: UnsignedChannelAnnouncement {ref features, ref excess_data, ..}, .. },
- 2 + 2*32 + 4*33 + features.flags.len() + excess_data.len() + 4*64 }
- }, {
+impl_writeable!(ChannelAnnouncement, {
node_signature_1,
node_signature_2,
bitcoin_signature_1,
});
impl Writeable for UnsignedChannelUpdate {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(64 + self.excess_data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ // `message_flags` used to indicate presence of `htlc_maximum_msat`, but was deprecated in the spec.
+ const MESSAGE_FLAGS: u8 = 1;
self.chain_hash.write(w)?;
self.short_channel_id.write(w)?;
self.timestamp.write(w)?;
- self.flags.write(w)?;
+ let all_flags = self.flags as u16 | ((MESSAGE_FLAGS as u16) << 8);
+ all_flags.write(w)?;
self.cltv_expiry_delta.write(w)?;
self.htlc_minimum_msat.write(w)?;
self.fee_base_msat.write(w)?;
self.fee_proportional_millionths.write(w)?;
+ self.htlc_maximum_msat.write(w)?;
w.write_all(&self.excess_data[..])?;
Ok(())
}
}
-impl<R: Read> Readable<R> for UnsignedChannelUpdate {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for UnsignedChannelUpdate {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Self {
chain_hash: Readable::read(r)?,
short_channel_id: Readable::read(r)?,
timestamp: Readable::read(r)?,
- flags: Readable::read(r)?,
+ flags: {
+ let flags: u16 = Readable::read(r)?;
+ // Note: we ignore the `message_flags` for now, since it was deprecated by the spec.
+ flags as u8
+ },
cltv_expiry_delta: Readable::read(r)?,
htlc_minimum_msat: Readable::read(r)?,
fee_base_msat: Readable::read(r)?,
fee_proportional_millionths: Readable::read(r)?,
- excess_data: {
- let mut excess_data = vec![];
- r.read_to_end(&mut excess_data)?;
- excess_data
- },
+ htlc_maximum_msat: Readable::read(r)?,
+ excess_data: read_to_end(r)?,
})
}
}
-impl_writeable_len_match!(ChannelUpdate, {
- { ChannelUpdate { contents: UnsignedChannelUpdate {ref excess_data, ..}, .. },
- 64 + excess_data.len() + 64 }
- }, {
+impl_writeable!(ChannelUpdate, {
signature,
contents
});
impl Writeable for ErrorMessage {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(32 + 2 + self.data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ self.channel_id.write(w)?;
+ (self.data.len() as u16).write(w)?;
+ w.write_all(self.data.as_bytes())?;
+ Ok(())
+ }
+}
+
+impl Readable for ErrorMessage {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ Ok(Self {
+ channel_id: Readable::read(r)?,
+ data: {
+ let sz: usize = <u16 as Readable>::read(r)? as usize;
+ let mut data = Vec::with_capacity(sz);
+ data.resize(sz, 0);
+ r.read_exact(&mut data)?;
+ match String::from_utf8(data) {
+ Ok(s) => s,
+ Err(_) => return Err(DecodeError::InvalidValue),
+ }
+ }
+ })
+ }
+}
+
+impl Writeable for WarningMessage {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.channel_id.write(w)?;
(self.data.len() as u16).write(w)?;
w.write_all(self.data.as_bytes())?;
}
}
-impl<R: Read> Readable<R> for ErrorMessage {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
+impl Readable for WarningMessage {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
Ok(Self {
channel_id: Readable::read(r)?,
data: {
- let mut sz: usize = <u16 as Readable<R>>::read(r)? as usize;
- let mut data = vec![];
- let data_len = r.read_to_end(&mut data)?;
- sz = cmp::min(data_len, sz);
- match String::from_utf8(data[..sz as usize].to_vec()) {
+ let sz: usize = <u16 as Readable>::read(r)? as usize;
+ let mut data = Vec::with_capacity(sz);
+ data.resize(sz, 0);
+ r.read_exact(&mut data)?;
+ match String::from_utf8(data) {
Ok(s) => s,
Err(_) => return Err(DecodeError::InvalidValue),
}
}
impl Writeable for UnsignedNodeAnnouncement {
- fn write<W: Writer>(&self, w: &mut W) -> Result<(), ::std::io::Error> {
- w.size_hint(64 + 76 + self.features.flags.len() + self.addresses.len()*38 + self.excess_address_data.len() + self.excess_data.len());
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
self.features.write(w)?;
self.timestamp.write(w)?;
self.node_id.write(w)?;
w.write_all(&self.rgb)?;
self.alias.write(w)?;
- let mut addrs_to_encode = self.addresses.clone();
- addrs_to_encode.sort_unstable_by(|a, b| { a.get_id().cmp(&b.get_id()) });
- addrs_to_encode.dedup_by(|a, b| { a.get_id() == b.get_id() });
let mut addr_len = 0;
- for addr in &addrs_to_encode {
+ for addr in self.addresses.iter() {
addr_len += 1 + addr.len();
}
(addr_len + self.excess_address_data.len() as u16).write(w)?;
- for addr in addrs_to_encode {
+ for addr in self.addresses.iter() {
addr.write(w)?;
}
w.write_all(&self.excess_address_data[..])?;
}
}
-impl<R: Read> Readable<R> for UnsignedNodeAnnouncement {
- fn read(r: &mut R) -> Result<Self, DecodeError> {
- let features: GlobalFeatures = Readable::read(r)?;
- if features.requires_unknown_bits() {
- return Err(DecodeError::UnknownRequiredFeature);
- }
+impl Readable for UnsignedNodeAnnouncement {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let features: NodeFeatures = Readable::read(r)?;
let timestamp: u32 = Readable::read(r)?;
let node_id: PublicKey = Readable::read(r)?;
let mut rgb = [0; 3];
let alias: [u8; 32] = Readable::read(r)?;
let addr_len: u16 = Readable::read(r)?;
- let mut addresses: Vec<NetAddress> = Vec::with_capacity(4);
+ let mut addresses: Vec<NetAddress> = Vec::new();
let mut addr_readpos = 0;
let mut excess = false;
let mut excess_byte = 0;
if addr_len <= addr_readpos { break; }
match Readable::read(r) {
Ok(Ok(addr)) => {
- match addr {
- NetAddress::IPv4 { .. } => {
- if addresses.len() > 0 {
- return Err(DecodeError::ExtraAddressesPerType);
- }
- },
- NetAddress::IPv6 { .. } => {
- if addresses.len() > 1 || (addresses.len() == 1 && addresses[0].get_id() != 1) {
- return Err(DecodeError::ExtraAddressesPerType);
- }
- },
- NetAddress::OnionV2 { .. } => {
- if addresses.len() > 2 || (addresses.len() > 0 && addresses.last().unwrap().get_id() > 2) {
- return Err(DecodeError::ExtraAddressesPerType);
- }
- },
- NetAddress::OnionV3 { .. } => {
- if addresses.len() > 3 || (addresses.len() > 0 && addresses.last().unwrap().get_id() > 3) {
- return Err(DecodeError::ExtraAddressesPerType);
- }
- },
- }
if addr_len < addr_readpos + 1 + addr.len() {
return Err(DecodeError::BadLengthDescriptor);
}
}
Vec::new()
};
- r.read_to_end(&mut excess_data)?;
+ excess_data.extend(read_to_end(r)?.iter());
Ok(UnsignedNodeAnnouncement {
features,
timestamp,
}
}
-impl_writeable_len_match!(NodeAnnouncement, {
- { NodeAnnouncement { contents: UnsignedNodeAnnouncement { ref features, ref addresses, ref excess_address_data, ref excess_data, ..}, .. },
- 64 + 76 + features.flags.len() + addresses.len()*38 + excess_address_data.len() + excess_data.len() }
- }, {
+impl_writeable!(NodeAnnouncement, {
signature,
contents
});
+impl Readable for QueryShortChannelIds {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let chain_hash: BlockHash = Readable::read(r)?;
+
+ let encoding_len: u16 = Readable::read(r)?;
+ let encoding_type: u8 = Readable::read(r)?;
+
+ // Must be encoding_type=0 uncompressed serialization. We do not
+ // support encoding_type=1 zlib serialization.
+ if encoding_type != EncodingType::Uncompressed as u8 {
+ return Err(DecodeError::UnsupportedCompression);
+ }
+
+ // We expect the encoding_len to always includes the 1-byte
+ // encoding_type and that short_channel_ids are 8-bytes each
+ if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
+ return Err(DecodeError::InvalidValue);
+ }
+
+ // Read short_channel_ids (8-bytes each), for the u16 encoding_len
+ // less the 1-byte encoding_type
+ let short_channel_id_count: u16 = (encoding_len - 1)/8;
+ let mut short_channel_ids = Vec::with_capacity(short_channel_id_count as usize);
+ for _ in 0..short_channel_id_count {
+ short_channel_ids.push(Readable::read(r)?);
+ }
+
+ Ok(QueryShortChannelIds {
+ chain_hash,
+ short_channel_ids,
+ })
+ }
+}
+
+impl Writeable for QueryShortChannelIds {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ // Calculated from 1-byte encoding_type plus 8-bytes per short_channel_id
+ let encoding_len: u16 = 1 + self.short_channel_ids.len() as u16 * 8;
+
+ self.chain_hash.write(w)?;
+ encoding_len.write(w)?;
+
+ // We only support type=0 uncompressed serialization
+ (EncodingType::Uncompressed as u8).write(w)?;
+
+ for scid in self.short_channel_ids.iter() {
+ scid.write(w)?;
+ }
+
+ Ok(())
+ }
+}
+
+impl_writeable_msg!(ReplyShortChannelIdsEnd, {
+ chain_hash,
+ full_information,
+}, {});
+
+impl QueryChannelRange {
+ /**
+ * Calculates the overflow safe ending block height for the query.
+ * Overflow returns `0xffffffff`, otherwise returns `first_blocknum + number_of_blocks`
+ */
+ pub fn end_blocknum(&self) -> u32 {
+ match self.first_blocknum.checked_add(self.number_of_blocks) {
+ Some(block) => block,
+ None => u32::max_value(),
+ }
+ }
+}
+
+impl_writeable_msg!(QueryChannelRange, {
+ chain_hash,
+ first_blocknum,
+ number_of_blocks
+}, {});
+
+impl Readable for ReplyChannelRange {
+ fn read<R: Read>(r: &mut R) -> Result<Self, DecodeError> {
+ let chain_hash: BlockHash = Readable::read(r)?;
+ let first_blocknum: u32 = Readable::read(r)?;
+ let number_of_blocks: u32 = Readable::read(r)?;
+ let sync_complete: bool = Readable::read(r)?;
+
+ let encoding_len: u16 = Readable::read(r)?;
+ let encoding_type: u8 = Readable::read(r)?;
+
+ // Must be encoding_type=0 uncompressed serialization. We do not
+ // support encoding_type=1 zlib serialization.
+ if encoding_type != EncodingType::Uncompressed as u8 {
+ return Err(DecodeError::UnsupportedCompression);
+ }
+
+ // We expect the encoding_len to always includes the 1-byte
+ // encoding_type and that short_channel_ids are 8-bytes each
+ if encoding_len == 0 || (encoding_len - 1) % 8 != 0 {
+ return Err(DecodeError::InvalidValue);
+ }
+
+ // Read short_channel_ids (8-bytes each), for the u16 encoding_len
+ // less the 1-byte encoding_type
+ let short_channel_id_count: u16 = (encoding_len - 1)/8;
+ let mut short_channel_ids = Vec::with_capacity(short_channel_id_count as usize);
+ for _ in 0..short_channel_id_count {
+ short_channel_ids.push(Readable::read(r)?);
+ }
+
+ Ok(ReplyChannelRange {
+ chain_hash,
+ first_blocknum,
+ number_of_blocks,
+ sync_complete,
+ short_channel_ids
+ })
+ }
+}
+
+impl Writeable for ReplyChannelRange {
+ fn write<W: Writer>(&self, w: &mut W) -> Result<(), io::Error> {
+ let encoding_len: u16 = 1 + self.short_channel_ids.len() as u16 * 8;
+ self.chain_hash.write(w)?;
+ self.first_blocknum.write(w)?;
+ self.number_of_blocks.write(w)?;
+ self.sync_complete.write(w)?;
+
+ encoding_len.write(w)?;
+ (EncodingType::Uncompressed as u8).write(w)?;
+ for scid in self.short_channel_ids.iter() {
+ scid.write(w)?;
+ }
+
+ Ok(())
+ }
+}
+
+impl_writeable_msg!(GossipTimestampFilter, {
+ chain_hash,
+ first_timestamp,
+ timestamp_range,
+}, {});
+
#[cfg(test)]
mod tests {
use hex;
+ use ln::{PaymentPreimage, PaymentHash, PaymentSecret};
+ use ln::features::{ChannelFeatures, ChannelTypeFeatures, InitFeatures, NodeFeatures};
use ln::msgs;
- use ln::msgs::{GlobalFeatures, LocalFeatures, OptionalField, OnionErrorPacket};
- use ln::channelmanager::{PaymentPreimage, PaymentHash};
- use util::ser::Writeable;
+ use ln::msgs::{FinalOnionHopData, OptionalField, OnionErrorPacket, OnionHopDataFormat};
+ use util::ser::{Writeable, Readable, Hostname};
- use bitcoin_hashes::sha256d::Hash as Sha256dHash;
- use bitcoin_hashes::hex::FromHex;
+ use bitcoin::hashes::hex::FromHex;
use bitcoin::util::address::Address;
use bitcoin::network::constants::Network;
use bitcoin::blockdata::script::Builder;
use bitcoin::blockdata::opcodes;
+ use bitcoin::hash_types::{Txid, BlockHash};
- use secp256k1::key::{PublicKey,SecretKey};
- use secp256k1::{Secp256k1, Message};
+ use bitcoin::secp256k1::{PublicKey,SecretKey};
+ use bitcoin::secp256k1::{Secp256k1, Message};
+
+ use io::{self, Cursor};
+ use prelude::*;
+ use core::convert::TryFrom;
#[test]
fn encoding_channel_reestablish_no_secret() {
($privkey: expr, $ctx: expr, $string: expr) => {
{
let sighash = Message::from_slice(&$string.into_bytes()[..]).unwrap();
- $ctx.sign(&sighash, &$privkey)
+ $ctx.sign_ecdsa(&sighash, &$privkey)
}
}
}
assert_eq!(encoded_value, hex::decode("040000000000000005000000000000000600000000000000070000000000000000083a840000034dd977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073acf9953cef4700860f5967838eba2bae89288ad188ebf8b20bf995c3ea53a26df1876d0a3a0e13172ba286a673140190c02ba9da60a2e43a745188c8a83c7f3ef").unwrap());
}
- fn do_encoding_channel_announcement(unknown_features_bits: bool, non_bitcoin_chain_hash: bool, excess_data: bool) {
+ fn do_encoding_channel_announcement(unknown_features_bits: bool, excess_data: bool) {
let secp_ctx = Secp256k1::new();
let (privkey_1, pubkey_1) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
let (privkey_2, pubkey_2) = get_keys_from!("0202020202020202020202020202020202020202020202020202020202020202", secp_ctx);
let sig_2 = get_sig_on!(privkey_2, secp_ctx, String::from("01010101010101010101010101010101"));
let sig_3 = get_sig_on!(privkey_3, secp_ctx, String::from("01010101010101010101010101010101"));
let sig_4 = get_sig_on!(privkey_4, secp_ctx, String::from("01010101010101010101010101010101"));
- let mut features = GlobalFeatures::new();
+ let mut features = ChannelFeatures::known();
if unknown_features_bits {
- features.flags = vec![0xFF, 0xFF];
+ features = ChannelFeatures::from_le_bytes(vec![0xFF, 0xFF]);
}
let unsigned_channel_announcement = msgs::UnsignedChannelAnnouncement {
features,
- chain_hash: if !non_bitcoin_chain_hash { Sha256dHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap() } else { Sha256dHash::from_hex("000000000933ea01ad0ee984209779baaec3ced90fa3f408719526f8d77f4943").unwrap() },
+ chain_hash: BlockHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap(),
short_channel_id: 2316138423780173,
node_id_1: pubkey_1,
node_id_2: pubkey_2,
} else {
target_value.append(&mut hex::decode("0000").unwrap());
}
- if non_bitcoin_chain_hash {
- target_value.append(&mut hex::decode("43497fd7f826957108f4a30fd9cec3aeba79972084e90ead01ea330900000000").unwrap());
- } else {
- target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
- }
+ target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
target_value.append(&mut hex::decode("00083a840000034d031b84c5567b126440995d3ed5aaba0565d71e1834604819ff9c17f5e9d5dd078f024d4b6cd1361032ca9bd2aeb9d900aa4d45d9ead80ac9423374c451a7254d076602531fe6068134503d2723133227c867ac8fa6c83c537e9a44c3c5bdbdcb1fe33703462779ad4aad39514614751a71085f2f10e1c7a593e4e030efb5b8721ce55b0b").unwrap());
if excess_data {
target_value.append(&mut hex::decode("0a00001400001e000028").unwrap());
#[test]
fn encoding_channel_announcement() {
- do_encoding_channel_announcement(false, false, false);
- do_encoding_channel_announcement(true, false, false);
- do_encoding_channel_announcement(true, true, false);
- do_encoding_channel_announcement(true, true, true);
- do_encoding_channel_announcement(false, true, true);
- do_encoding_channel_announcement(false, false, true);
- do_encoding_channel_announcement(false, true, false);
- do_encoding_channel_announcement(true, false, true);
+ do_encoding_channel_announcement(true, false);
+ do_encoding_channel_announcement(false, true);
+ do_encoding_channel_announcement(false, false);
+ do_encoding_channel_announcement(true, true);
}
- fn do_encoding_node_announcement(unknown_features_bits: bool, ipv4: bool, ipv6: bool, onionv2: bool, onionv3: bool, excess_address_data: bool, excess_data: bool) {
+ fn do_encoding_node_announcement(unknown_features_bits: bool, ipv4: bool, ipv6: bool, onionv2: bool, onionv3: bool, hostname: bool, excess_address_data: bool, excess_data: bool) {
let secp_ctx = Secp256k1::new();
let (privkey_1, pubkey_1) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
let sig_1 = get_sig_on!(privkey_1, secp_ctx, String::from("01010101010101010101010101010101"));
- let mut features = GlobalFeatures::new();
- if unknown_features_bits {
- features.flags = vec![0xFF, 0xFF];
- }
+ let features = if unknown_features_bits {
+ NodeFeatures::from_le_bytes(vec![0xFF, 0xFF])
+ } else {
+ // Set to some features we may support
+ NodeFeatures::from_le_bytes(vec![2 | 1 << 5])
+ };
let mut addresses = Vec::new();
if ipv4 {
addresses.push(msgs::NetAddress::IPv4 {
});
}
if onionv2 {
- addresses.push(msgs::NetAddress::OnionV2 {
- addr: [255, 254, 253, 252, 251, 250, 249, 248, 247, 246],
- port: 9735
- });
+ addresses.push(msgs::NetAddress::OnionV2(
+ [255, 254, 253, 252, 251, 250, 249, 248, 247, 246, 38, 7]
+ ));
}
if onionv3 {
addresses.push(msgs::NetAddress::OnionV3 {
port: 9735
});
}
+ if hostname {
+ addresses.push(msgs::NetAddress::Hostname {
+ hostname: Hostname::try_from(String::from("host")).unwrap(),
+ port: 9735,
+ });
+ }
let mut addr_len = 0;
for addr in &addresses {
addr_len += addr.len() + 1;
if unknown_features_bits {
target_value.append(&mut hex::decode("0002ffff").unwrap());
} else {
- target_value.append(&mut hex::decode("0000").unwrap());
+ target_value.append(&mut hex::decode("000122").unwrap());
}
target_value.append(&mut hex::decode("013413a7031b84c5567b126440995d3ed5aaba0565d71e1834604819ff9c17f5e9d5dd078f2020201010101010101010101010101010101010101010101010101010101010101010").unwrap());
target_value.append(&mut vec![(addr_len >> 8) as u8, addr_len as u8]);
if onionv3 {
target_value.append(&mut hex::decode("04fffefdfcfbfaf9f8f7f6f5f4f3f2f1f0efeeedecebeae9e8e7e6e5e4e3e2e1e00020102607").unwrap());
}
+ if hostname {
+ target_value.append(&mut hex::decode("0504686f73742607").unwrap());
+ }
if excess_address_data {
target_value.append(&mut hex::decode("216c280b5395a2546e7e4b2663e04f811622f15a4f92e83aa2e92ba2a573c139142c54ae63072a1ec1ee7dc0c04bde5c847806172aa05c92c22ae8e308d1d269").unwrap());
}
#[test]
fn encoding_node_announcement() {
- do_encoding_node_announcement(true, true, true, true, true, true, true);
- do_encoding_node_announcement(false, false, false, false, false, false, false);
- do_encoding_node_announcement(false, true, false, false, false, false, false);
- do_encoding_node_announcement(false, false, true, false, false, false, false);
- do_encoding_node_announcement(false, false, false, true, false, false, false);
- do_encoding_node_announcement(false, false, false, false, true, false, false);
- do_encoding_node_announcement(false, false, false, false, false, true, false);
- do_encoding_node_announcement(false, true, false, true, false, true, false);
- do_encoding_node_announcement(false, false, true, false, true, false, false);
- }
-
- fn do_encoding_channel_update(non_bitcoin_chain_hash: bool, direction: bool, disable: bool, htlc_maximum_msat: bool) {
+ do_encoding_node_announcement(true, true, true, true, true, true, true, true);
+ do_encoding_node_announcement(false, false, false, false, false, false, false, false);
+ do_encoding_node_announcement(false, true, false, false, false, false, false, false);
+ do_encoding_node_announcement(false, false, true, false, false, false, false, false);
+ do_encoding_node_announcement(false, false, false, true, false, false, false, false);
+ do_encoding_node_announcement(false, false, false, false, true, false, false, false);
+ do_encoding_node_announcement(false, false, false, false, false, true, false, false);
+ do_encoding_node_announcement(false, false, false, false, false, false, true, false);
+ do_encoding_node_announcement(false, true, false, true, false, false, true, false);
+ do_encoding_node_announcement(false, false, true, false, true, false, false, false);
+ }
+
+ fn do_encoding_channel_update(direction: bool, disable: bool, excess_data: bool) {
let secp_ctx = Secp256k1::new();
let (privkey_1, _) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
let sig_1 = get_sig_on!(privkey_1, secp_ctx, String::from("01010101010101010101010101010101"));
let unsigned_channel_update = msgs::UnsignedChannelUpdate {
- chain_hash: if !non_bitcoin_chain_hash { Sha256dHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap() } else { Sha256dHash::from_hex("000000000933ea01ad0ee984209779baaec3ced90fa3f408719526f8d77f4943").unwrap() },
+ chain_hash: BlockHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap(),
short_channel_id: 2316138423780173,
timestamp: 20190119,
- flags: if direction { 1 } else { 0 } | if disable { 1 << 1 } else { 0 } | if htlc_maximum_msat { 1 << 8 } else { 0 },
+ flags: if direction { 1 } else { 0 } | if disable { 1 << 1 } else { 0 },
cltv_expiry_delta: 144,
htlc_minimum_msat: 1000000,
+ htlc_maximum_msat: 131355275467161,
fee_base_msat: 10000,
fee_proportional_millionths: 20,
- excess_data: if htlc_maximum_msat { vec![0, 0, 0, 0, 59, 154, 202, 0] } else { Vec::new() }
+ excess_data: if excess_data { vec![0, 0, 0, 0, 59, 154, 202, 0] } else { Vec::new() }
};
let channel_update = msgs::ChannelUpdate {
signature: sig_1,
};
let encoded_value = channel_update.encode();
let mut target_value = hex::decode("d977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073a").unwrap();
- if non_bitcoin_chain_hash {
- target_value.append(&mut hex::decode("43497fd7f826957108f4a30fd9cec3aeba79972084e90ead01ea330900000000").unwrap());
- } else {
- target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
- }
+ target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
target_value.append(&mut hex::decode("00083a840000034d013413a7").unwrap());
- if htlc_maximum_msat {
- target_value.append(&mut hex::decode("01").unwrap());
- } else {
- target_value.append(&mut hex::decode("00").unwrap());
- }
+ target_value.append(&mut hex::decode("01").unwrap());
target_value.append(&mut hex::decode("00").unwrap());
if direction {
let flag = target_value.last_mut().unwrap();
*flag = *flag | 1 << 1;
}
target_value.append(&mut hex::decode("009000000000000f42400000271000000014").unwrap());
- if htlc_maximum_msat {
+ target_value.append(&mut hex::decode("0000777788889999").unwrap());
+ if excess_data {
target_value.append(&mut hex::decode("000000003b9aca00").unwrap());
}
assert_eq!(encoded_value, target_value);
#[test]
fn encoding_channel_update() {
- do_encoding_channel_update(false, false, false, false);
- do_encoding_channel_update(true, false, false, false);
- do_encoding_channel_update(false, true, false, false);
- do_encoding_channel_update(false, false, true, false);
- do_encoding_channel_update(false, false, false, true);
- do_encoding_channel_update(true, true, true, true);
+ do_encoding_channel_update(false, false, false);
+ do_encoding_channel_update(false, false, true);
+ do_encoding_channel_update(true, false, false);
+ do_encoding_channel_update(true, false, true);
+ do_encoding_channel_update(false, true, false);
+ do_encoding_channel_update(false, true, true);
+ do_encoding_channel_update(true, true, false);
+ do_encoding_channel_update(true, true, true);
}
- fn do_encoding_open_channel(non_bitcoin_chain_hash: bool, random_bit: bool, shutdown: bool) {
+ fn do_encoding_open_channel(random_bit: bool, shutdown: bool, incl_chan_type: bool) {
let secp_ctx = Secp256k1::new();
let (_, pubkey_1) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
let (_, pubkey_2) = get_keys_from!("0202020202020202020202020202020202020202020202020202020202020202", secp_ctx);
let (_, pubkey_5) = get_keys_from!("0505050505050505050505050505050505050505050505050505050505050505", secp_ctx);
let (_, pubkey_6) = get_keys_from!("0606060606060606060606060606060606060606060606060606060606060606", secp_ctx);
let open_channel = msgs::OpenChannel {
- chain_hash: if !non_bitcoin_chain_hash { Sha256dHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap() } else { Sha256dHash::from_hex("000000000933ea01ad0ee984209779baaec3ced90fa3f408719526f8d77f4943").unwrap() },
+ chain_hash: BlockHash::from_hex("6fe28c0ab6f1b372c1a6a246ae63f74f931e8365e15a089c68d6190000000000").unwrap(),
temporary_channel_id: [2; 32],
funding_satoshis: 1311768467284833366,
push_msat: 2536655962884945560,
max_accepted_htlcs: 49340,
funding_pubkey: pubkey_1,
revocation_basepoint: pubkey_2,
- payment_basepoint: pubkey_3,
+ payment_point: pubkey_3,
delayed_payment_basepoint: pubkey_4,
htlc_basepoint: pubkey_5,
first_per_commitment_point: pubkey_6,
channel_flags: if random_bit { 1 << 5 } else { 0 },
- shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent }
+ shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, inner: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent },
+ channel_type: if incl_chan_type { Some(ChannelTypeFeatures::empty()) } else { None },
};
let encoded_value = open_channel.encode();
let mut target_value = Vec::new();
- if non_bitcoin_chain_hash {
- target_value.append(&mut hex::decode("43497fd7f826957108f4a30fd9cec3aeba79972084e90ead01ea330900000000").unwrap());
- } else {
- target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
- }
+ target_value.append(&mut hex::decode("000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f").unwrap());
target_value.append(&mut hex::decode("02020202020202020202020202020202020202020202020202020202020202021234567890123456233403289122369832144668701144767633030896203198784335490624111800083a840000034d000c89d4c0bcc0bc031b84c5567b126440995d3ed5aaba0565d71e1834604819ff9c17f5e9d5dd078f024d4b6cd1361032ca9bd2aeb9d900aa4d45d9ead80ac9423374c451a7254d076602531fe6068134503d2723133227c867ac8fa6c83c537e9a44c3c5bdbdcb1fe33703462779ad4aad39514614751a71085f2f10e1c7a593e4e030efb5b8721ce55b0b0362c0a046dacce86ddd0343c6d3c7c79c2208ba0d9c9cf24a6d046d21d21f90f703f006a18d5653c4edf5391ff23a61f03ff83d237e880ee61187fa9f379a028e0a").unwrap());
if random_bit {
target_value.append(&mut hex::decode("20").unwrap());
if shutdown {
target_value.append(&mut hex::decode("001976a91479b000887626b294a914501a4cd226b58b23598388ac").unwrap());
}
+ if incl_chan_type {
+ target_value.append(&mut hex::decode("0100").unwrap());
+ }
assert_eq!(encoded_value, target_value);
}
#[test]
fn encoding_open_channel() {
do_encoding_open_channel(false, false, false);
- do_encoding_open_channel(true, false, false);
- do_encoding_open_channel(false, true, false);
do_encoding_open_channel(false, false, true);
+ do_encoding_open_channel(false, true, false);
+ do_encoding_open_channel(false, true, true);
+ do_encoding_open_channel(true, false, false);
+ do_encoding_open_channel(true, false, true);
+ do_encoding_open_channel(true, true, false);
do_encoding_open_channel(true, true, true);
}
max_accepted_htlcs: 49340,
funding_pubkey: pubkey_1,
revocation_basepoint: pubkey_2,
- payment_basepoint: pubkey_3,
+ payment_point: pubkey_3,
delayed_payment_basepoint: pubkey_4,
htlc_basepoint: pubkey_5,
first_per_commitment_point: pubkey_6,
- shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent }
+ shutdown_scriptpubkey: if shutdown { OptionalField::Present(Address::p2pkh(&::bitcoin::PublicKey{compressed: true, inner: pubkey_1}, Network::Testnet).script_pubkey()) } else { OptionalField::Absent },
+ channel_type: None,
};
let encoded_value = accept_channel.encode();
let mut target_value = hex::decode("020202020202020202020202020202020202020202020202020202020202020212345678901234562334032891223698321446687011447600083a840000034d000c89d4c0bcc0bc031b84c5567b126440995d3ed5aaba0565d71e1834604819ff9c17f5e9d5dd078f024d4b6cd1361032ca9bd2aeb9d900aa4d45d9ead80ac9423374c451a7254d076602531fe6068134503d2723133227c867ac8fa6c83c537e9a44c3c5bdbdcb1fe33703462779ad4aad39514614751a71085f2f10e1c7a593e4e030efb5b8721ce55b0b0362c0a046dacce86ddd0343c6d3c7c79c2208ba0d9c9cf24a6d046d21d21f90f703f006a18d5653c4edf5391ff23a61f03ff83d237e880ee61187fa9f379a028e0a").unwrap();
let sig_1 = get_sig_on!(privkey_1, secp_ctx, String::from("01010101010101010101010101010101"));
let funding_created = msgs::FundingCreated {
temporary_channel_id: [2; 32],
- funding_txid: Sha256dHash::from_hex("c2d4449afa8d26140898dd54d3390b057ba2a5afcf03ba29d7dc0d8b9ffe966e").unwrap(),
+ funding_txid: Txid::from_hex("c2d4449afa8d26140898dd54d3390b057ba2a5afcf03ba29d7dc0d8b9ffe966e").unwrap(),
funding_output_index: 255,
signature: sig_1,
};
}
#[test]
- fn encoding_funding_locked() {
+ fn encoding_channel_ready() {
let secp_ctx = Secp256k1::new();
let (_, pubkey_1,) = get_keys_from!("0101010101010101010101010101010101010101010101010101010101010101", secp_ctx);
- let funding_locked = msgs::FundingLocked {
+ let channel_ready = msgs::ChannelReady {
channel_id: [2; 32],
next_per_commitment_point: pubkey_1,
+ short_channel_id_alias: None,
};
- let encoded_value = funding_locked.encode();
+ let encoded_value = channel_ready.encode();
let target_value = hex::decode("0202020202020202020202020202020202020202020202020202020202020202031b84c5567b126440995d3ed5aaba0565d71e1834604819ff9c17f5e9d5dd078f").unwrap();
assert_eq!(encoded_value, target_value);
}
let script = Builder::new().push_opcode(opcodes::OP_TRUE).into_script();
let shutdown = msgs::Shutdown {
channel_id: [2; 32],
- scriptpubkey: if script_type == 1 { Address::p2pkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey() } else if script_type == 2 { Address::p2sh(&script, Network::Testnet).script_pubkey() } else if script_type == 3 { Address::p2wpkh(&::bitcoin::PublicKey{compressed: true, key: pubkey_1}, Network::Testnet).script_pubkey() } else { Address::p2wsh(&script, Network::Testnet).script_pubkey() },
+ scriptpubkey:
+ if script_type == 1 { Address::p2pkh(&::bitcoin::PublicKey{compressed: true, inner: pubkey_1}, Network::Testnet).script_pubkey() }
+ else if script_type == 2 { Address::p2sh(&script, Network::Testnet).unwrap().script_pubkey() }
+ else if script_type == 3 { Address::p2wpkh(&::bitcoin::PublicKey{compressed: true, inner: pubkey_1}, Network::Testnet).unwrap().script_pubkey() }
+ else { Address::p2wsh(&script, Network::Testnet).script_pubkey() },
};
let encoded_value = shutdown.encode();
let mut target_value = hex::decode("0202020202020202020202020202020202020202020202020202020202020202").unwrap();
channel_id: [2; 32],
fee_satoshis: 2316138423780173,
signature: sig_1,
+ fee_range: None,
};
let encoded_value = closing_signed.encode();
let target_value = hex::decode("020202020202020202020202020202020202020202020202020202020202020200083a840000034dd977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073a").unwrap();
assert_eq!(encoded_value, target_value);
+ assert_eq!(msgs::ClosingSigned::read(&mut Cursor::new(&target_value)).unwrap(), closing_signed);
+
+ let closing_signed_with_range = msgs::ClosingSigned {
+ channel_id: [2; 32],
+ fee_satoshis: 2316138423780173,
+ signature: sig_1,
+ fee_range: Some(msgs::ClosingSignedFeeRange {
+ min_fee_satoshis: 0xdeadbeef,
+ max_fee_satoshis: 0x1badcafe01234567,
+ }),
+ };
+ let encoded_value_with_range = closing_signed_with_range.encode();
+ let target_value_with_range = hex::decode("020202020202020202020202020202020202020202020202020202020202020200083a840000034dd977cb9b53d93a6ff64bb5f1e158b4094b66e798fb12911168a3ccdf80a83096340a6a95da0ae8d9f776528eecdbb747eb6b545495a4319ed5378e35b21e073a011000000000deadbeef1badcafe01234567").unwrap();
+ assert_eq!(encoded_value_with_range, target_value_with_range);
+ assert_eq!(msgs::ClosingSigned::read(&mut Cursor::new(&target_value_with_range)).unwrap(),
+ closing_signed_with_range);
}
#[test]
assert_eq!(encoded_value, target_value);
}
- fn do_encoding_init(unknown_global_bits: bool, initial_routing_sync: bool) {
- let mut global = GlobalFeatures::new();
- if unknown_global_bits {
- global.flags = vec![0xFF, 0xFF];
- }
- let mut local = LocalFeatures::new();
- if initial_routing_sync {
- local.set_initial_routing_sync();
- }
- let init = msgs::Init {
- global_features: global,
- local_features: local,
+ #[test]
+ fn encoding_init() {
+ assert_eq!(msgs::Init {
+ features: InitFeatures::from_le_bytes(vec![0xFF, 0xFF, 0xFF]),
+ remote_network_address: None,
+ }.encode(), hex::decode("00023fff0003ffffff").unwrap());
+ assert_eq!(msgs::Init {
+ features: InitFeatures::from_le_bytes(vec![0xFF]),
+ remote_network_address: None,
+ }.encode(), hex::decode("0001ff0001ff").unwrap());
+ assert_eq!(msgs::Init {
+ features: InitFeatures::from_le_bytes(vec![]),
+ remote_network_address: None,
+ }.encode(), hex::decode("00000000").unwrap());
+
+ let init_msg = msgs::Init { features: InitFeatures::from_le_bytes(vec![]),
+ remote_network_address: Some(msgs::NetAddress::IPv4 {
+ addr: [127, 0, 0, 1],
+ port: 1000,
+ }),
};
- let encoded_value = init.encode();
- let mut target_value = Vec::new();
- if unknown_global_bits {
- target_value.append(&mut hex::decode("0002ffff").unwrap());
- } else {
- target_value.append(&mut hex::decode("0000").unwrap());
- }
- if initial_routing_sync {
- target_value.append(&mut hex::decode("00012a").unwrap());
- } else {
- target_value.append(&mut hex::decode("000122").unwrap());
- }
+ let encoded_value = init_msg.encode();
+ let target_value = hex::decode("000000000307017f00000103e8").unwrap();
assert_eq!(encoded_value, target_value);
+ assert_eq!(msgs::Init::read(&mut Cursor::new(&target_value)).unwrap(), init_msg);
}
#[test]
- fn encoding_init() {
- do_encoding_init(false, false);
- do_encoding_init(true, false);
- do_encoding_init(false, true);
- do_encoding_init(true, true);
+ fn encoding_error() {
+ let error = msgs::ErrorMessage {
+ channel_id: [2; 32],
+ data: String::from("rust-lightning"),
+ };
+ let encoded_value = error.encode();
+ let target_value = hex::decode("0202020202020202020202020202020202020202020202020202020202020202000e727573742d6c696768746e696e67").unwrap();
+ assert_eq!(encoded_value, target_value);
}
#[test]
- fn encoding_error() {
- let error = msgs::ErrorMessage {
+ fn encoding_warning() {
+ let error = msgs::WarningMessage {
channel_id: [2; 32],
data: String::from("rust-lightning"),
};
let target_value = hex::decode("004000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000").unwrap();
assert_eq!(encoded_value, target_value);
}
+
+ #[test]
+ fn encoding_legacy_onion_hop_data() {
+ let msg = msgs::OnionHopData {
+ format: OnionHopDataFormat::Legacy {
+ short_channel_id: 0xdeadbeef1bad1dea,
+ },
+ amt_to_forward: 0x0badf00d01020304,
+ outgoing_cltv_value: 0xffffffff,
+ };
+ let encoded_value = msg.encode();
+ let target_value = hex::decode("00deadbeef1bad1dea0badf00d01020304ffffffff000000000000000000000000").unwrap();
+ assert_eq!(encoded_value, target_value);
+ }
+
+ #[test]
+ fn encoding_nonfinal_onion_hop_data() {
+ let mut msg = msgs::OnionHopData {
+ format: OnionHopDataFormat::NonFinalNode {
+ short_channel_id: 0xdeadbeef1bad1dea,
+ },
+ amt_to_forward: 0x0badf00d01020304,
+ outgoing_cltv_value: 0xffffffff,
+ };
+ let encoded_value = msg.encode();
+ let target_value = hex::decode("1a02080badf00d010203040404ffffffff0608deadbeef1bad1dea").unwrap();
+ assert_eq!(encoded_value, target_value);
+ msg = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ if let OnionHopDataFormat::NonFinalNode { short_channel_id } = msg.format {
+ assert_eq!(short_channel_id, 0xdeadbeef1bad1dea);
+ } else { panic!(); }
+ assert_eq!(msg.amt_to_forward, 0x0badf00d01020304);
+ assert_eq!(msg.outgoing_cltv_value, 0xffffffff);
+ }
+
+ #[test]
+ fn encoding_final_onion_hop_data() {
+ let mut msg = msgs::OnionHopData {
+ format: OnionHopDataFormat::FinalNode {
+ payment_data: None,
+ keysend_preimage: None,
+ },
+ amt_to_forward: 0x0badf00d01020304,
+ outgoing_cltv_value: 0xffffffff,
+ };
+ let encoded_value = msg.encode();
+ let target_value = hex::decode("1002080badf00d010203040404ffffffff").unwrap();
+ assert_eq!(encoded_value, target_value);
+ msg = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ if let OnionHopDataFormat::FinalNode { payment_data: None, .. } = msg.format { } else { panic!(); }
+ assert_eq!(msg.amt_to_forward, 0x0badf00d01020304);
+ assert_eq!(msg.outgoing_cltv_value, 0xffffffff);
+ }
+
+ #[test]
+ fn encoding_final_onion_hop_data_with_secret() {
+ let expected_payment_secret = PaymentSecret([0x42u8; 32]);
+ let mut msg = msgs::OnionHopData {
+ format: OnionHopDataFormat::FinalNode {
+ payment_data: Some(FinalOnionHopData {
+ payment_secret: expected_payment_secret,
+ total_msat: 0x1badca1f
+ }),
+ keysend_preimage: None,
+ },
+ amt_to_forward: 0x0badf00d01020304,
+ outgoing_cltv_value: 0xffffffff,
+ };
+ let encoded_value = msg.encode();
+ let target_value = hex::decode("3602080badf00d010203040404ffffffff082442424242424242424242424242424242424242424242424242424242424242421badca1f").unwrap();
+ assert_eq!(encoded_value, target_value);
+ msg = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ if let OnionHopDataFormat::FinalNode {
+ payment_data: Some(FinalOnionHopData {
+ payment_secret,
+ total_msat: 0x1badca1f
+ }),
+ keysend_preimage: None,
+ } = msg.format {
+ assert_eq!(payment_secret, expected_payment_secret);
+ } else { panic!(); }
+ assert_eq!(msg.amt_to_forward, 0x0badf00d01020304);
+ assert_eq!(msg.outgoing_cltv_value, 0xffffffff);
+ }
+
+ #[test]
+ fn query_channel_range_end_blocknum() {
+ let tests: Vec<(u32, u32, u32)> = vec![
+ (10000, 1500, 11500),
+ (0, 0xffffffff, 0xffffffff),
+ (1, 0xffffffff, 0xffffffff),
+ ];
+
+ for (first_blocknum, number_of_blocks, expected) in tests.into_iter() {
+ let sut = msgs::QueryChannelRange {
+ chain_hash: BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap(),
+ first_blocknum,
+ number_of_blocks,
+ };
+ assert_eq!(sut.end_blocknum(), expected);
+ }
+ }
+
+ #[test]
+ fn encoding_query_channel_range() {
+ let mut query_channel_range = msgs::QueryChannelRange {
+ chain_hash: BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap(),
+ first_blocknum: 100000,
+ number_of_blocks: 1500,
+ };
+ let encoded_value = query_channel_range.encode();
+ let target_value = hex::decode("0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e2206000186a0000005dc").unwrap();
+ assert_eq!(encoded_value, target_value);
+
+ query_channel_range = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ assert_eq!(query_channel_range.first_blocknum, 100000);
+ assert_eq!(query_channel_range.number_of_blocks, 1500);
+ }
+
+ #[test]
+ fn encoding_reply_channel_range() {
+ do_encoding_reply_channel_range(0);
+ do_encoding_reply_channel_range(1);
+ }
+
+ fn do_encoding_reply_channel_range(encoding_type: u8) {
+ let mut target_value = hex::decode("0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e2206000b8a06000005dc01").unwrap();
+ let expected_chain_hash = BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap();
+ let mut reply_channel_range = msgs::ReplyChannelRange {
+ chain_hash: expected_chain_hash,
+ first_blocknum: 756230,
+ number_of_blocks: 1500,
+ sync_complete: true,
+ short_channel_ids: vec![0x000000000000008e, 0x0000000000003c69, 0x000000000045a6c4],
+ };
+
+ if encoding_type == 0 {
+ target_value.append(&mut hex::decode("001900000000000000008e0000000000003c69000000000045a6c4").unwrap());
+ let encoded_value = reply_channel_range.encode();
+ assert_eq!(encoded_value, target_value);
+
+ reply_channel_range = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ assert_eq!(reply_channel_range.chain_hash, expected_chain_hash);
+ assert_eq!(reply_channel_range.first_blocknum, 756230);
+ assert_eq!(reply_channel_range.number_of_blocks, 1500);
+ assert_eq!(reply_channel_range.sync_complete, true);
+ assert_eq!(reply_channel_range.short_channel_ids[0], 0x000000000000008e);
+ assert_eq!(reply_channel_range.short_channel_ids[1], 0x0000000000003c69);
+ assert_eq!(reply_channel_range.short_channel_ids[2], 0x000000000045a6c4);
+ } else {
+ target_value.append(&mut hex::decode("001601789c636000833e08659309a65878be010010a9023a").unwrap());
+ let result: Result<msgs::ReplyChannelRange, msgs::DecodeError> = Readable::read(&mut Cursor::new(&target_value[..]));
+ assert!(result.is_err(), "Expected decode failure with unsupported zlib encoding");
+ }
+ }
+
+ #[test]
+ fn encoding_query_short_channel_ids() {
+ do_encoding_query_short_channel_ids(0);
+ do_encoding_query_short_channel_ids(1);
+ }
+
+ fn do_encoding_query_short_channel_ids(encoding_type: u8) {
+ let mut target_value = hex::decode("0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e2206").unwrap();
+ let expected_chain_hash = BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap();
+ let mut query_short_channel_ids = msgs::QueryShortChannelIds {
+ chain_hash: expected_chain_hash,
+ short_channel_ids: vec![0x0000000000008e, 0x0000000000003c69, 0x000000000045a6c4],
+ };
+
+ if encoding_type == 0 {
+ target_value.append(&mut hex::decode("001900000000000000008e0000000000003c69000000000045a6c4").unwrap());
+ let encoded_value = query_short_channel_ids.encode();
+ assert_eq!(encoded_value, target_value);
+
+ query_short_channel_ids = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ assert_eq!(query_short_channel_ids.chain_hash, expected_chain_hash);
+ assert_eq!(query_short_channel_ids.short_channel_ids[0], 0x000000000000008e);
+ assert_eq!(query_short_channel_ids.short_channel_ids[1], 0x0000000000003c69);
+ assert_eq!(query_short_channel_ids.short_channel_ids[2], 0x000000000045a6c4);
+ } else {
+ target_value.append(&mut hex::decode("001601789c636000833e08659309a65878be010010a9023a").unwrap());
+ let result: Result<msgs::QueryShortChannelIds, msgs::DecodeError> = Readable::read(&mut Cursor::new(&target_value[..]));
+ assert!(result.is_err(), "Expected decode failure with unsupported zlib encoding");
+ }
+ }
+
+ #[test]
+ fn encoding_reply_short_channel_ids_end() {
+ let expected_chain_hash = BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap();
+ let mut reply_short_channel_ids_end = msgs::ReplyShortChannelIdsEnd {
+ chain_hash: expected_chain_hash,
+ full_information: true,
+ };
+ let encoded_value = reply_short_channel_ids_end.encode();
+ let target_value = hex::decode("0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e220601").unwrap();
+ assert_eq!(encoded_value, target_value);
+
+ reply_short_channel_ids_end = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ assert_eq!(reply_short_channel_ids_end.chain_hash, expected_chain_hash);
+ assert_eq!(reply_short_channel_ids_end.full_information, true);
+ }
+
+ #[test]
+ fn encoding_gossip_timestamp_filter(){
+ let expected_chain_hash = BlockHash::from_hex("06226e46111a0b59caaf126043eb5bbf28c34f3a5e332a1fc7b2b73cf188910f").unwrap();
+ let mut gossip_timestamp_filter = msgs::GossipTimestampFilter {
+ chain_hash: expected_chain_hash,
+ first_timestamp: 1590000000,
+ timestamp_range: 0xffff_ffff,
+ };
+ let encoded_value = gossip_timestamp_filter.encode();
+ let target_value = hex::decode("0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e22065ec57980ffffffff").unwrap();
+ assert_eq!(encoded_value, target_value);
+
+ gossip_timestamp_filter = Readable::read(&mut Cursor::new(&target_value[..])).unwrap();
+ assert_eq!(gossip_timestamp_filter.chain_hash, expected_chain_hash);
+ assert_eq!(gossip_timestamp_filter.first_timestamp, 1590000000);
+ assert_eq!(gossip_timestamp_filter.timestamp_range, 0xffff_ffff);
+ }
+
+ #[test]
+ fn decode_onion_hop_data_len_as_bigsize() {
+ // Tests that we can decode an onion payload that is >253 bytes.
+ // Previously, receiving a payload of this size could've caused us to fail to decode a valid
+ // payload, because we were decoding the length (a BigSize, big-endian) as a VarInt
+ // (little-endian).
+
+ // Encode a test onion payload with a big custom TLV such that it's >253 bytes, forcing the
+ // payload length to be encoded over multiple bytes rather than a single u8.
+ let big_payload = encode_big_payload().unwrap();
+ let mut rd = Cursor::new(&big_payload[..]);
+ <msgs::OnionHopData as Readable>::read(&mut rd).unwrap();
+ }
+ // see above test, needs to be a separate method for use of the serialization macros.
+ fn encode_big_payload() -> Result<Vec<u8>, io::Error> {
+ use util::ser::HighZeroBytesDroppedBigSize;
+ let payload = msgs::OnionHopData {
+ format: OnionHopDataFormat::NonFinalNode {
+ short_channel_id: 0xdeadbeef1bad1dea,
+ },
+ amt_to_forward: 1000,
+ outgoing_cltv_value: 0xffffffff,
+ };
+ let mut encoded_payload = Vec::new();
+ let test_bytes = vec![42u8; 1000];
+ if let OnionHopDataFormat::NonFinalNode { short_channel_id } = payload.format {
+ encode_varint_length_prefixed_tlv!(&mut encoded_payload, {
+ (1, test_bytes, vec_type),
+ (2, HighZeroBytesDroppedBigSize(payload.amt_to_forward), required),
+ (4, HighZeroBytesDroppedBigSize(payload.outgoing_cltv_value), required),
+ (6, short_channel_id, required)
+ });
+ }
+ Ok(encoded_payload)
+ }
}