WIP
[rust-lightning] / src / lib.rs
1 //! Utilities that handle persisting Rust-Lightning data to disk via standard filesystem APIs.
2
3 // Prefix these with `rustdoc::` when we update our MSRV to be >= 1.52 to remove warnings.
4 #![deny(broken_intra_doc_links)]
5 #![deny(private_intra_doc_links)]
6
7 #![deny(missing_docs)]
8
9 #![cfg_attr(docsrs, feature(doc_auto_cfg))]
10
11 #![cfg_attr(all(test, feature = "_bench_unstable"), feature(test))]
12 #[cfg(all(test, feature = "_bench_unstable"))] extern crate test;
13
14 mod util;
15
16 extern crate lightning;
17 extern crate bitcoin;
18 extern crate libc;
19
20 use bitcoin::hash_types::{BlockHash, Txid};
21 use bitcoin::hashes::hex::FromHex;
22 use lightning::chain::channelmonitor::ChannelMonitor;
23 use lightning::chain::keysinterface::{EntropySource, SignerProvider, WriteableEcdsaChannelSigner};
24 use lightning::util::ser::{ReadableArgs, Writeable};
25 use lightning::util::persist::KVStorePersister;
26 use std::fs;
27 use std::io::Cursor;
28 use std::ops::Deref;
29 use std::path::{Path, PathBuf};
30
31 /// FilesystemPersister persists channel data on disk, where each channel's
32 /// data is stored in a file named after its funding outpoint.
33 ///
34 /// Warning: this module does the best it can with calls to persist data, but it
35 /// can only guarantee that the data is passed to the drive. It is up to the
36 /// drive manufacturers to do the actual persistence properly, which they often
37 /// don't (especially on consumer-grade hardware). Therefore, it is up to the
38 /// user to validate their entire storage stack, to ensure the writes are
39 /// persistent.
40 /// Corollary: especially when dealing with larger amounts of money, it is best
41 /// practice to have multiple channel data backups and not rely only on one
42 /// FilesystemPersister.
43 pub struct FilesystemPersister {
44         path_to_channel_data: String,
45 }
46
47 impl FilesystemPersister {
48         /// Initialize a new FilesystemPersister and set the path to the individual channels'
49         /// files.
50         pub fn new(path_to_channel_data: String) -> Self {
51                 Self {
52                         path_to_channel_data,
53                 }
54         }
55
56         /// Get the directory which was provided when this persister was initialized.
57         pub fn get_data_dir(&self) -> String {
58                 self.path_to_channel_data.clone()
59         }
60
61         /// Read `ChannelMonitor`s from disk.
62         pub fn read_channelmonitors<ES: Deref, WES: WriteableEcdsaChannelSigner, SP: SignerProvider<Signer = WES> + Sized, SPD: Deref<Target=SP>> (
63                 &self, entropy_source: ES, signer_provider: SPD
64         ) -> Result<Vec<(BlockHash, ChannelMonitor<WES>)>, std::io::Error>
65                 where
66                         ES::Target: EntropySource + Sized,
67         {
68                 let mut path = PathBuf::from(&self.path_to_channel_data);
69                 path.push("monitors");
70                 if !Path::new(&path).exists() {
71                         return Ok(Vec::new());
72                 }
73                 let mut res = Vec::new();
74                 for file_option in fs::read_dir(path)? {
75                         let file = file_option.unwrap();
76                         let owned_file_name = file.file_name();
77                         let filename = owned_file_name.to_str()
78                                 .ok_or_else(|| std::io::Error::new(std::io::ErrorKind::InvalidData,
79                                         "File name is not a valid utf8 string"))?;
80                         if !filename.is_ascii() || filename.len() < 65 {
81                                 return Err(std::io::Error::new(
82                                         std::io::ErrorKind::InvalidData,
83                                         "Invalid ChannelMonitor file name",
84                                 ));
85                         }
86                         if filename.ends_with(".tmp") {
87                                 // If we were in the middle of committing an new update and crashed, it should be
88                                 // safe to ignore the update - we should never have returned to the caller and
89                                 // irrevocably committed to the new state in any way.
90                                 continue;
91                         }
92
93                         let txid = Txid::from_hex(filename.split_at(64).0)
94                                 .map_err(|_| std::io::Error::new(
95                                         std::io::ErrorKind::InvalidData,
96                                         "Invalid tx ID in filename",
97                                 ))?;
98
99                         let index = filename.split_at(65).1.parse()
100                                 .map_err(|_| std::io::Error::new(
101                                         std::io::ErrorKind::InvalidData,
102                                         "Invalid tx index in filename",
103                                 ))?;
104
105                         let contents = fs::read(&file.path())?;
106                         let mut buffer = Cursor::new(&contents);
107                         match <(BlockHash, ChannelMonitor<WES>)>::read(&mut buffer, (&*entropy_source, &*signer_provider)) {
108                                 Ok((blockhash, channel_monitor)) => {
109                                         if channel_monitor.get_funding_txo().0.txid != txid || channel_monitor.get_funding_txo().0.index != index {
110                                                 return Err(std::io::Error::new(std::io::ErrorKind::InvalidData,
111                                                                                "ChannelMonitor was stored in the wrong file"));
112                                         }
113                                         res.push((blockhash, channel_monitor));
114                                 }
115                                 Err(e) => return Err(std::io::Error::new(
116                                         std::io::ErrorKind::InvalidData,
117                                         format!("Failed to deserialize ChannelMonitor: {}", e),
118                                 ))
119                         }
120                 }
121                 Ok(res)
122         }
123 }
124
125 impl KVStorePersister for FilesystemPersister {
126         fn persist<W: Writeable>(&self, key: &str, object: &W) -> std::io::Result<()> {
127                 let mut dest_file = PathBuf::from(self.path_to_channel_data.clone());
128                 dest_file.push(key);
129                 util::write_to_file(dest_file, object)
130         }
131 }
132
133 #[cfg(test)]
134 mod tests {
135         extern crate lightning;
136         extern crate bitcoin;
137         use crate::FilesystemPersister;
138         use bitcoin::blockdata::block::{Block, BlockHeader};
139         use bitcoin::hashes::hex::FromHex;
140         use bitcoin::{Txid, TxMerkleNode};
141         use lightning::chain::ChannelMonitorUpdateStatus;
142         use lightning::chain::chainmonitor::Persist;
143         use lightning::chain::transaction::OutPoint;
144         use lightning::{check_closed_broadcast, check_closed_event, check_added_monitors};
145         use lightning::ln::functional_test_utils::*;
146         use lightning::util::events::{ClosureReason, MessageSendEventsProvider};
147         use lightning::util::test_utils;
148         use std::fs;
149         use bitcoin::hashes::Hash;
150         #[cfg(target_os = "windows")]
151         use {
152                 lightning::get_event_msg,
153                 lightning::ln::msgs::ChannelMessageHandler,
154         };
155
156         impl Drop for FilesystemPersister {
157                 fn drop(&mut self) {
158                         // We test for invalid directory names, so it's OK if directory removal
159                         // fails.
160                         match fs::remove_dir_all(&self.path_to_channel_data) {
161                                 Err(e) => println!("Failed to remove test persister directory: {}", e),
162                                 _ => {}
163                         }
164                 }
165         }
166
167         #[test]
168         fn test_if_monitors_is_not_dir() {
169                 let persister = FilesystemPersister::new("test_monitors_is_not_dir".to_string());
170
171                 fs::create_dir_all(&persister.path_to_channel_data).unwrap();
172                 let mut path = std::path::PathBuf::from(&persister.path_to_channel_data);
173                 path.push("monitors");
174                 fs::File::create(path).unwrap();
175
176                 let chanmon_cfgs = create_chanmon_cfgs(1);
177                 let mut node_cfgs = create_node_cfgs(1, &chanmon_cfgs);
178                 let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, &persister, &node_cfgs[0].keys_manager);
179                 node_cfgs[0].chain_monitor = chain_mon_0;
180                 let node_chanmgrs = create_node_chanmgrs(1, &node_cfgs, &[None]);
181                 let nodes = create_network(1, &node_cfgs, &node_chanmgrs);
182
183                 // Check that read_channelmonitors() returns error if monitors/ is not a
184                 // directory.
185                 assert!(persister.read_channelmonitors(nodes[0].keys_manager, nodes[0].keys_manager).is_err());
186         }
187
188         // Integration-test the FilesystemPersister. Test relaying a few payments
189         // and check that the persisted data is updated the appropriate number of
190         // times.
191         #[test]
192         fn test_filesystem_persister() {
193                 // Create the nodes, giving them FilesystemPersisters for data persisters.
194                 let persister_0 = FilesystemPersister::new("test_filesystem_persister_0".to_string());
195                 let persister_1 = FilesystemPersister::new("test_filesystem_persister_1".to_string());
196                 let chanmon_cfgs = create_chanmon_cfgs(2);
197                 let mut node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
198                 let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, &persister_0, &node_cfgs[0].keys_manager);
199                 let chain_mon_1 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[1].chain_source), &chanmon_cfgs[1].tx_broadcaster, &chanmon_cfgs[1].logger, &chanmon_cfgs[1].fee_estimator, &persister_1, &node_cfgs[1].keys_manager);
200                 node_cfgs[0].chain_monitor = chain_mon_0;
201                 node_cfgs[1].chain_monitor = chain_mon_1;
202                 let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
203                 let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
204
205                 // Check that the persisted channel data is empty before any channels are
206                 // open.
207                 let mut persisted_chan_data_0 = persister_0.read_channelmonitors(nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
208                 assert_eq!(persisted_chan_data_0.len(), 0);
209                 let mut persisted_chan_data_1 = persister_1.read_channelmonitors(nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
210                 assert_eq!(persisted_chan_data_1.len(), 0);
211
212                 // Helper to make sure the channel is on the expected update ID.
213                 macro_rules! check_persisted_data {
214                         ($expected_update_id: expr) => {
215                                 persisted_chan_data_0 = persister_0.read_channelmonitors(nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
216                                 assert_eq!(persisted_chan_data_0.len(), 1);
217                                 for (_, mon) in persisted_chan_data_0.iter() {
218                                         assert_eq!(mon.get_latest_update_id(), $expected_update_id);
219                                 }
220                                 persisted_chan_data_1 = persister_1.read_channelmonitors(nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
221                                 assert_eq!(persisted_chan_data_1.len(), 1);
222                                 for (_, mon) in persisted_chan_data_1.iter() {
223                                         assert_eq!(mon.get_latest_update_id(), $expected_update_id);
224                                 }
225                         }
226                 }
227
228                 // Create some initial channel and check that a channel was persisted.
229                 let _ = create_announced_chan_between_nodes(&nodes, 0, 1);
230                 check_persisted_data!(0);
231
232                 // Send a few payments and make sure the monitors are updated to the latest.
233                 send_payment(&nodes[0], &vec!(&nodes[1])[..], 8000000);
234                 check_persisted_data!(5);
235                 send_payment(&nodes[1], &vec!(&nodes[0])[..], 4000000);
236                 check_persisted_data!(10);
237
238                 // Force close because cooperative close doesn't result in any persisted
239                 // updates.
240                 nodes[0].node.force_close_broadcasting_latest_txn(&nodes[0].node.list_channels()[0].channel_id, &nodes[1].node.get_our_node_id()).unwrap();
241                 check_closed_event!(nodes[0], 1, ClosureReason::HolderForceClosed);
242                 check_closed_broadcast!(nodes[0], true);
243                 check_added_monitors!(nodes[0], 1);
244
245                 let node_txn = nodes[0].tx_broadcaster.txn_broadcasted.lock().unwrap();
246                 assert_eq!(node_txn.len(), 1);
247
248                 let header = BlockHeader { version: 0x20000000, prev_blockhash: nodes[0].best_block_hash(), merkle_root: TxMerkleNode::all_zeros(), time: 42, bits: 42, nonce: 42 };
249                 connect_block(&nodes[1], &Block { header, txdata: vec![node_txn[0].clone(), node_txn[0].clone()]});
250                 check_closed_broadcast!(nodes[1], true);
251                 check_closed_event!(nodes[1], 1, ClosureReason::CommitmentTxConfirmed);
252                 check_added_monitors!(nodes[1], 1);
253
254                 // Make sure everything is persisted as expected after close.
255                 check_persisted_data!(11);
256         }
257
258         // Test that if the persister's path to channel data is read-only, writing a
259         // monitor to it results in the persister returning a PermanentFailure.
260         // Windows ignores the read-only flag for folders, so this test is Unix-only.
261         #[cfg(not(target_os = "windows"))]
262         #[test]
263         fn test_readonly_dir_perm_failure() {
264                 let persister = FilesystemPersister::new("test_readonly_dir_perm_failure".to_string());
265                 fs::create_dir_all(&persister.path_to_channel_data).unwrap();
266
267                 // Set up a dummy channel and force close. This will produce a monitor
268                 // that we can then use to test persistence.
269                 let chanmon_cfgs = create_chanmon_cfgs(2);
270                 let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
271                 let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
272                 let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
273                 let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
274                 nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
275                 check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed);
276                 let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
277                 let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
278                 let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
279
280                 // Set the persister's directory to read-only, which should result in
281                 // returning a permanent failure when we then attempt to persist a
282                 // channel update.
283                 let path = &persister.path_to_channel_data;
284                 let mut perms = fs::metadata(path).unwrap().permissions();
285                 perms.set_readonly(true);
286                 fs::set_permissions(path, perms).unwrap();
287
288                 let test_txo = OutPoint {
289                         txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
290                         index: 0
291                 };
292                 match persister.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
293                         ChannelMonitorUpdateStatus::PermanentFailure => {},
294                         _ => panic!("unexpected result from persisting new channel")
295                 }
296
297                 nodes[1].node.get_and_clear_pending_msg_events();
298                 added_monitors.clear();
299         }
300
301         // Test that if a persister's directory name is invalid, monitor persistence
302         // will fail.
303         #[cfg(target_os = "windows")]
304         #[test]
305         fn test_fail_on_open() {
306                 // Set up a dummy channel and force close. This will produce a monitor
307                 // that we can then use to test persistence.
308                 let chanmon_cfgs = create_chanmon_cfgs(2);
309                 let mut node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
310                 let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
311                 let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
312                 let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
313                 nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
314                 check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed);
315                 let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
316                 let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
317                 let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
318
319                 // Create the persister with an invalid directory name and test that the
320                 // channel fails to open because the directories fail to be created. There
321                 // don't seem to be invalid filename characters on Unix that Rust doesn't
322                 // handle, hence why the test is Windows-only.
323                 let persister = FilesystemPersister::new(":<>/".to_string());
324
325                 let test_txo = OutPoint {
326                         txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
327                         index: 0
328                 };
329                 match persister.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
330                         ChannelMonitorUpdateStatus::PermanentFailure => {},
331                         _ => panic!("unexpected result from persisting new channel")
332                 }
333
334                 nodes[1].node.get_and_clear_pending_msg_events();
335                 added_monitors.clear();
336         }
337 }
338
339 #[cfg(all(test, feature = "_bench_unstable"))]
340 pub mod bench {
341         use test::Bencher;
342
343         #[bench]
344         fn bench_sends(bench: &mut Bencher) {
345                 let persister_a = super::FilesystemPersister::new("bench_filesystem_persister_a".to_string());
346                 let persister_b = super::FilesystemPersister::new("bench_filesystem_persister_b".to_string());
347                 lightning::ln::channelmanager::bench::bench_two_sends(bench, persister_a, persister_b);
348         }
349 }