/**
- * BackgroundProcessor takes care of tasks that (1) need to happen periodically to keep
+ * `BackgroundProcessor` takes care of tasks that (1) need to happen periodically to keep
* Rust-Lightning running properly, and (2) either can or should be run in the background. Its
* responsibilities are:
- * Monitoring whether the ChannelManager needs to be re-persisted to disk, and if so,
+ * Processing [`Event`]s with a user-provided [`EventHandler`].
+ * Monitoring whether the [`ChannelManager`] needs to be re-persisted to disk, and if so,
* writing it to disk/backups by invoking the callback given to it at startup.
- * ChannelManager persistence should be done in the background.
- * Calling `ChannelManager::timer_tick_occurred()` and
- * `PeerManager::timer_tick_occurred()` every minute (can be done in the
- * background).
+ * [`ChannelManager`] persistence should be done in the background.
+ * Calling [`ChannelManager::timer_tick_occurred`] and [`PeerManager::timer_tick_occurred`]
+ * at the appropriate intervals.
*
- * Note that if ChannelManager persistence fails and the persisted manager becomes out-of-date,
- * then there is a risk of channels force-closing on startup when the manager realizes it's
- * outdated. However, as long as `ChannelMonitor` backups are sound, no funds besides those used
- * for unilateral chain closure fees are at risk.
+ * It will also call [`PeerManager::process_events`] periodically though this shouldn't be relied
+ * upon as doing so may result in high latency.
+ *
+ * # Note
+ *
+ * If [`ChannelManager`] persistence fails and the persisted manager becomes out-of-date, then
+ * there is a risk of channels force-closing on startup when the manager realizes it's outdated.
+ * However, as long as [`ChannelMonitor`] backups are sound, no funds besides those used for
+ * unilateral chain closure fees are at risk.
+ *
+ * [`ChannelMonitor`]: lightning::chain::channelmonitor::ChannelMonitor
+ * [`Event`]: lightning::util::events::Event
* BackgroundProcessor will immediately stop on drop. It should be stored until shutdown.
*/
@SuppressWarnings("unchecked") // We correctly assign various generic arrays
* `persist_manager` returns an error. In case of an error, the error is retrieved by calling
* either [`join`] or [`stop`].
*
- * Typically, users should either implement [`ChannelManagerPersister`] to never return an
- * error or call [`join`] and handle any error that may arise. For the latter case, the
- * `BackgroundProcessor` must be restarted by calling `start` again after handling the error.
+ * # Data Persistence
*
* `persist_manager` is responsible for writing out the [`ChannelManager`] to disk, and/or
* uploading to one or more backup services. See [`ChannelManager::write`] for writing out a
* [`ChannelManager`]. See [`FilesystemPersister::persist_manager`] for Rust-Lightning's
* provided implementation.
*
+ * Typically, users should either implement [`ChannelManagerPersister`] to never return an
+ * error or call [`join`] and handle any error that may arise. For the latter case,
+ * `BackgroundProcessor` must be restarted by calling `start` again after handling the error.
+ *
+ * # Event Handling
+ *
+ * `event_handler` is responsible for handling events that users should be notified of (e.g.,
+ * payment failed). [`BackgroundProcessor`] may decorate the given [`EventHandler`] with common
+ * functionality implemented by other handlers.
+ * [`NetGraphMsgHandler`] if given will update the [`NetworkGraph`] based on payment failures.
+ *
* [top-level documentation]: Self
* [`join`]: Self::join
* [`stop`]: Self::stop
* [`ChannelManager`]: lightning::ln::channelmanager::ChannelManager
* [`ChannelManager::write`]: lightning::ln::channelmanager::ChannelManager#impl-Writeable
* [`FilesystemPersister::persist_manager`]: lightning_persister::FilesystemPersister::persist_manager
+ * [`NetworkGraph`]: lightning::routing::network_graph::NetworkGraph
+ *
+ * Note that net_graph_msg_handler (or a relevant inner pointer) may be NULL or all-0s to represent None
*/
- public static BackgroundProcessor start(ChannelManagerPersister persister, EventHandler event_handler, ChainMonitor chain_monitor, ChannelManager channel_manager, PeerManager peer_manager, Logger logger) {
- long ret = bindings.BackgroundProcessor_start(persister == null ? 0 : persister.ptr, event_handler == null ? 0 : event_handler.ptr, chain_monitor == null ? 0 : chain_monitor.ptr & ~1, channel_manager == null ? 0 : channel_manager.ptr & ~1, peer_manager == null ? 0 : peer_manager.ptr & ~1, logger == null ? 0 : logger.ptr);
- if (ret < 1024) { return null; }
+ public static BackgroundProcessor start(ChannelManagerPersister persister, EventHandler event_handler, ChainMonitor chain_monitor, ChannelManager channel_manager, NetGraphMsgHandler net_graph_msg_handler, PeerManager peer_manager, Logger logger) {
+ long ret = bindings.BackgroundProcessor_start(persister == null ? 0 : persister.ptr, event_handler == null ? 0 : event_handler.ptr, chain_monitor == null ? 0 : chain_monitor.ptr & ~1, channel_manager == null ? 0 : channel_manager.ptr & ~1, net_graph_msg_handler == null ? 0 : net_graph_msg_handler.ptr & ~1, peer_manager == null ? 0 : peer_manager.ptr & ~1, logger == null ? 0 : logger.ptr);
+ if (ret >= 0 && ret < 1024) { return null; }
BackgroundProcessor ret_hu_conv = new BackgroundProcessor(null, ret);
ret_hu_conv.ptrs_to.add(ret_hu_conv);
ret_hu_conv.ptrs_to.add(persister);
ret_hu_conv.ptrs_to.add(event_handler);
ret_hu_conv.ptrs_to.add(chain_monitor);
ret_hu_conv.ptrs_to.add(channel_manager);
+ ret_hu_conv.ptrs_to.add(net_graph_msg_handler);
ret_hu_conv.ptrs_to.add(peer_manager);
ret_hu_conv.ptrs_to.add(logger);
return ret_hu_conv;
*/
public Result_NoneErrorZ join() {
long ret = bindings.BackgroundProcessor_join(this.ptr);
- if (ret < 1024) { return null; }
+ if (ret >= 0 && ret < 1024) { return null; }
Result_NoneErrorZ ret_hu_conv = Result_NoneErrorZ.constr_from_ptr(ret);
this.ptrs_to.add(this);
// Due to rust's strict-ownership memory model, in some cases we need to "move"
*/
public Result_NoneErrorZ stop() {
long ret = bindings.BackgroundProcessor_stop(this.ptr);
- if (ret < 1024) { return null; }
+ if (ret >= 0 && ret < 1024) { return null; }
Result_NoneErrorZ ret_hu_conv = Result_NoneErrorZ.constr_from_ptr(ret);
this.ptrs_to.add(this);
// Due to rust's strict-ownership memory model, in some cases we need to "move"