2020-09-22 14:53:52 +02:00
|
|
|
// Copyright 2015-2020 Parity Technologies (UK) Ltd.
|
|
|
|
// This file is part of OpenEthereum.
|
2016-04-21 13:57:27 +02:00
|
|
|
|
2020-09-22 14:53:52 +02:00
|
|
|
// OpenEthereum is free software: you can redistribute it and/or modify
|
2016-04-21 13:57:27 +02:00
|
|
|
// it under the terms of the GNU General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
|
2020-09-22 14:53:52 +02:00
|
|
|
// OpenEthereum is distributed in the hope that it will be useful,
|
2016-04-21 13:57:27 +02:00
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU General Public License for more details.
|
|
|
|
|
|
|
|
// You should have received a copy of the GNU General Public License
|
2020-09-22 14:53:52 +02:00
|
|
|
// along with OpenEthereum. If not, see <http://www.gnu.org/licenses/>.
|
2016-04-21 13:57:27 +02:00
|
|
|
|
2016-05-24 20:30:21 +02:00
|
|
|
extern crate ansi_term;
|
2020-08-05 06:08:03 +02:00
|
|
|
use self::ansi_term::{
|
|
|
|
Colour,
|
|
|
|
Colour::{Blue, Cyan, Green, White, Yellow},
|
|
|
|
Style,
|
|
|
|
};
|
2016-05-24 20:30:21 +02:00
|
|
|
|
2020-08-05 06:08:03 +02:00
|
|
|
use std::{
|
|
|
|
sync::{
|
|
|
|
atomic::{AtomicBool, AtomicUsize, Ordering as AtomicOrdering},
|
|
|
|
Arc,
|
|
|
|
},
|
|
|
|
time::{Duration, Instant},
|
|
|
|
};
|
2017-07-10 13:21:11 +02:00
|
|
|
|
2018-04-11 12:56:37 +02:00
|
|
|
use atty;
|
2020-08-05 06:08:03 +02:00
|
|
|
use ethcore::{
|
|
|
|
client::{
|
|
|
|
BlockChainClient, BlockChainInfo, BlockId, BlockInfo, BlockQueueInfo, ChainInfo,
|
|
|
|
ChainNotify, Client, ClientIoMessage, ClientReport, NewBlocks,
|
|
|
|
},
|
|
|
|
snapshot::{service::Service as SnapshotService, RestorationStatus, SnapshotService as SS},
|
2018-03-13 11:49:57 +01:00
|
|
|
};
|
2020-08-05 06:08:03 +02:00
|
|
|
use io::{IoContext, IoHandler, TimerToken};
|
|
|
|
use number_prefix::{binary_prefix, Prefixed, Standalone};
|
|
|
|
use parity_rpc::{informant::RpcStats, is_major_importing_or_waiting};
|
|
|
|
use parking_lot::{Mutex, RwLock};
|
2020-08-13 18:25:19 +02:00
|
|
|
use sync::{ManageNetwork, SyncProvider};
|
2020-08-05 06:08:03 +02:00
|
|
|
use types::BlockNumber;
|
2016-04-21 13:57:27 +02:00
|
|
|
|
2016-10-24 18:27:23 +02:00
|
|
|
/// Format byte counts to standard denominations.
|
|
|
|
pub fn format_bytes(b: usize) -> String {
|
2020-08-05 06:08:03 +02:00
|
|
|
match binary_prefix(b as f64) {
|
|
|
|
Standalone(bytes) => format!("{} bytes", bytes),
|
|
|
|
Prefixed(prefix, n) => format!("{:.0} {}B", n, prefix),
|
|
|
|
}
|
2016-10-24 18:27:23 +02:00
|
|
|
}
|
|
|
|
|
2016-09-16 12:18:27 +02:00
|
|
|
/// Something that can be converted to milliseconds.
|
2016-09-15 16:56:10 +02:00
|
|
|
pub trait MillisecondDuration {
|
2020-08-05 06:08:03 +02:00
|
|
|
/// Get the value in milliseconds.
|
|
|
|
fn as_milliseconds(&self) -> u64;
|
2016-05-25 09:57:31 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
impl MillisecondDuration for Duration {
|
2020-08-05 06:08:03 +02:00
|
|
|
fn as_milliseconds(&self) -> u64 {
|
|
|
|
self.as_secs() * 1000 + self.subsec_nanos() as u64 / 1_000_000
|
|
|
|
}
|
2016-05-25 09:57:31 +02:00
|
|
|
}
|
|
|
|
|
2017-07-10 13:21:11 +02:00
|
|
|
#[derive(Default)]
|
|
|
|
struct CacheSizes {
|
2020-08-05 06:08:03 +02:00
|
|
|
sizes: ::std::collections::BTreeMap<&'static str, usize>,
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
impl CacheSizes {
|
2020-08-05 06:08:03 +02:00
|
|
|
fn insert(&mut self, key: &'static str, bytes: usize) {
|
|
|
|
self.sizes.insert(key, bytes);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn display<F>(&self, style: Style, paint: F) -> String
|
|
|
|
where
|
|
|
|
F: Fn(Style, String) -> String,
|
|
|
|
{
|
|
|
|
use std::fmt::Write;
|
|
|
|
|
|
|
|
let mut buf = String::new();
|
|
|
|
for (name, &size) in &self.sizes {
|
|
|
|
write!(buf, " {:>8} {}", paint(style, format_bytes(size)), name)
|
|
|
|
.expect("writing to string won't fail unless OOM; qed")
|
|
|
|
}
|
|
|
|
|
|
|
|
buf
|
|
|
|
}
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
pub struct SyncInfo {
|
2020-08-05 06:08:03 +02:00
|
|
|
last_imported_block_number: BlockNumber,
|
2020-12-15 15:59:08 +01:00
|
|
|
last_imported_ancient_number: Option<BlockNumber>,
|
2020-08-05 06:08:03 +02:00
|
|
|
num_peers: usize,
|
|
|
|
max_peers: u32,
|
|
|
|
snapshot_sync: bool,
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
pub struct Report {
|
2020-08-05 06:08:03 +02:00
|
|
|
importing: bool,
|
|
|
|
chain_info: BlockChainInfo,
|
|
|
|
client_report: ClientReport,
|
|
|
|
queue_info: BlockQueueInfo,
|
|
|
|
cache_sizes: CacheSizes,
|
|
|
|
sync_info: Option<SyncInfo>,
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Something which can provide data to the informant.
|
|
|
|
pub trait InformantData: Send + Sync {
|
2020-08-05 06:08:03 +02:00
|
|
|
/// Whether it executes transactions
|
|
|
|
fn executes_transactions(&self) -> bool;
|
2017-07-10 13:21:11 +02:00
|
|
|
|
2020-08-05 06:08:03 +02:00
|
|
|
/// Whether it is currently importing (also included in `Report`)
|
|
|
|
fn is_major_importing(&self) -> bool;
|
2017-07-10 13:21:11 +02:00
|
|
|
|
2020-08-05 06:08:03 +02:00
|
|
|
/// Generate a report of blockchain status, memory usage, and sync info.
|
|
|
|
fn report(&self) -> Report;
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Informant data for a full node.
|
|
|
|
pub struct FullNodeInformantData {
|
2020-08-05 06:08:03 +02:00
|
|
|
pub client: Arc<Client>,
|
2020-07-29 10:36:15 +02:00
|
|
|
pub sync: Option<Arc<dyn SyncProvider>>,
|
|
|
|
pub net: Option<Arc<dyn ManageNetwork>>,
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
impl InformantData for FullNodeInformantData {
|
2020-08-05 06:08:03 +02:00
|
|
|
fn executes_transactions(&self) -> bool {
|
|
|
|
true
|
|
|
|
}
|
|
|
|
|
|
|
|
fn is_major_importing(&self) -> bool {
|
|
|
|
let state = self.sync.as_ref().map(|sync| sync.status().state);
|
|
|
|
is_major_importing_or_waiting(state, self.client.queue_info(), false)
|
|
|
|
}
|
|
|
|
|
|
|
|
fn report(&self) -> Report {
|
|
|
|
let (client_report, queue_info, blockchain_cache_info) = (
|
|
|
|
self.client.report(),
|
|
|
|
self.client.queue_info(),
|
|
|
|
self.client.blockchain_cache_info(),
|
|
|
|
);
|
|
|
|
|
|
|
|
let chain_info = self.client.chain_info();
|
|
|
|
|
|
|
|
let mut cache_sizes = CacheSizes::default();
|
|
|
|
cache_sizes.insert("queue", queue_info.mem_used);
|
|
|
|
cache_sizes.insert("chain", blockchain_cache_info.total());
|
|
|
|
|
|
|
|
let importing = self.is_major_importing();
|
|
|
|
let sync_info = match (self.sync.as_ref(), self.net.as_ref()) {
|
|
|
|
(Some(sync), Some(net)) => {
|
|
|
|
let status = sync.status();
|
|
|
|
let num_peers_range = net.num_peers_range();
|
|
|
|
debug_assert!(num_peers_range.end() >= num_peers_range.start());
|
|
|
|
|
|
|
|
Some(SyncInfo {
|
|
|
|
last_imported_block_number: status
|
|
|
|
.last_imported_block_number
|
|
|
|
.unwrap_or(chain_info.best_block_number),
|
2020-12-15 15:59:08 +01:00
|
|
|
last_imported_ancient_number: status.last_imported_old_block_number,
|
2020-08-05 06:08:03 +02:00
|
|
|
num_peers: status.num_peers,
|
|
|
|
max_peers: status
|
|
|
|
.current_max_peers(*num_peers_range.start(), *num_peers_range.end()),
|
|
|
|
snapshot_sync: status.is_snapshot_syncing(),
|
|
|
|
})
|
|
|
|
}
|
|
|
|
_ => None,
|
|
|
|
};
|
|
|
|
|
|
|
|
Report {
|
|
|
|
importing,
|
|
|
|
chain_info,
|
|
|
|
client_report,
|
|
|
|
queue_info,
|
|
|
|
cache_sizes,
|
|
|
|
sync_info,
|
|
|
|
}
|
|
|
|
}
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
pub struct Informant<T> {
|
2020-08-05 06:08:03 +02:00
|
|
|
last_tick: RwLock<Instant>,
|
|
|
|
with_color: bool,
|
|
|
|
target: T,
|
|
|
|
snapshot: Option<Arc<SnapshotService>>,
|
|
|
|
rpc_stats: Option<Arc<RpcStats>>,
|
|
|
|
last_import: Mutex<Instant>,
|
|
|
|
skipped: AtomicUsize,
|
|
|
|
skipped_txs: AtomicUsize,
|
|
|
|
in_shutdown: AtomicBool,
|
|
|
|
last_report: Mutex<ClientReport>,
|
2017-07-10 13:21:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
impl<T: InformantData> Informant<T> {
|
2020-08-05 06:08:03 +02:00
|
|
|
/// Make a new instance potentially `with_color` output.
|
|
|
|
pub fn new(
|
|
|
|
target: T,
|
|
|
|
snapshot: Option<Arc<SnapshotService>>,
|
|
|
|
rpc_stats: Option<Arc<RpcStats>>,
|
|
|
|
with_color: bool,
|
|
|
|
) -> Self {
|
|
|
|
Informant {
|
|
|
|
last_tick: RwLock::new(Instant::now()),
|
|
|
|
with_color: with_color,
|
|
|
|
target: target,
|
|
|
|
snapshot: snapshot,
|
|
|
|
rpc_stats: rpc_stats,
|
|
|
|
last_import: Mutex::new(Instant::now()),
|
|
|
|
skipped: AtomicUsize::new(0),
|
|
|
|
skipped_txs: AtomicUsize::new(0),
|
|
|
|
in_shutdown: AtomicBool::new(false),
|
|
|
|
last_report: Mutex::new(Default::default()),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Signal that we're shutting down; no more output necessary.
|
|
|
|
pub fn shutdown(&self) {
|
|
|
|
self.in_shutdown
|
|
|
|
.store(true, ::std::sync::atomic::Ordering::SeqCst);
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn tick(&self) {
|
|
|
|
let now = Instant::now();
|
2021-01-14 15:55:52 +01:00
|
|
|
let elapsed;
|
|
|
|
{
|
|
|
|
let last_tick = self.last_tick.read();
|
|
|
|
if now < *last_tick + Duration::from_millis(1500) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
elapsed = now - *last_tick;
|
|
|
|
}
|
2020-08-05 06:08:03 +02:00
|
|
|
|
|
|
|
let (client_report, full_report) = {
|
2020-07-29 11:00:04 +02:00
|
|
|
let last_report = self.last_report.lock();
|
2020-08-05 06:08:03 +02:00
|
|
|
let full_report = self.target.report();
|
|
|
|
let diffed = full_report.client_report.clone() - &*last_report;
|
|
|
|
(diffed, full_report)
|
|
|
|
};
|
|
|
|
|
|
|
|
let Report {
|
|
|
|
importing,
|
|
|
|
chain_info,
|
|
|
|
queue_info,
|
|
|
|
cache_sizes,
|
|
|
|
sync_info,
|
|
|
|
..
|
|
|
|
} = full_report;
|
|
|
|
|
|
|
|
let rpc_stats = self.rpc_stats.as_ref();
|
|
|
|
let snapshot_sync = sync_info.as_ref().map_or(false, |s| s.snapshot_sync)
|
2020-09-14 16:08:57 +02:00
|
|
|
&& self
|
|
|
|
.snapshot
|
|
|
|
.as_ref()
|
|
|
|
.map_or(false, |s| match s.restoration_status() {
|
|
|
|
RestorationStatus::Ongoing { .. } | RestorationStatus::Initializing { .. } => {
|
|
|
|
true
|
|
|
|
}
|
|
|
|
_ => false,
|
|
|
|
});
|
2020-08-05 06:08:03 +02:00
|
|
|
if !importing && !snapshot_sync && elapsed < Duration::from_secs(30) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
*self.last_tick.write() = now;
|
|
|
|
*self.last_report.lock() = full_report.client_report.clone();
|
|
|
|
|
|
|
|
let paint = |c: Style, t: String| match self.with_color && atty::is(atty::Stream::Stdout) {
|
|
|
|
true => format!("{}", c.paint(t)),
|
|
|
|
false => t,
|
|
|
|
};
|
|
|
|
|
2020-12-10 17:57:26 +01:00
|
|
|
info!(target: "import", "{}{} {} {} {}",
|
2020-08-05 06:08:03 +02:00
|
|
|
match importing {
|
|
|
|
true => match snapshot_sync {
|
|
|
|
false => format!("Syncing {} {} {} {}+{} Qed",
|
|
|
|
paint(White.bold(), format!("{:>8}", format!("#{}", chain_info.best_block_number))),
|
|
|
|
paint(White.bold(), format!("{}", chain_info.best_block_hash)),
|
|
|
|
if self.target.executes_transactions() {
|
|
|
|
format!("{} blk/s {} tx/s {} Mgas/s",
|
|
|
|
paint(Yellow.bold(), format!("{:7.2}", (client_report.blocks_imported * 1000) as f64 / elapsed.as_milliseconds() as f64)),
|
|
|
|
paint(Yellow.bold(), format!("{:6.1}", (client_report.transactions_applied * 1000) as f64 / elapsed.as_milliseconds() as f64)),
|
|
|
|
paint(Yellow.bold(), format!("{:6.1}", (client_report.gas_processed / 1000).low_u64() as f64 / elapsed.as_milliseconds() as f64))
|
|
|
|
)
|
|
|
|
} else {
|
|
|
|
format!("{} hdr/s",
|
|
|
|
paint(Yellow.bold(), format!("{:6.1}", (client_report.blocks_imported * 1000) as f64 / elapsed.as_milliseconds() as f64))
|
|
|
|
)
|
|
|
|
},
|
|
|
|
paint(Green.bold(), format!("{:5}", queue_info.unverified_queue_size)),
|
|
|
|
paint(Green.bold(), format!("{:5}", queue_info.verified_queue_size))
|
|
|
|
),
|
|
|
|
true => {
|
|
|
|
self.snapshot.as_ref().map_or(String::new(), |s|
|
2020-09-14 16:08:57 +02:00
|
|
|
match s.restoration_status() {
|
|
|
|
RestorationStatus::Ongoing { state_chunks, block_chunks, state_chunks_done, block_chunks_done, .. } => {
|
2020-08-05 06:08:03 +02:00
|
|
|
format!("Syncing snapshot {}/{}", state_chunks_done + block_chunks_done, state_chunks + block_chunks)
|
|
|
|
},
|
|
|
|
RestorationStatus::Initializing { chunks_done } => {
|
|
|
|
format!("Snapshot initializing ({} chunks restored)", chunks_done)
|
|
|
|
},
|
|
|
|
_ => String::new(),
|
|
|
|
}
|
|
|
|
)
|
|
|
|
},
|
|
|
|
},
|
|
|
|
false => String::new(),
|
|
|
|
},
|
2020-12-10 17:57:26 +01:00
|
|
|
match chain_info.ancient_block_number {
|
2020-12-15 15:59:08 +01:00
|
|
|
Some(ancient_number) => format!(" Ancient:#{}", ancient_number),
|
2020-12-10 17:57:26 +01:00
|
|
|
None => String::new(),
|
|
|
|
},
|
2020-08-05 06:08:03 +02:00
|
|
|
match sync_info.as_ref() {
|
|
|
|
Some(ref sync_info) => format!("{}{}/{} peers",
|
|
|
|
match importing {
|
|
|
|
true => format!("{}",
|
|
|
|
if self.target.executes_transactions() {
|
2020-12-10 17:57:26 +01:00
|
|
|
paint(Green.bold(), format!("{:>8} ", format!("LI:#{}", sync_info.last_imported_block_number)))
|
2020-08-05 06:08:03 +02:00
|
|
|
} else {
|
|
|
|
String::new()
|
|
|
|
}
|
|
|
|
),
|
2020-12-15 15:59:08 +01:00
|
|
|
false => match sync_info.last_imported_ancient_number {
|
2020-12-10 17:57:26 +01:00
|
|
|
Some(number) => format!("{} ", paint(Yellow.bold(), format!("{:>8}", format!("AB:#{}", number)))),
|
2020-08-05 06:08:03 +02:00
|
|
|
None => String::new(),
|
|
|
|
}
|
|
|
|
},
|
|
|
|
paint(Cyan.bold(), format!("{:2}", sync_info.num_peers)),
|
|
|
|
paint(Cyan.bold(), format!("{:2}", sync_info.max_peers)),
|
|
|
|
),
|
|
|
|
_ => String::new(),
|
|
|
|
},
|
|
|
|
cache_sizes.display(Blue.bold(), &paint),
|
|
|
|
match rpc_stats {
|
|
|
|
Some(ref rpc_stats) => format!(
|
|
|
|
"RPC: {} conn, {} req/s, {} µs",
|
|
|
|
paint(Blue.bold(), format!("{:2}", rpc_stats.sessions())),
|
|
|
|
paint(Blue.bold(), format!("{:4}", rpc_stats.requests_rate())),
|
|
|
|
paint(Blue.bold(), format!("{:4}", rpc_stats.approximated_roundtrip())),
|
|
|
|
),
|
|
|
|
_ => String::new(),
|
|
|
|
},
|
|
|
|
);
|
|
|
|
}
|
2016-04-21 13:57:27 +02:00
|
|
|
}
|
|
|
|
|
2017-07-10 13:21:11 +02:00
|
|
|
impl ChainNotify for Informant<FullNodeInformantData> {
|
2020-12-02 11:31:11 +01:00
|
|
|
// t_nb 11.2 Informant. Prints new block inclusiong to console/log.
|
2020-08-05 06:08:03 +02:00
|
|
|
fn new_blocks(&self, new_blocks: NewBlocks) {
|
|
|
|
if new_blocks.has_more_blocks_to_import {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
let mut last_import = self.last_import.lock();
|
|
|
|
let client = &self.target.client;
|
|
|
|
|
|
|
|
let importing = self.target.is_major_importing();
|
|
|
|
let ripe = Instant::now() > *last_import + Duration::from_secs(1) && !importing;
|
|
|
|
let txs_imported = new_blocks
|
|
|
|
.imported
|
|
|
|
.iter()
|
|
|
|
.take(
|
|
|
|
new_blocks
|
|
|
|
.imported
|
|
|
|
.len()
|
|
|
|
.saturating_sub(if ripe { 1 } else { 0 }),
|
|
|
|
)
|
|
|
|
.filter_map(|h| client.block(BlockId::Hash(*h)))
|
|
|
|
.map(|b| b.transactions_count())
|
|
|
|
.sum();
|
|
|
|
|
|
|
|
if ripe {
|
|
|
|
if let Some(block) = new_blocks
|
|
|
|
.imported
|
|
|
|
.last()
|
|
|
|
.and_then(|h| client.block(BlockId::Hash(*h)))
|
|
|
|
{
|
|
|
|
let header_view = block.header_view();
|
|
|
|
let size = block.rlp().as_raw().len();
|
|
|
|
let (skipped, skipped_txs) = (
|
|
|
|
self.skipped.load(AtomicOrdering::Relaxed) + new_blocks.imported.len() - 1,
|
|
|
|
self.skipped_txs.load(AtomicOrdering::Relaxed) + txs_imported,
|
|
|
|
);
|
|
|
|
info!(target: "import", "Imported {} {} ({} txs, {} Mgas, {} ms, {} KiB){}",
|
|
|
|
Colour::White.bold().paint(format!("#{}", header_view.number())),
|
|
|
|
Colour::White.bold().paint(format!("{}", header_view.hash())),
|
|
|
|
Colour::Yellow.bold().paint(format!("{}", block.transactions_count())),
|
|
|
|
Colour::Yellow.bold().paint(format!("{:.2}", header_view.gas_used().low_u64() as f32 / 1000000f32)),
|
|
|
|
Colour::Purple.bold().paint(format!("{}", new_blocks.duration.as_milliseconds())),
|
|
|
|
Colour::Blue.bold().paint(format!("{:.2}", size as f32 / 1024f32)),
|
|
|
|
if skipped > 0 {
|
|
|
|
format!(" + another {} block(s) containing {} tx(s)",
|
|
|
|
Colour::Red.bold().paint(format!("{}", skipped)),
|
|
|
|
Colour::Red.bold().paint(format!("{}", skipped_txs))
|
|
|
|
)
|
|
|
|
} else {
|
|
|
|
String::new()
|
|
|
|
}
|
|
|
|
);
|
|
|
|
self.skipped.store(0, AtomicOrdering::Relaxed);
|
|
|
|
self.skipped_txs.store(0, AtomicOrdering::Relaxed);
|
|
|
|
*last_import = Instant::now();
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
self.skipped
|
|
|
|
.fetch_add(new_blocks.imported.len(), AtomicOrdering::Relaxed);
|
|
|
|
self.skipped_txs
|
|
|
|
.fetch_add(txs_imported, AtomicOrdering::Relaxed);
|
|
|
|
}
|
|
|
|
}
|
2016-07-20 12:36:20 +02:00
|
|
|
}
|
|
|
|
|
2016-12-11 02:02:40 +01:00
|
|
|
const INFO_TIMER: TimerToken = 0;
|
|
|
|
|
2017-07-10 13:21:11 +02:00
|
|
|
impl<T: InformantData> IoHandler<ClientIoMessage> for Informant<T> {
|
2020-08-05 06:08:03 +02:00
|
|
|
fn initialize(&self, io: &IoContext<ClientIoMessage>) {
|
|
|
|
io.register_timer(INFO_TIMER, Duration::from_secs(5))
|
|
|
|
.expect("Error registering timer");
|
|
|
|
}
|
|
|
|
|
|
|
|
fn timeout(&self, _io: &IoContext<ClientIoMessage>, timer: TimerToken) {
|
|
|
|
if timer == INFO_TIMER && !self.in_shutdown.load(AtomicOrdering::SeqCst) {
|
|
|
|
self.tick();
|
|
|
|
}
|
|
|
|
}
|
2016-12-10 23:58:39 +01:00
|
|
|
}
|