2016-02-05 13:40:41 +01:00
// Copyright 2015, 2016 Ethcore (UK) Ltd.
// This file is part of Parity.
// Parity is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
// Parity is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
// You should have received a copy of the GNU General Public License
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
2016-02-15 00:51:50 +01:00
///
2016-04-06 10:07:24 +02:00
/// `BlockChain` synchronization strategy.
2016-02-15 00:51:50 +01:00
/// Syncs to peers and keeps up to date.
2016-01-09 18:40:13 +01:00
/// This implementation uses ethereum protocol v63
///
2016-05-16 14:41:41 +02:00
/// Syncing strategy summary.
/// Split the chain into ranges of N blocks each. Download ranges sequentially. Split each range into subchains of M blocks. Download subchains in parallel.
/// State.
/// Sync state consists of the following data:
2016-05-25 17:03:58 +02:00
/// - s: State enum which can be one of the following values: `ChainHead`, `Blocks`, `Idle`
2016-05-16 14:41:41 +02:00
/// - H: A set of downloaded block headers
/// - B: A set of downloaded block bodies
/// - S: Set of block subchain start block hashes to download.
/// - l: Last imported / common block hash
/// - P: A set of connected peers. For each peer we maintain its last known total difficulty and starting block hash being requested if any.
/// General behaviour.
2016-05-25 17:03:58 +02:00
/// We start with all sets empty, l is set to the best block in the block chain, s is set to `ChainHead`.
/// If at any moment a bad block is reported by the block queue, we set s to `ChainHead`, reset l to the best block in the block chain and clear H, B and S.
/// If at any moment P becomes empty, we set s to `ChainHead`, and clear H, B and S.
2016-05-16 14:41:41 +02:00
///
2016-05-25 17:03:58 +02:00
/// Workflow for `ChainHead` state.
/// In this state we try to get subchain headers with a single `GetBlockHeaders` request.
/// On `NewPeer` / On `Restart`:
2016-05-16 14:41:41 +02:00
/// If peer's total difficulty is higher, request N/M headers with interval M+1 starting from l
2016-05-25 17:03:58 +02:00
/// On `BlockHeaders(R)`:
2016-05-16 14:41:41 +02:00
/// If R is empty:
/// If l is equal to genesis block hash or l is more than 1000 blocks behind our best hash:
/// Remove current peer from P. set l to the best block in the block chain. Select peer with maximum total difficulty from P and restart.
/// Else
/// Set l to l’ s parent and restart.
/// Else if we already have all the headers in the block chain or the block queue:
2016-05-25 17:03:58 +02:00
/// Set s to `Idle`,
2016-05-16 14:41:41 +02:00
/// Else
2016-05-25 17:03:58 +02:00
/// Set S to R, set s to `Blocks`.
2016-05-16 14:41:41 +02:00
///
///
/// All other messages are ignored.
2016-05-25 17:03:58 +02:00
/// Workflow for `Blocks` state.
2016-05-16 14:41:41 +02:00
/// In this state we download block headers and bodies from multiple peers.
2016-05-25 17:03:58 +02:00
/// On `NewPeer` / On `Restart`:
2016-05-16 14:41:41 +02:00
/// For all idle peers:
/// Find a set of 256 or less block hashes in H which are not in B and not being downloaded by other peers. If the set is not empty:
/// Request block bodies for the hashes in the set.
/// Else
/// Find an element in S which is not being downloaded by other peers. If found: Request M headers starting from the element.
///
2016-05-25 17:03:58 +02:00
/// On `BlockHeaders(R)`:
2016-05-16 14:41:41 +02:00
/// If R is empty remove current peer from P and restart.
/// Validate received headers. For each header find a parent in H or R or the blockchain. Restart if there is a block with unknown parent.
2016-05-25 17:03:58 +02:00
/// Go to `CollectBlocks`.
2016-05-16 14:41:41 +02:00
///
2016-05-25 17:03:58 +02:00
/// On `BlockBodies(R)`:
2016-05-16 14:41:41 +02:00
/// If R is empty remove current peer from P and restart.
/// Add bodies with a matching header in H to B.
2016-05-25 17:03:58 +02:00
/// Go to `CollectBlocks`.
2016-01-09 18:40:13 +01:00
///
2016-05-25 17:03:58 +02:00
/// `CollectBlocks`:
2016-05-16 14:41:41 +02:00
/// Find a chain of blocks C in H starting from h where h’ s parent equals to l. The chain ends with the first block which does not have a body in B.
/// Add all blocks from the chain to the block queue. Remove them from H and B. Set l to the hash of the last block from C.
/// Update and merge subchain heads in S. For each h in S find a chain of blocks in B starting from h. Remove h from S. if the chain does not include an element from S add the end of the chain to S.
2016-05-25 17:03:58 +02:00
/// If H is empty and S contains a single element set s to `ChainHead`.
2016-05-16 14:41:41 +02:00
/// Restart.
///
/// All other messages are ignored.
/// Workflow for Idle state.
2016-05-25 17:03:58 +02:00
/// On `NewBlock`:
/// Import the block. If the block is unknown set s to `ChainHead` and restart.
/// On `NewHashes`:
/// Set s to `ChainHead` and restart.
2016-05-16 14:41:41 +02:00
///
/// All other messages are ignored.
2016-01-09 18:40:13 +01:00
///
2016-01-10 14:11:23 +01:00
use util ::* ;
2016-09-01 14:49:12 +02:00
use rlp ::* ;
2016-08-05 10:32:04 +02:00
use network ::* ;
2016-10-18 18:16:00 +02:00
use ethcore ::views ::{ HeaderView } ;
2016-02-03 21:42:30 +01:00
use ethcore ::header ::{ BlockNumber , Header as BlockHeader } ;
2016-07-01 21:40:54 +02:00
use ethcore ::client ::{ BlockChainClient , BlockStatus , BlockID , BlockChainInfo , BlockImportError } ;
2016-02-03 21:42:30 +01:00
use ethcore ::error ::* ;
2016-09-06 15:31:13 +02:00
use ethcore ::snapshot ::{ ManifestData , RestorationStatus } ;
2016-08-05 10:32:04 +02:00
use sync_io ::SyncIo ;
2016-02-03 21:42:30 +01:00
use time ;
2016-02-24 21:23:58 +01:00
use super ::SyncConfig ;
2016-10-18 18:16:00 +02:00
use block_sync ::{ BlockDownloader , BlockRequest , BlockDownloaderImportError as DownloaderImportError } ;
2016-09-06 15:31:13 +02:00
use snapshot ::{ Snapshot , ChunkType } ;
2016-06-22 12:10:26 +02:00
use rand ::{ thread_rng , Rng } ;
2016-10-24 16:24:35 +02:00
use api ::{ PeerInfo as PeerInfoDigest , WARP_SYNC_PROTOCOL_ID } ;
2015-12-22 22:19:50 +01:00
2016-05-16 14:41:41 +02:00
known_heap_size! ( 0 , PeerInfo ) ;
2015-12-25 14:55:55 +01:00
2016-01-08 17:52:25 +01:00
type PacketDecodeError = DecoderError ;
2016-10-18 18:16:00 +02:00
const PROTOCOL_VERSION_63 : u8 = 63 ;
2016-10-24 16:24:35 +02:00
const PROTOCOL_VERSION_1 : u8 = 1 ;
2015-12-24 17:18:47 +01:00
const MAX_BODIES_TO_SEND : usize = 256 ;
2015-12-27 00:48:03 +01:00
const MAX_HEADERS_TO_SEND : usize = 512 ;
2015-12-24 17:18:47 +01:00
const MAX_NODE_DATA_TO_SEND : usize = 1024 ;
const MAX_RECEIPTS_TO_SEND : usize = 1024 ;
2016-03-13 15:59:25 +01:00
const MAX_RECEIPTS_HEADERS_TO_SEND : usize = 256 ;
2016-02-05 18:34:08 +01:00
const MIN_PEERS_PROPAGATION : usize = 4 ;
const MAX_PEERS_PROPAGATION : usize = 128 ;
2016-02-06 18:56:21 +01:00
const MAX_PEER_LAG_PROPAGATION : BlockNumber = 20 ;
2016-06-20 00:40:11 +02:00
const MAX_NEW_HASHES : usize = 64 ;
const MAX_TX_TO_IMPORT : usize = 512 ;
2016-07-23 17:27:23 +02:00
const MAX_NEW_BLOCK_AGE : BlockNumber = 20 ;
2016-10-08 10:35:54 +02:00
const MAX_TRANSACTION_SIZE : usize = 300 * 1024 ;
2016-10-18 18:16:00 +02:00
// Min number of blocks to be behind for a snapshot sync
const SNAPSHOT_RESTORE_THRESHOLD : BlockNumber = 100000 ;
2016-10-29 13:07:06 +02:00
const SNAPSHOT_MIN_PEERS : usize = 3 ;
2015-12-24 17:18:47 +01:00
2015-12-22 22:19:50 +01:00
const STATUS_PACKET : u8 = 0x00 ;
const NEW_BLOCK_HASHES_PACKET : u8 = 0x01 ;
const TRANSACTIONS_PACKET : u8 = 0x02 ;
const GET_BLOCK_HEADERS_PACKET : u8 = 0x03 ;
const BLOCK_HEADERS_PACKET : u8 = 0x04 ;
const GET_BLOCK_BODIES_PACKET : u8 = 0x05 ;
const BLOCK_BODIES_PACKET : u8 = 0x06 ;
const NEW_BLOCK_PACKET : u8 = 0x07 ;
const GET_NODE_DATA_PACKET : u8 = 0x0d ;
const NODE_DATA_PACKET : u8 = 0x0e ;
const GET_RECEIPTS_PACKET : u8 = 0x0f ;
const RECEIPTS_PACKET : u8 = 0x10 ;
2016-10-24 16:24:35 +02:00
pub const ETH_PACKET_COUNT : u8 = 0x11 ;
2016-09-06 15:31:13 +02:00
const GET_SNAPSHOT_MANIFEST_PACKET : u8 = 0x11 ;
const SNAPSHOT_MANIFEST_PACKET : u8 = 0x12 ;
const GET_SNAPSHOT_DATA_PACKET : u8 = 0x13 ;
const SNAPSHOT_DATA_PACKET : u8 = 0x14 ;
2015-12-22 22:19:50 +01:00
2016-10-24 16:24:35 +02:00
pub const SNAPSHOT_SYNC_PACKET_COUNT : u8 = 0x15 ;
2016-10-29 13:07:06 +02:00
const MAX_SNAPSHOT_CHUNKS_DOWNLOAD_AHEAD : usize = 3 ;
const WAIT_PEERS_TIMEOUT_SEC : u64 = 5 ;
const STATUS_TIMEOUT_SEC : u64 = 5 ;
const HEADERS_TIMEOUT_SEC : u64 = 15 ;
const BODIES_TIMEOUT_SEC : u64 = 10 ;
const RECEIPTS_TIMEOUT_SEC : u64 = 10 ;
const FORK_HEADER_TIMEOUT_SEC : u64 = 3 ;
const SNAPSHOT_MANIFEST_TIMEOUT_SEC : u64 = 3 ;
const SNAPSHOT_DATA_TIMEOUT_SEC : u64 = 60 ;
2016-02-03 21:42:30 +01:00
2015-12-22 22:19:50 +01:00
#[ derive(Copy, Clone, Eq, PartialEq, Debug) ]
2016-02-03 21:05:04 +01:00
/// Sync state
2015-12-22 22:19:50 +01:00
pub enum SyncState {
2016-10-29 13:07:06 +02:00
/// Collecting enough peers to start syncing.
WaitingPeers ,
/// Waiting for snapshot manifest download
2016-09-06 15:31:13 +02:00
SnapshotManifest ,
/// Downloading snapshot data
SnapshotData ,
2016-10-29 13:07:06 +02:00
/// Waiting for snapshot restoration progress.
2016-09-06 15:31:13 +02:00
SnapshotWaiting ,
2016-10-18 18:16:00 +02:00
/// Downloading new blocks
Blocks ,
2015-12-22 22:19:50 +01:00
/// Initial chain sync complete. Waiting for new packets
Idle ,
2016-01-10 23:37:09 +01:00
/// Block downloading paused. Waiting for block queue to process blocks and free some space
2015-12-22 22:19:50 +01:00
Waiting ,
2016-05-02 13:13:12 +02:00
/// Downloading blocks learned from `NewHashes` packet
2015-12-22 22:19:50 +01:00
NewBlocks ,
}
2016-01-09 18:40:13 +01:00
/// Syncing status and statistics
2016-09-01 12:23:31 +02:00
#[ derive(Clone, Copy) ]
2015-12-22 22:19:50 +01:00
pub struct SyncStatus {
2016-01-09 18:40:13 +01:00
/// State
2015-12-25 14:55:55 +01:00
pub state : SyncState ,
2016-01-09 18:40:13 +01:00
/// Syncing protocol version. That's the maximum protocol version we connect to.
2015-12-25 14:55:55 +01:00
pub protocol_version : u8 ,
2016-03-26 00:22:09 +01:00
/// The underlying p2p network version.
2016-11-03 22:22:25 +01:00
pub network_id : usize ,
2016-04-06 10:07:24 +02:00
/// `BlockChain` height for the moment the sync started.
2015-12-25 14:55:55 +01:00
pub start_block_number : BlockNumber ,
2016-02-03 21:05:04 +01:00
/// Last fully downloaded and imported block number (if any).
pub last_imported_block_number : Option < BlockNumber > ,
/// Highest block number in the download queue (if any).
pub highest_block_number : Option < BlockNumber > ,
2016-01-09 18:40:13 +01:00
/// Total number of blocks for the sync process.
2016-02-03 21:05:04 +01:00
pub blocks_total : BlockNumber ,
2016-01-09 18:40:13 +01:00
/// Number of blocks downloaded so far.
2016-02-03 21:05:04 +01:00
pub blocks_received : BlockNumber ,
2016-01-22 04:54:38 +01:00
/// Total number of connected peers
pub num_peers : usize ,
2016-09-06 15:31:13 +02:00
/// Total number of active peers.
2016-01-22 04:54:38 +01:00
pub num_active_peers : usize ,
2016-09-06 15:31:13 +02:00
/// Heap memory used in bytes.
2016-02-24 22:37:28 +01:00
pub mem_used : usize ,
2016-09-06 15:31:13 +02:00
/// Snapshot chunks
pub num_snapshot_chunks : usize ,
/// Snapshot chunks downloaded
pub snapshot_chunks_done : usize ,
2016-10-18 18:16:00 +02:00
/// Last fully downloaded and imported ancient block number (if any).
pub last_imported_old_block_number : Option < BlockNumber > ,
2015-12-22 22:19:50 +01:00
}
2016-07-20 12:36:20 +02:00
impl SyncStatus {
2016-10-18 18:16:00 +02:00
/// Indicates if snapshot download is in progress
pub fn is_snapshot_syncing ( & self ) -> bool {
2016-10-27 08:28:12 +02:00
self . state = = SyncState ::SnapshotManifest
| | self . state = = SyncState ::SnapshotData
2016-10-18 18:16:00 +02:00
| | self . state = = SyncState ::SnapshotWaiting
}
2016-08-11 17:19:20 +02:00
/// Returns max no of peers to display in informants
pub fn current_max_peers ( & self , min_peers : u32 , max_peers : u32 ) -> u32 {
if self . num_peers as u32 > min_peers {
max_peers
} else {
min_peers
}
}
2016-07-20 12:36:20 +02:00
}
2016-02-06 18:56:21 +01:00
#[ derive(PartialEq, Eq, Debug, Clone) ]
2016-01-10 23:37:09 +01:00
/// Peer data type requested
2016-01-08 16:26:00 +01:00
enum PeerAsking {
2015-12-22 22:19:50 +01:00
Nothing ,
2016-07-27 21:38:22 +02:00
ForkHeader ,
2015-12-22 22:19:50 +01:00
BlockHeaders ,
BlockBodies ,
2016-10-18 18:16:00 +02:00
BlockReceipts ,
2016-09-06 15:31:13 +02:00
SnapshotManifest ,
SnapshotData ,
2015-12-22 22:19:50 +01:00
}
2016-10-18 18:16:00 +02:00
#[ derive(PartialEq, Eq, Debug, Clone, Copy) ]
/// Block downloader channel.
enum BlockSet {
/// New blocks better than out best blocks
NewBlocks ,
/// Missing old blocks
OldBlocks ,
}
2016-08-07 22:42:35 +02:00
#[ derive(Clone, Eq, PartialEq) ]
enum ForkConfirmation {
/// Fork block confirmation pending.
Unconfirmed ,
/// Peers chain is too short to confirm the fork.
TooShort ,
/// Fork is confurmed.
Confirmed ,
}
2016-02-06 18:56:21 +01:00
#[ derive(Clone) ]
2016-01-10 23:37:09 +01:00
/// Syncing peer information
2015-12-22 22:19:50 +01:00
struct PeerInfo {
2016-01-10 23:37:09 +01:00
/// eth protocol version
2016-10-18 18:16:00 +02:00
protocol_version : u8 ,
2016-01-10 23:37:09 +01:00
/// Peer chain genesis hash
2015-12-22 22:19:50 +01:00
genesis : H256 ,
2016-02-03 21:05:04 +01:00
/// Peer network id
2016-11-03 22:22:25 +01:00
network_id : usize ,
2016-01-10 23:37:09 +01:00
/// Peer best block hash
2016-02-14 17:10:55 +01:00
latest_hash : H256 ,
2016-05-16 14:41:41 +02:00
/// Peer total difficulty if known
difficulty : Option < U256 > ,
2016-01-10 23:37:09 +01:00
/// Type of data currenty being requested from peer.
2015-12-22 22:19:50 +01:00
asking : PeerAsking ,
2016-01-10 23:37:09 +01:00
/// A set of block numbers being requested
2016-05-16 14:41:41 +02:00
asking_blocks : Vec < H256 > ,
2016-02-12 13:07:02 +01:00
/// Holds requested header hash if currently requesting block header by hash
asking_hash : Option < H256 > ,
2016-09-06 15:31:13 +02:00
/// Holds requested snapshot chunk hash if any.
asking_snapshot_data : Option < H256 > ,
2016-02-03 21:42:30 +01:00
/// Request timestamp
2016-10-29 13:07:06 +02:00
ask_time : u64 ,
2016-09-01 19:07:58 +02:00
/// Holds a set of transactions recently sent to this peer to avoid spamming.
last_sent_transactions : HashSet < H256 > ,
/// Pending request is expired and result should be ignored
2016-07-23 17:27:23 +02:00
expired : bool ,
2016-08-07 22:42:35 +02:00
/// Peer fork confirmation status
confirmation : ForkConfirmation ,
2016-09-06 15:31:13 +02:00
/// Best snapshot hash
snapshot_hash : Option < H256 > ,
/// Best snapshot block number
snapshot_number : Option < BlockNumber > ,
2016-10-18 18:16:00 +02:00
/// Block set requested
block_set : Option < BlockSet > ,
2016-07-27 21:38:22 +02:00
}
impl PeerInfo {
2016-08-07 22:42:35 +02:00
fn can_sync ( & self ) -> bool {
self . confirmation = = ForkConfirmation ::Confirmed & & ! self . expired
}
fn is_allowed ( & self ) -> bool {
self . confirmation ! = ForkConfirmation ::Unconfirmed & & ! self . expired
2016-07-27 21:38:22 +02:00
}
2015-12-22 22:19:50 +01:00
}
2016-01-10 23:37:09 +01:00
/// Blockchain sync handler.
/// See module documentation for more details.
2015-12-22 22:19:50 +01:00
pub struct ChainSync {
/// Sync state
state : SyncState ,
/// Last block number for the start of sync
starting_block : BlockNumber ,
/// Highest block number seen
2016-02-03 21:05:04 +01:00
highest_block : Option < BlockNumber > ,
2016-05-16 14:41:41 +02:00
/// All connected peers
2015-12-22 22:19:50 +01:00
peers : HashMap < PeerId , PeerInfo > ,
2016-05-16 14:41:41 +02:00
/// Peers active for current sync round
active_peers : HashSet < PeerId > ,
2016-10-18 18:16:00 +02:00
/// Block download process for new blocks
new_blocks : BlockDownloader ,
/// Block download process for ancient blocks
old_blocks : Option < BlockDownloader > ,
2016-02-12 13:07:02 +01:00
/// Last propagated block number
2016-03-07 12:16:37 +01:00
last_sent_block_number : BlockNumber ,
2016-02-24 21:23:58 +01:00
/// Network ID
2016-11-03 22:22:25 +01:00
network_id : usize ,
2016-07-27 21:38:22 +02:00
/// Optional fork block to check
fork_block : Option < ( BlockNumber , H256 ) > ,
2016-09-06 15:31:13 +02:00
/// Snapshot downloader.
snapshot : Snapshot ,
2016-10-29 13:07:06 +02:00
/// Connected peers pending Status message.
/// Value is request timestamp.
handshaking_peers : HashMap < PeerId , u64 > ,
/// Sync start timestamp. Measured when first peer is connected
sync_start_time : Option < u64 > ,
2015-12-22 22:19:50 +01:00
}
2016-02-04 23:24:36 +01:00
type RlpResponseResult = Result < Option < ( PacketId , RlpStream ) > , PacketDecodeError > ;
2015-12-22 22:19:50 +01:00
impl ChainSync {
2016-01-09 19:13:58 +01:00
/// Create a new instance of syncing strategy.
2016-05-31 20:54:02 +02:00
pub fn new ( config : SyncConfig , chain : & BlockChainClient ) -> ChainSync {
2016-10-18 18:16:00 +02:00
let chain_info = chain . chain_info ( ) ;
let mut sync = ChainSync {
2016-10-29 13:07:06 +02:00
state : if config . warp_sync { SyncState ::WaitingPeers } else { SyncState ::Idle } ,
2016-10-18 18:16:00 +02:00
starting_block : chain . chain_info ( ) . best_block_number ,
2016-02-03 21:05:04 +01:00
highest_block : None ,
2015-12-22 22:19:50 +01:00
peers : HashMap ::new ( ) ,
2016-10-29 13:07:06 +02:00
handshaking_peers : HashMap ::new ( ) ,
2016-05-16 14:41:41 +02:00
active_peers : HashSet ::new ( ) ,
2016-10-18 18:16:00 +02:00
new_blocks : BlockDownloader ::new ( false , & chain_info . best_block_hash , chain_info . best_block_number ) ,
old_blocks : None ,
2016-03-07 12:16:37 +01:00
last_sent_block_number : 0 ,
2016-02-24 21:23:58 +01:00
network_id : config . network_id ,
2016-07-27 21:38:22 +02:00
fork_block : config . fork_block ,
2016-09-06 15:31:13 +02:00
snapshot : Snapshot ::new ( ) ,
2016-10-29 13:07:06 +02:00
sync_start_time : None ,
2016-10-18 18:16:00 +02:00
} ;
2016-10-29 13:07:06 +02:00
sync . update_targets ( chain ) ;
2016-10-18 18:16:00 +02:00
sync
2015-12-22 22:19:50 +01:00
}
2016-10-12 20:18:59 +02:00
/// Returns synchonization status
2015-12-22 22:19:50 +01:00
pub fn status ( & self ) -> SyncStatus {
2016-10-18 18:16:00 +02:00
let last_imported_number = self . new_blocks . last_imported_block_number ( ) ;
2015-12-22 22:19:50 +01:00
SyncStatus {
state : self . state . clone ( ) ,
2016-10-24 16:24:35 +02:00
protocol_version : PROTOCOL_VERSION_63 ,
2016-03-26 00:22:09 +01:00
network_id : self . network_id ,
2015-12-22 22:19:50 +01:00
start_block_number : self . starting_block ,
2016-10-18 18:16:00 +02:00
last_imported_block_number : Some ( last_imported_number ) ,
last_imported_old_block_number : self . old_blocks . as_ref ( ) . map ( | d | d . last_imported_block_number ( ) ) ,
highest_block_number : self . highest_block . map ( | n | max ( n , last_imported_number ) ) ,
blocks_received : if last_imported_number > self . starting_block { last_imported_number - self . starting_block } else { 0 } ,
2016-02-29 14:59:10 +01:00
blocks_total : match self . highest_block { Some ( x ) if x > self . starting_block = > x - self . starting_block , _ = > 0 } ,
2016-08-07 22:42:35 +02:00
num_peers : self . peers . values ( ) . filter ( | p | p . is_allowed ( ) ) . count ( ) ,
num_active_peers : self . peers . values ( ) . filter ( | p | p . is_allowed ( ) & & p . asking ! = PeerAsking ::Nothing ) . count ( ) ,
2016-09-06 15:31:13 +02:00
num_snapshot_chunks : self . snapshot . total_chunks ( ) ,
snapshot_chunks_done : self . snapshot . done_chunks ( ) ,
2016-03-01 22:30:23 +01:00
mem_used :
2016-10-18 18:16:00 +02:00
self . new_blocks . heap_size ( )
+ self . old_blocks . as_ref ( ) . map_or ( 0 , | d | d . heap_size ( ) )
+ self . peers . heap_size_of_children ( ) ,
2015-12-22 22:19:50 +01:00
}
}
2016-10-12 20:18:59 +02:00
/// Returns information on peers connections
pub fn peers ( & self , io : & SyncIo ) -> Vec < PeerInfoDigest > {
self . peers . iter ( )
2016-10-27 08:28:12 +02:00
. filter_map ( | ( & peer_id , peer_data ) |
2016-10-12 20:18:59 +02:00
io . peer_session_info ( peer_id ) . map ( | session_info |
PeerInfoDigest {
id : session_info . id . map ( | id | id . hex ( ) ) ,
client_version : session_info . client_version ,
capabilities : session_info . peer_capabilities . into_iter ( ) . map ( | c | c . to_string ( ) ) . collect ( ) ,
remote_address : session_info . remote_address ,
local_address : session_info . local_address ,
2016-10-18 18:16:00 +02:00
eth_version : peer_data . protocol_version as u32 ,
2016-10-12 20:18:59 +02:00
eth_difficulty : peer_data . difficulty ,
eth_head : peer_data . latest_hash ,
} )
)
. collect ( )
}
2015-12-22 22:19:50 +01:00
/// Abort all sync activity
pub fn abort ( & mut self , io : & mut SyncIo ) {
2016-10-18 18:16:00 +02:00
self . reset_and_continue ( io ) ;
2015-12-22 22:19:50 +01:00
self . peers . clear ( ) ;
}
2016-03-11 11:16:49 +01:00
#[ cfg_attr(feature= " dev " , allow(for_kv_map)) ] // Because it's not possible to get `values_mut()`
2016-05-16 14:41:41 +02:00
/// Reset sync. Clear all downloaded data but keep the queue
2016-09-06 15:31:13 +02:00
fn reset ( & mut self , io : & mut SyncIo ) {
2016-10-18 18:16:00 +02:00
self . new_blocks . reset ( ) ;
2016-09-06 15:31:13 +02:00
self . snapshot . clear ( ) ;
if self . state = = SyncState ::SnapshotData {
debug! ( target :" sync " , " Aborting snapshot restore " ) ;
io . snapshot_service ( ) . abort_restore ( ) ;
}
2016-01-17 15:56:09 +01:00
for ( _ , ref mut p ) in & mut self . peers {
2016-10-18 18:16:00 +02:00
if p . block_set ! = Some ( BlockSet ::OldBlocks ) {
p . asking_blocks . clear ( ) ;
p . asking_hash = None ;
// mark any pending requests as expired
if p . asking ! = PeerAsking ::Nothing & & p . is_allowed ( ) {
p . expired = true ;
}
2016-07-23 17:27:23 +02:00
}
2015-12-22 22:19:50 +01:00
}
self . state = SyncState ::Idle ;
2016-10-10 23:05:41 +02:00
// Reactivate peers only if some progress has been made
// since the last sync round of if starting fresh.
2016-05-16 14:41:41 +02:00
self . active_peers = self . peers . keys ( ) . cloned ( ) . collect ( ) ;
2015-12-22 22:19:50 +01:00
}
/// Restart sync
2016-10-18 18:16:00 +02:00
pub fn reset_and_continue ( & mut self , io : & mut SyncIo ) {
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " Restarting " ) ;
2016-09-06 15:31:13 +02:00
self . reset ( io ) ;
2016-05-16 14:41:41 +02:00
self . continue_sync ( io ) ;
}
2016-10-10 23:05:41 +02:00
/// Remove peer from active peer set. Peer will be reactivated on the next sync
/// round.
2016-10-18 18:16:00 +02:00
fn deactivate_peer ( & mut self , _io : & mut SyncIo , peer_id : PeerId ) {
2016-05-31 15:34:25 +02:00
trace! ( target : " sync " , " Deactivating peer {} " , peer_id ) ;
2016-05-16 14:41:41 +02:00
self . active_peers . remove ( & peer_id ) ;
2015-12-22 22:19:50 +01:00
}
2016-10-29 13:07:06 +02:00
fn maybe_start_snapshot_sync ( & mut self , io : & mut SyncIo ) {
if self . state ! = SyncState ::WaitingPeers {
return ;
}
2016-11-09 19:40:36 +01:00
// Make sure the snapshot block is not too far away from best block and network best block and
// that it is higher than fork detection block
let our_best_block = io . chain ( ) . chain_info ( ) . best_block_number ;
let fork_block = self . fork_block . as_ref ( ) . map ( | & ( n , _ ) | n ) . unwrap_or ( 0 ) ;
2016-10-29 13:07:06 +02:00
let ( best_hash , max_peers , snapshot_peers ) = {
//collect snapshot infos from peers
let snapshots = self . peers . iter ( )
2016-11-09 19:40:36 +01:00
. filter ( | & ( _ , p ) | p . is_allowed ( ) & & p . snapshot_number . map_or ( false , | sn |
our_best_block < sn & & ( sn - our_best_block ) > SNAPSHOT_RESTORE_THRESHOLD & &
sn > fork_block & &
self . highest_block . map_or ( true , | highest | highest > = sn & & ( highest - sn ) < = SNAPSHOT_RESTORE_THRESHOLD )
) )
2016-10-29 13:07:06 +02:00
. filter_map ( | ( p , peer ) | peer . snapshot_hash . map ( | hash | ( p , hash . clone ( ) ) ) ) ;
let mut snapshot_peers = HashMap ::new ( ) ;
let mut max_peers : usize = 0 ;
let mut best_hash = None ;
for ( p , hash ) in snapshots {
let peers = snapshot_peers . entry ( hash ) . or_insert_with ( Vec ::new ) ;
peers . push ( * p ) ;
if peers . len ( ) > max_peers {
max_peers = peers . len ( ) ;
best_hash = Some ( hash ) ;
}
}
( best_hash , max_peers , snapshot_peers )
} ;
let timeout = self . sync_start_time . map_or ( false , | t | ( ( time ::precise_time_ns ( ) - t ) / 1_000_000_000 ) > WAIT_PEERS_TIMEOUT_SEC ) ;
if let ( Some ( hash ) , Some ( peers ) ) = ( best_hash , best_hash . map_or ( None , | h | snapshot_peers . get ( & h ) ) ) {
if max_peers > = SNAPSHOT_MIN_PEERS {
trace! ( target : " sync " , " Starting confirmed snapshot sync {:?} with {:?} " , hash , peers ) ;
self . start_snapshot_sync ( io , peers ) ;
} else if timeout {
trace! ( target : " sync " , " Starting unconfirmed snapshot sync {:?} with {:?} " , hash , peers ) ;
self . start_snapshot_sync ( io , peers ) ;
}
} else if timeout {
trace! ( target : " sync " , " No snapshots found, starting full sync " ) ;
self . state = SyncState ::Idle ;
self . continue_sync ( io ) ;
}
}
fn start_snapshot_sync ( & mut self , io : & mut SyncIo , peers : & [ PeerId ] ) {
2016-09-06 15:31:13 +02:00
self . snapshot . clear ( ) ;
2016-10-29 13:07:06 +02:00
for p in peers {
if self . peers . get ( p ) . map_or ( false , | p | p . asking = = PeerAsking ::Nothing ) {
self . request_snapshot_manifest ( io , * p ) ;
}
}
2016-09-06 15:31:13 +02:00
self . state = SyncState ::SnapshotManifest ;
}
2016-09-16 23:03:26 +02:00
2016-10-18 18:16:00 +02:00
/// Restart sync disregarding the block queue status. May end up re-downloading up to QUEUE_SIZE blocks
pub fn restart ( & mut self , io : & mut SyncIo ) {
2016-10-29 13:07:06 +02:00
self . update_targets ( io . chain ( ) ) ;
2016-10-18 18:16:00 +02:00
self . reset_and_continue ( io ) ;
}
2016-10-27 08:28:12 +02:00
2016-10-29 13:07:06 +02:00
/// Update sync after the blockchain has been changed externally.
pub fn update_targets ( & mut self , chain : & BlockChainClient ) {
2016-03-26 12:00:05 +01:00
// Do not assume that the block queue/chain still has our last_imported_block
2016-10-18 18:16:00 +02:00
let chain = chain . chain_info ( ) ;
self . new_blocks = BlockDownloader ::new ( false , & chain . best_block_hash , chain . best_block_number ) ;
if let ( Some ( ancient_block_hash ) , Some ( ancient_block_number ) ) = ( chain . ancient_block_hash , chain . ancient_block_number ) {
trace! ( target : " sync " , " Downloading old blocks from {:?} (#{}) till {:?} (#{:?}) " , ancient_block_hash , ancient_block_number , chain . first_block_hash , chain . first_block_number ) ;
let mut downloader = BlockDownloader ::new ( true , & ancient_block_hash , ancient_block_number ) ;
if let Some ( hash ) = chain . first_block_hash {
trace! ( target : " sync " , " Downloader target set to {:?} " , hash ) ;
downloader . set_target ( & hash ) ;
}
self . old_blocks = Some ( downloader ) ;
} else {
self . old_blocks = None ;
}
2016-03-26 12:00:05 +01:00
}
2016-05-16 14:41:41 +02:00
2015-12-22 22:19:50 +01:00
/// Called by peer to report status
2016-01-14 19:03:48 +01:00
fn on_peer_status ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-10-29 13:07:06 +02:00
self . handshaking_peers . remove ( & peer_id ) ;
2016-10-18 18:16:00 +02:00
let protocol_version : u8 = try ! ( r . val_at ( 0 ) ) ;
2016-10-24 16:24:35 +02:00
let warp_protocol = io . protocol_version ( & WARP_SYNC_PROTOCOL_ID , peer_id ) ! = 0 ;
2016-02-06 21:08:20 +01:00
let peer = PeerInfo {
2016-09-06 15:31:13 +02:00
protocol_version : protocol_version ,
2016-01-08 17:52:25 +01:00
network_id : try ! ( r . val_at ( 1 ) ) ,
2016-05-16 14:41:41 +02:00
difficulty : Some ( try ! ( r . val_at ( 2 ) ) ) ,
2016-02-14 17:10:55 +01:00
latest_hash : try ! ( r . val_at ( 3 ) ) ,
2016-01-08 17:52:25 +01:00
genesis : try ! ( r . val_at ( 4 ) ) ,
2015-12-22 22:19:50 +01:00
asking : PeerAsking ::Nothing ,
asking_blocks : Vec ::new ( ) ,
2016-02-12 13:07:02 +01:00
asking_hash : None ,
2016-10-29 13:07:06 +02:00
ask_time : 0 ,
2016-09-01 19:07:58 +02:00
last_sent_transactions : HashSet ::new ( ) ,
2016-07-23 17:27:23 +02:00
expired : false ,
2016-08-07 22:42:35 +02:00
confirmation : if self . fork_block . is_none ( ) { ForkConfirmation ::Confirmed } else { ForkConfirmation ::Unconfirmed } ,
2016-09-06 15:31:13 +02:00
asking_snapshot_data : None ,
2016-10-24 16:24:35 +02:00
snapshot_hash : if warp_protocol { Some ( try ! ( r . val_at ( 5 ) ) ) } else { None } ,
snapshot_number : if warp_protocol { Some ( try ! ( r . val_at ( 6 ) ) ) } else { None } ,
2016-10-18 18:16:00 +02:00
block_set : None ,
2015-12-22 22:19:50 +01:00
} ;
2015-12-25 14:55:55 +01:00
2016-10-29 13:07:06 +02:00
if self . sync_start_time . is_none ( ) {
self . sync_start_time = Some ( time ::precise_time_ns ( ) ) ;
}
trace! ( target : " sync " , " New peer {} (protocol: {}, network: {:?}, difficulty: {:?}, latest:{}, genesis:{}, snapshot:{:?}) " ,
peer_id , peer . protocol_version , peer . network_id , peer . difficulty , peer . latest_hash , peer . genesis , peer . snapshot_number ) ;
2016-06-17 16:01:33 +02:00
if io . is_expired ( ) {
2016-06-20 01:26:40 +02:00
trace! ( target : " sync " , " Status packet from expired session {}:{} " , peer_id , io . peer_info ( peer_id ) ) ;
2016-06-17 16:01:33 +02:00
return Ok ( ( ) ) ;
}
2016-02-03 21:05:04 +01:00
2016-02-12 13:07:02 +01:00
if self . peers . contains_key ( & peer_id ) {
2016-06-20 01:26:40 +02:00
debug! ( target : " sync " , " Unexpected status packet from {}:{} " , peer_id , io . peer_info ( peer_id ) ) ;
2016-02-12 13:07:02 +01:00
return Ok ( ( ) ) ;
}
2016-01-10 14:11:23 +01:00
let chain_info = io . chain ( ) . chain_info ( ) ;
if peer . genesis ! = chain_info . genesis_hash {
io . disable_peer ( peer_id ) ;
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " Peer {} genesis hash mismatch (ours: {}, theirs: {}) " , peer_id , chain_info . genesis_hash , peer . genesis ) ;
2016-01-10 14:11:23 +01:00
return Ok ( ( ) ) ;
}
2016-02-24 21:23:58 +01:00
if peer . network_id ! = self . network_id {
2016-01-10 14:11:23 +01:00
io . disable_peer ( peer_id ) ;
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " Peer {} network id mismatch (ours: {}, theirs: {}) " , peer_id , self . network_id , peer . network_id ) ;
2016-01-10 14:11:23 +01:00
return Ok ( ( ) ) ;
}
2016-10-24 16:24:35 +02:00
if ( warp_protocol & & peer . protocol_version ! = PROTOCOL_VERSION_1 ) | | ( ! warp_protocol & & peer . protocol_version ! = PROTOCOL_VERSION_63 ) {
2016-10-18 18:16:00 +02:00
io . disable_peer ( peer_id ) ;
trace! ( target : " sync " , " Peer {} unsupported eth protocol ({}) " , peer_id , peer . protocol_version ) ;
return Ok ( ( ) ) ;
}
2015-12-25 14:55:55 +01:00
2016-02-12 13:07:02 +01:00
self . peers . insert ( peer_id . clone ( ) , peer ) ;
2016-10-18 18:16:00 +02:00
self . active_peers . insert ( peer_id . clone ( ) ) ;
2016-02-14 01:03:48 +01:00
debug! ( target : " sync " , " Connected {}:{} " , peer_id , io . peer_info ( peer_id ) ) ;
2016-07-27 21:38:22 +02:00
if let Some ( ( fork_block , _ ) ) = self . fork_block {
2016-10-18 18:16:00 +02:00
self . request_fork_header_by_number ( io , peer_id , fork_block ) ;
2016-07-27 21:38:22 +02:00
} else {
self . sync_peer ( io , peer_id , false ) ;
}
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
2015-12-22 22:19:50 +01:00
}
2016-03-11 11:16:49 +01:00
#[ cfg_attr(feature= " dev " , allow(cyclomatic_complexity)) ]
2015-12-22 22:19:50 +01:00
/// Called by peer once it has new block headers during sync
2016-01-14 19:03:48 +01:00
fn on_peer_block_headers ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-07-27 21:38:22 +02:00
let confirmed = match self . peers . get_mut ( & peer_id ) {
Some ( ref mut peer ) if peer . asking = = PeerAsking ::ForkHeader = > {
2016-10-18 18:16:00 +02:00
peer . asking = PeerAsking ::Nothing ;
2016-07-27 21:38:22 +02:00
let item_count = r . item_count ( ) ;
2016-10-18 18:16:00 +02:00
let ( fork_number , fork_hash ) = self . fork_block . expect ( " ForkHeader request is sent only fork block is Some; qed " ) . clone ( ) ;
if item_count = = 0 | | item_count ! = 1 {
trace! ( target : " sync " , " {}: Chain is too short to confirm the block " , peer_id ) ;
peer . confirmation = ForkConfirmation ::TooShort ;
} else {
let header = try ! ( r . at ( 0 ) ) . as_raw ( ) ;
if header . sha3 ( ) = = fork_hash {
2016-08-07 22:42:35 +02:00
trace! ( target : " sync " , " {}: Confirmed peer " , peer_id ) ;
peer . confirmation = ForkConfirmation ::Confirmed ;
2016-10-18 18:16:00 +02:00
if ! io . chain_overlay ( ) . read ( ) . contains_key ( & fork_number ) {
io . chain_overlay ( ) . write ( ) . insert ( fork_number , header . to_vec ( ) ) ;
}
} else {
trace! ( target : " sync " , " {}: Fork mismatch " , peer_id ) ;
io . disconnect_peer ( peer_id ) ;
return Ok ( ( ) ) ;
2016-08-07 22:42:35 +02:00
}
2016-07-27 21:38:22 +02:00
}
2016-10-18 18:16:00 +02:00
true
2016-07-27 21:38:22 +02:00
} ,
_ = > false ,
} ;
if confirmed {
self . sync_peer ( io , peer_id , false ) ;
return Ok ( ( ) ) ;
}
2016-05-16 14:41:41 +02:00
self . clear_peer_download ( peer_id ) ;
2016-07-27 21:38:22 +02:00
let expected_hash = self . peers . get ( & peer_id ) . and_then ( | p | p . asking_hash ) ;
2016-10-18 18:16:00 +02:00
let block_set = self . peers . get ( & peer_id ) . and_then ( | p | p . block_set ) . unwrap_or ( BlockSet ::NewBlocks ) ;
if ! self . reset_peer_asking ( peer_id , PeerAsking ::BlockHeaders ) | | expected_hash . is_none ( ) {
trace! ( target : " sync " , " {}: Ignored unexpected headers, expected_hash = {:?} " , peer_id , expected_hash ) ;
2016-05-16 14:41:41 +02:00
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
2015-12-22 22:19:50 +01:00
let item_count = r . item_count ( ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} -> BlockHeaders ({} entries), state = {:?}, set = {:?} " , peer_id , item_count , self . state , block_set ) ;
2016-10-29 13:07:06 +02:00
if ( self . state = = SyncState ::Idle | | self . state = = SyncState ::WaitingPeers ) & & self . old_blocks . is_none ( ) {
2015-12-22 22:19:50 +01:00
trace! ( target : " sync " , " Ignored unexpected block headers " ) ;
2016-05-16 14:41:41 +02:00
self . continue_sync ( io ) ;
2016-01-08 17:52:25 +01:00
return Ok ( ( ) ) ;
2015-12-22 22:19:50 +01:00
}
2016-01-08 16:26:00 +01:00
if self . state = = SyncState ::Waiting {
2015-12-22 22:19:50 +01:00
trace! ( target : " sync " , " Ignored block headers while waiting " ) ;
2016-05-16 14:41:41 +02:00
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
2016-02-03 21:05:04 +01:00
2016-10-18 18:16:00 +02:00
let result = {
let mut downloader = match block_set {
BlockSet ::NewBlocks = > & mut self . new_blocks ,
BlockSet ::OldBlocks = > {
match self . old_blocks {
None = > {
trace! ( target : " sync " , " Ignored block headers while block download is inactive " ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
Some ( ref mut blocks ) = > blocks ,
2016-03-11 18:50:29 +01:00
}
2015-12-24 17:18:47 +01:00
}
2016-10-18 18:16:00 +02:00
} ;
downloader . import_headers ( io , r , expected_hash )
} ;
2016-06-20 00:10:34 +02:00
2016-10-18 18:16:00 +02:00
match result {
Err ( DownloaderImportError ::Useless ) = > {
self . deactivate_peer ( io , peer_id ) ;
2016-06-14 13:25:06 +02:00
} ,
2016-10-18 18:16:00 +02:00
Err ( DownloaderImportError ::Invalid ) = > {
io . disable_peer ( peer_id ) ;
self . deactivate_peer ( io , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
2016-06-14 13:25:06 +02:00
} ,
2016-10-18 18:16:00 +02:00
Ok ( ( ) ) = > ( ) ,
2016-05-16 14:41:41 +02:00
}
2016-10-18 18:16:00 +02:00
self . collect_blocks ( io , block_set ) ;
2016-07-23 17:27:23 +02:00
// give a task to the same peer first if received valuable headers.
self . sync_peer ( io , peer_id , false ) ;
// give tasks to other peers
2015-12-27 00:48:03 +01:00
self . continue_sync ( io ) ;
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
2015-12-22 22:19:50 +01:00
}
/// Called by peer once it has new block bodies
2016-01-14 19:03:48 +01:00
fn on_peer_block_bodies ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-05-16 14:41:41 +02:00
self . clear_peer_download ( peer_id ) ;
2016-10-18 18:16:00 +02:00
let block_set = self . peers . get ( & peer_id ) . and_then ( | p | p . block_set ) . unwrap_or ( BlockSet ::NewBlocks ) ;
if ! self . reset_peer_asking ( peer_id , PeerAsking ::BlockBodies ) {
trace! ( target : " sync " , " {}: Ignored unexpected bodies " , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
2015-12-24 17:18:47 +01:00
let item_count = r . item_count ( ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} -> BlockBodies ({} entries), set = {:?} " , peer_id , item_count , block_set ) ;
2016-05-16 14:41:41 +02:00
if item_count = = 0 {
self . deactivate_peer ( io , peer_id ) ;
}
else if self . state = = SyncState ::Waiting {
2015-12-24 17:18:47 +01:00
trace! ( target : " sync " , " Ignored block bodies while waiting " ) ;
2016-03-29 23:29:43 +02:00
}
2016-05-16 14:41:41 +02:00
else
{
2016-10-18 18:16:00 +02:00
let result = {
let mut downloader = match block_set {
BlockSet ::NewBlocks = > & mut self . new_blocks ,
BlockSet ::OldBlocks = > match self . old_blocks {
None = > {
trace! ( target : " sync " , " Ignored block headers while block download is inactive " ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
Some ( ref mut blocks ) = > blocks ,
}
} ;
downloader . import_bodies ( io , r )
} ;
match result {
Err ( DownloaderImportError ::Invalid ) = > {
io . disable_peer ( peer_id ) ;
self . deactivate_peer ( io , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
Err ( DownloaderImportError ::Useless ) = > {
self . deactivate_peer ( io , peer_id ) ;
} ,
Ok ( ( ) ) = > ( ) ,
2015-12-24 17:18:47 +01:00
}
2016-10-18 18:16:00 +02:00
self . collect_blocks ( io , block_set ) ;
self . sync_peer ( io , peer_id , false ) ;
}
self . continue_sync ( io ) ;
Ok ( ( ) )
}
/// Called by peer once it has new block receipts
fn on_peer_block_receipts ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
self . clear_peer_download ( peer_id ) ;
let block_set = self . peers . get ( & peer_id ) . and_then ( | p | p . block_set ) . unwrap_or ( BlockSet ::NewBlocks ) ;
if ! self . reset_peer_asking ( peer_id , PeerAsking ::BlockReceipts ) {
trace! ( target : " sync " , " {}: Ignored unexpected receipts " , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
let item_count = r . item_count ( ) ;
trace! ( target : " sync " , " {} -> BlockReceipts ({} entries) " , peer_id , item_count ) ;
if item_count = = 0 {
self . deactivate_peer ( io , peer_id ) ;
}
else if self . state = = SyncState ::Waiting {
trace! ( target : " sync " , " Ignored block receipts while waiting " ) ;
}
else
{
let result = {
let mut downloader = match block_set {
BlockSet ::NewBlocks = > & mut self . new_blocks ,
BlockSet ::OldBlocks = > match self . old_blocks {
None = > {
trace! ( target : " sync " , " Ignored block headers while block download is inactive " ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
Some ( ref mut blocks ) = > blocks ,
}
} ;
downloader . import_receipts ( io , r )
} ;
match result {
Err ( DownloaderImportError ::Invalid ) = > {
io . disable_peer ( peer_id ) ;
self . deactivate_peer ( io , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
Err ( DownloaderImportError ::Useless ) = > {
self . deactivate_peer ( io , peer_id ) ;
} ,
Ok ( ( ) ) = > ( ) ,
2016-06-20 11:07:22 +02:00
}
2016-10-18 18:16:00 +02:00
self . collect_blocks ( io , block_set ) ;
self . sync_peer ( io , peer_id , false ) ;
2015-12-24 17:18:47 +01:00
}
self . continue_sync ( io ) ;
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
2015-12-22 22:19:50 +01:00
}
/// Called by peer once it has new block bodies
2016-03-11 11:16:49 +01:00
#[ cfg_attr(feature= " dev " , allow(cyclomatic_complexity)) ]
2016-01-14 19:03:48 +01:00
fn on_peer_new_block ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-08-07 22:42:35 +02:00
if ! self . peers . get ( & peer_id ) . map_or ( false , | p | p . can_sync ( ) ) {
2016-08-05 17:01:08 +02:00
trace! ( target : " sync " , " Ignoring new block from unconfirmed peer {} " , peer_id ) ;
return Ok ( ( ) ) ;
}
2016-01-08 17:52:25 +01:00
let block_rlp = try ! ( r . at ( 0 ) ) ;
let header_rlp = try ! ( block_rlp . at ( 0 ) ) ;
2016-01-08 16:00:32 +01:00
let h = header_rlp . as_raw ( ) . sha3 ( ) ;
2016-01-08 16:26:00 +01:00
trace! ( target : " sync " , " {} -> NewBlock ({}) " , peer_id , h ) ;
2016-02-07 22:08:15 +01:00
let header : BlockHeader = try ! ( header_rlp . as_val ( ) ) ;
2016-06-23 22:40:44 +02:00
if header . number ( ) > self . highest_block . unwrap_or ( 0 ) {
self . highest_block = Some ( header . number ( ) ) ;
}
2016-05-16 14:41:41 +02:00
let mut unknown = false ;
2016-02-12 13:07:02 +01:00
{
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
peer . latest_hash = header . hash ( ) ;
}
2016-02-12 13:07:02 +01:00
}
2016-10-18 18:16:00 +02:00
let last_imported_number = self . new_blocks . last_imported_block_number ( ) ;
if last_imported_number > header . number ( ) & & last_imported_number - header . number ( ) > MAX_NEW_BLOCK_AGE {
2016-07-23 17:27:23 +02:00
trace! ( target : " sync " , " Ignored ancient new block {:?} " , h ) ;
io . disable_peer ( peer_id ) ;
return Ok ( ( ) ) ;
}
2016-06-22 12:10:26 +02:00
match io . chain ( ) . import_block ( block_rlp . as_raw ( ) . to_vec ( ) ) {
2016-07-01 21:40:54 +02:00
Err ( BlockImportError ::Import ( ImportError ::AlreadyInChain ) ) = > {
2016-06-22 12:10:26 +02:00
trace! ( target : " sync " , " New block already in chain {:?} " , h ) ;
} ,
2016-07-01 21:40:54 +02:00
Err ( BlockImportError ::Import ( ImportError ::AlreadyQueued ) ) = > {
2016-06-22 12:10:26 +02:00
trace! ( target : " sync " , " New block already queued {:?} " , h ) ;
} ,
Ok ( _ ) = > {
2016-10-18 18:16:00 +02:00
self . new_blocks . mark_as_known ( & header . hash ( ) , header . number ( ) ) ;
2016-08-29 11:35:24 +02:00
trace! ( target : " sync " , " New block queued {:?} ({}) " , h , header . number ( ) ) ;
2016-06-22 12:10:26 +02:00
} ,
2016-07-01 21:40:54 +02:00
Err ( BlockImportError ::Block ( BlockError ::UnknownParent ( p ) ) ) = > {
2016-06-22 12:10:26 +02:00
unknown = true ;
trace! ( target : " sync " , " New block with unknown parent ({:?}) {:?} " , p , h ) ;
} ,
Err ( e ) = > {
debug! ( target : " sync " , " Bad new block {:?} : {:?} " , h , e ) ;
io . disable_peer ( peer_id ) ;
}
} ;
2016-02-03 21:42:30 +01:00
if unknown {
2016-06-22 12:10:26 +02:00
if self . state ! = SyncState ::Idle {
trace! ( target : " sync " , " NewBlock ignored while seeking " ) ;
} else {
trace! ( target : " sync " , " New unknown block {:?} " , h ) ;
//TODO: handle too many unknown blocks
let difficulty : U256 = try ! ( r . val_at ( 1 ) ) ;
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
if peer . difficulty . map_or ( true , | pd | difficulty > pd ) {
peer . difficulty = Some ( difficulty ) ;
trace! ( target : " sync " , " Received block {:?} with no known parent. Peer needs syncing... " , h ) ;
}
2016-05-16 14:41:41 +02:00
}
2016-06-22 12:10:26 +02:00
self . sync_peer ( io , peer_id , true ) ;
2016-01-10 23:37:09 +01:00
}
}
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
2015-12-22 22:19:50 +01:00
}
2016-05-02 13:13:12 +02:00
/// Handles `NewHashes` packet. Initiates headers download for any unknown hashes.
2016-01-14 19:03:48 +01:00
fn on_peer_new_hashes ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-08-07 22:42:35 +02:00
if ! self . peers . get ( & peer_id ) . map_or ( false , | p | p . can_sync ( ) ) {
2016-08-05 17:01:08 +02:00
trace! ( target : " sync " , " Ignoring new hashes from unconfirmed peer {} " , peer_id ) ;
return Ok ( ( ) ) ;
}
2016-05-16 14:41:41 +02:00
if self . state ! = SyncState ::Idle {
2015-12-24 17:18:47 +01:00
trace! ( target : " sync " , " Ignoring new hashes since we're already downloading. " ) ;
2016-06-23 22:40:44 +02:00
let max = r . iter ( ) . take ( MAX_NEW_HASHES ) . map ( | item | item . val_at ::< BlockNumber > ( 1 ) . unwrap_or ( 0 ) ) . fold ( 0 u64 , max ) ;
if max > self . highest_block . unwrap_or ( 0 ) {
self . highest_block = Some ( max ) ;
}
2016-01-08 17:52:25 +01:00
return Ok ( ( ) ) ;
2015-12-24 17:18:47 +01:00
}
2016-01-08 16:26:00 +01:00
trace! ( target : " sync " , " {} -> NewHashes ({} entries) " , peer_id , r . item_count ( ) ) ;
2016-06-20 00:40:11 +02:00
let hashes = r . iter ( ) . take ( MAX_NEW_HASHES ) . map ( | item | ( item . val_at ::< H256 > ( 0 ) , item . val_at ::< BlockNumber > ( 1 ) ) ) ;
2016-02-12 13:07:02 +01:00
let mut max_height : BlockNumber = 0 ;
2016-05-16 14:41:41 +02:00
let mut new_hashes = Vec ::new ( ) ;
2016-10-18 18:16:00 +02:00
let last_imported_number = self . new_blocks . last_imported_block_number ( ) ;
2016-07-23 17:27:23 +02:00
for ( rh , rn ) in hashes {
let hash = try ! ( rh ) ;
let number = try ! ( rn ) ;
if number > self . highest_block . unwrap_or ( 0 ) {
self . highest_block = Some ( number ) ;
2016-06-23 22:40:44 +02:00
}
2016-10-18 18:16:00 +02:00
if self . new_blocks . is_downloading ( & hash ) {
2016-02-12 13:07:02 +01:00
continue ;
}
2016-10-18 18:16:00 +02:00
if last_imported_number > number & & last_imported_number - number > MAX_NEW_BLOCK_AGE {
2016-07-23 17:27:23 +02:00
trace! ( target : " sync " , " Ignored ancient new block hash {:?} " , hash ) ;
io . disable_peer ( peer_id ) ;
continue ;
}
match io . chain ( ) . block_status ( BlockID ::Hash ( hash . clone ( ) ) ) {
2015-12-24 17:18:47 +01:00
BlockStatus ::InChain = > {
2016-07-23 17:27:23 +02:00
trace! ( target : " sync " , " New block hash already in chain {:?} " , hash ) ;
2015-12-24 17:18:47 +01:00
} ,
2016-01-10 23:37:09 +01:00
BlockStatus ::Queued = > {
2016-07-23 17:27:23 +02:00
trace! ( target : " sync " , " New hash block already queued {:?} " , hash ) ;
2015-12-24 17:18:47 +01:00
} ,
BlockStatus ::Unknown = > {
2016-07-23 17:27:23 +02:00
new_hashes . push ( hash . clone ( ) ) ;
if number > max_height {
trace! ( target : " sync " , " New unknown block hash {:?} " , hash ) ;
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
peer . latest_hash = hash . clone ( ) ;
}
2016-07-23 17:27:23 +02:00
max_height = number ;
2015-12-24 17:18:47 +01:00
}
} ,
2016-05-16 14:41:41 +02:00
BlockStatus ::Bad = > {
2016-07-23 17:27:23 +02:00
debug! ( target : " sync " , " Bad new block hash {:?} " , hash ) ;
2015-12-24 17:18:47 +01:00
io . disable_peer ( peer_id ) ;
2016-01-08 17:52:25 +01:00
return Ok ( ( ) ) ;
2015-12-24 17:18:47 +01:00
}
}
2016-01-08 17:52:25 +01:00
} ;
2016-02-12 13:07:02 +01:00
if max_height ! = 0 {
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " Downloading blocks for new hashes " ) ;
2016-10-18 18:16:00 +02:00
self . new_blocks . reset_to ( new_hashes ) ;
2016-05-16 14:41:41 +02:00
self . state = SyncState ::NewBlocks ;
2016-01-23 18:44:45 +01:00
self . sync_peer ( io , peer_id , true ) ;
}
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
2015-12-22 22:19:50 +01:00
}
2016-09-06 15:31:13 +02:00
/// Called when snapshot manifest is downloaded from a peer.
fn on_snapshot_manifest ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
if ! self . peers . get ( & peer_id ) . map_or ( false , | p | p . can_sync ( ) ) {
trace! ( target : " sync " , " Ignoring snapshot manifest from unconfirmed peer {} " , peer_id ) ;
return Ok ( ( ) ) ;
}
self . clear_peer_download ( peer_id ) ;
if ! self . reset_peer_asking ( peer_id , PeerAsking ::SnapshotManifest ) | | self . state ! = SyncState ::SnapshotManifest {
2016-10-29 13:07:06 +02:00
trace! ( target : " sync " , " {}: Ignored unexpected/expired manifest " , peer_id ) ;
2016-09-06 15:31:13 +02:00
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
let manifest_rlp = try ! ( r . at ( 0 ) ) ;
2016-09-16 23:03:26 +02:00
let manifest = match ManifestData ::from_rlp ( manifest_rlp . as_raw ( ) ) {
2016-09-06 15:31:13 +02:00
Err ( e ) = > {
trace! ( target : " sync " , " {}: Ignored bad manifest: {:?} " , peer_id , e ) ;
io . disconnect_peer ( peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
Ok ( manifest ) = > manifest ,
} ;
self . snapshot . reset_to ( & manifest , & manifest_rlp . as_raw ( ) . sha3 ( ) ) ;
io . snapshot_service ( ) . begin_restore ( manifest ) ;
self . state = SyncState ::SnapshotData ;
// give a task to the same peer first.
self . sync_peer ( io , peer_id , false ) ;
// give tasks to other peers
self . continue_sync ( io ) ;
Ok ( ( ) )
}
/// Called when snapshot data is downloaded from a peer.
fn on_snapshot_data ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
if ! self . peers . get ( & peer_id ) . map_or ( false , | p | p . can_sync ( ) ) {
trace! ( target : " sync " , " Ignoring snapshot data from unconfirmed peer {} " , peer_id ) ;
return Ok ( ( ) ) ;
}
self . clear_peer_download ( peer_id ) ;
if ! self . reset_peer_asking ( peer_id , PeerAsking ::SnapshotData ) | | self . state ! = SyncState ::SnapshotData {
trace! ( target : " sync " , " {}: Ignored unexpected snapshot data " , peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
// check service status
match io . snapshot_service ( ) . status ( ) {
RestorationStatus ::Inactive | RestorationStatus ::Failed = > {
trace! ( target : " sync " , " {}: Snapshot restoration aborted " , peer_id ) ;
2016-10-29 13:07:06 +02:00
self . state = SyncState ::WaitingPeers ;
2016-09-06 15:31:13 +02:00
self . snapshot . clear ( ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
} ,
RestorationStatus ::Ongoing { .. } = > {
trace! ( target : " sync " , " {}: Snapshot restoration is ongoing " , peer_id ) ;
} ,
}
let snapshot_data : Bytes = try ! ( r . val_at ( 0 ) ) ;
match self . snapshot . validate_chunk ( & snapshot_data ) {
Ok ( ChunkType ::Block ( hash ) ) = > {
trace! ( target : " sync " , " {}: Processing block chunk " , peer_id ) ;
io . snapshot_service ( ) . restore_block_chunk ( hash , snapshot_data ) ;
}
Ok ( ChunkType ::State ( hash ) ) = > {
trace! ( target : " sync " , " {}: Processing state chunk " , peer_id ) ;
io . snapshot_service ( ) . restore_state_chunk ( hash , snapshot_data ) ;
}
Err ( ( ) ) = > {
trace! ( target : " sync " , " {}: Got bad snapshot chunk " , peer_id ) ;
io . disconnect_peer ( peer_id ) ;
self . continue_sync ( io ) ;
return Ok ( ( ) ) ;
}
}
if self . snapshot . is_complete ( ) {
// wait for snapshot restoration process to complete
self . state = SyncState ::SnapshotWaiting ;
}
// give a task to the same peer first.
self . sync_peer ( io , peer_id , false ) ;
// give tasks to other peers
self . continue_sync ( io ) ;
Ok ( ( ) )
}
2015-12-22 22:19:50 +01:00
/// Called by peer when it is disconnecting
2016-01-14 19:03:48 +01:00
pub fn on_peer_aborting ( & mut self , io : & mut SyncIo , peer : PeerId ) {
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " == Disconnecting {}: {} " , peer , io . peer_info ( peer ) ) ;
2016-10-29 13:07:06 +02:00
self . handshaking_peers . remove ( & peer ) ;
2016-01-10 14:11:23 +01:00
if self . peers . contains_key ( & peer ) {
2016-02-14 01:03:48 +01:00
debug! ( target : " sync " , " Disconnected {} " , peer ) ;
2016-01-10 14:11:23 +01:00
self . clear_peer_download ( peer ) ;
2016-01-15 12:26:04 +01:00
self . peers . remove ( & peer ) ;
2016-05-16 14:41:41 +02:00
self . active_peers . remove ( & peer ) ;
2016-01-10 14:11:23 +01:00
self . continue_sync ( io ) ;
}
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Called when a new peer is connected
2016-01-14 19:03:48 +01:00
pub fn on_peer_connected ( & mut self , io : & mut SyncIo , peer : PeerId ) {
2016-05-16 14:41:41 +02:00
trace! ( target : " sync " , " == Connected {}: {} " , peer , io . peer_info ( peer ) ) ;
2016-09-06 15:31:13 +02:00
if let Err ( e ) = self . send_status ( io , peer ) {
2016-03-13 10:33:55 +01:00
debug! ( target :" sync " , " Error sending status request: {:?} " , e ) ;
2016-02-26 11:37:06 +01:00
io . disable_peer ( peer ) ;
2016-10-29 13:07:06 +02:00
} else {
self . handshaking_peers . insert ( peer , time ::precise_time_ns ( ) ) ;
2016-02-26 11:37:06 +01:00
}
2015-12-22 22:19:50 +01:00
}
2015-12-27 02:27:15 +01:00
/// Resume downloading
2015-12-24 17:18:47 +01:00
fn continue_sync ( & mut self , io : & mut SyncIo ) {
2016-10-18 18:16:00 +02:00
let mut peers : Vec < ( PeerId , U256 , u8 ) > = self . peers . iter ( ) . filter_map ( | ( k , p ) |
2016-09-06 15:31:13 +02:00
if p . can_sync ( ) { Some ( ( * k , p . difficulty . unwrap_or_else ( U256 ::zero ) , p . protocol_version ) ) } else { None } ) . collect ( ) ;
2016-06-22 12:10:26 +02:00
thread_rng ( ) . shuffle ( & mut peers ) ; //TODO: sort by rating
2016-09-06 15:31:13 +02:00
// prefer peers with higher protocol version
peers . sort_by ( | & ( _ , _ , ref v1 ) , & ( _ , _ , ref v2 ) | v1 . cmp ( v2 ) ) ;
2016-06-18 15:10:24 +02:00
trace! ( target : " sync " , " Syncing with {}/{} peers " , self . active_peers . len ( ) , peers . len ( ) ) ;
2016-09-06 15:31:13 +02:00
for ( p , _ , _ ) in peers {
2016-05-16 14:41:41 +02:00
if self . active_peers . contains ( & p ) {
self . sync_peer ( io , p , false ) ;
}
}
2016-10-31 14:59:41 +01:00
if ( self . state ! = SyncState ::WaitingPeers & & self . state ! = SyncState ::SnapshotWaiting & & self . state ! = SyncState ::Waiting & & self . state ! = SyncState ::Idle )
& & ! self . peers . values ( ) . any ( | p | p . asking ! = PeerAsking ::Nothing & & p . block_set ! = Some ( BlockSet ::OldBlocks ) & & p . can_sync ( ) ) {
self . complete_sync ( io ) ;
}
2015-12-22 22:19:50 +01:00
}
2016-05-16 14:41:41 +02:00
/// Called after all blocks have been downloaded
2016-09-06 15:31:13 +02:00
fn complete_sync ( & mut self , io : & mut SyncIo ) {
2015-12-27 00:48:03 +01:00
trace! ( target : " sync " , " Sync complete " ) ;
2016-09-06 15:31:13 +02:00
self . reset ( io ) ;
2015-12-24 17:18:47 +01:00
self . state = SyncState ::Idle ;
2015-12-22 22:19:50 +01:00
}
/// Enter waiting state
fn pause_sync ( & mut self ) {
2015-12-25 14:55:55 +01:00
trace! ( target : " sync " , " Block queue full, pausing sync " ) ;
2015-12-24 17:18:47 +01:00
self . state = SyncState ::Waiting ;
2015-12-22 22:19:50 +01:00
}
2016-07-19 20:42:23 +02:00
/// Find something to do for a peer. Called for a new peer or when a peer is done with its task.
2016-06-22 12:10:26 +02:00
fn sync_peer ( & mut self , io : & mut SyncIo , peer_id : PeerId , force : bool ) {
if ! self . active_peers . contains ( & peer_id ) {
trace! ( target : " sync " , " Skipping deactivated peer " ) ;
return ;
}
2016-09-06 15:31:13 +02:00
let ( peer_latest , peer_difficulty , peer_snapshot_number , peer_snapshot_hash ) = {
2016-10-27 08:28:12 +02:00
if let Some ( peer ) = self . peers . get_mut ( & peer_id ) {
2016-10-10 17:43:05 +02:00
if peer . asking ! = PeerAsking ::Nothing | | ! peer . can_sync ( ) {
return ;
}
if self . state = = SyncState ::Waiting {
trace! ( target : " sync " , " Waiting for the block queue " ) ;
return ;
}
if self . state = = SyncState ::SnapshotWaiting {
trace! ( target : " sync " , " Waiting for the snapshot restoration " ) ;
return ;
}
2016-10-18 18:16:00 +02:00
( peer . latest_hash . clone ( ) , peer . difficulty . clone ( ) , peer . snapshot_number . as_ref ( ) . cloned ( ) . unwrap_or ( 0 ) , peer . snapshot_hash . as_ref ( ) . cloned ( ) )
2016-10-10 17:43:05 +02:00
} else {
2016-09-06 15:31:13 +02:00
return ;
}
2015-12-22 22:19:50 +01:00
} ;
2016-05-16 14:41:41 +02:00
let chain_info = io . chain ( ) . chain_info ( ) ;
2016-10-18 18:16:00 +02:00
let syncing_difficulty = chain_info . pending_total_difficulty ;
2016-05-16 14:41:41 +02:00
2016-10-18 18:16:00 +02:00
let higher_difficulty = peer_difficulty . map_or ( true , | pd | pd > syncing_difficulty ) ;
if force | | self . state = = SyncState ::NewBlocks | | higher_difficulty | | self . old_blocks . is_some ( ) {
2016-05-16 14:41:41 +02:00
match self . state {
2016-10-29 13:07:06 +02:00
SyncState ::WaitingPeers = > {
trace! ( target : " sync " , " Checking snapshot sync: {} vs {} " , peer_snapshot_number , chain_info . best_block_number ) ;
self . maybe_start_snapshot_sync ( io ) ;
2016-10-18 18:16:00 +02:00
} ,
SyncState ::Idle | SyncState ::Blocks | SyncState ::NewBlocks = > {
if io . chain ( ) . queue_info ( ) . is_full ( ) {
self . pause_sync ( ) ;
return ;
}
let have_latest = io . chain ( ) . block_status ( BlockID ::Hash ( peer_latest ) ) ! = BlockStatus ::Unknown ;
2016-10-20 12:56:37 +02:00
if ! have_latest & & ( higher_difficulty | | force | | self . state = = SyncState ::NewBlocks ) {
2016-10-18 18:16:00 +02:00
// check if got new blocks to download
if let Some ( request ) = self . new_blocks . request_blocks ( io ) {
self . request_blocks ( io , peer_id , request , BlockSet ::NewBlocks ) ;
if self . state = = SyncState ::Idle {
self . state = SyncState ::Blocks ;
}
return ;
2016-09-06 15:31:13 +02:00
}
2016-05-16 14:41:41 +02:00
}
2016-10-18 18:16:00 +02:00
if let Some ( request ) = self . old_blocks . as_mut ( ) . and_then ( | d | d . request_blocks ( io ) ) {
self . request_blocks ( io , peer_id , request , BlockSet ::OldBlocks ) ;
return ;
2016-05-16 14:41:41 +02:00
}
2016-09-06 15:31:13 +02:00
} ,
SyncState ::SnapshotData = > {
2016-10-29 13:07:06 +02:00
if let RestorationStatus ::Ongoing { state_chunks : _ , block_chunks : _ , state_chunks_done , block_chunks_done , } = io . snapshot_service ( ) . status ( ) {
if self . snapshot . done_chunks ( ) - ( state_chunks_done + block_chunks_done ) as usize > MAX_SNAPSHOT_CHUNKS_DOWNLOAD_AHEAD {
trace! ( target : " sync " , " Snapshot queue full, pausing sync " ) ;
self . state = SyncState ::SnapshotWaiting ;
return ;
}
}
2016-09-06 15:31:13 +02:00
if peer_snapshot_hash . is_some ( ) & & peer_snapshot_hash = = self . snapshot . snapshot_hash ( ) {
self . request_snapshot_data ( io , peer_id ) ;
}
} ,
2016-09-16 23:03:26 +02:00
SyncState ::SnapshotManifest | //already downloading from other peer
SyncState ::Waiting | SyncState ::SnapshotWaiting = > ( )
2015-12-22 22:19:50 +01:00
}
}
2016-02-03 21:05:04 +01:00
}
2016-10-18 18:16:00 +02:00
/// Perofrm block download request`
fn request_blocks ( & mut self , io : & mut SyncIo , peer_id : PeerId , request : BlockRequest , block_set : BlockSet ) {
match request {
BlockRequest ::Headers { start , count , skip } = > {
self . request_headers_by_hash ( io , peer_id , & start , count , skip , false , block_set ) ;
} ,
BlockRequest ::Bodies { hashes } = > {
self . request_bodies ( io , peer_id , hashes , block_set ) ;
} ,
BlockRequest ::Receipts { hashes } = > {
self . request_receipts ( io , peer_id , hashes , block_set ) ;
2016-06-20 00:40:11 +02:00
} ,
2016-03-01 22:03:29 +01:00
}
2015-12-22 22:19:50 +01:00
}
2016-09-06 15:31:13 +02:00
/// Find some headers or blocks to download for a peer.
fn request_snapshot_data ( & mut self , io : & mut SyncIo , peer_id : PeerId ) {
self . clear_peer_download ( peer_id ) ;
// find chunk data to download
if let Some ( hash ) = self . snapshot . needed_chunk ( ) {
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
peer . asking_snapshot_data = Some ( hash . clone ( ) ) ;
}
2016-09-06 15:31:13 +02:00
self . request_snapshot_chunk ( io , peer_id , & hash ) ;
}
}
2016-01-09 18:40:13 +01:00
/// Clear all blocks/headers marked as being downloaded by a peer.
2016-01-14 19:03:48 +01:00
fn clear_peer_download ( & mut self , peer_id : PeerId ) {
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
match peer . asking {
2016-10-18 18:16:00 +02:00
PeerAsking ::BlockHeaders = > {
if let Some ( ref hash ) = peer . asking_hash {
self . new_blocks . clear_header_download ( hash ) ;
if let Some ( ref mut old ) = self . old_blocks {
old . clear_header_download ( hash ) ;
}
2016-10-10 17:43:05 +02:00
}
} ,
PeerAsking ::BlockBodies = > {
2016-10-18 18:16:00 +02:00
self . new_blocks . clear_body_download ( & peer . asking_blocks ) ;
if let Some ( ref mut old ) = self . old_blocks {
old . clear_body_download ( & peer . asking_blocks ) ;
}
} ,
PeerAsking ::BlockReceipts = > {
self . new_blocks . clear_receipt_download ( & peer . asking_blocks ) ;
if let Some ( ref mut old ) = self . old_blocks {
old . clear_receipt_download ( & peer . asking_blocks ) ;
2016-10-10 17:43:05 +02:00
}
} ,
PeerAsking ::SnapshotData = > {
if let Some ( hash ) = peer . asking_snapshot_data {
self . snapshot . clear_chunk_download ( & hash ) ;
}
} ,
_ = > ( ) ,
}
2016-06-20 00:40:11 +02:00
}
}
2016-01-09 18:40:13 +01:00
/// Checks if there are blocks fully downloaded that can be imported into the blockchain and does the import.
2016-10-27 08:28:12 +02:00
#[ cfg_attr(feature= " dev " , allow(block_in_if_condition_stmt)) ]
2016-10-18 18:16:00 +02:00
fn collect_blocks ( & mut self , io : & mut SyncIo , block_set : BlockSet ) {
match block_set {
BlockSet ::NewBlocks = > {
if self . new_blocks . collect_blocks ( io , self . state = = SyncState ::NewBlocks ) = = Err ( DownloaderImportError ::Invalid ) {
self . restart ( io ) ;
}
} ,
BlockSet ::OldBlocks = > {
2016-10-27 08:28:12 +02:00
if self . old_blocks . as_mut ( ) . map_or ( false , | downloader | { downloader . collect_blocks ( io , false ) = = Err ( DownloaderImportError ::Invalid ) } ) {
self . restart ( io ) ;
} else if self . old_blocks . as_ref ( ) . map_or ( false , | downloader | { downloader . is_complete ( ) } ) {
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " Background block download is complete " ) ;
self . old_blocks = None ;
2016-02-03 21:42:30 +01:00
}
2015-12-24 17:18:47 +01:00
}
}
2015-12-22 22:19:50 +01:00
}
2015-12-24 17:18:47 +01:00
2016-01-09 18:40:13 +01:00
/// Request headers from a peer by block hash
2016-05-25 17:03:58 +02:00
#[ cfg_attr(feature= " dev " , allow(too_many_arguments)) ]
2016-10-18 18:16:00 +02:00
fn request_headers_by_hash ( & mut self , sync : & mut SyncIo , peer_id : PeerId , h : & H256 , count : u64 , skip : u64 , reverse : bool , set : BlockSet ) {
trace! ( target : " sync " , " {} <- GetBlockHeaders: {} entries starting from {}, set = {:?} " , peer_id , count , h , set ) ;
2015-12-22 22:19:50 +01:00
let mut rlp = RlpStream ::new_list ( 4 ) ;
rlp . append ( h ) ;
rlp . append ( & count ) ;
rlp . append ( & skip ) ;
rlp . append ( & if reverse { 1 u32 } else { 0 u32 } ) ;
2016-10-18 18:16:00 +02:00
self . send_request ( sync , peer_id , PeerAsking ::BlockHeaders , GET_BLOCK_HEADERS_PACKET , rlp . out ( ) ) ;
let peer = self . peers . get_mut ( & peer_id ) . expect ( " peer_id may originate either from on_packet, where it is already validated or from enumerating self.peers. qed " ) ;
peer . asking_hash = Some ( h . clone ( ) ) ;
peer . block_set = Some ( set ) ;
2015-12-22 22:19:50 +01:00
}
2016-07-27 21:38:22 +02:00
/// Request headers from a peer by block number
#[ cfg_attr(feature= " dev " , allow(too_many_arguments)) ]
2016-10-18 18:16:00 +02:00
fn request_fork_header_by_number ( & mut self , sync : & mut SyncIo , peer_id : PeerId , n : BlockNumber ) {
trace! ( target : " sync " , " {} <- GetForkHeader: at {} " , peer_id , n ) ;
2016-07-27 21:38:22 +02:00
let mut rlp = RlpStream ::new_list ( 4 ) ;
rlp . append ( & n ) ;
2016-10-18 18:16:00 +02:00
rlp . append ( & 1 u32 ) ;
rlp . append ( & 0 u32 ) ;
rlp . append ( & 0 u32 ) ;
self . send_request ( sync , peer_id , PeerAsking ::ForkHeader , GET_BLOCK_HEADERS_PACKET , rlp . out ( ) ) ;
2016-07-27 21:38:22 +02:00
}
2016-09-06 15:31:13 +02:00
/// Request snapshot manifest from a peer.
fn request_snapshot_manifest ( & mut self , sync : & mut SyncIo , peer_id : PeerId ) {
trace! ( target : " sync " , " {} <- GetSnapshotManifest " , peer_id ) ;
let rlp = RlpStream ::new_list ( 0 ) ;
self . send_request ( sync , peer_id , PeerAsking ::SnapshotManifest , GET_SNAPSHOT_MANIFEST_PACKET , rlp . out ( ) ) ;
}
/// Request snapshot chunk from a peer.
fn request_snapshot_chunk ( & mut self , sync : & mut SyncIo , peer_id : PeerId , chunk : & H256 ) {
trace! ( target : " sync " , " {} <- GetSnapshotData {:?} " , peer_id , chunk ) ;
let mut rlp = RlpStream ::new_list ( 1 ) ;
rlp . append ( chunk ) ;
self . send_request ( sync , peer_id , PeerAsking ::SnapshotData , GET_SNAPSHOT_DATA_PACKET , rlp . out ( ) ) ;
}
2016-01-09 18:40:13 +01:00
/// Request block bodies from a peer
2016-10-18 18:16:00 +02:00
fn request_bodies ( & mut self , sync : & mut SyncIo , peer_id : PeerId , hashes : Vec < H256 > , set : BlockSet ) {
2015-12-22 22:19:50 +01:00
let mut rlp = RlpStream ::new_list ( hashes . len ( ) ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} <- GetBlockBodies: {} entries starting from {:?}, set = {:?} " , peer_id , hashes . len ( ) , hashes . first ( ) , set ) ;
for h in & hashes {
rlp . append ( & h . clone ( ) ) ;
2015-12-22 22:19:50 +01:00
}
2015-12-25 14:55:55 +01:00
self . send_request ( sync , peer_id , PeerAsking ::BlockBodies , GET_BLOCK_BODIES_PACKET , rlp . out ( ) ) ;
2016-10-18 18:16:00 +02:00
let peer = self . peers . get_mut ( & peer_id ) . expect ( " peer_id may originate either from on_packet, where it is already validated or from enumerating self.peers. qed " ) ;
peer . asking_blocks = hashes ;
peer . block_set = Some ( set ) ;
}
/// Request block receipts from a peer
fn request_receipts ( & mut self , sync : & mut SyncIo , peer_id : PeerId , hashes : Vec < H256 > , set : BlockSet ) {
let mut rlp = RlpStream ::new_list ( hashes . len ( ) ) ;
trace! ( target : " sync " , " {} <- GetBlockReceipts: {} entries starting from {:?}, set = {:?} " , peer_id , hashes . len ( ) , hashes . first ( ) , set ) ;
for h in & hashes {
rlp . append ( & h . clone ( ) ) ;
}
self . send_request ( sync , peer_id , PeerAsking ::BlockReceipts , GET_RECEIPTS_PACKET , rlp . out ( ) ) ;
let peer = self . peers . get_mut ( & peer_id ) . expect ( " peer_id may originate either from on_packet, where it is already validated or from enumerating self.peers. qed " ) ;
peer . asking_blocks = hashes ;
peer . block_set = Some ( set ) ;
2015-12-22 22:19:50 +01:00
}
2016-01-09 18:40:13 +01:00
/// Reset peer status after request is complete.
2016-05-16 14:41:41 +02:00
fn reset_peer_asking ( & mut self , peer_id : PeerId , asking : PeerAsking ) -> bool {
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
peer . expired = false ;
2016-10-18 18:16:00 +02:00
peer . block_set = None ;
2016-10-10 17:43:05 +02:00
if peer . asking ! = asking {
trace! ( target :" sync " , " Asking {:?} while expected {:?} " , peer . asking , asking ) ;
peer . asking = PeerAsking ::Nothing ;
2016-10-18 18:16:00 +02:00
return false ;
} else {
2016-10-10 17:43:05 +02:00
peer . asking = PeerAsking ::Nothing ;
2016-10-18 18:16:00 +02:00
return true ;
2016-10-10 17:43:05 +02:00
}
2015-12-27 00:48:03 +01:00
}
2016-10-27 08:28:12 +02:00
false
2015-12-27 00:48:03 +01:00
}
2016-01-09 18:40:13 +01:00
/// Generic request sender
2016-01-14 19:03:48 +01:00
fn send_request ( & mut self , sync : & mut SyncIo , peer_id : PeerId , asking : PeerAsking , packet_id : PacketId , packet : Bytes ) {
2016-10-10 17:43:05 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( & peer_id ) {
if peer . asking ! = PeerAsking ::Nothing {
warn! ( target :" sync " , " Asking {:?} while requesting {:?} " , peer . asking , asking ) ;
}
peer . asking = asking ;
2016-10-29 13:07:06 +02:00
peer . ask_time = time ::precise_time_ns ( ) ;
2016-10-25 18:40:01 +02:00
let result = if packet_id > = ETH_PACKET_COUNT {
sync . send_protocol ( WARP_SYNC_PROTOCOL_ID , peer_id , packet_id , packet )
} else {
sync . send ( peer_id , packet_id , packet )
} ;
if let Err ( e ) = result {
2016-10-10 17:43:05 +02:00
debug! ( target :" sync " , " Error sending request: {:?} " , e ) ;
sync . disable_peer ( peer_id ) ;
}
2015-12-22 22:19:50 +01:00
}
}
2015-12-24 17:18:47 +01:00
2016-02-12 13:07:02 +01:00
/// Generic packet sender
fn send_packet ( & mut self , sync : & mut SyncIo , peer_id : PeerId , packet_id : PacketId , packet : Bytes ) {
if let Err ( e ) = sync . send ( peer_id , packet_id , packet ) {
2016-03-07 11:34:07 +01:00
debug! ( target :" sync " , " Error sending packet: {:?} " , e ) ;
2016-02-12 13:07:02 +01:00
sync . disable_peer ( peer_id ) ;
}
}
2016-03-28 18:11:00 +02:00
2016-01-09 18:40:13 +01:00
/// Called when peer sends us new transactions
2016-03-05 16:46:04 +01:00
fn on_peer_transactions ( & mut self , io : & mut SyncIo , peer_id : PeerId , r : & UntrustedRlp ) -> Result < ( ) , PacketDecodeError > {
2016-10-25 18:40:01 +02:00
// Accept transactions only when fully synced
2016-10-30 12:19:20 +01:00
if ! io . is_chain_queue_empty ( ) | | ( self . state ! = SyncState ::Idle & & self . state ! = SyncState ::NewBlocks ) {
2016-10-25 18:40:01 +02:00
trace! ( target : " sync " , " {} Ignoring transactions while syncing " , peer_id ) ;
2016-03-17 12:17:53 +01:00
return Ok ( ( ) ) ;
}
2016-08-07 22:42:35 +02:00
if ! self . peers . get ( & peer_id ) . map_or ( false , | p | p . can_sync ( ) ) {
2016-07-27 21:38:22 +02:00
trace! ( target : " sync " , " {} Ignoring transactions from unconfirmed/unknown peer " , peer_id ) ;
}
2016-03-17 12:17:53 +01:00
2016-06-19 14:35:42 +02:00
let mut item_count = r . item_count ( ) ;
2016-03-05 16:46:04 +01:00
trace! ( target : " sync " , " {} -> Transactions ({} entries) " , peer_id , item_count ) ;
2016-06-19 14:35:42 +02:00
item_count = min ( item_count , MAX_TX_TO_IMPORT ) ;
2016-03-08 15:46:44 +01:00
let mut transactions = Vec ::with_capacity ( item_count ) ;
2016-06-19 14:35:42 +02:00
for i in 0 .. item_count {
2016-10-08 10:35:54 +02:00
let rlp = try ! ( r . at ( i ) ) ;
if rlp . as_raw ( ) . len ( ) > MAX_TRANSACTION_SIZE {
debug! ( " Skipped oversized transaction of {} bytes " , rlp . as_raw ( ) . len ( ) ) ;
continue ;
}
let tx = rlp . as_raw ( ) . to_vec ( ) ;
2016-03-08 15:46:44 +01:00
transactions . push ( tx ) ;
2016-03-05 16:46:04 +01:00
}
2016-06-23 12:04:54 +02:00
io . chain ( ) . queue_transactions ( transactions ) ;
2016-05-16 14:41:41 +02:00
Ok ( ( ) )
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Send Status message
2016-09-06 15:31:13 +02:00
fn send_status ( & mut self , io : & mut SyncIo , peer : PeerId ) -> Result < ( ) , NetworkError > {
2016-10-24 16:24:35 +02:00
let warp_protocol = io . protocol_version ( & WARP_SYNC_PROTOCOL_ID , peer ) ! = 0 ;
let protocol = if warp_protocol { PROTOCOL_VERSION_1 } else { PROTOCOL_VERSION_63 } ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " Sending status to {}, protocol version {} " , peer , protocol ) ;
2016-10-24 16:24:35 +02:00
let mut packet = RlpStream ::new_list ( if warp_protocol { 7 } else { 5 } ) ;
2016-01-07 20:43:37 +01:00
let chain = io . chain ( ) . chain_info ( ) ;
2016-10-18 18:16:00 +02:00
packet . append ( & ( protocol as u32 ) ) ;
2016-02-24 21:23:58 +01:00
packet . append ( & self . network_id ) ;
2015-12-24 17:18:47 +01:00
packet . append ( & chain . total_difficulty ) ;
2016-01-07 16:08:12 +01:00
packet . append ( & chain . best_block_hash ) ;
2015-12-24 17:18:47 +01:00
packet . append ( & chain . genesis_hash ) ;
2016-10-24 16:24:35 +02:00
if warp_protocol {
2016-09-06 15:31:13 +02:00
let manifest = io . snapshot_service ( ) . manifest ( ) ;
let block_number = manifest . as_ref ( ) . map_or ( 0 , | m | m . block_number ) ;
let manifest_hash = manifest . map_or ( H256 ::new ( ) , | m | m . into_rlp ( ) . sha3 ( ) ) ;
packet . append ( & manifest_hash ) ;
packet . append ( & block_number ) ;
}
2016-02-26 11:37:06 +01:00
io . respond ( STATUS_PACKET , packet . out ( ) )
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Respond to GetBlockHeaders request
2016-06-13 18:49:40 +02:00
fn return_block_headers ( io : & SyncIo , r : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
2015-12-24 17:18:47 +01:00
// Packet layout:
// [ block: { P , B_32 }, maxHeaders: P, skip: P, reverse: P in { 0 , 1 } ]
2016-01-08 17:52:25 +01:00
let max_headers : usize = try ! ( r . val_at ( 1 ) ) ;
let skip : usize = try ! ( r . val_at ( 2 ) ) ;
let reverse : bool = try ! ( r . val_at ( 3 ) ) ;
2016-01-07 20:43:37 +01:00
let last = io . chain ( ) . chain_info ( ) . best_block_number ;
2016-06-03 11:36:30 +02:00
let number = if try ! ( r . at ( 0 ) ) . size ( ) = = 32 {
2015-12-24 17:18:47 +01:00
// id is a hash
2016-01-08 17:52:25 +01:00
let hash : H256 = try ! ( r . val_at ( 0 ) ) ;
2016-06-13 18:49:40 +02:00
trace! ( target : " sync " , " {} -> GetBlockHeaders (hash: {}, max: {}, skip: {}, reverse:{}) " , peer_id , hash , max_headers , skip , reverse ) ;
2016-05-19 11:00:32 +02:00
match io . chain ( ) . block_header ( BlockID ::Hash ( hash ) ) {
2016-06-10 17:12:51 +02:00
Some ( hdr ) = > {
let number = From ::from ( HeaderView ::new ( & hdr ) . number ( ) ) ;
2016-06-14 12:18:20 +02:00
debug_assert_eq! ( HeaderView ::new ( & hdr ) . sha3 ( ) , hash ) ;
2016-06-13 18:49:40 +02:00
if max_headers = = 1 | | io . chain ( ) . block_hash ( BlockID ::Number ( number ) ) ! = Some ( hash ) {
// Non canonical header or single header requested
2016-06-14 12:18:20 +02:00
// TODO: handle single-step reverse hashchains of non-canon hashes
trace! ( target :" sync " , " Returning single header: {:?} " , hash ) ;
2016-06-10 17:12:51 +02:00
let mut rlp = RlpStream ::new_list ( 1 ) ;
rlp . append_raw ( & hdr , 1 ) ;
return Ok ( Some ( ( BLOCK_HEADERS_PACKET , rlp ) ) ) ;
}
number
}
2016-05-16 14:41:41 +02:00
None = > return Ok ( Some ( ( BLOCK_HEADERS_PACKET , RlpStream ::new_list ( 0 ) ) ) ) //no such header, return nothing
2015-12-24 17:18:47 +01:00
}
2016-05-16 14:41:41 +02:00
} else {
2016-06-13 18:49:40 +02:00
trace! ( target : " sync " , " {} -> GetBlockHeaders (number: {}, max: {}, skip: {}, reverse:{}) " , peer_id , try ! ( r . val_at ::< BlockNumber > ( 0 ) ) , max_headers , skip , reverse ) ;
2016-01-08 17:52:25 +01:00
try ! ( r . val_at ( 0 ) )
2015-12-24 17:18:47 +01:00
} ;
2016-06-03 11:36:30 +02:00
let mut number = if reverse {
min ( last , number )
2015-12-27 00:48:03 +01:00
} else {
2016-06-03 11:36:30 +02:00
max ( 0 , number )
} ;
2015-12-24 17:18:47 +01:00
let max_count = min ( MAX_HEADERS_TO_SEND , max_headers ) ;
let mut count = 0 ;
let mut data = Bytes ::new ( ) ;
2016-01-08 16:26:00 +01:00
let inc = ( skip + 1 ) as BlockNumber ;
2016-10-18 18:16:00 +02:00
let overlay = io . chain_overlay ( ) . read ( ) ;
2016-10-24 16:24:35 +02:00
2016-02-24 13:01:45 +01:00
while number < = last & & count < max_count {
2016-10-18 18:16:00 +02:00
if let Some ( hdr ) = overlay . get ( & number ) {
trace! ( target : " sync " , " {}: Returning cached fork header " , peer_id ) ;
2016-10-27 08:28:12 +02:00
data . extend_from_slice ( hdr ) ;
2016-10-18 18:16:00 +02:00
count + = 1 ;
} else if let Some ( mut hdr ) = io . chain ( ) . block_header ( BlockID ::Number ( number ) ) {
2016-01-17 15:56:09 +01:00
data . append ( & mut hdr ) ;
count + = 1 ;
2016-10-24 16:24:35 +02:00
} else {
// No required block.
break ;
2015-12-24 17:18:47 +01:00
}
2015-12-25 14:55:55 +01:00
if reverse {
2016-02-24 13:01:45 +01:00
if number < = inc | | number = = 0 {
2016-01-08 16:26:00 +01:00
break ;
}
number - = inc ;
2015-12-25 14:55:55 +01:00
}
else {
2016-01-08 16:26:00 +01:00
number + = inc ;
2015-12-25 14:55:55 +01:00
}
2015-12-24 17:18:47 +01:00
}
let mut rlp = RlpStream ::new_list ( count as usize ) ;
rlp . append_raw ( & data , count as usize ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} -> GetBlockHeaders: returned {} entries " , peer_id , count ) ;
2016-02-04 23:24:36 +01:00
Ok ( Some ( ( BLOCK_HEADERS_PACKET , rlp ) ) )
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Respond to GetBlockBodies request
2016-06-13 18:49:40 +02:00
fn return_block_bodies ( io : & SyncIo , r : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
2015-12-24 17:18:47 +01:00
let mut count = r . item_count ( ) ;
if count = = 0 {
debug! ( target : " sync " , " Empty GetBlockBodies request, ignoring. " ) ;
2016-02-04 23:24:36 +01:00
return Ok ( None ) ;
2015-12-24 17:18:47 +01:00
}
count = min ( count , MAX_BODIES_TO_SEND ) ;
let mut added = 0 usize ;
let mut data = Bytes ::new ( ) ;
for i in 0 .. count {
2016-05-19 11:00:32 +02:00
if let Some ( mut hdr ) = io . chain ( ) . block_body ( BlockID ::Hash ( try ! ( r . val_at ::< H256 > ( i ) ) ) ) {
2016-01-19 11:10:38 +01:00
data . append ( & mut hdr ) ;
added + = 1 ;
2015-12-24 17:18:47 +01:00
}
}
let mut rlp = RlpStream ::new_list ( added ) ;
rlp . append_raw ( & data , added ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} -> GetBlockBodies: returned {} entries " , peer_id , added ) ;
2016-02-04 23:24:36 +01:00
Ok ( Some ( ( BLOCK_BODIES_PACKET , rlp ) ) )
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Respond to GetNodeData request
2016-06-13 18:49:40 +02:00
fn return_node_data ( io : & SyncIo , r : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
2015-12-24 17:18:47 +01:00
let mut count = r . item_count ( ) ;
2016-06-13 18:49:40 +02:00
trace! ( target : " sync " , " {} -> GetNodeData: {} entries " , peer_id , count ) ;
2015-12-24 17:18:47 +01:00
if count = = 0 {
debug! ( target : " sync " , " Empty GetNodeData request, ignoring. " ) ;
2016-02-04 23:24:36 +01:00
return Ok ( None ) ;
2015-12-24 17:18:47 +01:00
}
count = min ( count , MAX_NODE_DATA_TO_SEND ) ;
let mut added = 0 usize ;
2016-10-27 15:25:54 +02:00
let mut data = Vec ::new ( ) ;
2015-12-24 17:18:47 +01:00
for i in 0 .. count {
2016-10-27 15:25:54 +02:00
if let Some ( hdr ) = io . chain ( ) . state_data ( & try ! ( r . val_at ::< H256 > ( i ) ) ) {
data . push ( hdr ) ;
2016-01-19 11:10:38 +01:00
added + = 1 ;
2015-12-24 17:18:47 +01:00
}
}
2016-06-13 18:49:40 +02:00
trace! ( target : " sync " , " {} -> GetNodeData: return {} entries " , peer_id , added ) ;
2015-12-24 17:18:47 +01:00
let mut rlp = RlpStream ::new_list ( added ) ;
2016-10-29 13:08:49 +02:00
for d in data {
2016-10-27 15:25:54 +02:00
rlp . append ( & d ) ;
}
2016-02-04 23:24:36 +01:00
Ok ( Some ( ( NODE_DATA_PACKET , rlp ) ) )
2015-12-24 17:18:47 +01:00
}
2016-06-13 18:49:40 +02:00
fn return_receipts ( io : & SyncIo , rlp : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
2016-02-04 23:24:36 +01:00
let mut count = rlp . item_count ( ) ;
2016-06-13 18:49:40 +02:00
trace! ( target : " sync " , " {} -> GetReceipts: {} entries " , peer_id , count ) ;
2015-12-24 17:18:47 +01:00
if count = = 0 {
debug! ( target : " sync " , " Empty GetReceipts request, ignoring. " ) ;
2016-02-04 23:24:36 +01:00
return Ok ( None ) ;
2015-12-24 17:18:47 +01:00
}
2016-03-12 19:23:17 +01:00
count = min ( count , MAX_RECEIPTS_HEADERS_TO_SEND ) ;
let mut added_headers = 0 usize ;
let mut added_receipts = 0 usize ;
2015-12-24 17:18:47 +01:00
let mut data = Bytes ::new ( ) ;
for i in 0 .. count {
2016-03-12 19:23:17 +01:00
if let Some ( mut receipts_bytes ) = io . chain ( ) . block_receipts ( & try ! ( rlp . val_at ::< H256 > ( i ) ) ) {
data . append ( & mut receipts_bytes ) ;
added_receipts + = receipts_bytes . len ( ) ;
added_headers + = 1 ;
if added_receipts > MAX_RECEIPTS_TO_SEND { break ; }
2015-12-24 17:18:47 +01:00
}
}
2016-03-12 19:23:17 +01:00
let mut rlp_result = RlpStream ::new_list ( added_headers ) ;
rlp_result . append_raw ( & data , added_headers ) ;
2016-02-04 23:24:36 +01:00
Ok ( Some ( ( RECEIPTS_PACKET , rlp_result ) ) )
}
2016-09-06 15:31:13 +02:00
/// Respond to GetSnapshotManifest request
fn return_snapshot_manifest ( io : & SyncIo , r : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
let count = r . item_count ( ) ;
trace! ( target : " sync " , " {} -> GetSnapshotManifest " , peer_id ) ;
if count ! = 0 {
debug! ( target : " sync " , " Invalid GetSnapshotManifest request, ignoring. " ) ;
return Ok ( None ) ;
}
let rlp = match io . snapshot_service ( ) . manifest ( ) {
Some ( manifest ) = > {
trace! ( target : " sync " , " {} <- SnapshotManifest " , peer_id ) ;
let mut rlp = RlpStream ::new_list ( 1 ) ;
rlp . append_raw ( & manifest . into_rlp ( ) , 1 ) ;
rlp
} ,
None = > {
trace! ( target : " sync " , " {}: No manifest to return " , peer_id ) ;
2016-09-16 23:03:26 +02:00
RlpStream ::new_list ( 0 )
2016-09-06 15:31:13 +02:00
}
} ;
Ok ( Some ( ( SNAPSHOT_MANIFEST_PACKET , rlp ) ) )
}
2016-10-24 16:24:35 +02:00
/// Respond to GetSnapshotData request
2016-09-06 15:31:13 +02:00
fn return_snapshot_data ( io : & SyncIo , r : & UntrustedRlp , peer_id : PeerId ) -> RlpResponseResult {
let hash : H256 = try ! ( r . val_at ( 0 ) ) ;
trace! ( target : " sync " , " {} -> GetSnapshotData {:?} " , peer_id , hash ) ;
let rlp = match io . snapshot_service ( ) . chunk ( hash ) {
Some ( data ) = > {
let mut rlp = RlpStream ::new_list ( 1 ) ;
2016-10-18 18:16:00 +02:00
trace! ( target : " sync " , " {} <- SnapshotData " , peer_id ) ;
2016-09-06 15:31:13 +02:00
rlp . append ( & data ) ;
rlp
} ,
None = > {
2016-09-16 23:03:26 +02:00
RlpStream ::new_list ( 0 )
2016-09-06 15:31:13 +02:00
}
} ;
Ok ( Some ( ( SNAPSHOT_DATA_PACKET , rlp ) ) )
}
2016-06-20 17:28:48 +02:00
fn return_rlp < FRlp , FError > ( io : & mut SyncIo , rlp : & UntrustedRlp , peer : PeerId , rlp_func : FRlp , error_func : FError ) -> Result < ( ) , PacketDecodeError >
2016-06-13 18:49:40 +02:00
where FRlp : Fn ( & SyncIo , & UntrustedRlp , PeerId ) -> RlpResponseResult ,
2016-08-05 10:32:04 +02:00
FError : FnOnce ( NetworkError ) -> String
2016-02-04 23:24:36 +01:00
{
2016-06-13 18:49:40 +02:00
let response = rlp_func ( io , rlp , peer ) ;
2016-02-04 23:24:36 +01:00
match response {
Err ( e ) = > Err ( e ) ,
Ok ( Some ( ( packet_id , rlp_stream ) ) ) = > {
io . respond ( packet_id , rlp_stream . out ( ) ) . unwrap_or_else (
| e | debug! ( target : " sync " , " {:?} " , error_func ( e ) ) ) ;
Ok ( ( ) )
}
_ = > Ok ( ( ) )
}
2015-12-24 17:18:47 +01:00
}
2016-01-09 18:40:13 +01:00
/// Dispatch incoming requests and responses
2016-06-20 17:28:48 +02:00
pub fn dispatch_packet ( sync : & RwLock < ChainSync > , io : & mut SyncIo , peer : PeerId , packet_id : u8 , data : & [ u8 ] ) {
2016-01-08 17:52:25 +01:00
let rlp = UntrustedRlp ::new ( data ) ;
let result = match packet_id {
2016-06-20 17:28:48 +02:00
GET_BLOCK_BODIES_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
2016-02-04 23:24:36 +01:00
ChainSync ::return_block_bodies ,
| e | format! ( " Error sending block bodies: {:?} " , e ) ) ,
2016-06-20 17:28:48 +02:00
GET_BLOCK_HEADERS_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
2016-02-04 23:24:36 +01:00
ChainSync ::return_block_headers ,
| e | format! ( " Error sending block headers: {:?} " , e ) ) ,
2016-06-20 17:28:48 +02:00
GET_RECEIPTS_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
2016-02-05 01:01:50 +01:00
ChainSync ::return_receipts ,
| e | format! ( " Error sending receipts: {:?} " , e ) ) ,
2016-02-04 23:24:36 +01:00
2016-06-20 17:28:48 +02:00
GET_NODE_DATA_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
2016-02-05 01:01:50 +01:00
ChainSync ::return_node_data ,
| e | format! ( " Error sending nodes: {:?} " , e ) ) ,
2016-02-15 00:51:50 +01:00
2016-09-06 15:31:13 +02:00
GET_SNAPSHOT_MANIFEST_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
ChainSync ::return_snapshot_manifest ,
| e | format! ( " Error sending snapshot manifest: {:?} " , e ) ) ,
GET_SNAPSHOT_DATA_PACKET = > ChainSync ::return_rlp ( io , & rlp , peer ,
ChainSync ::return_snapshot_data ,
| e | format! ( " Error sending snapshot data: {:?} " , e ) ) ,
2016-06-20 17:28:48 +02:00
_ = > {
2016-07-13 19:59:59 +02:00
sync . write ( ) . on_packet ( io , peer , packet_id , data ) ;
2016-06-20 17:28:48 +02:00
Ok ( ( ) )
}
} ;
result . unwrap_or_else ( | e | {
debug! ( target :" sync " , " {} -> Malformed packet {} : {} " , peer , packet_id , e ) ;
} )
}
pub fn on_packet ( & mut self , io : & mut SyncIo , peer : PeerId , packet_id : u8 , data : & [ u8 ] ) {
if packet_id ! = STATUS_PACKET & & ! self . peers . contains_key ( & peer ) {
debug! ( target :" sync " , " Unexpected packet from unregistered peer: {}:{} " , peer , io . peer_info ( peer ) ) ;
return ;
}
let rlp = UntrustedRlp ::new ( data ) ;
let result = match packet_id {
STATUS_PACKET = > self . on_peer_status ( io , peer , & rlp ) ,
TRANSACTIONS_PACKET = > self . on_peer_transactions ( io , peer , & rlp ) ,
BLOCK_HEADERS_PACKET = > self . on_peer_block_headers ( io , peer , & rlp ) ,
BLOCK_BODIES_PACKET = > self . on_peer_block_bodies ( io , peer , & rlp ) ,
2016-10-18 18:16:00 +02:00
RECEIPTS_PACKET = > self . on_peer_block_receipts ( io , peer , & rlp ) ,
2016-06-20 17:28:48 +02:00
NEW_BLOCK_PACKET = > self . on_peer_new_block ( io , peer , & rlp ) ,
NEW_BLOCK_HASHES_PACKET = > self . on_peer_new_hashes ( io , peer , & rlp ) ,
2016-09-06 15:31:13 +02:00
SNAPSHOT_MANIFEST_PACKET = > self . on_snapshot_manifest ( io , peer , & rlp ) ,
SNAPSHOT_DATA_PACKET = > self . on_snapshot_data ( io , peer , & rlp ) ,
2016-02-03 21:05:04 +01:00
_ = > {
2016-10-25 18:40:01 +02:00
debug! ( target : " sync " , " {}: Unknown packet {} " , peer , packet_id ) ;
2016-01-08 17:52:25 +01:00
Ok ( ( ) )
}
} ;
result . unwrap_or_else ( | e | {
debug! ( target :" sync " , " {} -> Malformed packet {} : {} " , peer , packet_id , e ) ;
} )
2015-12-24 17:18:47 +01:00
}
2015-12-22 22:19:50 +01:00
2016-09-16 23:03:26 +02:00
#[ cfg_attr(feature= " dev " , allow(match_same_arms)) ]
2016-06-18 15:10:24 +02:00
pub fn maintain_peers ( & mut self , io : & mut SyncIo ) {
2016-10-29 13:07:06 +02:00
let tick = time ::precise_time_ns ( ) ;
2016-06-18 15:10:24 +02:00
let mut aborting = Vec ::new ( ) ;
2016-02-03 21:42:30 +01:00
for ( peer_id , peer ) in & self . peers {
2016-10-29 13:07:06 +02:00
let elapsed = ( tick - peer . ask_time ) / 1_000_000_000 ;
2016-07-23 17:27:23 +02:00
let timeout = match peer . asking {
2016-10-29 13:07:06 +02:00
PeerAsking ::BlockHeaders = > elapsed > HEADERS_TIMEOUT_SEC ,
PeerAsking ::BlockBodies = > elapsed > BODIES_TIMEOUT_SEC ,
PeerAsking ::BlockReceipts = > elapsed > RECEIPTS_TIMEOUT_SEC ,
2016-07-23 17:27:23 +02:00
PeerAsking ::Nothing = > false ,
2016-10-29 13:07:06 +02:00
PeerAsking ::ForkHeader = > elapsed > FORK_HEADER_TIMEOUT_SEC ,
PeerAsking ::SnapshotManifest = > elapsed > SNAPSHOT_MANIFEST_TIMEOUT_SEC ,
PeerAsking ::SnapshotData = > elapsed > SNAPSHOT_DATA_TIMEOUT_SEC ,
2016-07-23 17:27:23 +02:00
} ;
if timeout {
2016-06-13 18:49:40 +02:00
trace! ( target :" sync " , " Timeout {} " , peer_id ) ;
2016-02-03 21:42:30 +01:00
io . disconnect_peer ( * peer_id ) ;
2016-06-18 15:10:24 +02:00
aborting . push ( * peer_id ) ;
2016-02-03 21:42:30 +01:00
}
}
2016-06-18 15:10:24 +02:00
for p in aborting {
self . on_peer_aborting ( io , p ) ;
}
2016-10-29 13:07:06 +02:00
// Check for handshake timeouts
for ( peer , ask_time ) in & self . handshaking_peers {
let elapsed = ( tick - ask_time ) / 1_000_000_000 ;
if elapsed > STATUS_TIMEOUT_SEC {
trace! ( target :" sync " , " Status timeout {} " , peer ) ;
io . disconnect_peer ( * peer ) ;
}
}
2016-02-03 21:42:30 +01:00
}
2016-02-05 18:34:08 +01:00
fn check_resume ( & mut self , io : & mut SyncIo ) {
2016-09-06 15:31:13 +02:00
if self . state = = SyncState ::Waiting & & ! io . chain ( ) . queue_info ( ) . is_full ( ) & & self . state = = SyncState ::Waiting {
2016-02-11 21:10:41 +01:00
self . state = SyncState ::Blocks ;
2016-02-04 02:28:16 +01:00
self . continue_sync ( io ) ;
2016-10-29 13:07:06 +02:00
} else if self . state = = SyncState ::SnapshotWaiting {
match io . snapshot_service ( ) . status ( ) {
RestorationStatus ::Inactive = > {
trace! ( target :" sync " , " Snapshot restoration is complete " ) ;
self . restart ( io ) ;
self . continue_sync ( io ) ;
} ,
RestorationStatus ::Ongoing { state_chunks : _ , block_chunks : _ , state_chunks_done , block_chunks_done , } = > {
if ! self . snapshot . is_complete ( ) & & self . snapshot . done_chunks ( ) - ( state_chunks_done + block_chunks_done ) as usize < = MAX_SNAPSHOT_CHUNKS_DOWNLOAD_AHEAD {
trace! ( target :" sync " , " Resuming snapshot sync " ) ;
self . state = SyncState ::SnapshotData ;
self . continue_sync ( io ) ;
}
} ,
RestorationStatus ::Failed = > {
trace! ( target : " sync " , " Snapshot restoration aborted " ) ;
self . state = SyncState ::WaitingPeers ;
self . snapshot . clear ( ) ;
self . continue_sync ( io ) ;
} ,
}
2016-02-04 02:28:16 +01:00
}
2015-12-22 22:19:50 +01:00
}
2016-02-05 18:34:08 +01:00
2016-02-07 22:52:56 +01:00
/// creates rlp to send for the tree defined by 'from' and 'to' hashes
2016-02-05 18:34:08 +01:00
fn create_new_hashes_rlp ( chain : & BlockChainClient , from : & H256 , to : & H256 ) -> Option < Bytes > {
match chain . tree_route ( from , to ) {
Some ( route ) = > {
2016-05-24 21:56:17 +02:00
let uncles = chain . find_uncles ( from ) . unwrap_or_else ( Vec ::new ) ;
2016-02-05 18:34:08 +01:00
match route . blocks . len ( ) {
0 = > None ,
_ = > {
2016-05-24 21:56:17 +02:00
let mut blocks = route . blocks ;
blocks . extend ( uncles ) ;
let mut rlp_stream = RlpStream ::new_list ( blocks . len ( ) ) ;
for block_hash in blocks {
2016-02-07 23:39:02 +01:00
let mut hash_rlp = RlpStream ::new_list ( 2 ) ;
2016-06-14 12:18:20 +02:00
let number = HeaderView ::new ( & chain . block_header ( BlockID ::Hash ( block_hash . clone ( ) ) )
. expect ( " chain.tree_route and chain.find_uncles only return hahses of blocks that are in the blockchain. qed. " ) ) . number ( ) ;
2016-02-07 23:39:02 +01:00
hash_rlp . append ( & block_hash ) ;
2016-06-13 18:49:40 +02:00
hash_rlp . append ( & number ) ;
2016-06-18 15:11:10 +02:00
rlp_stream . append_raw ( hash_rlp . as_raw ( ) , 1 ) ;
2016-02-07 23:39:02 +01:00
}
Some ( rlp_stream . out ( ) )
2016-02-05 18:34:08 +01:00
}
}
} ,
None = > None
}
}
2016-02-07 22:52:56 +01:00
/// creates latest block rlp for the given client
2016-02-06 21:00:52 +01:00
fn create_latest_block_rlp ( chain : & BlockChainClient ) -> Bytes {
2016-02-07 23:39:02 +01:00
let mut rlp_stream = RlpStream ::new_list ( 2 ) ;
2016-10-10 17:43:05 +02:00
rlp_stream . append_raw ( & chain . block ( BlockID ::Hash ( chain . chain_info ( ) . best_block_hash ) ) . expect ( " Best block always exists " ) , 1 ) ;
2016-02-07 23:39:02 +01:00
rlp_stream . append ( & chain . chain_info ( ) . total_difficulty ) ;
rlp_stream . out ( )
2016-02-06 21:00:52 +01:00
}
2016-06-29 21:49:12 +02:00
/// creates latest block rlp for the given client
fn create_new_block_rlp ( chain : & BlockChainClient , hash : & H256 ) -> Bytes {
let mut rlp_stream = RlpStream ::new_list ( 2 ) ;
rlp_stream . append_raw ( & chain . block ( BlockID ::Hash ( hash . clone ( ) ) ) . expect ( " Block has just been sealed; qed " ) , 1 ) ;
rlp_stream . append ( & chain . block_total_difficulty ( BlockID ::Hash ( hash . clone ( ) ) ) . expect ( " Block has just been sealed; qed. " ) ) ;
rlp_stream . out ( )
}
2016-02-07 22:52:56 +01:00
/// returns peer ids that have less blocks than our chain
2016-10-10 17:43:05 +02:00
fn get_lagging_peers ( & mut self , chain_info : & BlockChainInfo , io : & SyncIo ) -> Vec < PeerId > {
2016-02-05 23:45:25 +01:00
let latest_hash = chain_info . best_block_hash ;
2016-02-14 17:10:55 +01:00
self . peers . iter_mut ( ) . filter_map ( | ( & id , ref mut peer_info ) |
2016-05-19 11:00:32 +02:00
match io . chain ( ) . block_status ( BlockID ::Hash ( peer_info . latest_hash . clone ( ) ) ) {
2016-02-06 20:16:59 +01:00
BlockStatus ::InChain = > {
2016-10-10 17:43:05 +02:00
if peer_info . latest_hash ! = latest_hash {
Some ( id )
} else {
None
2016-02-14 17:10:55 +01:00
}
2016-02-06 20:16:59 +01:00
} ,
2016-02-14 17:10:55 +01:00
_ = > None
2016-02-06 18:56:21 +01:00
} )
2016-02-14 17:10:55 +01:00
. collect ::< Vec < _ > > ( )
2016-02-05 23:45:25 +01:00
}
2016-10-10 17:43:05 +02:00
fn select_random_lagging_peers ( & mut self , peers : & [ PeerId ] ) -> Vec < PeerId > {
2016-05-29 00:38:10 +02:00
use rand ::Rng ;
// take sqrt(x) peers
2016-08-08 17:18:10 +02:00
let mut peers = peers . to_vec ( ) ;
2016-05-29 00:38:10 +02:00
let mut count = ( self . peers . len ( ) as f64 ) . powf ( 0.5 ) . round ( ) as usize ;
count = min ( count , MAX_PEERS_PROPAGATION ) ;
count = max ( count , MIN_PEERS_PROPAGATION ) ;
2016-08-08 17:18:10 +02:00
::rand ::thread_rng ( ) . shuffle ( & mut peers ) ;
peers . truncate ( count ) ;
peers
2016-05-29 00:38:10 +02:00
}
2016-02-05 18:34:08 +01:00
2016-05-29 00:38:10 +02:00
/// propagates latest block to lagging peers
2016-10-10 17:43:05 +02:00
fn propagate_blocks ( & mut self , chain_info : & BlockChainInfo , io : & mut SyncIo , sealed : & [ H256 ] , peers : & [ PeerId ] ) -> usize {
2016-08-08 17:18:10 +02:00
trace! ( target : " sync " , " Sending NewBlocks to {:?} " , peers ) ;
2016-02-06 21:00:52 +01:00
let mut sent = 0 ;
2016-10-10 17:43:05 +02:00
for peer_id in peers {
2016-06-29 21:49:12 +02:00
if sealed . is_empty ( ) {
let rlp = ChainSync ::create_latest_block_rlp ( io . chain ( ) ) ;
2016-10-10 17:43:05 +02:00
self . send_packet ( io , * peer_id , NEW_BLOCK_PACKET , rlp ) ;
2016-06-29 21:49:12 +02:00
} else {
for h in sealed {
let rlp = ChainSync ::create_new_block_rlp ( io . chain ( ) , h ) ;
2016-10-10 17:43:05 +02:00
self . send_packet ( io , * peer_id , NEW_BLOCK_PACKET , rlp ) ;
2016-06-29 21:49:12 +02:00
}
}
2016-10-27 08:28:12 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( peer_id ) {
2016-10-10 17:43:05 +02:00
peer . latest_hash = chain_info . best_block_hash . clone ( ) ;
}
2016-05-17 10:32:05 +02:00
sent + = 1 ;
2016-02-06 21:00:52 +01:00
}
sent
}
2016-03-01 21:31:58 +01:00
/// propagates new known hashes to all peers
2016-10-10 17:43:05 +02:00
fn propagate_new_hashes ( & mut self , chain_info : & BlockChainInfo , io : & mut SyncIo , peers : & [ PeerId ] ) -> usize {
2016-08-08 17:18:10 +02:00
trace! ( target : " sync " , " Sending NewHashes to {:?} " , peers ) ;
2016-02-05 18:34:08 +01:00
let mut sent = 0 ;
2016-10-10 17:43:05 +02:00
let last_parent = HeaderView ::new ( & io . chain ( ) . block_header ( BlockID ::Hash ( chain_info . best_block_hash . clone ( ) ) )
. expect ( " Best block always exists " ) ) . parent_hash ( ) ;
for peer_id in peers {
sent + = match ChainSync ::create_new_hashes_rlp ( io . chain ( ) , & last_parent , & chain_info . best_block_hash ) {
2016-02-05 18:34:08 +01:00
Some ( rlp ) = > {
2016-02-06 20:23:25 +01:00
{
2016-10-27 08:28:12 +02:00
if let Some ( ref mut peer ) = self . peers . get_mut ( peer_id ) {
2016-10-10 17:43:05 +02:00
peer . latest_hash = chain_info . best_block_hash . clone ( ) ;
}
2016-02-06 20:23:25 +01:00
}
2016-10-10 17:43:05 +02:00
self . send_packet ( io , * peer_id , NEW_BLOCK_HASHES_PACKET , rlp ) ;
2016-02-05 18:34:08 +01:00
1
} ,
None = > 0
}
}
sent
}
2016-03-28 18:11:00 +02:00
/// propagates new transactions to all peers
2016-09-01 19:07:58 +02:00
pub fn propagate_new_transactions ( & mut self , io : & mut SyncIo ) -> usize {
2016-03-28 18:11:00 +02:00
// Early out of nobody to send to.
2016-04-06 10:07:24 +02:00
if self . peers . is_empty ( ) {
2016-03-28 18:11:00 +02:00
return 0 ;
}
2016-09-01 19:07:58 +02:00
let transactions = io . chain ( ) . pending_transactions ( ) ;
2016-04-06 23:03:07 +02:00
if transactions . is_empty ( ) {
return 0 ;
}
2016-09-16 23:03:26 +02:00
let all_transactions_hashes = transactions . iter ( ) . map ( | tx | tx . hash ( ) ) . collect ::< HashSet < H256 > > ( ) ;
2016-09-01 19:07:58 +02:00
let all_transactions_rlp = {
let mut packet = RlpStream ::new_list ( transactions . len ( ) ) ;
for tx in & transactions { packet . append ( tx ) ; }
packet . out ( )
} ;
2016-03-28 18:11:00 +02:00
2016-09-01 19:07:58 +02:00
// sqrt(x)/x scaled to max u32
let fraction = ( self . peers . len ( ) as f64 ) . powf ( - 0.5 ) . mul ( u32 ::max_value ( ) as f64 ) . round ( ) as u32 ;
let small = self . peers . len ( ) < MIN_PEERS_PROPAGATION ;
let lucky_peers = self . peers . iter_mut ( )
. filter ( | _ | small | | ::rand ::random ::< u32 > ( ) < fraction )
. take ( MAX_PEERS_PROPAGATION )
. filter_map ( | ( peer_id , mut peer_info ) | {
// Send all transactions
if peer_info . last_sent_transactions . is_empty ( ) {
peer_info . last_sent_transactions = all_transactions_hashes . clone ( ) ;
return Some ( ( * peer_id , all_transactions_rlp . clone ( ) ) ) ;
}
2016-03-28 18:11:00 +02:00
2016-09-01 19:07:58 +02:00
// Get hashes of all transactions to send to this peer
let to_send = all_transactions_hashes . difference ( & peer_info . last_sent_transactions ) . cloned ( ) . collect ::< HashSet < _ > > ( ) ;
if to_send . is_empty ( ) {
return None ;
}
2016-03-28 18:11:00 +02:00
2016-09-01 19:07:58 +02:00
// Construct RLP
let mut packet = RlpStream ::new_list ( to_send . len ( ) ) ;
for tx in & transactions {
if to_send . contains ( & tx . hash ( ) ) {
packet . append ( tx ) ;
}
}
2016-09-02 11:37:45 +02:00
peer_info . last_sent_transactions = all_transactions_hashes . clone ( ) ;
2016-09-01 19:07:58 +02:00
Some ( ( * peer_id , packet . out ( ) ) )
} )
. collect ::< Vec < _ > > ( ) ;
// Send RLPs
2016-03-28 18:11:00 +02:00
let sent = lucky_peers . len ( ) ;
2016-09-03 01:33:34 +02:00
if sent > 0 {
2016-10-27 08:28:12 +02:00
for ( peer_id , rlp ) in lucky_peers {
2016-09-03 01:33:34 +02:00
self . send_packet ( io , peer_id , TRANSACTIONS_PACKET , rlp ) ;
}
2016-09-01 19:07:58 +02:00
2016-09-03 01:33:34 +02:00
trace! ( target : " sync " , " Sent up to {} transactions to {} peers. " , transactions . len ( ) , sent ) ;
}
2016-03-28 18:11:00 +02:00
sent
}
2016-06-29 21:49:12 +02:00
fn propagate_latest_blocks ( & mut self , io : & mut SyncIo , sealed : & [ H256 ] ) {
2016-03-07 12:16:37 +01:00
let chain_info = io . chain ( ) . chain_info ( ) ;
if ( ( ( chain_info . best_block_number as i64 ) - ( self . last_sent_block_number as i64 ) ) . abs ( ) as BlockNumber ) < MAX_PEER_LAG_PROPAGATION {
2016-08-08 17:18:10 +02:00
let mut peers = self . get_lagging_peers ( & chain_info , io ) ;
if sealed . is_empty ( ) {
let hashes = self . propagate_new_hashes ( & chain_info , io , & peers ) ;
peers = self . select_random_lagging_peers ( & peers ) ;
let blocks = self . propagate_blocks ( & chain_info , io , sealed , & peers ) ;
if blocks ! = 0 | | hashes ! = 0 {
trace! ( target : " sync " , " Sent latest {} blocks and {} hashes to peers. " , blocks , hashes ) ;
}
} else {
self . propagate_blocks ( & chain_info , io , sealed , & peers ) ;
self . propagate_new_hashes ( & chain_info , io , & peers ) ;
trace! ( target : " sync " , " Sent sealed block to all peers " ) ;
} ;
2016-02-12 13:07:02 +01:00
}
2016-05-27 17:26:50 +02:00
self . propagate_new_transactions ( io ) ;
2016-03-07 12:16:37 +01:00
self . last_sent_block_number = chain_info . best_block_number ;
2016-02-05 18:34:08 +01:00
}
2016-03-05 16:46:04 +01:00
2016-03-07 12:16:37 +01:00
/// Maintain other peers. Send out any new blocks and transactions
pub fn maintain_sync ( & mut self , io : & mut SyncIo ) {
2016-10-29 13:07:06 +02:00
self . maybe_start_snapshot_sync ( io ) ;
2016-03-07 12:16:37 +01:00
self . check_resume ( io ) ;
}
2016-05-31 20:58:33 +02:00
2016-09-01 19:07:58 +02:00
/// called when block is imported to chain - propagates the blocks and updates transactions sent to peers
2016-06-29 21:49:12 +02:00
pub fn chain_new_blocks ( & mut self , io : & mut SyncIo , _imported : & [ H256 ] , invalid : & [ H256 ] , _enacted : & [ H256 ] , _retracted : & [ H256 ] , sealed : & [ H256 ] ) {
2016-05-31 20:58:33 +02:00
if io . is_chain_queue_empty ( ) {
2016-06-29 21:49:12 +02:00
self . propagate_latest_blocks ( io , sealed ) ;
2016-05-31 20:58:33 +02:00
}
if ! invalid . is_empty ( ) {
trace! ( target : " sync " , " Bad blocks in the queue, restarting " ) ;
2016-10-18 18:16:00 +02:00
self . restart ( io ) ;
2016-05-31 20:58:33 +02:00
}
2016-09-01 19:07:58 +02:00
for peer_info in self . peers . values_mut ( ) {
peer_info . last_sent_transactions . clear ( ) ;
}
2016-05-31 20:58:33 +02:00
}
2015-12-22 22:19:50 +01:00
}
2016-02-04 19:30:31 +01:00
#[ cfg(test) ]
mod tests {
2016-02-04 20:03:14 +01:00
use tests ::helpers ::* ;
2016-09-06 15:31:13 +02:00
use tests ::snapshot ::TestSnapshotService ;
2016-02-04 20:03:14 +01:00
use super ::* ;
2016-02-24 21:23:58 +01:00
use ::SyncConfig ;
2016-02-04 20:03:14 +01:00
use util ::* ;
2016-09-01 14:49:12 +02:00
use rlp ::* ;
2016-02-05 23:45:25 +01:00
use super ::{ PeerInfo , PeerAsking } ;
2016-03-17 10:20:35 +01:00
use ethcore ::views ::BlockView ;
2016-02-07 22:08:15 +01:00
use ethcore ::header ::* ;
use ethcore ::client ::* ;
2016-05-31 22:24:32 +02:00
use ethcore ::miner ::MinerService ;
2016-02-07 22:08:15 +01:00
fn get_dummy_block ( order : u32 , parent_hash : H256 ) -> Bytes {
let mut header = Header ::new ( ) ;
2016-08-29 11:35:24 +02:00
header . set_gas_limit ( 0. into ( ) ) ;
header . set_difficulty ( ( order * 100 ) . into ( ) ) ;
header . set_timestamp ( ( order * 10 ) as u64 ) ;
header . set_number ( order as u64 ) ;
header . set_parent_hash ( parent_hash ) ;
header . set_state_root ( H256 ::zero ( ) ) ;
2016-02-07 22:08:15 +01:00
let mut rlp = RlpStream ::new_list ( 3 ) ;
rlp . append ( & header ) ;
2016-09-01 14:49:12 +02:00
rlp . append_raw ( & ::rlp ::EMPTY_LIST_RLP , 1 ) ;
rlp . append_raw ( & ::rlp ::EMPTY_LIST_RLP , 1 ) ;
2016-02-07 22:08:15 +01:00
rlp . out ( )
}
fn get_dummy_blocks ( order : u32 , parent_hash : H256 ) -> Bytes {
let mut rlp = RlpStream ::new_list ( 1 ) ;
rlp . append_raw ( & get_dummy_block ( order , parent_hash ) , 1 ) ;
2016-05-31 17:25:25 +02:00
let difficulty : U256 = ( 100 * order ) . into ( ) ;
2016-02-07 22:08:15 +01:00
rlp . append ( & difficulty ) ;
rlp . out ( )
}
2016-02-04 20:03:14 +01:00
2016-02-07 22:20:59 +01:00
fn get_dummy_hashes ( ) -> Bytes {
let mut rlp = RlpStream ::new_list ( 5 ) ;
for _ in 0 .. 5 {
let mut hash_d_rlp = RlpStream ::new_list ( 2 ) ;
let hash : H256 = H256 ::from ( 0 u64 ) ;
let diff : U256 = U256 ::from ( 1 u64 ) ;
hash_d_rlp . append ( & hash ) ;
hash_d_rlp . append ( & diff ) ;
rlp . append_raw ( & hash_d_rlp . out ( ) , 1 ) ;
}
rlp . out ( )
}
2016-02-04 20:03:14 +01:00
2016-02-04 23:24:36 +01:00
#[ test ]
fn return_receipts_empty ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-04 23:24:36 +01:00
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_receipts ( & io , & UntrustedRlp ::new ( & [ 0xc0 ] ) , 0 ) ;
2016-02-04 23:24:36 +01:00
assert! ( result . is_ok ( ) ) ;
}
2016-02-04 20:03:14 +01:00
#[ test ]
fn return_receipts ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-02-04 23:24:36 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-06-20 17:28:48 +02:00
let sync = dummy_sync_with_peer ( H256 ::new ( ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-04 20:03:14 +01:00
2016-02-04 23:24:36 +01:00
let mut receipt_list = RlpStream ::new_list ( 4 ) ;
receipt_list . append ( & H256 ::from ( " 0000000000000000000000000000000000000000000000005555555555555555 " ) ) ;
receipt_list . append ( & H256 ::from ( " ff00000000000000000000000000000000000000000000000000000000000000 " ) ) ;
receipt_list . append ( & H256 ::from ( " fff0000000000000000000000000000000000000000000000000000000000000 " ) ) ;
receipt_list . append ( & H256 ::from ( " aff0000000000000000000000000000000000000000000000000000000000000 " ) ) ;
2016-02-04 20:03:14 +01:00
2016-02-04 23:58:32 +01:00
let receipts_request = receipt_list . out ( ) ;
2016-02-04 23:24:36 +01:00
// it returns rlp ONLY for hashes started with "f"
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_receipts ( & io , & UntrustedRlp ::new ( & receipts_request . clone ( ) ) , 0 ) ;
2016-02-04 20:03:14 +01:00
assert! ( result . is_ok ( ) ) ;
2016-02-04 23:24:36 +01:00
let rlp_result = result . unwrap ( ) ;
assert! ( rlp_result . is_some ( ) ) ;
// the length of two rlp-encoded receipts
2016-03-12 19:23:17 +01:00
assert_eq! ( 603 , rlp_result . unwrap ( ) . 1. out ( ) . len ( ) ) ;
2016-02-04 23:58:32 +01:00
io . sender = Some ( 2 usize ) ;
2016-06-20 17:28:48 +02:00
ChainSync ::dispatch_packet ( & RwLock ::new ( sync ) , & mut io , 0 usize , super ::GET_RECEIPTS_PACKET , & receipts_request ) ;
2016-02-05 00:57:51 +01:00
assert_eq! ( 1 , io . queue . len ( ) ) ;
2016-02-04 20:03:14 +01:00
}
2016-02-04 23:58:32 +01:00
2016-05-16 14:41:41 +02:00
#[ test ]
fn return_block_headers ( ) {
use ethcore ::views ::HeaderView ;
fn make_hash_req ( h : & H256 , count : usize , skip : usize , reverse : bool ) -> Bytes {
let mut rlp = RlpStream ::new_list ( 4 ) ;
rlp . append ( h ) ;
rlp . append ( & count ) ;
rlp . append ( & skip ) ;
rlp . append ( & if reverse { 1 u32 } else { 0 u32 } ) ;
rlp . out ( )
}
fn make_num_req ( n : usize , count : usize , skip : usize , reverse : bool ) -> Bytes {
let mut rlp = RlpStream ::new_list ( 4 ) ;
rlp . append ( & n ) ;
rlp . append ( & count ) ;
rlp . append ( & skip ) ;
rlp . append ( & if reverse { 1 u32 } else { 0 u32 } ) ;
rlp . out ( )
}
fn to_header_vec ( rlp : ::chain ::RlpResponseResult ) -> Vec < Bytes > {
Rlp ::new ( & rlp . unwrap ( ) . unwrap ( ) . 1. out ( ) ) . iter ( ) . map ( | r | r . as_raw ( ) . to_vec ( ) ) . collect ( )
}
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 100 , EachBlockWith ::Nothing ) ;
let blocks : Vec < _ > = ( 0 .. 100 ) . map ( | i | ( & client as & BlockChainClient ) . block ( BlockID ::Number ( i as BlockNumber ) ) . unwrap ( ) ) . collect ( ) ;
let headers : Vec < _ > = blocks . iter ( ) . map ( | b | Rlp ::new ( b ) . at ( 0 ) . as_raw ( ) . to_vec ( ) ) . collect ( ) ;
let hashes : Vec < _ > = headers . iter ( ) . map ( | h | HeaderView ::new ( h ) . sha3 ( ) ) . collect ( ) ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-05-16 14:41:41 +02:00
let unknown : H256 = H256 ::new ( ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & unknown , 1 , 0 , false ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert! ( to_header_vec ( result ) . is_empty ( ) ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & unknown , 1 , 0 , true ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert! ( to_header_vec ( result ) . is_empty ( ) ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & hashes [ 2 ] , 1 , 0 , true ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 2 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & hashes [ 2 ] , 1 , 0 , false ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 2 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & hashes [ 50 ] , 3 , 5 , false ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 50 ] . clone ( ) , headers [ 56 ] . clone ( ) , headers [ 62 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_hash_req ( & hashes [ 50 ] , 3 , 5 , true ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 50 ] . clone ( ) , headers [ 44 ] . clone ( ) , headers [ 38 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_num_req ( 2 , 1 , 0 , true ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 2 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_num_req ( 2 , 1 , 0 , false ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 2 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_num_req ( 50 , 3 , 5 , false ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 50 ] . clone ( ) , headers [ 56 ] . clone ( ) , headers [ 62 ] . clone ( ) ] ) ;
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_block_headers ( & io , & UntrustedRlp ::new ( & make_num_req ( 50 , 3 , 5 , true ) ) , 0 ) ;
2016-05-16 14:41:41 +02:00
assert_eq! ( to_header_vec ( result ) , vec! [ headers [ 50 ] . clone ( ) , headers [ 44 ] . clone ( ) , headers [ 38 ] . clone ( ) ] ) ;
}
2016-02-05 00:57:51 +01:00
#[ test ]
fn return_nodes ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
let mut queue = VecDeque ::new ( ) ;
2016-06-20 17:28:48 +02:00
let sync = dummy_sync_with_peer ( H256 ::new ( ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-04 23:58:32 +01:00
2016-02-05 00:57:51 +01:00
let mut node_list = RlpStream ::new_list ( 3 ) ;
node_list . append ( & H256 ::from ( " 0000000000000000000000000000000000000000000000005555555555555555 " ) ) ;
node_list . append ( & H256 ::from ( " ffffffffffffffffffffffffffffffffffffffffffffaaaaaaaaaaaaaaaaaaaa " ) ) ;
node_list . append ( & H256 ::from ( " aff0000000000000000000000000000000000000000000000000000000000000 " ) ) ;
let node_request = node_list . out ( ) ;
// it returns rlp ONLY for hashes started with "f"
2016-06-13 18:49:40 +02:00
let result = ChainSync ::return_node_data ( & io , & UntrustedRlp ::new ( & node_request . clone ( ) ) , 0 ) ;
2016-02-05 00:57:51 +01:00
assert! ( result . is_ok ( ) ) ;
let rlp_result = result . unwrap ( ) ;
assert! ( rlp_result . is_some ( ) ) ;
// the length of one rlp-encoded hashe
2016-10-27 15:25:54 +02:00
let rlp = rlp_result . unwrap ( ) . 1. out ( ) ;
let rlp = Rlp ::new ( & rlp ) ;
assert_eq! ( 1 , rlp . item_count ( ) ) ;
2016-02-05 00:57:51 +01:00
io . sender = Some ( 2 usize ) ;
2016-06-20 17:28:48 +02:00
ChainSync ::dispatch_packet ( & RwLock ::new ( sync ) , & mut io , 0 usize , super ::GET_NODE_DATA_PACKET , & node_request ) ;
2016-02-05 00:57:51 +01:00
assert_eq! ( 1 , io . queue . len ( ) ) ;
}
2016-02-05 23:45:25 +01:00
2016-05-16 14:41:41 +02:00
fn dummy_sync_with_peer ( peer_latest_hash : H256 , client : & BlockChainClient ) -> ChainSync {
2016-05-31 21:31:42 +02:00
let mut sync = ChainSync ::new ( SyncConfig ::default ( ) , client ) ;
2016-02-05 23:45:25 +01:00
sync . peers . insert ( 0 ,
2016-05-16 14:41:41 +02:00
PeerInfo {
2016-02-06 18:56:21 +01:00
protocol_version : 0 ,
genesis : H256 ::zero ( ) ,
2016-11-03 22:22:25 +01:00
network_id : 0 ,
2016-02-14 17:10:55 +01:00
latest_hash : peer_latest_hash ,
2016-05-16 14:41:41 +02:00
difficulty : None ,
2016-02-06 18:56:21 +01:00
asking : PeerAsking ::Nothing ,
2016-05-16 14:41:41 +02:00
asking_blocks : Vec ::new ( ) ,
2016-02-12 13:07:02 +01:00
asking_hash : None ,
2016-10-29 13:07:06 +02:00
ask_time : 0 ,
2016-09-01 19:07:58 +02:00
last_sent_transactions : HashSet ::new ( ) ,
2016-07-23 17:27:23 +02:00
expired : false ,
2016-08-07 22:42:35 +02:00
confirmation : super ::ForkConfirmation ::Confirmed ,
2016-09-06 15:31:13 +02:00
snapshot_number : None ,
snapshot_hash : None ,
asking_snapshot_data : None ,
2016-10-18 18:16:00 +02:00
block_set : None ,
2016-05-16 14:41:41 +02:00
} ) ;
2016-02-05 23:45:25 +01:00
sync
}
#[ test ]
fn finds_lagging_peers ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
2016-02-05 23:45:25 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 10 ) , & client ) ;
2016-03-04 21:52:03 +01:00
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-05 23:45:25 +01:00
2016-03-04 21:52:03 +01:00
let lagging_peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-02-05 23:45:25 +01:00
assert_eq! ( 1 , lagging_peers . len ( ) )
}
#[ test ]
fn calculates_tree_for_lagging_peer ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 15 , EachBlockWith ::Uncle ) ;
2016-02-05 23:45:25 +01:00
let start = client . block_hash_delta_minus ( 4 ) ;
let end = client . block_hash_delta_minus ( 2 ) ;
// wrong way end -> start, should be None
let rlp = ChainSync ::create_new_hashes_rlp ( & client , & end , & start ) ;
assert! ( rlp . is_none ( ) ) ;
let rlp = ChainSync ::create_new_hashes_rlp ( & client , & start , & end ) . unwrap ( ) ;
2016-02-07 23:39:02 +01:00
// size of three rlp encoded hash-difficulty
assert_eq! ( 107 , rlp . len ( ) ) ;
2016-02-05 23:45:25 +01:00
}
#[ test ]
2016-02-06 21:00:52 +01:00
fn sends_new_hashes_to_lagging_peer ( ) {
2016-02-05 23:45:25 +01:00
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
2016-02-05 23:45:25 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-04 21:52:03 +01:00
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-05 23:45:25 +01:00
2016-08-10 16:29:40 +02:00
let peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-08-08 17:18:10 +02:00
let peer_count = sync . propagate_new_hashes ( & chain_info , & mut io , & peers ) ;
2016-02-05 23:45:25 +01:00
// 1 message should be send
assert_eq! ( 1 , io . queue . len ( ) ) ;
// 1 peer should be updated
2016-02-06 18:56:21 +01:00
assert_eq! ( 1 , peer_count ) ;
2016-02-05 23:45:25 +01:00
// NEW_BLOCK_HASHES_PACKET
assert_eq! ( 0x01 , io . queue [ 0 ] . packet_id ) ;
}
2016-02-06 21:00:52 +01:00
#[ test ]
fn sends_latest_block_to_lagging_peer ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
2016-02-06 21:00:52 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-04 21:52:03 +01:00
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-08-10 16:29:40 +02:00
let peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-08-08 17:18:10 +02:00
let peer_count = sync . propagate_blocks ( & chain_info , & mut io , & [ ] , & peers ) ;
2016-06-29 21:49:12 +02:00
// 1 message should be send
assert_eq! ( 1 , io . queue . len ( ) ) ;
// 1 peer should be updated
assert_eq! ( 1 , peer_count ) ;
// NEW_BLOCK_PACKET
assert_eq! ( 0x07 , io . queue [ 0 ] . packet_id ) ;
}
#[ test ]
fn sends_sealed_block ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
let mut queue = VecDeque ::new ( ) ;
let hash = client . block_hash ( BlockID ::Number ( 99 ) ) . unwrap ( ) ;
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-08-10 16:29:40 +02:00
let peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-08-08 17:18:10 +02:00
let peer_count = sync . propagate_blocks ( & chain_info , & mut io , & [ hash . clone ( ) ] , & peers ) ;
2016-02-06 21:00:52 +01:00
// 1 message should be send
assert_eq! ( 1 , io . queue . len ( ) ) ;
// 1 peer should be updated
assert_eq! ( 1 , peer_count ) ;
// NEW_BLOCK_PACKET
assert_eq! ( 0x07 , io . queue [ 0 ] . packet_id ) ;
}
2016-02-07 21:01:09 +01:00
2016-09-01 19:07:58 +02:00
#[ test ]
fn propagates_transactions ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
client . insert_transaction_to_queue ( ) ;
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 1 ) , & client ) ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-09-01 19:07:58 +02:00
let peer_count = sync . propagate_new_transactions ( & mut io ) ;
// Try to propagate same transactions for the second time
let peer_count2 = sync . propagate_new_transactions ( & mut io ) ;
// 1 message should be send
assert_eq! ( 1 , io . queue . len ( ) ) ;
// 1 peer should be updated but only once
assert_eq! ( 1 , peer_count ) ;
assert_eq! ( 0 , peer_count2 ) ;
// TRANSACTIONS_PACKET
assert_eq! ( 0x02 , io . queue [ 0 ] . packet_id ) ;
}
#[ test ]
fn propagates_transactions_again_after_new_block ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
client . insert_transaction_to_queue ( ) ;
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 1 ) , & client ) ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-09-01 19:07:58 +02:00
let peer_count = sync . propagate_new_transactions ( & mut io ) ;
sync . chain_new_blocks ( & mut io , & [ ] , & [ ] , & [ ] , & [ ] , & [ ] ) ;
// Try to propagate same transactions for the second time
let peer_count2 = sync . propagate_new_transactions ( & mut io ) ;
2016-09-02 11:37:45 +02:00
// 2 message should be send
2016-09-01 19:07:58 +02:00
assert_eq! ( 2 , io . queue . len ( ) ) ;
2016-09-02 11:37:45 +02:00
// 1 peer should be updated twice
2016-09-01 19:07:58 +02:00
assert_eq! ( 1 , peer_count ) ;
assert_eq! ( 1 , peer_count2 ) ;
// TRANSACTIONS_PACKET
assert_eq! ( 0x02 , io . queue [ 0 ] . packet_id ) ;
assert_eq! ( 0x02 , io . queue [ 1 ] . packet_id ) ;
}
2016-09-02 11:37:45 +02:00
#[ test ]
fn propagates_transactions_without_alternating ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
client . insert_transaction_to_queue ( ) ;
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 1 ) , & client ) ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
2016-09-02 11:37:45 +02:00
// should sent some
{
2016-09-16 23:03:26 +02:00
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-09-02 11:37:45 +02:00
let peer_count = sync . propagate_new_transactions ( & mut io ) ;
assert_eq! ( 1 , io . queue . len ( ) ) ;
assert_eq! ( 1 , peer_count ) ;
}
// Insert some more
client . insert_transaction_to_queue ( ) ;
2016-09-16 23:03:26 +02:00
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-09-02 11:37:45 +02:00
// Propagate new transactions
let peer_count2 = sync . propagate_new_transactions ( & mut io ) ;
// And now the peer should have all transactions
let peer_count3 = sync . propagate_new_transactions ( & mut io ) ;
// 2 message should be send (in total)
assert_eq! ( 2 , io . queue . len ( ) ) ;
// 1 peer should be updated but only once after inserting new transaction
assert_eq! ( 1 , peer_count2 ) ;
assert_eq! ( 0 , peer_count3 ) ;
// TRANSACTIONS_PACKET
assert_eq! ( 0x02 , io . queue [ 0 ] . packet_id ) ;
assert_eq! ( 0x02 , io . queue [ 1 ] . packet_id ) ;
}
2016-09-01 19:07:58 +02:00
2016-02-07 21:01:09 +01:00
#[ test ]
2016-05-14 14:28:44 +02:00
fn handles_peer_new_block_malformed ( ) {
2016-02-07 22:08:15 +01:00
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 10 , EachBlockWith ::Uncle ) ;
2016-02-07 22:08:15 +01:00
let block_data = get_dummy_block ( 11 , client . chain_info ( ) . best_block_hash ) ;
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
//sync.have_common_block = true;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 22:08:15 +01:00
let block = UntrustedRlp ::new ( & block_data ) ;
let result = sync . on_peer_new_block ( & mut io , 0 , & block ) ;
assert! ( result . is_err ( ) ) ;
}
#[ test ]
fn handles_peer_new_block ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 10 , EachBlockWith ::Uncle ) ;
2016-02-07 22:08:15 +01:00
let block_data = get_dummy_blocks ( 11 , client . chain_info ( ) . best_block_hash ) ;
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 22:08:15 +01:00
let block = UntrustedRlp ::new ( & block_data ) ;
let result = sync . on_peer_new_block ( & mut io , 0 , & block ) ;
assert! ( result . is_ok ( ) ) ;
}
#[ test ]
fn handles_peer_new_block_empty ( ) {
2016-02-07 21:01:09 +01:00
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 10 , EachBlockWith ::Uncle ) ;
2016-02-07 21:01:09 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 21:01:09 +01:00
let empty_data = vec! [ ] ;
let block = UntrustedRlp ::new ( & empty_data ) ;
let result = sync . on_peer_new_block ( & mut io , 0 , & block ) ;
assert! ( result . is_err ( ) ) ;
}
2016-02-07 22:20:59 +01:00
#[ test ]
fn handles_peer_new_hashes ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 10 , EachBlockWith ::Uncle ) ;
2016-02-07 22:20:59 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 22:20:59 +01:00
let hashes_data = get_dummy_hashes ( ) ;
let hashes_rlp = UntrustedRlp ::new ( & hashes_data ) ;
let result = sync . on_peer_new_hashes ( & mut io , 0 , & hashes_rlp ) ;
assert! ( result . is_ok ( ) ) ;
}
#[ test ]
fn handles_peer_new_hashes_empty ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 10 , EachBlockWith ::Uncle ) ;
2016-02-07 22:20:59 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 22:20:59 +01:00
let empty_hashes_data = vec! [ ] ;
let hashes_rlp = UntrustedRlp ::new ( & empty_hashes_data ) ;
let result = sync . on_peer_new_hashes ( & mut io , 0 , & hashes_rlp ) ;
assert! ( result . is_ok ( ) ) ;
}
2016-02-07 23:39:02 +01:00
2016-02-08 12:14:39 +01:00
// idea is that what we produce when propagading latest hashes should be accepted in
// on_peer_new_hashes in our code as well
2016-02-07 23:39:02 +01:00
#[ test ]
fn hashes_rlp_mutually_acceptable ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
2016-02-07 23:39:02 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-04 21:52:03 +01:00
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 23:39:02 +01:00
2016-08-10 16:29:40 +02:00
let peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-08-08 17:18:10 +02:00
sync . propagate_new_hashes ( & chain_info , & mut io , & peers ) ;
2016-02-07 23:39:02 +01:00
let data = & io . queue [ 0 ] . data . clone ( ) ;
2016-05-17 10:32:05 +02:00
let result = sync . on_peer_new_hashes ( & mut io , 0 , & UntrustedRlp ::new ( data ) ) ;
2016-02-07 23:39:02 +01:00
assert! ( result . is_ok ( ) ) ;
}
2016-02-08 12:14:39 +01:00
// idea is that what we produce when propagading latest block should be accepted in
// on_peer_new_block in our code as well
2016-02-07 23:39:02 +01:00
#[ test ]
fn block_rlp_mutually_acceptable ( ) {
let mut client = TestBlockChainClient ::new ( ) ;
2016-03-05 16:46:04 +01:00
client . add_blocks ( 100 , EachBlockWith ::Uncle ) ;
2016-02-07 23:39:02 +01:00
let mut queue = VecDeque ::new ( ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-04 21:52:03 +01:00
let chain_info = client . chain_info ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-02-07 23:39:02 +01:00
2016-08-10 16:29:40 +02:00
let peers = sync . get_lagging_peers ( & chain_info , & io ) ;
2016-08-08 17:18:10 +02:00
sync . propagate_blocks ( & chain_info , & mut io , & [ ] , & peers ) ;
2016-02-07 23:39:02 +01:00
let data = & io . queue [ 0 ] . data . clone ( ) ;
2016-05-17 10:32:05 +02:00
let result = sync . on_peer_new_block ( & mut io , 0 , & UntrustedRlp ::new ( data ) ) ;
2016-02-07 23:39:02 +01:00
assert! ( result . is_ok ( ) ) ;
}
2016-02-24 13:01:45 +01:00
2016-03-05 16:46:04 +01:00
#[ test ]
fn should_add_transactions_to_queue ( ) {
// given
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 98 , EachBlockWith ::Uncle ) ;
client . add_blocks ( 1 , EachBlockWith ::UncleAndTransaction ) ;
client . add_blocks ( 1 , EachBlockWith ::Transaction ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-05 16:46:04 +01:00
let good_blocks = vec! [ client . block_hash_delta_minus ( 2 ) ] ;
let retracted_blocks = vec! [ client . block_hash_delta_minus ( 1 ) ] ;
2016-04-15 07:38:23 +02:00
// Add some balance to clients and reset nonces
2016-04-06 10:07:24 +02:00
for h in & [ good_blocks [ 0 ] , retracted_blocks [ 0 ] ] {
2016-05-19 11:00:32 +02:00
let block = client . block ( BlockID ::Hash ( * h ) ) . unwrap ( ) ;
2016-03-17 10:20:35 +01:00
let view = BlockView ::new ( & block ) ;
2016-03-17 11:27:38 +01:00
client . set_balance ( view . transactions ( ) [ 0 ] . sender ( ) . unwrap ( ) , U256 ::from ( 1_000_000_000 ) ) ;
2016-04-15 07:38:23 +02:00
client . set_nonce ( view . transactions ( ) [ 0 ] . sender ( ) . unwrap ( ) , U256 ::from ( 0 ) ) ;
2016-03-17 10:20:35 +01:00
}
2016-03-05 16:46:04 +01:00
// when
2016-04-15 07:38:23 +02:00
{
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-06-01 12:44:11 +02:00
io . chain . miner . chain_new_blocks ( io . chain , & [ ] , & [ ] , & [ ] , & good_blocks ) ;
2016-06-29 21:49:12 +02:00
sync . chain_new_blocks ( & mut io , & [ ] , & [ ] , & [ ] , & good_blocks , & [ ] ) ;
2016-05-31 22:24:32 +02:00
assert_eq! ( io . chain . miner . status ( ) . transactions_in_future_queue , 0 ) ;
assert_eq! ( io . chain . miner . status ( ) . transactions_in_pending_queue , 1 ) ;
2016-04-15 07:38:23 +02:00
}
// We need to update nonce status (because we say that the block has been imported)
for h in & [ good_blocks [ 0 ] ] {
2016-05-19 11:00:32 +02:00
let block = client . block ( BlockID ::Hash ( * h ) ) . unwrap ( ) ;
2016-04-15 07:38:23 +02:00
let view = BlockView ::new ( & block ) ;
client . set_nonce ( view . transactions ( ) [ 0 ] . sender ( ) . unwrap ( ) , U256 ::from ( 1 ) ) ;
}
{
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-06-01 12:44:11 +02:00
io . chain . miner . chain_new_blocks ( io . chain , & [ ] , & [ ] , & good_blocks , & retracted_blocks ) ;
2016-06-29 21:49:12 +02:00
sync . chain_new_blocks ( & mut io , & [ ] , & [ ] , & good_blocks , & retracted_blocks , & [ ] ) ;
2016-04-15 07:38:23 +02:00
}
2016-03-05 16:46:04 +01:00
// then
2016-05-31 22:24:32 +02:00
let status = client . miner . status ( ) ;
2016-03-17 11:19:12 +01:00
assert_eq! ( status . transactions_in_pending_queue , 1 ) ;
assert_eq! ( status . transactions_in_future_queue , 0 ) ;
2016-03-05 16:46:04 +01:00
}
2016-03-17 12:17:53 +01:00
#[ test ]
fn should_not_add_transactions_to_queue_if_not_synced ( ) {
// given
let mut client = TestBlockChainClient ::new ( ) ;
client . add_blocks ( 98 , EachBlockWith ::Uncle ) ;
client . add_blocks ( 1 , EachBlockWith ::UncleAndTransaction ) ;
client . add_blocks ( 1 , EachBlockWith ::Transaction ) ;
2016-05-16 14:41:41 +02:00
let mut sync = dummy_sync_with_peer ( client . block_hash_delta_minus ( 5 ) , & client ) ;
2016-03-17 12:17:53 +01:00
let good_blocks = vec! [ client . block_hash_delta_minus ( 2 ) ] ;
let retracted_blocks = vec! [ client . block_hash_delta_minus ( 1 ) ] ;
let mut queue = VecDeque ::new ( ) ;
2016-09-16 23:03:26 +02:00
let ss = TestSnapshotService ::new ( ) ;
let mut io = TestIo ::new ( & mut client , & ss , & mut queue , None ) ;
2016-03-17 12:17:53 +01:00
// when
2016-06-29 21:49:12 +02:00
sync . chain_new_blocks ( & mut io , & [ ] , & [ ] , & [ ] , & good_blocks , & [ ] ) ;
2016-05-31 22:24:32 +02:00
assert_eq! ( io . chain . miner . status ( ) . transactions_in_future_queue , 0 ) ;
assert_eq! ( io . chain . miner . status ( ) . transactions_in_pending_queue , 0 ) ;
2016-06-29 21:49:12 +02:00
sync . chain_new_blocks ( & mut io , & [ ] , & [ ] , & good_blocks , & retracted_blocks , & [ ] ) ;
2016-03-17 12:17:53 +01:00
// then
2016-05-31 22:24:32 +02:00
let status = io . chain . miner . status ( ) ;
2016-03-17 14:19:12 +01:00
assert_eq! ( status . transactions_in_pending_queue , 0 ) ;
assert_eq! ( status . transactions_in_future_queue , 0 ) ;
2016-03-17 12:17:53 +01:00
}
2016-02-11 18:36:26 +01:00
}