Merge pull request #560 from ethcore/clippy_warnings
Fixing clippy warnings = small refactoring of `request_blocks`
This commit is contained in:
commit
1a7ef8cc85
@ -635,16 +635,7 @@ impl ChainSync {
|
|||||||
match self.last_imported_block { None => 0, Some(x) => x }
|
match self.last_imported_block { None => 0, Some(x) => x }
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Find some headers or blocks to download for a peer.
|
fn find_block_bodies_hashes_to_request(&self, ignore_others: bool) -> (Vec<H256>, Vec<BlockNumber>) {
|
||||||
fn request_blocks(&mut self, io: &mut SyncIo, peer_id: PeerId, ignore_others: bool) {
|
|
||||||
self.clear_peer_download(peer_id);
|
|
||||||
|
|
||||||
if io.chain().queue_info().is_full() {
|
|
||||||
self.pause_sync();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// check to see if we need to download any block bodies first
|
|
||||||
let mut needed_bodies: Vec<H256> = Vec::new();
|
let mut needed_bodies: Vec<H256> = Vec::new();
|
||||||
let mut needed_numbers: Vec<BlockNumber> = Vec::new();
|
let mut needed_numbers: Vec<BlockNumber> = Vec::new();
|
||||||
|
|
||||||
@ -664,74 +655,88 @@ impl ChainSync {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
(needed_bodies, needed_numbers)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Find some headers or blocks to download for a peer.
|
||||||
|
fn request_blocks(&mut self, io: &mut SyncIo, peer_id: PeerId, ignore_others: bool) {
|
||||||
|
self.clear_peer_download(peer_id);
|
||||||
|
|
||||||
|
if io.chain().queue_info().is_full() {
|
||||||
|
self.pause_sync();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// check to see if we need to download any block bodies first
|
||||||
|
let (needed_bodies, needed_numbers) = self.find_block_bodies_hashes_to_request(ignore_others);
|
||||||
if !needed_bodies.is_empty() {
|
if !needed_bodies.is_empty() {
|
||||||
let (head, _) = self.headers.range_iter().next().unwrap();
|
let (head, _) = self.headers.range_iter().next().unwrap();
|
||||||
if needed_numbers.first().unwrap() - head > self.max_download_ahead_blocks as BlockNumber {
|
if needed_numbers.first().unwrap() - head > self.max_download_ahead_blocks as BlockNumber {
|
||||||
trace!(target: "sync", "{}: Stalled download ({} vs {}), helping with downloading block bodies", peer_id, needed_numbers.first().unwrap(), head);
|
trace!(target: "sync", "{}: Stalled download ({} vs {}), helping with downloading block bodies", peer_id, needed_numbers.first().unwrap(), head);
|
||||||
self.request_blocks(io, peer_id, true);
|
self.request_blocks(io, peer_id, true);
|
||||||
return;
|
} else {
|
||||||
|
self.downloading_bodies.extend(needed_numbers.iter());
|
||||||
|
replace(&mut self.peers.get_mut(&peer_id).unwrap().asking_blocks, needed_numbers);
|
||||||
|
self.request_bodies(io, peer_id, needed_bodies);
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// check if need to download headers
|
||||||
|
let mut start = 0;
|
||||||
|
if !self.have_common_block {
|
||||||
|
// download backwards until common block is found 1 header at a time
|
||||||
|
let chain_info = io.chain().chain_info();
|
||||||
|
start = chain_info.best_block_number;
|
||||||
|
if !self.headers.is_empty() {
|
||||||
|
start = min(start, self.headers.range_iter().next().unwrap().0 - 1);
|
||||||
|
}
|
||||||
|
if start == 0 {
|
||||||
|
self.have_common_block = true; //reached genesis
|
||||||
|
self.last_imported_hash = Some(chain_info.genesis_hash);
|
||||||
|
self.last_imported_block = Some(0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if self.have_common_block {
|
||||||
|
let mut headers: Vec<BlockNumber> = Vec::new();
|
||||||
|
let mut prev = self.current_base_block() + 1;
|
||||||
|
let head = self.headers.range_iter().next().map(|(h, _)| h);
|
||||||
|
for (next, ref items) in self.headers.range_iter() {
|
||||||
|
if !headers.is_empty() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if next <= prev {
|
||||||
|
prev = next + items.len() as BlockNumber;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let mut block = prev;
|
||||||
|
while block < next && headers.len() < MAX_HEADERS_TO_REQUEST {
|
||||||
|
if ignore_others || !self.downloading_headers.contains(&(block as BlockNumber)) {
|
||||||
|
headers.push(block as BlockNumber);
|
||||||
|
}
|
||||||
|
block += 1;
|
||||||
|
}
|
||||||
|
prev = next + items.len() as BlockNumber;
|
||||||
|
}
|
||||||
|
|
||||||
|
if !headers.is_empty() {
|
||||||
|
start = headers[0];
|
||||||
|
if head.is_some() && start > head.unwrap() && start - head.unwrap() > self.max_download_ahead_blocks as BlockNumber {
|
||||||
|
trace!(target: "sync", "{}: Stalled download ({} vs {}), helping with downloading headers", peer_id, start, head.unwrap());
|
||||||
|
self.request_blocks(io, peer_id, true);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
let count = headers.len();
|
||||||
|
self.downloading_headers.extend(headers.iter());
|
||||||
|
replace(&mut self.peers.get_mut(&peer_id).unwrap().asking_blocks, headers);
|
||||||
|
assert!(!self.headers.have_item(&start));
|
||||||
|
self.request_headers_by_number(io, peer_id, start, count, 0, false);
|
||||||
}
|
}
|
||||||
self.downloading_bodies.extend(needed_numbers.iter());
|
|
||||||
replace(&mut self.peers.get_mut(&peer_id).unwrap().asking_blocks, needed_numbers);
|
|
||||||
self.request_bodies(io, peer_id, needed_bodies);
|
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
// check if need to download headers
|
// continue search for common block
|
||||||
let mut start = 0;
|
self.downloading_headers.insert(start);
|
||||||
if !self.have_common_block {
|
self.request_headers_by_number(io, peer_id, start, 1, 0, false);
|
||||||
// download backwards until common block is found 1 header at a time
|
|
||||||
let chain_info = io.chain().chain_info();
|
|
||||||
start = chain_info.best_block_number;
|
|
||||||
if !self.headers.is_empty() {
|
|
||||||
start = min(start, self.headers.range_iter().next().unwrap().0 - 1);
|
|
||||||
}
|
|
||||||
if start == 0 {
|
|
||||||
self.have_common_block = true; //reached genesis
|
|
||||||
self.last_imported_hash = Some(chain_info.genesis_hash);
|
|
||||||
self.last_imported_block = Some(0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if self.have_common_block {
|
|
||||||
let mut headers: Vec<BlockNumber> = Vec::new();
|
|
||||||
let mut prev = self.current_base_block() + 1;
|
|
||||||
let head = self.headers.range_iter().next().map(|(h, _)| h);
|
|
||||||
for (next, ref items) in self.headers.range_iter() {
|
|
||||||
if !headers.is_empty() {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
if next <= prev {
|
|
||||||
prev = next + items.len() as BlockNumber;
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let mut block = prev;
|
|
||||||
while block < next && headers.len() < MAX_HEADERS_TO_REQUEST {
|
|
||||||
if ignore_others || !self.downloading_headers.contains(&(block as BlockNumber)) {
|
|
||||||
headers.push(block as BlockNumber);
|
|
||||||
}
|
|
||||||
block += 1;
|
|
||||||
}
|
|
||||||
prev = next + items.len() as BlockNumber;
|
|
||||||
}
|
|
||||||
|
|
||||||
if !headers.is_empty() {
|
|
||||||
start = headers[0];
|
|
||||||
if head.is_some() && start > head.unwrap() && start - head.unwrap() > self.max_download_ahead_blocks as BlockNumber {
|
|
||||||
trace!(target: "sync", "{}: Stalled download ({} vs {}), helping with downloading headers", peer_id, start, head.unwrap());
|
|
||||||
self.request_blocks(io, peer_id, true);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
let count = headers.len();
|
|
||||||
self.downloading_headers.extend(headers.iter());
|
|
||||||
replace(&mut self.peers.get_mut(&peer_id).unwrap().asking_blocks, headers);
|
|
||||||
assert!(!self.headers.have_item(&start));
|
|
||||||
self.request_headers_by_number(io, peer_id, start, count, 0, false);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
// continue search for common block
|
|
||||||
self.downloading_headers.insert(start);
|
|
||||||
self.request_headers_by_number(io, peer_id, start, 1, 0, false);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2003,6 +2003,7 @@ mod tests {
|
|||||||
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
#[cfg_attr(feature = "dev", allow(cyclomatic_complexity))]
|
||||||
fn u256_multi_full_mul() {
|
fn u256_multi_full_mul() {
|
||||||
let result = U256([0, 0, 0, 0]).full_mul(U256([0, 0, 0, 0]));
|
let result = U256([0, 0, 0, 0]).full_mul(U256([0, 0, 0, 0]));
|
||||||
assert_eq!(U512([0, 0, 0, 0, 0, 0, 0, 0]), result);
|
assert_eq!(U512([0, 0, 0, 0, 0, 0, 0, 0]), result);
|
||||||
|
@ -111,7 +111,7 @@ impl<Row, Col, Val> Table<Row, Col, Val>
|
|||||||
///
|
///
|
||||||
/// Returns previous value (if any)
|
/// Returns previous value (if any)
|
||||||
pub fn insert(&mut self, row: Row, col: Col, val: Val) -> Option<Val> {
|
pub fn insert(&mut self, row: Row, col: Col, val: Val) -> Option<Val> {
|
||||||
self.map.entry(row).or_insert_with(|| HashMap::new()).insert(col, val)
|
self.map.entry(row).or_insert_with(HashMap::new).insert(col, val)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user