replace synchronization primitives with those from parking_lot (#1593)
* parking_lot in cargo.toml * replace all lock invocations with parking_lot ones * use parking_lot synchronization primitives
This commit is contained in:
parent
4226c0f631
commit
36d3d0d7d7
18
Cargo.lock
generated
18
Cargo.lock
generated
@ -232,6 +232,7 @@ name = "ethash"
|
|||||||
version = "1.3.0"
|
version = "1.3.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"log 0.3.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
"log 0.3.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"parking_lot 0.2.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"primal 0.2.3 (registry+https://github.com/rust-lang/crates.io-index)",
|
"primal 0.2.3 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"sha3 0.1.0",
|
"sha3 0.1.0",
|
||||||
]
|
]
|
||||||
@ -403,6 +404,7 @@ dependencies = [
|
|||||||
"log 0.3.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
"log 0.3.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"mio 0.5.1 (git+https://github.com/ethcore/mio?branch=v0.5.x)",
|
"mio 0.5.1 (git+https://github.com/ethcore/mio?branch=v0.5.x)",
|
||||||
"nix 0.5.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
"nix 0.5.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"parking_lot 0.2.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"rand 0.3.14 (registry+https://github.com/rust-lang/crates.io-index)",
|
"rand 0.3.14 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"rocksdb 0.4.5 (git+https://github.com/ethcore/rust-rocksdb)",
|
"rocksdb 0.4.5 (git+https://github.com/ethcore/rust-rocksdb)",
|
||||||
"rust-crypto 0.2.36 (registry+https://github.com/rust-lang/crates.io-index)",
|
"rust-crypto 0.2.36 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
@ -907,6 +909,17 @@ dependencies = [
|
|||||||
"parity-dapps 0.3.0 (git+https://github.com/ethcore/parity-ui.git)",
|
"parity-dapps 0.3.0 (git+https://github.com/ethcore/parity-ui.git)",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "parking_lot"
|
||||||
|
version = "0.2.6"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
dependencies = [
|
||||||
|
"kernel32-sys 0.2.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"libc 0.2.12 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"smallvec 0.1.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"winapi 0.2.6 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "phf"
|
name = "phf"
|
||||||
version = "0.7.14"
|
version = "0.7.14"
|
||||||
@ -1174,6 +1187,11 @@ name = "slab"
|
|||||||
version = "0.2.0"
|
version = "0.2.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "smallvec"
|
||||||
|
version = "0.1.8"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "solicit"
|
name = "solicit"
|
||||||
version = "0.4.4"
|
version = "0.4.4"
|
||||||
|
@ -70,7 +70,7 @@ mod url;
|
|||||||
use std::sync::{Arc, Mutex};
|
use std::sync::{Arc, Mutex};
|
||||||
use std::net::SocketAddr;
|
use std::net::SocketAddr;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use ethcore_util::misc::Lockable;
|
|
||||||
use jsonrpc_core::{IoHandler, IoDelegate};
|
use jsonrpc_core::{IoHandler, IoDelegate};
|
||||||
use router::auth::{Authorization, NoAuth, HttpBasicAuth};
|
use router::auth::{Authorization, NoAuth, HttpBasicAuth};
|
||||||
use ethcore_rpc::Extendable;
|
use ethcore_rpc::Extendable;
|
||||||
@ -153,7 +153,7 @@ impl Server {
|
|||||||
|
|
||||||
/// Set callback for panics.
|
/// Set callback for panics.
|
||||||
pub fn set_panic_handler<F>(&self, handler: F) where F : Fn() -> () + Send + 'static {
|
pub fn set_panic_handler<F>(&self, handler: F) where F : Fn() -> () + Send + 'static {
|
||||||
*self.panic_handler.locked() = Some(Box::new(handler));
|
*self.panic_handler.lock().unwrap() = Some(Box::new(handler));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -17,7 +17,6 @@
|
|||||||
//! Ethcore rocksdb ipc service
|
//! Ethcore rocksdb ipc service
|
||||||
|
|
||||||
use traits::*;
|
use traits::*;
|
||||||
use misc::RwLockable;
|
|
||||||
use rocksdb::{DB, Writable, WriteBatch, IteratorMode, DBIterator, IndexType, Options, DBCompactionStyle, BlockBasedOptions, Direction};
|
use rocksdb::{DB, Writable, WriteBatch, IteratorMode, DBIterator, IndexType, Options, DBCompactionStyle, BlockBasedOptions, Direction};
|
||||||
use std::sync::{RwLock, Arc};
|
use std::sync::{RwLock, Arc};
|
||||||
use std::convert::From;
|
use std::convert::From;
|
||||||
@ -137,8 +136,8 @@ impl Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn flush(&self) -> Result<(), Error> {
|
pub fn flush(&self) -> Result<(), Error> {
|
||||||
let mut cache_lock = self.write_cache.unwrapped_write();
|
let mut cache_lock = self.write_cache.write();
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
if db_lock.is_none() { return Ok(()); }
|
if db_lock.is_none() { return Ok(()); }
|
||||||
let db = db_lock.as_ref().unwrap();
|
let db = db_lock.as_ref().unwrap();
|
||||||
|
|
||||||
@ -147,8 +146,8 @@ impl Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn flush_all(&self) -> Result<(), Error> {
|
pub fn flush_all(&self) -> Result<(), Error> {
|
||||||
let mut cache_lock = self.write_cache.unwrapped_write();
|
let mut cache_lock = self.write_cache.write();
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
if db_lock.is_none() { return Ok(()); }
|
if db_lock.is_none() { return Ok(()); }
|
||||||
let db = db_lock.as_ref().expect("we should have exited with Ok(()) on the previous step");
|
let db = db_lock.as_ref().expect("we should have exited with Ok(()) on the previous step");
|
||||||
|
|
||||||
@ -167,7 +166,7 @@ impl Drop for Database {
|
|||||||
#[derive(Ipc)]
|
#[derive(Ipc)]
|
||||||
impl DatabaseService for Database {
|
impl DatabaseService for Database {
|
||||||
fn open(&self, config: DatabaseConfig, path: String) -> Result<(), Error> {
|
fn open(&self, config: DatabaseConfig, path: String) -> Result<(), Error> {
|
||||||
let mut db = self.db.unwrapped_write();
|
let mut db = self.db.write();
|
||||||
if db.is_some() { return Err(Error::AlreadyOpen); }
|
if db.is_some() { return Err(Error::AlreadyOpen); }
|
||||||
|
|
||||||
let mut opts = Options::new();
|
let mut opts = Options::new();
|
||||||
@ -194,7 +193,7 @@ impl DatabaseService for Database {
|
|||||||
fn close(&self) -> Result<(), Error> {
|
fn close(&self) -> Result<(), Error> {
|
||||||
try!(self.flush_all());
|
try!(self.flush_all());
|
||||||
|
|
||||||
let mut db = self.db.unwrapped_write();
|
let mut db = self.db.write();
|
||||||
if db.is_none() { return Err(Error::IsClosed); }
|
if db.is_none() { return Err(Error::IsClosed); }
|
||||||
|
|
||||||
*db = None;
|
*db = None;
|
||||||
@ -202,19 +201,19 @@ impl DatabaseService for Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn put(&self, key: &[u8], value: &[u8]) -> Result<(), Error> {
|
fn put(&self, key: &[u8], value: &[u8]) -> Result<(), Error> {
|
||||||
let mut cache_lock = self.write_cache.unwrapped_write();
|
let mut cache_lock = self.write_cache.write();
|
||||||
cache_lock.write(key.to_vec(), value.to_vec());
|
cache_lock.write(key.to_vec(), value.to_vec());
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn delete(&self, key: &[u8]) -> Result<(), Error> {
|
fn delete(&self, key: &[u8]) -> Result<(), Error> {
|
||||||
let mut cache_lock = self.write_cache.unwrapped_write();
|
let mut cache_lock = self.write_cache.write();
|
||||||
cache_lock.remove(key.to_vec());
|
cache_lock.remove(key.to_vec());
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn write(&self, transaction: DBTransaction) -> Result<(), Error> {
|
fn write(&self, transaction: DBTransaction) -> Result<(), Error> {
|
||||||
let mut cache_lock = self.write_cache.unwrapped_write();
|
let mut cache_lock = self.write_cache.write();
|
||||||
|
|
||||||
let mut writes = transaction.writes.borrow_mut();
|
let mut writes = transaction.writes.borrow_mut();
|
||||||
for kv in writes.drain(..) {
|
for kv in writes.drain(..) {
|
||||||
@ -231,13 +230,13 @@ impl DatabaseService for Database {
|
|||||||
fn get(&self, key: &[u8]) -> Result<Option<Vec<u8>>, Error> {
|
fn get(&self, key: &[u8]) -> Result<Option<Vec<u8>>, Error> {
|
||||||
{
|
{
|
||||||
let key_vec = key.to_vec();
|
let key_vec = key.to_vec();
|
||||||
let cache_hit = self.write_cache.unwrapped_read().get(&key_vec);
|
let cache_hit = self.write_cache.read().get(&key_vec);
|
||||||
|
|
||||||
if cache_hit.is_some() {
|
if cache_hit.is_some() {
|
||||||
return Ok(Some(cache_hit.expect("cache_hit.is_some() = true, still there is none somehow here")))
|
return Ok(Some(cache_hit.expect("cache_hit.is_some() = true, still there is none somehow here")))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
||||||
|
|
||||||
match try!(db.get(key)) {
|
match try!(db.get(key)) {
|
||||||
@ -249,7 +248,7 @@ impl DatabaseService for Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_by_prefix(&self, prefix: &[u8]) -> Result<Option<Vec<u8>>, Error> {
|
fn get_by_prefix(&self, prefix: &[u8]) -> Result<Option<Vec<u8>>, Error> {
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
||||||
|
|
||||||
let mut iter = db.iterator(IteratorMode::From(prefix, Direction::Forward));
|
let mut iter = db.iterator(IteratorMode::From(prefix, Direction::Forward));
|
||||||
@ -261,17 +260,17 @@ impl DatabaseService for Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn is_empty(&self) -> Result<bool, Error> {
|
fn is_empty(&self) -> Result<bool, Error> {
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
||||||
|
|
||||||
Ok(db.iterator(IteratorMode::Start).next().is_none())
|
Ok(db.iterator(IteratorMode::Start).next().is_none())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn iter(&self) -> Result<IteratorHandle, Error> {
|
fn iter(&self) -> Result<IteratorHandle, Error> {
|
||||||
let db_lock = self.db.unwrapped_read();
|
let db_lock = self.db.read();
|
||||||
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
let db = try!(db_lock.as_ref().ok_or(Error::IsClosed));
|
||||||
|
|
||||||
let mut iterators = self.iterators.unwrapped_write();
|
let mut iterators = self.iterators.write();
|
||||||
let next_iterator = iterators.keys().last().unwrap_or(&0) + 1;
|
let next_iterator = iterators.keys().last().unwrap_or(&0) + 1;
|
||||||
iterators.insert(next_iterator, db.iterator(IteratorMode::Start));
|
iterators.insert(next_iterator, db.iterator(IteratorMode::Start));
|
||||||
Ok(next_iterator)
|
Ok(next_iterator)
|
||||||
@ -279,7 +278,7 @@ impl DatabaseService for Database {
|
|||||||
|
|
||||||
fn iter_next(&self, handle: IteratorHandle) -> Option<KeyValue>
|
fn iter_next(&self, handle: IteratorHandle) -> Option<KeyValue>
|
||||||
{
|
{
|
||||||
let mut iterators = self.iterators.unwrapped_write();
|
let mut iterators = self.iterators.write();
|
||||||
let mut iterator = match iterators.get_mut(&handle) {
|
let mut iterator = match iterators.get_mut(&handle) {
|
||||||
Some(some_iterator) => some_iterator,
|
Some(some_iterator) => some_iterator,
|
||||||
None => { return None; },
|
None => { return None; },
|
||||||
@ -294,7 +293,7 @@ impl DatabaseService for Database {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn dispose_iter(&self, handle: IteratorHandle) -> Result<(), Error> {
|
fn dispose_iter(&self, handle: IteratorHandle) -> Result<(), Error> {
|
||||||
let mut iterators = self.iterators.unwrapped_write();
|
let mut iterators = self.iterators.write();
|
||||||
iterators.remove(&handle);
|
iterators.remove(&handle);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -9,3 +9,4 @@ authors = ["arkpar <arkadiy@ethcore.io"]
|
|||||||
log = "0.3"
|
log = "0.3"
|
||||||
sha3 = { path = "../util/sha3" }
|
sha3 = { path = "../util/sha3" }
|
||||||
primal = "0.2.3"
|
primal = "0.2.3"
|
||||||
|
parking_lot = "0.2.6"
|
@ -21,7 +21,6 @@
|
|||||||
|
|
||||||
use primal::is_prime;
|
use primal::is_prime;
|
||||||
use std::cell::Cell;
|
use std::cell::Cell;
|
||||||
use std::sync::Mutex;
|
|
||||||
use std::mem;
|
use std::mem;
|
||||||
use std::ptr;
|
use std::ptr;
|
||||||
use sha3;
|
use sha3;
|
||||||
@ -30,6 +29,8 @@ use std::path::PathBuf;
|
|||||||
use std::io::{self, Read, Write};
|
use std::io::{self, Read, Write};
|
||||||
use std::fs::{self, File};
|
use std::fs::{self, File};
|
||||||
|
|
||||||
|
use parking_lot::Mutex;
|
||||||
|
|
||||||
pub const ETHASH_EPOCH_LENGTH: u64 = 30000;
|
pub const ETHASH_EPOCH_LENGTH: u64 = 30000;
|
||||||
pub const ETHASH_CACHE_ROUNDS: usize = 3;
|
pub const ETHASH_CACHE_ROUNDS: usize = 3;
|
||||||
pub const ETHASH_MIX_BYTES: usize = 128;
|
pub const ETHASH_MIX_BYTES: usize = 128;
|
||||||
@ -134,7 +135,7 @@ impl Light {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn to_file(&self) -> io::Result<()> {
|
pub fn to_file(&self) -> io::Result<()> {
|
||||||
let seed_compute = self.seed_compute.lock().unwrap();
|
let seed_compute = self.seed_compute.lock();
|
||||||
let path = Light::file_path(seed_compute.get_seedhash(self.block_number));
|
let path = Light::file_path(seed_compute.get_seedhash(self.block_number));
|
||||||
try!(fs::create_dir_all(path.parent().unwrap()));
|
try!(fs::create_dir_all(path.parent().unwrap()));
|
||||||
let mut file = try!(File::create(path));
|
let mut file = try!(File::create(path));
|
||||||
|
@ -18,6 +18,8 @@
|
|||||||
//! See https://github.com/ethereum/wiki/wiki/Ethash
|
//! See https://github.com/ethereum/wiki/wiki/Ethash
|
||||||
extern crate primal;
|
extern crate primal;
|
||||||
extern crate sha3;
|
extern crate sha3;
|
||||||
|
extern crate parking_lot;
|
||||||
|
|
||||||
#[macro_use]
|
#[macro_use]
|
||||||
extern crate log;
|
extern crate log;
|
||||||
mod compute;
|
mod compute;
|
||||||
@ -26,7 +28,8 @@ use std::mem;
|
|||||||
use compute::Light;
|
use compute::Light;
|
||||||
pub use compute::{ETHASH_EPOCH_LENGTH, H256, ProofOfWork, SeedHashCompute, quick_get_difficulty};
|
pub use compute::{ETHASH_EPOCH_LENGTH, H256, ProofOfWork, SeedHashCompute, quick_get_difficulty};
|
||||||
|
|
||||||
use std::sync::{Arc, Mutex};
|
use std::sync::Arc;
|
||||||
|
use parking_lot::Mutex;
|
||||||
|
|
||||||
struct LightCache {
|
struct LightCache {
|
||||||
recent_epoch: Option<u64>,
|
recent_epoch: Option<u64>,
|
||||||
@ -61,7 +64,7 @@ impl EthashManager {
|
|||||||
pub fn compute_light(&self, block_number: u64, header_hash: &H256, nonce: u64) -> ProofOfWork {
|
pub fn compute_light(&self, block_number: u64, header_hash: &H256, nonce: u64) -> ProofOfWork {
|
||||||
let epoch = block_number / ETHASH_EPOCH_LENGTH;
|
let epoch = block_number / ETHASH_EPOCH_LENGTH;
|
||||||
let light = {
|
let light = {
|
||||||
let mut lights = self.cache.lock().unwrap();
|
let mut lights = self.cache.lock();
|
||||||
let light = match lights.recent_epoch.clone() {
|
let light = match lights.recent_epoch.clone() {
|
||||||
Some(ref e) if *e == epoch => lights.recent.clone(),
|
Some(ref e) if *e == epoch => lights.recent.clone(),
|
||||||
_ => match lights.prev_epoch.clone() {
|
_ => match lights.prev_epoch.clone() {
|
||||||
@ -108,12 +111,12 @@ fn test_lru() {
|
|||||||
let hash = [0u8; 32];
|
let hash = [0u8; 32];
|
||||||
ethash.compute_light(1, &hash, 1);
|
ethash.compute_light(1, &hash, 1);
|
||||||
ethash.compute_light(50000, &hash, 1);
|
ethash.compute_light(50000, &hash, 1);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().recent_epoch.unwrap(), 1);
|
assert_eq!(ethash.cache.lock().recent_epoch.unwrap(), 1);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().prev_epoch.unwrap(), 0);
|
assert_eq!(ethash.cache.lock().prev_epoch.unwrap(), 0);
|
||||||
ethash.compute_light(1, &hash, 1);
|
ethash.compute_light(1, &hash, 1);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().recent_epoch.unwrap(), 0);
|
assert_eq!(ethash.cache.lock().recent_epoch.unwrap(), 0);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().prev_epoch.unwrap(), 1);
|
assert_eq!(ethash.cache.lock().prev_epoch.unwrap(), 1);
|
||||||
ethash.compute_light(70000, &hash, 1);
|
ethash.compute_light(70000, &hash, 1);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().recent_epoch.unwrap(), 2);
|
assert_eq!(ethash.cache.lock().recent_epoch.unwrap(), 2);
|
||||||
assert_eq!(ethash.cache.lock().unwrap().prev_epoch.unwrap(), 0);
|
assert_eq!(ethash.cache.lock().prev_epoch.unwrap(), 0);
|
||||||
}
|
}
|
||||||
|
@ -17,13 +17,13 @@
|
|||||||
//! Account management.
|
//! Account management.
|
||||||
|
|
||||||
use std::fmt;
|
use std::fmt;
|
||||||
use std::sync::RwLock;
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use util::{Address as H160, H256, H520, RwLockable};
|
use util::{Address as H160, H256, H520, RwLock};
|
||||||
use ethstore::{SecretStore, Error as SSError, SafeAccount, EthStore};
|
use ethstore::{SecretStore, Error as SSError, SafeAccount, EthStore};
|
||||||
use ethstore::dir::{KeyDirectory};
|
use ethstore::dir::{KeyDirectory};
|
||||||
use ethstore::ethkey::{Address as SSAddress, Message as SSMessage, Secret as SSSecret, Random, Generator};
|
use ethstore::ethkey::{Address as SSAddress, Message as SSMessage, Secret as SSSecret, Random, Generator};
|
||||||
|
|
||||||
|
|
||||||
/// Type of unlock.
|
/// Type of unlock.
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
enum Unlock {
|
enum Unlock {
|
||||||
@ -177,7 +177,7 @@ impl AccountProvider {
|
|||||||
|
|
||||||
// check if account is already unlocked pernamently, if it is, do nothing
|
// check if account is already unlocked pernamently, if it is, do nothing
|
||||||
{
|
{
|
||||||
let unlocked = self.unlocked.unwrapped_read();
|
let unlocked = self.unlocked.read();
|
||||||
if let Some(data) = unlocked.get(&account) {
|
if let Some(data) = unlocked.get(&account) {
|
||||||
if let Unlock::Perm = data.unlock {
|
if let Unlock::Perm = data.unlock {
|
||||||
return Ok(())
|
return Ok(())
|
||||||
@ -190,7 +190,7 @@ impl AccountProvider {
|
|||||||
password: password,
|
password: password,
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut unlocked = self.unlocked.unwrapped_write();
|
let mut unlocked = self.unlocked.write();
|
||||||
unlocked.insert(account, data);
|
unlocked.insert(account, data);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@ -208,7 +208,7 @@ impl AccountProvider {
|
|||||||
/// Checks if given account is unlocked
|
/// Checks if given account is unlocked
|
||||||
pub fn is_unlocked<A>(&self, account: A) -> bool where Address: From<A> {
|
pub fn is_unlocked<A>(&self, account: A) -> bool where Address: From<A> {
|
||||||
let account = Address::from(account).into();
|
let account = Address::from(account).into();
|
||||||
let unlocked = self.unlocked.unwrapped_read();
|
let unlocked = self.unlocked.read();
|
||||||
unlocked.get(&account).is_some()
|
unlocked.get(&account).is_some()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -218,12 +218,12 @@ impl AccountProvider {
|
|||||||
let message = Message::from(message).into();
|
let message = Message::from(message).into();
|
||||||
|
|
||||||
let data = {
|
let data = {
|
||||||
let unlocked = self.unlocked.unwrapped_read();
|
let unlocked = self.unlocked.read();
|
||||||
try!(unlocked.get(&account).ok_or(Error::NotUnlocked)).clone()
|
try!(unlocked.get(&account).ok_or(Error::NotUnlocked)).clone()
|
||||||
};
|
};
|
||||||
|
|
||||||
if let Unlock::Temp = data.unlock {
|
if let Unlock::Temp = data.unlock {
|
||||||
let mut unlocked = self.unlocked.unwrapped_write();
|
let mut unlocked = self.unlocked.write();
|
||||||
unlocked.remove(&account).expect("data exists: so key must exist: qed");
|
unlocked.remove(&account).expect("data exists: so key must exist: qed");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -193,14 +193,14 @@ impl BlockQueue {
|
|||||||
fn verify(verification: Arc<Verification>, engine: Arc<Box<Engine>>, wait: Arc<Condvar>, ready: Arc<QueueSignal>, deleting: Arc<AtomicBool>, empty: Arc<Condvar>) {
|
fn verify(verification: Arc<Verification>, engine: Arc<Box<Engine>>, wait: Arc<Condvar>, ready: Arc<QueueSignal>, deleting: Arc<AtomicBool>, empty: Arc<Condvar>) {
|
||||||
while !deleting.load(AtomicOrdering::Acquire) {
|
while !deleting.load(AtomicOrdering::Acquire) {
|
||||||
{
|
{
|
||||||
let mut unverified = verification.unverified.locked();
|
let mut unverified = verification.unverified.lock();
|
||||||
|
|
||||||
if unverified.is_empty() && verification.verifying.locked().is_empty() {
|
if unverified.is_empty() && verification.verifying.lock().is_empty() {
|
||||||
empty.notify_all();
|
empty.notify_all();
|
||||||
}
|
}
|
||||||
|
|
||||||
while unverified.is_empty() && !deleting.load(AtomicOrdering::Acquire) {
|
while unverified.is_empty() && !deleting.load(AtomicOrdering::Acquire) {
|
||||||
unverified = wait.wait(unverified).unwrap();
|
wait.wait(&mut unverified);
|
||||||
}
|
}
|
||||||
|
|
||||||
if deleting.load(AtomicOrdering::Acquire) {
|
if deleting.load(AtomicOrdering::Acquire) {
|
||||||
@ -209,11 +209,11 @@ impl BlockQueue {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let block = {
|
let block = {
|
||||||
let mut unverified = verification.unverified.locked();
|
let mut unverified = verification.unverified.lock();
|
||||||
if unverified.is_empty() {
|
if unverified.is_empty() {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
let mut verifying = verification.verifying.locked();
|
let mut verifying = verification.verifying.lock();
|
||||||
let block = unverified.pop_front().unwrap();
|
let block = unverified.pop_front().unwrap();
|
||||||
verifying.push_back(VerifyingBlock{ hash: block.header.hash(), block: None });
|
verifying.push_back(VerifyingBlock{ hash: block.header.hash(), block: None });
|
||||||
block
|
block
|
||||||
@ -222,7 +222,7 @@ impl BlockQueue {
|
|||||||
let block_hash = block.header.hash();
|
let block_hash = block.header.hash();
|
||||||
match verify_block_unordered(block.header, block.bytes, engine.deref().deref()) {
|
match verify_block_unordered(block.header, block.bytes, engine.deref().deref()) {
|
||||||
Ok(verified) => {
|
Ok(verified) => {
|
||||||
let mut verifying = verification.verifying.locked();
|
let mut verifying = verification.verifying.lock();
|
||||||
for e in verifying.iter_mut() {
|
for e in verifying.iter_mut() {
|
||||||
if e.hash == block_hash {
|
if e.hash == block_hash {
|
||||||
e.block = Some(verified);
|
e.block = Some(verified);
|
||||||
@ -231,16 +231,16 @@ impl BlockQueue {
|
|||||||
}
|
}
|
||||||
if !verifying.is_empty() && verifying.front().unwrap().hash == block_hash {
|
if !verifying.is_empty() && verifying.front().unwrap().hash == block_hash {
|
||||||
// we're next!
|
// we're next!
|
||||||
let mut verified = verification.verified.locked();
|
let mut verified = verification.verified.lock();
|
||||||
let mut bad = verification.bad.locked();
|
let mut bad = verification.bad.lock();
|
||||||
BlockQueue::drain_verifying(&mut verifying, &mut verified, &mut bad);
|
BlockQueue::drain_verifying(&mut verifying, &mut verified, &mut bad);
|
||||||
ready.set();
|
ready.set();
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
let mut verifying = verification.verifying.locked();
|
let mut verifying = verification.verifying.lock();
|
||||||
let mut verified = verification.verified.locked();
|
let mut verified = verification.verified.lock();
|
||||||
let mut bad = verification.bad.locked();
|
let mut bad = verification.bad.lock();
|
||||||
warn!(target: "client", "Stage 2 block verification failed for {}\nError: {:?}", block_hash, err);
|
warn!(target: "client", "Stage 2 block verification failed for {}\nError: {:?}", block_hash, err);
|
||||||
bad.insert(block_hash.clone());
|
bad.insert(block_hash.clone());
|
||||||
verifying.retain(|e| e.hash != block_hash);
|
verifying.retain(|e| e.hash != block_hash);
|
||||||
@ -265,29 +265,29 @@ impl BlockQueue {
|
|||||||
|
|
||||||
/// Clear the queue and stop verification activity.
|
/// Clear the queue and stop verification activity.
|
||||||
pub fn clear(&self) {
|
pub fn clear(&self) {
|
||||||
let mut unverified = self.verification.unverified.locked();
|
let mut unverified = self.verification.unverified.lock();
|
||||||
let mut verifying = self.verification.verifying.locked();
|
let mut verifying = self.verification.verifying.lock();
|
||||||
let mut verified = self.verification.verified.locked();
|
let mut verified = self.verification.verified.lock();
|
||||||
unverified.clear();
|
unverified.clear();
|
||||||
verifying.clear();
|
verifying.clear();
|
||||||
verified.clear();
|
verified.clear();
|
||||||
self.processing.unwrapped_write().clear();
|
self.processing.write().clear();
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Wait for unverified queue to be empty
|
/// Wait for unverified queue to be empty
|
||||||
pub fn flush(&self) {
|
pub fn flush(&self) {
|
||||||
let mut unverified = self.verification.unverified.locked();
|
let mut unverified = self.verification.unverified.lock();
|
||||||
while !unverified.is_empty() || !self.verification.verifying.locked().is_empty() {
|
while !unverified.is_empty() || !self.verification.verifying.lock().is_empty() {
|
||||||
unverified = self.empty.wait(unverified).unwrap();
|
self.empty.wait(&mut unverified);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Check if the block is currently in the queue
|
/// Check if the block is currently in the queue
|
||||||
pub fn block_status(&self, hash: &H256) -> BlockStatus {
|
pub fn block_status(&self, hash: &H256) -> BlockStatus {
|
||||||
if self.processing.unwrapped_read().contains(&hash) {
|
if self.processing.read().contains(&hash) {
|
||||||
return BlockStatus::Queued;
|
return BlockStatus::Queued;
|
||||||
}
|
}
|
||||||
if self.verification.bad.locked().contains(&hash) {
|
if self.verification.bad.lock().contains(&hash) {
|
||||||
return BlockStatus::Bad;
|
return BlockStatus::Bad;
|
||||||
}
|
}
|
||||||
BlockStatus::Unknown
|
BlockStatus::Unknown
|
||||||
@ -298,11 +298,11 @@ impl BlockQueue {
|
|||||||
let header = BlockView::new(&bytes).header();
|
let header = BlockView::new(&bytes).header();
|
||||||
let h = header.hash();
|
let h = header.hash();
|
||||||
{
|
{
|
||||||
if self.processing.unwrapped_read().contains(&h) {
|
if self.processing.read().contains(&h) {
|
||||||
return Err(ImportError::AlreadyQueued.into());
|
return Err(ImportError::AlreadyQueued.into());
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut bad = self.verification.bad.locked();
|
let mut bad = self.verification.bad.lock();
|
||||||
if bad.contains(&h) {
|
if bad.contains(&h) {
|
||||||
return Err(ImportError::KnownBad.into());
|
return Err(ImportError::KnownBad.into());
|
||||||
}
|
}
|
||||||
@ -315,14 +315,14 @@ impl BlockQueue {
|
|||||||
|
|
||||||
match verify_block_basic(&header, &bytes, self.engine.deref().deref()) {
|
match verify_block_basic(&header, &bytes, self.engine.deref().deref()) {
|
||||||
Ok(()) => {
|
Ok(()) => {
|
||||||
self.processing.unwrapped_write().insert(h.clone());
|
self.processing.write().insert(h.clone());
|
||||||
self.verification.unverified.locked().push_back(UnverifiedBlock { header: header, bytes: bytes });
|
self.verification.unverified.lock().push_back(UnverifiedBlock { header: header, bytes: bytes });
|
||||||
self.more_to_verify.notify_all();
|
self.more_to_verify.notify_all();
|
||||||
Ok(h)
|
Ok(h)
|
||||||
},
|
},
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
warn!(target: "client", "Stage 1 block verification failed for {}\nError: {:?}", BlockView::new(&bytes).header_view().sha3(), err);
|
warn!(target: "client", "Stage 1 block verification failed for {}\nError: {:?}", BlockView::new(&bytes).header_view().sha3(), err);
|
||||||
self.verification.bad.locked().insert(h.clone());
|
self.verification.bad.lock().insert(h.clone());
|
||||||
Err(err)
|
Err(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -333,10 +333,10 @@ impl BlockQueue {
|
|||||||
if block_hashes.is_empty() {
|
if block_hashes.is_empty() {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
let mut verified_lock = self.verification.verified.locked();
|
let mut verified_lock = self.verification.verified.lock();
|
||||||
let mut verified = verified_lock.deref_mut();
|
let mut verified = verified_lock.deref_mut();
|
||||||
let mut bad = self.verification.bad.locked();
|
let mut bad = self.verification.bad.lock();
|
||||||
let mut processing = self.processing.unwrapped_write();
|
let mut processing = self.processing.write();
|
||||||
bad.reserve(block_hashes.len());
|
bad.reserve(block_hashes.len());
|
||||||
for hash in block_hashes {
|
for hash in block_hashes {
|
||||||
bad.insert(hash.clone());
|
bad.insert(hash.clone());
|
||||||
@ -360,7 +360,7 @@ impl BlockQueue {
|
|||||||
if block_hashes.is_empty() {
|
if block_hashes.is_empty() {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
let mut processing = self.processing.unwrapped_write();
|
let mut processing = self.processing.write();
|
||||||
for hash in block_hashes {
|
for hash in block_hashes {
|
||||||
processing.remove(&hash);
|
processing.remove(&hash);
|
||||||
}
|
}
|
||||||
@ -368,7 +368,7 @@ impl BlockQueue {
|
|||||||
|
|
||||||
/// Removes up to `max` verified blocks from the queue
|
/// Removes up to `max` verified blocks from the queue
|
||||||
pub fn drain(&self, max: usize) -> Vec<PreverifiedBlock> {
|
pub fn drain(&self, max: usize) -> Vec<PreverifiedBlock> {
|
||||||
let mut verified = self.verification.verified.locked();
|
let mut verified = self.verification.verified.lock();
|
||||||
let count = min(max, verified.len());
|
let count = min(max, verified.len());
|
||||||
let mut result = Vec::with_capacity(count);
|
let mut result = Vec::with_capacity(count);
|
||||||
for _ in 0..count {
|
for _ in 0..count {
|
||||||
@ -385,15 +385,15 @@ impl BlockQueue {
|
|||||||
/// Get queue status.
|
/// Get queue status.
|
||||||
pub fn queue_info(&self) -> BlockQueueInfo {
|
pub fn queue_info(&self) -> BlockQueueInfo {
|
||||||
let (unverified_len, unverified_bytes) = {
|
let (unverified_len, unverified_bytes) = {
|
||||||
let v = self.verification.unverified.locked();
|
let v = self.verification.unverified.lock();
|
||||||
(v.len(), v.heap_size_of_children())
|
(v.len(), v.heap_size_of_children())
|
||||||
};
|
};
|
||||||
let (verifying_len, verifying_bytes) = {
|
let (verifying_len, verifying_bytes) = {
|
||||||
let v = self.verification.verifying.locked();
|
let v = self.verification.verifying.lock();
|
||||||
(v.len(), v.heap_size_of_children())
|
(v.len(), v.heap_size_of_children())
|
||||||
};
|
};
|
||||||
let (verified_len, verified_bytes) = {
|
let (verified_len, verified_bytes) = {
|
||||||
let v = self.verification.verified.locked();
|
let v = self.verification.verified.lock();
|
||||||
(v.len(), v.heap_size_of_children())
|
(v.len(), v.heap_size_of_children())
|
||||||
};
|
};
|
||||||
BlockQueueInfo {
|
BlockQueueInfo {
|
||||||
@ -407,18 +407,18 @@ impl BlockQueue {
|
|||||||
+ verifying_bytes
|
+ verifying_bytes
|
||||||
+ verified_bytes
|
+ verified_bytes
|
||||||
// TODO: https://github.com/servo/heapsize/pull/50
|
// TODO: https://github.com/servo/heapsize/pull/50
|
||||||
//+ self.processing.unwrapped_read().heap_size_of_children(),
|
//+ self.processing.read().heap_size_of_children(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Optimise memory footprint of the heap fields.
|
/// Optimise memory footprint of the heap fields.
|
||||||
pub fn collect_garbage(&self) {
|
pub fn collect_garbage(&self) {
|
||||||
{
|
{
|
||||||
self.verification.unverified.locked().shrink_to_fit();
|
self.verification.unverified.lock().shrink_to_fit();
|
||||||
self.verification.verifying.locked().shrink_to_fit();
|
self.verification.verifying.lock().shrink_to_fit();
|
||||||
self.verification.verified.locked().shrink_to_fit();
|
self.verification.verified.lock().shrink_to_fit();
|
||||||
}
|
}
|
||||||
self.processing.unwrapped_write().shrink_to_fit();
|
self.processing.write().shrink_to_fit();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -170,7 +170,7 @@ impl BlockProvider for BlockChain {
|
|||||||
/// Get raw block data
|
/// Get raw block data
|
||||||
fn block(&self, hash: &H256) -> Option<Bytes> {
|
fn block(&self, hash: &H256) -> Option<Bytes> {
|
||||||
{
|
{
|
||||||
let read = self.blocks.unwrapped_read();
|
let read = self.blocks.read();
|
||||||
if let Some(v) = read.get(hash) {
|
if let Some(v) = read.get(hash) {
|
||||||
return Some(v.clone());
|
return Some(v.clone());
|
||||||
}
|
}
|
||||||
@ -184,7 +184,7 @@ impl BlockProvider for BlockChain {
|
|||||||
match opt {
|
match opt {
|
||||||
Some(b) => {
|
Some(b) => {
|
||||||
let bytes: Bytes = b.to_vec();
|
let bytes: Bytes = b.to_vec();
|
||||||
let mut write = self.blocks.unwrapped_write();
|
let mut write = self.blocks.write();
|
||||||
write.insert(hash.clone(), bytes.clone());
|
write.insert(hash.clone(), bytes.clone());
|
||||||
Some(bytes)
|
Some(bytes)
|
||||||
},
|
},
|
||||||
@ -338,7 +338,7 @@ impl BlockChain {
|
|||||||
};
|
};
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut best_block = bc.best_block.unwrapped_write();
|
let mut best_block = bc.best_block.write();
|
||||||
best_block.number = bc.block_number(&best_block_hash).unwrap();
|
best_block.number = bc.block_number(&best_block_hash).unwrap();
|
||||||
best_block.total_difficulty = bc.block_details(&best_block_hash).unwrap().total_difficulty;
|
best_block.total_difficulty = bc.block_details(&best_block_hash).unwrap().total_difficulty;
|
||||||
best_block.hash = best_block_hash;
|
best_block.hash = best_block_hash;
|
||||||
@ -483,25 +483,25 @@ impl BlockChain {
|
|||||||
self.note_used(CacheID::BlockDetails(hash));
|
self.note_used(CacheID::BlockDetails(hash));
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut write_details = self.block_details.unwrapped_write();
|
let mut write_details = self.block_details.write();
|
||||||
batch.extend_with_cache(write_details.deref_mut(), update.block_details, CacheUpdatePolicy::Overwrite);
|
batch.extend_with_cache(write_details.deref_mut(), update.block_details, CacheUpdatePolicy::Overwrite);
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut write_receipts = self.block_receipts.unwrapped_write();
|
let mut write_receipts = self.block_receipts.write();
|
||||||
batch.extend_with_cache(write_receipts.deref_mut(), update.block_receipts, CacheUpdatePolicy::Remove);
|
batch.extend_with_cache(write_receipts.deref_mut(), update.block_receipts, CacheUpdatePolicy::Remove);
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut write_blocks_blooms = self.blocks_blooms.unwrapped_write();
|
let mut write_blocks_blooms = self.blocks_blooms.write();
|
||||||
batch.extend_with_cache(write_blocks_blooms.deref_mut(), update.blocks_blooms, CacheUpdatePolicy::Remove);
|
batch.extend_with_cache(write_blocks_blooms.deref_mut(), update.blocks_blooms, CacheUpdatePolicy::Remove);
|
||||||
}
|
}
|
||||||
|
|
||||||
// These cached values must be updated last and togeterh
|
// These cached values must be updated last and togeterh
|
||||||
{
|
{
|
||||||
let mut best_block = self.best_block.unwrapped_write();
|
let mut best_block = self.best_block.write();
|
||||||
let mut write_hashes = self.block_hashes.unwrapped_write();
|
let mut write_hashes = self.block_hashes.write();
|
||||||
let mut write_txs = self.transaction_addresses.unwrapped_write();
|
let mut write_txs = self.transaction_addresses.write();
|
||||||
|
|
||||||
// update best block
|
// update best block
|
||||||
match update.info.location {
|
match update.info.location {
|
||||||
@ -728,33 +728,33 @@ impl BlockChain {
|
|||||||
|
|
||||||
/// Get best block hash.
|
/// Get best block hash.
|
||||||
pub fn best_block_hash(&self) -> H256 {
|
pub fn best_block_hash(&self) -> H256 {
|
||||||
self.best_block.unwrapped_read().hash.clone()
|
self.best_block.read().hash.clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get best block number.
|
/// Get best block number.
|
||||||
pub fn best_block_number(&self) -> BlockNumber {
|
pub fn best_block_number(&self) -> BlockNumber {
|
||||||
self.best_block.unwrapped_read().number
|
self.best_block.read().number
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get best block total difficulty.
|
/// Get best block total difficulty.
|
||||||
pub fn best_block_total_difficulty(&self) -> U256 {
|
pub fn best_block_total_difficulty(&self) -> U256 {
|
||||||
self.best_block.unwrapped_read().total_difficulty
|
self.best_block.read().total_difficulty
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get current cache size.
|
/// Get current cache size.
|
||||||
pub fn cache_size(&self) -> CacheSize {
|
pub fn cache_size(&self) -> CacheSize {
|
||||||
CacheSize {
|
CacheSize {
|
||||||
blocks: self.blocks.unwrapped_read().heap_size_of_children(),
|
blocks: self.blocks.read().heap_size_of_children(),
|
||||||
block_details: self.block_details.unwrapped_read().heap_size_of_children(),
|
block_details: self.block_details.read().heap_size_of_children(),
|
||||||
transaction_addresses: self.transaction_addresses.unwrapped_read().heap_size_of_children(),
|
transaction_addresses: self.transaction_addresses.read().heap_size_of_children(),
|
||||||
blocks_blooms: self.blocks_blooms.unwrapped_read().heap_size_of_children(),
|
blocks_blooms: self.blocks_blooms.read().heap_size_of_children(),
|
||||||
block_receipts: self.block_receipts.unwrapped_read().heap_size_of_children(),
|
block_receipts: self.block_receipts.read().heap_size_of_children(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Let the cache system know that a cacheable item has been used.
|
/// Let the cache system know that a cacheable item has been used.
|
||||||
fn note_used(&self, id: CacheID) {
|
fn note_used(&self, id: CacheID) {
|
||||||
let mut cache_man = self.cache_man.unwrapped_write();
|
let mut cache_man = self.cache_man.write();
|
||||||
if !cache_man.cache_usage[0].contains(&id) {
|
if !cache_man.cache_usage[0].contains(&id) {
|
||||||
cache_man.cache_usage[0].insert(id.clone());
|
cache_man.cache_usage[0].insert(id.clone());
|
||||||
if cache_man.in_use.contains(&id) {
|
if cache_man.in_use.contains(&id) {
|
||||||
@ -773,13 +773,13 @@ impl BlockChain {
|
|||||||
|
|
||||||
for _ in 0..COLLECTION_QUEUE_SIZE {
|
for _ in 0..COLLECTION_QUEUE_SIZE {
|
||||||
{
|
{
|
||||||
let mut blocks = self.blocks.unwrapped_write();
|
let mut blocks = self.blocks.write();
|
||||||
let mut block_details = self.block_details.unwrapped_write();
|
let mut block_details = self.block_details.write();
|
||||||
let mut block_hashes = self.block_hashes.unwrapped_write();
|
let mut block_hashes = self.block_hashes.write();
|
||||||
let mut transaction_addresses = self.transaction_addresses.unwrapped_write();
|
let mut transaction_addresses = self.transaction_addresses.write();
|
||||||
let mut blocks_blooms = self.blocks_blooms.unwrapped_write();
|
let mut blocks_blooms = self.blocks_blooms.write();
|
||||||
let mut block_receipts = self.block_receipts.unwrapped_write();
|
let mut block_receipts = self.block_receipts.write();
|
||||||
let mut cache_man = self.cache_man.unwrapped_write();
|
let mut cache_man = self.cache_man.write();
|
||||||
|
|
||||||
for id in cache_man.cache_usage.pop_back().unwrap().into_iter() {
|
for id in cache_man.cache_usage.pop_back().unwrap().into_iter() {
|
||||||
cache_man.in_use.remove(&id);
|
cache_man.in_use.remove(&id);
|
||||||
|
@ -18,7 +18,7 @@ use std::collections::{HashSet, HashMap};
|
|||||||
use std::ops::Deref;
|
use std::ops::Deref;
|
||||||
use std::mem;
|
use std::mem;
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use std::sync::*;
|
use std::sync::{Arc, Weak};
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use std::fmt;
|
use std::fmt;
|
||||||
use std::sync::atomic::{AtomicUsize, AtomicBool, Ordering as AtomicOrdering};
|
use std::sync::atomic::{AtomicUsize, AtomicBool, Ordering as AtomicOrdering};
|
||||||
@ -30,12 +30,13 @@ use util::panics::*;
|
|||||||
use util::io::*;
|
use util::io::*;
|
||||||
use util::rlp;
|
use util::rlp;
|
||||||
use util::sha3::*;
|
use util::sha3::*;
|
||||||
use util::{Bytes, Lockable, RwLockable};
|
use util::Bytes;
|
||||||
use util::rlp::{RlpStream, Rlp, UntrustedRlp};
|
use util::rlp::{RlpStream, Rlp, UntrustedRlp};
|
||||||
use util::journaldb;
|
use util::journaldb;
|
||||||
use util::journaldb::JournalDB;
|
use util::journaldb::JournalDB;
|
||||||
use util::kvdb::*;
|
use util::kvdb::*;
|
||||||
use util::{Applyable, Stream, View, PerfTimer, Itertools, Colour};
|
use util::{Applyable, Stream, View, PerfTimer, Itertools, Colour};
|
||||||
|
use util::{Mutex, RwLock};
|
||||||
|
|
||||||
// other
|
// other
|
||||||
use views::BlockView;
|
use views::BlockView;
|
||||||
@ -236,12 +237,12 @@ impl Client {
|
|||||||
|
|
||||||
/// Sets the actor to be notified on certain events
|
/// Sets the actor to be notified on certain events
|
||||||
pub fn set_notify(&self, target: &Arc<ChainNotify>) {
|
pub fn set_notify(&self, target: &Arc<ChainNotify>) {
|
||||||
let mut write_lock = self.notify.unwrapped_write();
|
let mut write_lock = self.notify.write();
|
||||||
*write_lock = Some(Arc::downgrade(target));
|
*write_lock = Some(Arc::downgrade(target));
|
||||||
}
|
}
|
||||||
|
|
||||||
fn notify(&self) -> Option<Arc<ChainNotify>> {
|
fn notify(&self) -> Option<Arc<ChainNotify>> {
|
||||||
let read_lock = self.notify.unwrapped_read();
|
let read_lock = self.notify.read();
|
||||||
read_lock.as_ref().and_then(|weak| weak.upgrade())
|
read_lock.as_ref().and_then(|weak| weak.upgrade())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -293,7 +294,7 @@ impl Client {
|
|||||||
// Enact Verified Block
|
// Enact Verified Block
|
||||||
let parent = chain_has_parent.unwrap();
|
let parent = chain_has_parent.unwrap();
|
||||||
let last_hashes = self.build_last_hashes(header.parent_hash.clone());
|
let last_hashes = self.build_last_hashes(header.parent_hash.clone());
|
||||||
let db = self.state_db.locked().boxed_clone();
|
let db = self.state_db.lock().boxed_clone();
|
||||||
|
|
||||||
let enact_result = enact_verified(&block, engine, self.tracedb.tracing_enabled(), db, &parent, last_hashes, &self.vm_factory, self.trie_factory.clone());
|
let enact_result = enact_verified(&block, engine, self.tracedb.tracing_enabled(), db, &parent, last_hashes, &self.vm_factory, self.trie_factory.clone());
|
||||||
if let Err(e) = enact_result {
|
if let Err(e) = enact_result {
|
||||||
@ -369,7 +370,7 @@ impl Client {
|
|||||||
let route = self.commit_block(closed_block, &header.hash(), &block.bytes);
|
let route = self.commit_block(closed_block, &header.hash(), &block.bytes);
|
||||||
import_results.push(route);
|
import_results.push(route);
|
||||||
|
|
||||||
self.report.unwrapped_write().accrue_block(&block);
|
self.report.write().accrue_block(&block);
|
||||||
trace!(target: "client", "Imported #{} ({})", header.number(), header.hash());
|
trace!(target: "client", "Imported #{} ({})", header.number(), header.hash());
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -471,7 +472,7 @@ impl Client {
|
|||||||
};
|
};
|
||||||
|
|
||||||
self.block_header(id).and_then(|header| {
|
self.block_header(id).and_then(|header| {
|
||||||
let db = self.state_db.locked().boxed_clone();
|
let db = self.state_db.lock().boxed_clone();
|
||||||
|
|
||||||
// early exit for pruned blocks
|
// early exit for pruned blocks
|
||||||
if db.is_pruned() && self.chain.best_block_number() >= block_number + HISTORY {
|
if db.is_pruned() && self.chain.best_block_number() >= block_number + HISTORY {
|
||||||
@ -487,7 +488,7 @@ impl Client {
|
|||||||
/// Get a copy of the best block's state.
|
/// Get a copy of the best block's state.
|
||||||
pub fn state(&self) -> State {
|
pub fn state(&self) -> State {
|
||||||
State::from_existing(
|
State::from_existing(
|
||||||
self.state_db.locked().boxed_clone(),
|
self.state_db.lock().boxed_clone(),
|
||||||
HeaderView::new(&self.best_block_header()).state_root(),
|
HeaderView::new(&self.best_block_header()).state_root(),
|
||||||
self.engine.account_start_nonce(),
|
self.engine.account_start_nonce(),
|
||||||
self.trie_factory.clone())
|
self.trie_factory.clone())
|
||||||
@ -501,8 +502,8 @@ impl Client {
|
|||||||
|
|
||||||
/// Get the report.
|
/// Get the report.
|
||||||
pub fn report(&self) -> ClientReport {
|
pub fn report(&self) -> ClientReport {
|
||||||
let mut report = self.report.unwrapped_read().clone();
|
let mut report = self.report.read().clone();
|
||||||
report.state_db_mem = self.state_db.locked().mem_used();
|
report.state_db_mem = self.state_db.lock().mem_used();
|
||||||
report
|
report
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -514,7 +515,7 @@ impl Client {
|
|||||||
|
|
||||||
match self.mode {
|
match self.mode {
|
||||||
Mode::Dark(timeout) => {
|
Mode::Dark(timeout) => {
|
||||||
let mut ss = self.sleep_state.locked();
|
let mut ss = self.sleep_state.lock();
|
||||||
if let Some(t) = ss.last_activity {
|
if let Some(t) = ss.last_activity {
|
||||||
if Instant::now() > t + timeout {
|
if Instant::now() > t + timeout {
|
||||||
self.sleep();
|
self.sleep();
|
||||||
@ -523,7 +524,7 @@ impl Client {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
Mode::Passive(timeout, wakeup_after) => {
|
Mode::Passive(timeout, wakeup_after) => {
|
||||||
let mut ss = self.sleep_state.locked();
|
let mut ss = self.sleep_state.lock();
|
||||||
let now = Instant::now();
|
let now = Instant::now();
|
||||||
if let Some(t) = ss.last_activity {
|
if let Some(t) = ss.last_activity {
|
||||||
if now > t + timeout {
|
if now > t + timeout {
|
||||||
@ -600,14 +601,14 @@ impl Client {
|
|||||||
} else {
|
} else {
|
||||||
trace!(target: "mode", "sleep: Cannot sleep - syncing ongoing.");
|
trace!(target: "mode", "sleep: Cannot sleep - syncing ongoing.");
|
||||||
// TODO: Consider uncommenting.
|
// TODO: Consider uncommenting.
|
||||||
//*self.last_activity.locked() = Some(Instant::now());
|
//*self.last_activity.lock() = Some(Instant::now());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Notify us that the network has been started.
|
/// Notify us that the network has been started.
|
||||||
pub fn network_started(&self, url: &str) {
|
pub fn network_started(&self, url: &str) {
|
||||||
let mut previous_enode = self.previous_enode.locked();
|
let mut previous_enode = self.previous_enode.lock();
|
||||||
if let Some(ref u) = *previous_enode {
|
if let Some(ref u) = *previous_enode {
|
||||||
if u == url {
|
if u == url {
|
||||||
return;
|
return;
|
||||||
@ -661,7 +662,7 @@ impl BlockChainClient for Client {
|
|||||||
fn keep_alive(&self) {
|
fn keep_alive(&self) {
|
||||||
if self.mode != Mode::Active {
|
if self.mode != Mode::Active {
|
||||||
self.wake_up();
|
self.wake_up();
|
||||||
(*self.sleep_state.locked()).last_activity = Some(Instant::now());
|
(*self.sleep_state.lock()).last_activity = Some(Instant::now());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -785,7 +786,7 @@ impl BlockChainClient for Client {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn state_data(&self, hash: &H256) -> Option<Bytes> {
|
fn state_data(&self, hash: &H256) -> Option<Bytes> {
|
||||||
self.state_db.locked().state(hash)
|
self.state_db.lock().state(hash)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn block_receipts(&self, hash: &H256) -> Option<Bytes> {
|
fn block_receipts(&self, hash: &H256) -> Option<Bytes> {
|
||||||
@ -946,7 +947,7 @@ impl MiningBlockChainClient for Client {
|
|||||||
&self.vm_factory,
|
&self.vm_factory,
|
||||||
self.trie_factory.clone(),
|
self.trie_factory.clone(),
|
||||||
false, // TODO: this will need to be parameterised once we want to do immediate mining insertion.
|
false, // TODO: this will need to be parameterised once we want to do immediate mining insertion.
|
||||||
self.state_db.locked().boxed_clone(),
|
self.state_db.lock().boxed_clone(),
|
||||||
&self.chain.block_header(&h).expect("h is best block hash: so it's header must exist: qed"),
|
&self.chain.block_header(&h).expect("h is best block hash: so it's header must exist: qed"),
|
||||||
self.build_last_hashes(h.clone()),
|
self.build_last_hashes(h.clone()),
|
||||||
author,
|
author,
|
||||||
|
@ -115,38 +115,38 @@ impl TestBlockChainClient {
|
|||||||
vm_factory: EvmFactory::new(VMType::Interpreter),
|
vm_factory: EvmFactory::new(VMType::Interpreter),
|
||||||
};
|
};
|
||||||
client.add_blocks(1, EachBlockWith::Nothing); // add genesis block
|
client.add_blocks(1, EachBlockWith::Nothing); // add genesis block
|
||||||
client.genesis_hash = client.last_hash.unwrapped_read().clone();
|
client.genesis_hash = client.last_hash.read().clone();
|
||||||
client
|
client
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the transaction receipt result
|
/// Set the transaction receipt result
|
||||||
pub fn set_transaction_receipt(&self, id: TransactionID, receipt: LocalizedReceipt) {
|
pub fn set_transaction_receipt(&self, id: TransactionID, receipt: LocalizedReceipt) {
|
||||||
self.receipts.unwrapped_write().insert(id, receipt);
|
self.receipts.write().insert(id, receipt);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the execution result.
|
/// Set the execution result.
|
||||||
pub fn set_execution_result(&self, result: Executed) {
|
pub fn set_execution_result(&self, result: Executed) {
|
||||||
*self.execution_result.unwrapped_write() = Some(result);
|
*self.execution_result.write() = Some(result);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the balance of account `address` to `balance`.
|
/// Set the balance of account `address` to `balance`.
|
||||||
pub fn set_balance(&self, address: Address, balance: U256) {
|
pub fn set_balance(&self, address: Address, balance: U256) {
|
||||||
self.balances.unwrapped_write().insert(address, balance);
|
self.balances.write().insert(address, balance);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set nonce of account `address` to `nonce`.
|
/// Set nonce of account `address` to `nonce`.
|
||||||
pub fn set_nonce(&self, address: Address, nonce: U256) {
|
pub fn set_nonce(&self, address: Address, nonce: U256) {
|
||||||
self.nonces.unwrapped_write().insert(address, nonce);
|
self.nonces.write().insert(address, nonce);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set `code` at `address`.
|
/// Set `code` at `address`.
|
||||||
pub fn set_code(&self, address: Address, code: Bytes) {
|
pub fn set_code(&self, address: Address, code: Bytes) {
|
||||||
self.code.unwrapped_write().insert(address, code);
|
self.code.write().insert(address, code);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set storage `position` to `value` for account `address`.
|
/// Set storage `position` to `value` for account `address`.
|
||||||
pub fn set_storage(&self, address: Address, position: H256, value: H256) {
|
pub fn set_storage(&self, address: Address, position: H256, value: H256) {
|
||||||
self.storage.unwrapped_write().insert((address, position), value);
|
self.storage.write().insert((address, position), value);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set block queue size for testing
|
/// Set block queue size for testing
|
||||||
@ -156,11 +156,11 @@ impl TestBlockChainClient {
|
|||||||
|
|
||||||
/// Add blocks to test client.
|
/// Add blocks to test client.
|
||||||
pub fn add_blocks(&self, count: usize, with: EachBlockWith) {
|
pub fn add_blocks(&self, count: usize, with: EachBlockWith) {
|
||||||
let len = self.numbers.unwrapped_read().len();
|
let len = self.numbers.read().len();
|
||||||
for n in len..(len + count) {
|
for n in len..(len + count) {
|
||||||
let mut header = BlockHeader::new();
|
let mut header = BlockHeader::new();
|
||||||
header.difficulty = From::from(n);
|
header.difficulty = From::from(n);
|
||||||
header.parent_hash = self.last_hash.unwrapped_read().clone();
|
header.parent_hash = self.last_hash.read().clone();
|
||||||
header.number = n as BlockNumber;
|
header.number = n as BlockNumber;
|
||||||
header.gas_limit = U256::from(1_000_000);
|
header.gas_limit = U256::from(1_000_000);
|
||||||
let uncles = match with {
|
let uncles = match with {
|
||||||
@ -168,7 +168,7 @@ impl TestBlockChainClient {
|
|||||||
let mut uncles = RlpStream::new_list(1);
|
let mut uncles = RlpStream::new_list(1);
|
||||||
let mut uncle_header = BlockHeader::new();
|
let mut uncle_header = BlockHeader::new();
|
||||||
uncle_header.difficulty = From::from(n);
|
uncle_header.difficulty = From::from(n);
|
||||||
uncle_header.parent_hash = self.last_hash.unwrapped_read().clone();
|
uncle_header.parent_hash = self.last_hash.read().clone();
|
||||||
uncle_header.number = n as BlockNumber;
|
uncle_header.number = n as BlockNumber;
|
||||||
uncles.append(&uncle_header);
|
uncles.append(&uncle_header);
|
||||||
header.uncles_hash = uncles.as_raw().sha3();
|
header.uncles_hash = uncles.as_raw().sha3();
|
||||||
@ -181,7 +181,7 @@ impl TestBlockChainClient {
|
|||||||
let mut txs = RlpStream::new_list(1);
|
let mut txs = RlpStream::new_list(1);
|
||||||
let keypair = KeyPair::create().unwrap();
|
let keypair = KeyPair::create().unwrap();
|
||||||
// Update nonces value
|
// Update nonces value
|
||||||
self.nonces.unwrapped_write().insert(keypair.address(), U256::one());
|
self.nonces.write().insert(keypair.address(), U256::one());
|
||||||
let tx = Transaction {
|
let tx = Transaction {
|
||||||
action: Action::Create,
|
action: Action::Create,
|
||||||
value: U256::from(100),
|
value: U256::from(100),
|
||||||
@ -214,7 +214,7 @@ impl TestBlockChainClient {
|
|||||||
rlp.append(&header);
|
rlp.append(&header);
|
||||||
rlp.append_raw(&rlp::NULL_RLP, 1);
|
rlp.append_raw(&rlp::NULL_RLP, 1);
|
||||||
rlp.append_raw(&rlp::NULL_RLP, 1);
|
rlp.append_raw(&rlp::NULL_RLP, 1);
|
||||||
self.blocks.unwrapped_write().insert(hash, rlp.out());
|
self.blocks.write().insert(hash, rlp.out());
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Make a bad block by setting invalid parent hash.
|
/// Make a bad block by setting invalid parent hash.
|
||||||
@ -226,12 +226,12 @@ impl TestBlockChainClient {
|
|||||||
rlp.append(&header);
|
rlp.append(&header);
|
||||||
rlp.append_raw(&rlp::NULL_RLP, 1);
|
rlp.append_raw(&rlp::NULL_RLP, 1);
|
||||||
rlp.append_raw(&rlp::NULL_RLP, 1);
|
rlp.append_raw(&rlp::NULL_RLP, 1);
|
||||||
self.blocks.unwrapped_write().insert(hash, rlp.out());
|
self.blocks.write().insert(hash, rlp.out());
|
||||||
}
|
}
|
||||||
|
|
||||||
/// TODO:
|
/// TODO:
|
||||||
pub fn block_hash_delta_minus(&mut self, delta: usize) -> H256 {
|
pub fn block_hash_delta_minus(&mut self, delta: usize) -> H256 {
|
||||||
let blocks_read = self.numbers.unwrapped_read();
|
let blocks_read = self.numbers.read();
|
||||||
let index = blocks_read.len() - delta;
|
let index = blocks_read.len() - delta;
|
||||||
blocks_read[&index].clone()
|
blocks_read[&index].clone()
|
||||||
}
|
}
|
||||||
@ -239,9 +239,9 @@ impl TestBlockChainClient {
|
|||||||
fn block_hash(&self, id: BlockID) -> Option<H256> {
|
fn block_hash(&self, id: BlockID) -> Option<H256> {
|
||||||
match id {
|
match id {
|
||||||
BlockID::Hash(hash) => Some(hash),
|
BlockID::Hash(hash) => Some(hash),
|
||||||
BlockID::Number(n) => self.numbers.unwrapped_read().get(&(n as usize)).cloned(),
|
BlockID::Number(n) => self.numbers.read().get(&(n as usize)).cloned(),
|
||||||
BlockID::Earliest => self.numbers.unwrapped_read().get(&0).cloned(),
|
BlockID::Earliest => self.numbers.read().get(&0).cloned(),
|
||||||
BlockID::Latest => self.numbers.unwrapped_read().get(&(self.numbers.unwrapped_read().len() - 1)).cloned()
|
BlockID::Latest => self.numbers.read().get(&(self.numbers.read().len() - 1)).cloned()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -288,7 +288,7 @@ impl MiningBlockChainClient for TestBlockChainClient {
|
|||||||
|
|
||||||
impl BlockChainClient for TestBlockChainClient {
|
impl BlockChainClient for TestBlockChainClient {
|
||||||
fn call(&self, _t: &SignedTransaction, _analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
fn call(&self, _t: &SignedTransaction, _analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
||||||
Ok(self.execution_result.unwrapped_read().clone().unwrap())
|
Ok(self.execution_result.read().clone().unwrap())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn block_total_difficulty(&self, _id: BlockID) -> Option<U256> {
|
fn block_total_difficulty(&self, _id: BlockID) -> Option<U256> {
|
||||||
@ -301,7 +301,7 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
|
|
||||||
fn nonce(&self, address: &Address, id: BlockID) -> Option<U256> {
|
fn nonce(&self, address: &Address, id: BlockID) -> Option<U256> {
|
||||||
match id {
|
match id {
|
||||||
BlockID::Latest => Some(self.nonces.unwrapped_read().get(address).cloned().unwrap_or_else(U256::zero)),
|
BlockID::Latest => Some(self.nonces.read().get(address).cloned().unwrap_or_else(U256::zero)),
|
||||||
_ => None,
|
_ => None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -311,12 +311,12 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn code(&self, address: &Address) -> Option<Bytes> {
|
fn code(&self, address: &Address) -> Option<Bytes> {
|
||||||
self.code.unwrapped_read().get(address).cloned()
|
self.code.read().get(address).cloned()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn balance(&self, address: &Address, id: BlockID) -> Option<U256> {
|
fn balance(&self, address: &Address, id: BlockID) -> Option<U256> {
|
||||||
if let BlockID::Latest = id {
|
if let BlockID::Latest = id {
|
||||||
Some(self.balances.unwrapped_read().get(address).cloned().unwrap_or_else(U256::zero))
|
Some(self.balances.read().get(address).cloned().unwrap_or_else(U256::zero))
|
||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
@ -328,7 +328,7 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
|
|
||||||
fn storage_at(&self, address: &Address, position: &H256, id: BlockID) -> Option<H256> {
|
fn storage_at(&self, address: &Address, position: &H256, id: BlockID) -> Option<H256> {
|
||||||
if let BlockID::Latest = id {
|
if let BlockID::Latest = id {
|
||||||
Some(self.storage.unwrapped_read().get(&(address.clone(), position.clone())).cloned().unwrap_or_else(H256::new))
|
Some(self.storage.read().get(&(address.clone(), position.clone())).cloned().unwrap_or_else(H256::new))
|
||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
@ -343,7 +343,7 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn transaction_receipt(&self, id: TransactionID) -> Option<LocalizedReceipt> {
|
fn transaction_receipt(&self, id: TransactionID) -> Option<LocalizedReceipt> {
|
||||||
self.receipts.unwrapped_read().get(&id).cloned()
|
self.receipts.read().get(&id).cloned()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn blocks_with_bloom(&self, _bloom: &H2048, _from_block: BlockID, _to_block: BlockID) -> Option<Vec<BlockNumber>> {
|
fn blocks_with_bloom(&self, _bloom: &H2048, _from_block: BlockID, _to_block: BlockID) -> Option<Vec<BlockNumber>> {
|
||||||
@ -359,11 +359,11 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn block_header(&self, id: BlockID) -> Option<Bytes> {
|
fn block_header(&self, id: BlockID) -> Option<Bytes> {
|
||||||
self.block_hash(id).and_then(|hash| self.blocks.unwrapped_read().get(&hash).map(|r| Rlp::new(r).at(0).as_raw().to_vec()))
|
self.block_hash(id).and_then(|hash| self.blocks.read().get(&hash).map(|r| Rlp::new(r).at(0).as_raw().to_vec()))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn block_body(&self, id: BlockID) -> Option<Bytes> {
|
fn block_body(&self, id: BlockID) -> Option<Bytes> {
|
||||||
self.block_hash(id).and_then(|hash| self.blocks.unwrapped_read().get(&hash).map(|r| {
|
self.block_hash(id).and_then(|hash| self.blocks.read().get(&hash).map(|r| {
|
||||||
let mut stream = RlpStream::new_list(2);
|
let mut stream = RlpStream::new_list(2);
|
||||||
stream.append_raw(Rlp::new(&r).at(1).as_raw(), 1);
|
stream.append_raw(Rlp::new(&r).at(1).as_raw(), 1);
|
||||||
stream.append_raw(Rlp::new(&r).at(2).as_raw(), 1);
|
stream.append_raw(Rlp::new(&r).at(2).as_raw(), 1);
|
||||||
@ -372,13 +372,13 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn block(&self, id: BlockID) -> Option<Bytes> {
|
fn block(&self, id: BlockID) -> Option<Bytes> {
|
||||||
self.block_hash(id).and_then(|hash| self.blocks.unwrapped_read().get(&hash).cloned())
|
self.block_hash(id).and_then(|hash| self.blocks.read().get(&hash).cloned())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn block_status(&self, id: BlockID) -> BlockStatus {
|
fn block_status(&self, id: BlockID) -> BlockStatus {
|
||||||
match id {
|
match id {
|
||||||
BlockID::Number(number) if (number as usize) < self.blocks.unwrapped_read().len() => BlockStatus::InChain,
|
BlockID::Number(number) if (number as usize) < self.blocks.read().len() => BlockStatus::InChain,
|
||||||
BlockID::Hash(ref hash) if self.blocks.unwrapped_read().get(hash).is_some() => BlockStatus::InChain,
|
BlockID::Hash(ref hash) if self.blocks.read().get(hash).is_some() => BlockStatus::InChain,
|
||||||
_ => BlockStatus::Unknown
|
_ => BlockStatus::Unknown
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -389,7 +389,7 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
ancestor: H256::new(),
|
ancestor: H256::new(),
|
||||||
index: 0,
|
index: 0,
|
||||||
blocks: {
|
blocks: {
|
||||||
let numbers_read = self.numbers.unwrapped_read();
|
let numbers_read = self.numbers.read();
|
||||||
let mut adding = false;
|
let mut adding = false;
|
||||||
|
|
||||||
let mut blocks = Vec::new();
|
let mut blocks = Vec::new();
|
||||||
@ -446,11 +446,11 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
let header = Rlp::new(&b).val_at::<BlockHeader>(0);
|
let header = Rlp::new(&b).val_at::<BlockHeader>(0);
|
||||||
let h = header.hash();
|
let h = header.hash();
|
||||||
let number: usize = header.number as usize;
|
let number: usize = header.number as usize;
|
||||||
if number > self.blocks.unwrapped_read().len() {
|
if number > self.blocks.read().len() {
|
||||||
panic!("Unexpected block number. Expected {}, got {}", self.blocks.unwrapped_read().len(), number);
|
panic!("Unexpected block number. Expected {}, got {}", self.blocks.read().len(), number);
|
||||||
}
|
}
|
||||||
if number > 0 {
|
if number > 0 {
|
||||||
match self.blocks.unwrapped_read().get(&header.parent_hash) {
|
match self.blocks.read().get(&header.parent_hash) {
|
||||||
Some(parent) => {
|
Some(parent) => {
|
||||||
let parent = Rlp::new(parent).val_at::<BlockHeader>(0);
|
let parent = Rlp::new(parent).val_at::<BlockHeader>(0);
|
||||||
if parent.number != (header.number - 1) {
|
if parent.number != (header.number - 1) {
|
||||||
@ -462,27 +462,27 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
let len = self.numbers.unwrapped_read().len();
|
let len = self.numbers.read().len();
|
||||||
if number == len {
|
if number == len {
|
||||||
{
|
{
|
||||||
let mut difficulty = self.difficulty.unwrapped_write();
|
let mut difficulty = self.difficulty.write();
|
||||||
*difficulty.deref_mut() = *difficulty.deref() + header.difficulty;
|
*difficulty.deref_mut() = *difficulty.deref() + header.difficulty;
|
||||||
}
|
}
|
||||||
mem::replace(self.last_hash.unwrapped_write().deref_mut(), h.clone());
|
mem::replace(self.last_hash.write().deref_mut(), h.clone());
|
||||||
self.blocks.unwrapped_write().insert(h.clone(), b);
|
self.blocks.write().insert(h.clone(), b);
|
||||||
self.numbers.unwrapped_write().insert(number, h.clone());
|
self.numbers.write().insert(number, h.clone());
|
||||||
let mut parent_hash = header.parent_hash;
|
let mut parent_hash = header.parent_hash;
|
||||||
if number > 0 {
|
if number > 0 {
|
||||||
let mut n = number - 1;
|
let mut n = number - 1;
|
||||||
while n > 0 && self.numbers.unwrapped_read()[&n] != parent_hash {
|
while n > 0 && self.numbers.read()[&n] != parent_hash {
|
||||||
*self.numbers.unwrapped_write().get_mut(&n).unwrap() = parent_hash.clone();
|
*self.numbers.write().get_mut(&n).unwrap() = parent_hash.clone();
|
||||||
n -= 1;
|
n -= 1;
|
||||||
parent_hash = Rlp::new(&self.blocks.unwrapped_read()[&parent_hash]).val_at::<BlockHeader>(0).parent_hash;
|
parent_hash = Rlp::new(&self.blocks.read()[&parent_hash]).val_at::<BlockHeader>(0).parent_hash;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
self.blocks.unwrapped_write().insert(h.clone(), b.to_vec());
|
self.blocks.write().insert(h.clone(), b.to_vec());
|
||||||
}
|
}
|
||||||
Ok(h)
|
Ok(h)
|
||||||
}
|
}
|
||||||
@ -503,11 +503,11 @@ impl BlockChainClient for TestBlockChainClient {
|
|||||||
|
|
||||||
fn chain_info(&self) -> BlockChainInfo {
|
fn chain_info(&self) -> BlockChainInfo {
|
||||||
BlockChainInfo {
|
BlockChainInfo {
|
||||||
total_difficulty: *self.difficulty.unwrapped_read(),
|
total_difficulty: *self.difficulty.read(),
|
||||||
pending_total_difficulty: *self.difficulty.unwrapped_read(),
|
pending_total_difficulty: *self.difficulty.read(),
|
||||||
genesis_hash: self.genesis_hash.clone(),
|
genesis_hash: self.genesis_hash.clone(),
|
||||||
best_block_hash: self.last_hash.unwrapped_read().clone(),
|
best_block_hash: self.last_hash.read().clone(),
|
||||||
best_block_number: self.blocks.unwrapped_read().len() as BlockNumber - 1,
|
best_block_number: self.blocks.read().len() as BlockNumber - 1,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -18,11 +18,11 @@
|
|||||||
|
|
||||||
use std::ops::Deref;
|
use std::ops::Deref;
|
||||||
use std::hash::Hash;
|
use std::hash::Hash;
|
||||||
use std::sync::RwLock;
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use util::{DBTransaction, Database, RwLockable};
|
use util::{DBTransaction, Database, RwLock};
|
||||||
use util::rlp::{encode, Encodable, decode, Decodable};
|
use util::rlp::{encode, Encodable, decode, Decodable};
|
||||||
|
|
||||||
|
|
||||||
#[derive(Clone, Copy)]
|
#[derive(Clone, Copy)]
|
||||||
pub enum CacheUpdatePolicy {
|
pub enum CacheUpdatePolicy {
|
||||||
Overwrite,
|
Overwrite,
|
||||||
@ -115,14 +115,14 @@ pub trait Readable {
|
|||||||
T: Clone + Decodable,
|
T: Clone + Decodable,
|
||||||
C: Cache<K, T> {
|
C: Cache<K, T> {
|
||||||
{
|
{
|
||||||
let read = cache.unwrapped_read();
|
let read = cache.read();
|
||||||
if let Some(v) = read.get(key) {
|
if let Some(v) = read.get(key) {
|
||||||
return Some(v.clone());
|
return Some(v.clone());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
self.read(key).map(|value: T|{
|
self.read(key).map(|value: T|{
|
||||||
let mut write = cache.unwrapped_write();
|
let mut write = cache.write();
|
||||||
write.insert(key.clone(), value.clone());
|
write.insert(key.clone(), value.clone());
|
||||||
value
|
value
|
||||||
})
|
})
|
||||||
@ -137,7 +137,7 @@ pub trait Readable {
|
|||||||
R: Deref<Target = [u8]>,
|
R: Deref<Target = [u8]>,
|
||||||
C: Cache<K, T> {
|
C: Cache<K, T> {
|
||||||
{
|
{
|
||||||
let read = cache.unwrapped_read();
|
let read = cache.read();
|
||||||
if read.get(key).is_some() {
|
if read.get(key).is_some() {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
@ -97,7 +97,6 @@ extern crate ethash;
|
|||||||
pub extern crate ethstore;
|
pub extern crate ethstore;
|
||||||
extern crate semver;
|
extern crate semver;
|
||||||
extern crate ethcore_ipc_nano as nanoipc;
|
extern crate ethcore_ipc_nano as nanoipc;
|
||||||
|
|
||||||
extern crate ethcore_devtools as devtools;
|
extern crate ethcore_devtools as devtools;
|
||||||
|
|
||||||
#[cfg(feature = "jit" )] extern crate evmjit;
|
#[cfg(feature = "jit" )] extern crate evmjit;
|
||||||
|
@ -15,8 +15,8 @@
|
|||||||
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::sync::{Arc, RwLock};
|
use std::sync::Arc;
|
||||||
use util::{RwLockable, U256, H256};
|
use util::{RwLock, U256, H256};
|
||||||
|
|
||||||
/// External miner interface.
|
/// External miner interface.
|
||||||
pub trait ExternalMinerService: Send + Sync {
|
pub trait ExternalMinerService: Send + Sync {
|
||||||
@ -54,15 +54,15 @@ impl ExternalMiner {
|
|||||||
|
|
||||||
impl ExternalMinerService for ExternalMiner {
|
impl ExternalMinerService for ExternalMiner {
|
||||||
fn submit_hashrate(&self, hashrate: U256, id: H256) {
|
fn submit_hashrate(&self, hashrate: U256, id: H256) {
|
||||||
self.hashrates.unwrapped_write().insert(id, hashrate);
|
self.hashrates.write().insert(id, hashrate);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn hashrate(&self) -> U256 {
|
fn hashrate(&self) -> U256 {
|
||||||
self.hashrates.unwrapped_read().iter().fold(U256::from(0), |sum, (_, v)| sum + *v)
|
self.hashrates.read().iter().fold(U256::from(0), |sum, (_, v)| sum + *v)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn is_mining(&self) -> bool {
|
fn is_mining(&self) -> bool {
|
||||||
!self.hashrates.unwrapped_read().is_empty()
|
!self.hashrates.read().is_empty()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
use rayon::prelude::*;
|
use rayon::prelude::*;
|
||||||
use std::sync::atomic::AtomicBool;
|
use std::sync::atomic::{self, AtomicBool};
|
||||||
use std::time::{Instant, Duration};
|
use std::time::{Instant, Duration};
|
||||||
|
|
||||||
use util::*;
|
use util::*;
|
||||||
@ -26,7 +26,7 @@ use client::{MiningBlockChainClient, Executive, Executed, EnvInfo, TransactOptio
|
|||||||
use block::{ClosedBlock, IsBlock};
|
use block::{ClosedBlock, IsBlock};
|
||||||
use error::*;
|
use error::*;
|
||||||
use transaction::SignedTransaction;
|
use transaction::SignedTransaction;
|
||||||
use receipt::{Receipt};
|
use receipt::Receipt;
|
||||||
use spec::Spec;
|
use spec::Spec;
|
||||||
use engine::Engine;
|
use engine::Engine;
|
||||||
use miner::{MinerService, MinerStatus, TransactionQueue, AccountDetails, TransactionOrigin};
|
use miner::{MinerService, MinerStatus, TransactionQueue, AccountDetails, TransactionOrigin};
|
||||||
@ -34,7 +34,6 @@ use miner::work_notify::WorkPoster;
|
|||||||
use client::TransactionImportResult;
|
use client::TransactionImportResult;
|
||||||
use miner::price_info::PriceInfo;
|
use miner::price_info::PriceInfo;
|
||||||
|
|
||||||
|
|
||||||
/// Different possible definitions for pending transaction set.
|
/// Different possible definitions for pending transaction set.
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub enum PendingSet {
|
pub enum PendingSet {
|
||||||
@ -236,16 +235,16 @@ impl Miner {
|
|||||||
{
|
{
|
||||||
trace!(target: "miner", "recalibrating...");
|
trace!(target: "miner", "recalibrating...");
|
||||||
let txq = self.transaction_queue.clone();
|
let txq = self.transaction_queue.clone();
|
||||||
self.gas_pricer.lock().unwrap().recalibrate(move |price| {
|
self.gas_pricer.lock().recalibrate(move |price| {
|
||||||
trace!(target: "miner", "Got gas price! {}", price);
|
trace!(target: "miner", "Got gas price! {}", price);
|
||||||
txq.lock().unwrap().set_minimal_gas_price(price);
|
txq.lock().set_minimal_gas_price(price);
|
||||||
});
|
});
|
||||||
trace!(target: "miner", "done recalibration.");
|
trace!(target: "miner", "done recalibration.");
|
||||||
}
|
}
|
||||||
|
|
||||||
let (transactions, mut open_block, original_work_hash) = {
|
let (transactions, mut open_block, original_work_hash) = {
|
||||||
let transactions = {self.transaction_queue.locked().top_transactions()};
|
let transactions = {self.transaction_queue.lock().top_transactions()};
|
||||||
let mut sealing_work = self.sealing_work.locked();
|
let mut sealing_work = self.sealing_work.lock();
|
||||||
let last_work_hash = sealing_work.peek_last_ref().map(|pb| pb.block().fields().header.hash());
|
let last_work_hash = sealing_work.peek_last_ref().map(|pb| pb.block().fields().header.hash());
|
||||||
let best_hash = chain.best_block_header().sha3();
|
let best_hash = chain.best_block_header().sha3();
|
||||||
/*
|
/*
|
||||||
@ -315,7 +314,7 @@ impl Miner {
|
|||||||
};
|
};
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut queue = self.transaction_queue.locked();
|
let mut queue = self.transaction_queue.lock();
|
||||||
for hash in invalid_transactions.into_iter() {
|
for hash in invalid_transactions.into_iter() {
|
||||||
queue.remove_invalid(&hash, &fetch_account);
|
queue.remove_invalid(&hash, &fetch_account);
|
||||||
}
|
}
|
||||||
@ -346,7 +345,7 @@ impl Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let (work, is_new) = {
|
let (work, is_new) = {
|
||||||
let mut sealing_work = self.sealing_work.locked();
|
let mut sealing_work = self.sealing_work.lock();
|
||||||
let last_work_hash = sealing_work.peek_last_ref().map(|pb| pb.block().fields().header.hash());
|
let last_work_hash = sealing_work.peek_last_ref().map(|pb| pb.block().fields().header.hash());
|
||||||
trace!(target: "miner", "Checking whether we need to reseal: orig={:?} last={:?}, this={:?}", original_work_hash, last_work_hash, block.block().fields().header.hash());
|
trace!(target: "miner", "Checking whether we need to reseal: orig={:?} last={:?}, this={:?}", original_work_hash, last_work_hash, block.block().fields().header.hash());
|
||||||
let (work, is_new) = if last_work_hash.map_or(true, |h| h != block.block().fields().header.hash()) {
|
let (work, is_new) = if last_work_hash.map_or(true, |h| h != block.block().fields().header.hash()) {
|
||||||
@ -374,14 +373,14 @@ impl Miner {
|
|||||||
|
|
||||||
fn update_gas_limit(&self, chain: &MiningBlockChainClient) {
|
fn update_gas_limit(&self, chain: &MiningBlockChainClient) {
|
||||||
let gas_limit = HeaderView::new(&chain.best_block_header()).gas_limit();
|
let gas_limit = HeaderView::new(&chain.best_block_header()).gas_limit();
|
||||||
let mut queue = self.transaction_queue.locked();
|
let mut queue = self.transaction_queue.lock();
|
||||||
queue.set_gas_limit(gas_limit);
|
queue.set_gas_limit(gas_limit);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns true if we had to prepare new pending block
|
/// Returns true if we had to prepare new pending block
|
||||||
fn enable_and_prepare_sealing(&self, chain: &MiningBlockChainClient) -> bool {
|
fn enable_and_prepare_sealing(&self, chain: &MiningBlockChainClient) -> bool {
|
||||||
trace!(target: "miner", "enable_and_prepare_sealing: entering");
|
trace!(target: "miner", "enable_and_prepare_sealing: entering");
|
||||||
let have_work = self.sealing_work.locked().peek_last_ref().is_some();
|
let have_work = self.sealing_work.lock().peek_last_ref().is_some();
|
||||||
trace!(target: "miner", "enable_and_prepare_sealing: have_work={}", have_work);
|
trace!(target: "miner", "enable_and_prepare_sealing: have_work={}", have_work);
|
||||||
if !have_work {
|
if !have_work {
|
||||||
// --------------------------------------------------------------------------
|
// --------------------------------------------------------------------------
|
||||||
@ -391,7 +390,7 @@ impl Miner {
|
|||||||
self.sealing_enabled.store(true, atomic::Ordering::Relaxed);
|
self.sealing_enabled.store(true, atomic::Ordering::Relaxed);
|
||||||
self.prepare_sealing(chain);
|
self.prepare_sealing(chain);
|
||||||
}
|
}
|
||||||
let mut sealing_block_last_request = self.sealing_block_last_request.locked();
|
let mut sealing_block_last_request = self.sealing_block_last_request.lock();
|
||||||
let best_number = chain.chain_info().best_block_number;
|
let best_number = chain.chain_info().best_block_number;
|
||||||
if *sealing_block_last_request != best_number {
|
if *sealing_block_last_request != best_number {
|
||||||
trace!(target: "miner", "enable_and_prepare_sealing: Miner received request (was {}, now {}) - waking up.", *sealing_block_last_request, best_number);
|
trace!(target: "miner", "enable_and_prepare_sealing: Miner received request (was {}, now {}) - waking up.", *sealing_block_last_request, best_number);
|
||||||
@ -416,7 +415,7 @@ impl Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Are we allowed to do a non-mandatory reseal?
|
/// Are we allowed to do a non-mandatory reseal?
|
||||||
fn tx_reseal_allowed(&self) -> bool { Instant::now() > *self.next_allowed_reseal.locked() }
|
fn tx_reseal_allowed(&self) -> bool { Instant::now() > *self.next_allowed_reseal.lock() }
|
||||||
}
|
}
|
||||||
|
|
||||||
const SEALING_TIMEOUT_IN_BLOCKS : u64 = 5;
|
const SEALING_TIMEOUT_IN_BLOCKS : u64 = 5;
|
||||||
@ -424,7 +423,7 @@ const SEALING_TIMEOUT_IN_BLOCKS : u64 = 5;
|
|||||||
impl MinerService for Miner {
|
impl MinerService for Miner {
|
||||||
|
|
||||||
fn clear_and_reset(&self, chain: &MiningBlockChainClient) {
|
fn clear_and_reset(&self, chain: &MiningBlockChainClient) {
|
||||||
self.transaction_queue.locked().clear();
|
self.transaction_queue.lock().clear();
|
||||||
// --------------------------------------------------------------------------
|
// --------------------------------------------------------------------------
|
||||||
// | NOTE Code below requires transaction_queue and sealing_work locks. |
|
// | NOTE Code below requires transaction_queue and sealing_work locks. |
|
||||||
// | Make sure to release the locks before calling that method. |
|
// | Make sure to release the locks before calling that method. |
|
||||||
@ -433,8 +432,8 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn status(&self) -> MinerStatus {
|
fn status(&self) -> MinerStatus {
|
||||||
let status = self.transaction_queue.locked().status();
|
let status = self.transaction_queue.lock().status();
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
MinerStatus {
|
MinerStatus {
|
||||||
transactions_in_pending_queue: status.pending,
|
transactions_in_pending_queue: status.pending,
|
||||||
transactions_in_future_queue: status.future,
|
transactions_in_future_queue: status.future,
|
||||||
@ -443,7 +442,7 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn call(&self, chain: &MiningBlockChainClient, t: &SignedTransaction, analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
fn call(&self, chain: &MiningBlockChainClient, t: &SignedTransaction, analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
match sealing_work.peek_last_ref() {
|
match sealing_work.peek_last_ref() {
|
||||||
Some(work) => {
|
Some(work) => {
|
||||||
let block = work.block();
|
let block = work.block();
|
||||||
@ -490,7 +489,7 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn balance(&self, chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
fn balance(&self, chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
sealing_work.peek_last_ref().map_or_else(
|
sealing_work.peek_last_ref().map_or_else(
|
||||||
|| chain.latest_balance(address),
|
|| chain.latest_balance(address),
|
||||||
|b| b.block().fields().state.balance(address)
|
|b| b.block().fields().state.balance(address)
|
||||||
@ -498,7 +497,7 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn storage_at(&self, chain: &MiningBlockChainClient, address: &Address, position: &H256) -> H256 {
|
fn storage_at(&self, chain: &MiningBlockChainClient, address: &Address, position: &H256) -> H256 {
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
sealing_work.peek_last_ref().map_or_else(
|
sealing_work.peek_last_ref().map_or_else(
|
||||||
|| chain.latest_storage_at(address, position),
|
|| chain.latest_storage_at(address, position),
|
||||||
|b| b.block().fields().state.storage_at(address, position)
|
|b| b.block().fields().state.storage_at(address, position)
|
||||||
@ -506,79 +505,79 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn nonce(&self, chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
fn nonce(&self, chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
sealing_work.peek_last_ref().map_or_else(|| chain.latest_nonce(address), |b| b.block().fields().state.nonce(address))
|
sealing_work.peek_last_ref().map_or_else(|| chain.latest_nonce(address), |b| b.block().fields().state.nonce(address))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn code(&self, chain: &MiningBlockChainClient, address: &Address) -> Option<Bytes> {
|
fn code(&self, chain: &MiningBlockChainClient, address: &Address) -> Option<Bytes> {
|
||||||
let sealing_work = self.sealing_work.locked();
|
let sealing_work = self.sealing_work.lock();
|
||||||
sealing_work.peek_last_ref().map_or_else(|| chain.code(address), |b| b.block().fields().state.code(address))
|
sealing_work.peek_last_ref().map_or_else(|| chain.code(address), |b| b.block().fields().state.code(address))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_author(&self, author: Address) {
|
fn set_author(&self, author: Address) {
|
||||||
*self.author.unwrapped_write() = author;
|
*self.author.write() = author;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_extra_data(&self, extra_data: Bytes) {
|
fn set_extra_data(&self, extra_data: Bytes) {
|
||||||
*self.extra_data.unwrapped_write() = extra_data;
|
*self.extra_data.write() = extra_data;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the gas limit we wish to target when sealing a new block.
|
/// Set the gas limit we wish to target when sealing a new block.
|
||||||
fn set_gas_floor_target(&self, target: U256) {
|
fn set_gas_floor_target(&self, target: U256) {
|
||||||
self.gas_range_target.unwrapped_write().0 = target;
|
self.gas_range_target.write().0 = target;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_gas_ceil_target(&self, target: U256) {
|
fn set_gas_ceil_target(&self, target: U256) {
|
||||||
self.gas_range_target.unwrapped_write().1 = target;
|
self.gas_range_target.write().1 = target;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_minimal_gas_price(&self, min_gas_price: U256) {
|
fn set_minimal_gas_price(&self, min_gas_price: U256) {
|
||||||
self.transaction_queue.locked().set_minimal_gas_price(min_gas_price);
|
self.transaction_queue.lock().set_minimal_gas_price(min_gas_price);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn minimal_gas_price(&self) -> U256 {
|
fn minimal_gas_price(&self) -> U256 {
|
||||||
*self.transaction_queue.locked().minimal_gas_price()
|
*self.transaction_queue.lock().minimal_gas_price()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn sensible_gas_price(&self) -> U256 {
|
fn sensible_gas_price(&self) -> U256 {
|
||||||
// 10% above our minimum.
|
// 10% above our minimum.
|
||||||
*self.transaction_queue.locked().minimal_gas_price() * 110.into() / 100.into()
|
*self.transaction_queue.lock().minimal_gas_price() * 110.into() / 100.into()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn sensible_gas_limit(&self) -> U256 {
|
fn sensible_gas_limit(&self) -> U256 {
|
||||||
self.gas_range_target.unwrapped_read().0 / 5.into()
|
self.gas_range_target.read().0 / 5.into()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn transactions_limit(&self) -> usize {
|
fn transactions_limit(&self) -> usize {
|
||||||
self.transaction_queue.locked().limit()
|
self.transaction_queue.lock().limit()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_transactions_limit(&self, limit: usize) {
|
fn set_transactions_limit(&self, limit: usize) {
|
||||||
self.transaction_queue.locked().set_limit(limit)
|
self.transaction_queue.lock().set_limit(limit)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_tx_gas_limit(&self, limit: U256) {
|
fn set_tx_gas_limit(&self, limit: U256) {
|
||||||
self.transaction_queue.locked().set_tx_gas_limit(limit)
|
self.transaction_queue.lock().set_tx_gas_limit(limit)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the author that we will seal blocks as.
|
/// Get the author that we will seal blocks as.
|
||||||
fn author(&self) -> Address {
|
fn author(&self) -> Address {
|
||||||
*self.author.unwrapped_read()
|
*self.author.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the extra_data that we will seal blocks with.
|
/// Get the extra_data that we will seal blocks with.
|
||||||
fn extra_data(&self) -> Bytes {
|
fn extra_data(&self) -> Bytes {
|
||||||
self.extra_data.unwrapped_read().clone()
|
self.extra_data.read().clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the gas limit we wish to target when sealing a new block.
|
/// Get the gas limit we wish to target when sealing a new block.
|
||||||
fn gas_floor_target(&self) -> U256 {
|
fn gas_floor_target(&self) -> U256 {
|
||||||
self.gas_range_target.unwrapped_read().0
|
self.gas_range_target.read().0
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the gas limit we wish to target when sealing a new block.
|
/// Get the gas limit we wish to target when sealing a new block.
|
||||||
fn gas_ceil_target(&self) -> U256 {
|
fn gas_ceil_target(&self) -> U256 {
|
||||||
self.gas_range_target.unwrapped_read().1
|
self.gas_range_target.read().1
|
||||||
}
|
}
|
||||||
|
|
||||||
fn import_external_transactions(
|
fn import_external_transactions(
|
||||||
@ -588,7 +587,7 @@ impl MinerService for Miner {
|
|||||||
) -> Vec<Result<TransactionImportResult, Error>> {
|
) -> Vec<Result<TransactionImportResult, Error>> {
|
||||||
|
|
||||||
let results = {
|
let results = {
|
||||||
let mut transaction_queue = self.transaction_queue.locked();
|
let mut transaction_queue = self.transaction_queue.lock();
|
||||||
self.add_transactions_to_queue(
|
self.add_transactions_to_queue(
|
||||||
chain, transactions, TransactionOrigin::External, &mut transaction_queue
|
chain, transactions, TransactionOrigin::External, &mut transaction_queue
|
||||||
)
|
)
|
||||||
@ -615,7 +614,7 @@ impl MinerService for Miner {
|
|||||||
|
|
||||||
let imported = {
|
let imported = {
|
||||||
// Be sure to release the lock before we call enable_and_prepare_sealing
|
// Be sure to release the lock before we call enable_and_prepare_sealing
|
||||||
let mut transaction_queue = self.transaction_queue.locked();
|
let mut transaction_queue = self.transaction_queue.lock();
|
||||||
let import = self.add_transactions_to_queue(chain, vec![transaction], TransactionOrigin::Local, &mut transaction_queue).pop().unwrap();
|
let import = self.add_transactions_to_queue(chain, vec![transaction], TransactionOrigin::Local, &mut transaction_queue).pop().unwrap();
|
||||||
|
|
||||||
match import {
|
match import {
|
||||||
@ -651,13 +650,13 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn all_transactions(&self) -> Vec<SignedTransaction> {
|
fn all_transactions(&self) -> Vec<SignedTransaction> {
|
||||||
let queue = self.transaction_queue.locked();
|
let queue = self.transaction_queue.lock();
|
||||||
queue.top_transactions()
|
queue.top_transactions()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn pending_transactions(&self) -> Vec<SignedTransaction> {
|
fn pending_transactions(&self) -> Vec<SignedTransaction> {
|
||||||
let queue = self.transaction_queue.locked();
|
let queue = self.transaction_queue.lock();
|
||||||
let sw = self.sealing_work.locked();
|
let sw = self.sealing_work.lock();
|
||||||
// TODO: should only use the sealing_work when it's current (it could be an old block)
|
// TODO: should only use the sealing_work when it's current (it could be an old block)
|
||||||
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
||||||
true => sw.peek_last_ref(),
|
true => sw.peek_last_ref(),
|
||||||
@ -670,8 +669,8 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn pending_transactions_hashes(&self) -> Vec<H256> {
|
fn pending_transactions_hashes(&self) -> Vec<H256> {
|
||||||
let queue = self.transaction_queue.locked();
|
let queue = self.transaction_queue.lock();
|
||||||
let sw = self.sealing_work.locked();
|
let sw = self.sealing_work.lock();
|
||||||
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
||||||
true => sw.peek_last_ref(),
|
true => sw.peek_last_ref(),
|
||||||
false => None,
|
false => None,
|
||||||
@ -683,8 +682,8 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn transaction(&self, hash: &H256) -> Option<SignedTransaction> {
|
fn transaction(&self, hash: &H256) -> Option<SignedTransaction> {
|
||||||
let queue = self.transaction_queue.locked();
|
let queue = self.transaction_queue.lock();
|
||||||
let sw = self.sealing_work.locked();
|
let sw = self.sealing_work.lock();
|
||||||
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
let sealing_set = match self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
||||||
true => sw.peek_last_ref(),
|
true => sw.peek_last_ref(),
|
||||||
false => None,
|
false => None,
|
||||||
@ -696,7 +695,7 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn pending_receipts(&self) -> BTreeMap<H256, Receipt> {
|
fn pending_receipts(&self) -> BTreeMap<H256, Receipt> {
|
||||||
match (self.sealing_enabled.load(atomic::Ordering::Relaxed), self.sealing_work.locked().peek_last_ref()) {
|
match (self.sealing_enabled.load(atomic::Ordering::Relaxed), self.sealing_work.lock().peek_last_ref()) {
|
||||||
(true, Some(pending)) => {
|
(true, Some(pending)) => {
|
||||||
let hashes = pending.transactions()
|
let hashes = pending.transactions()
|
||||||
.iter()
|
.iter()
|
||||||
@ -711,14 +710,14 @@ impl MinerService for Miner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn last_nonce(&self, address: &Address) -> Option<U256> {
|
fn last_nonce(&self, address: &Address) -> Option<U256> {
|
||||||
self.transaction_queue.locked().last_nonce(address)
|
self.transaction_queue.lock().last_nonce(address)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn update_sealing(&self, chain: &MiningBlockChainClient) {
|
fn update_sealing(&self, chain: &MiningBlockChainClient) {
|
||||||
if self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
if self.sealing_enabled.load(atomic::Ordering::Relaxed) {
|
||||||
let current_no = chain.chain_info().best_block_number;
|
let current_no = chain.chain_info().best_block_number;
|
||||||
let has_local_transactions = self.transaction_queue.locked().has_local_pending_transactions();
|
let has_local_transactions = self.transaction_queue.lock().has_local_pending_transactions();
|
||||||
let last_request = *self.sealing_block_last_request.locked();
|
let last_request = *self.sealing_block_last_request.lock();
|
||||||
let should_disable_sealing = !self.forced_sealing()
|
let should_disable_sealing = !self.forced_sealing()
|
||||||
&& !has_local_transactions
|
&& !has_local_transactions
|
||||||
&& current_no > last_request
|
&& current_no > last_request
|
||||||
@ -727,9 +726,9 @@ impl MinerService for Miner {
|
|||||||
if should_disable_sealing {
|
if should_disable_sealing {
|
||||||
trace!(target: "miner", "Miner sleeping (current {}, last {})", current_no, last_request);
|
trace!(target: "miner", "Miner sleeping (current {}, last {})", current_no, last_request);
|
||||||
self.sealing_enabled.store(false, atomic::Ordering::Relaxed);
|
self.sealing_enabled.store(false, atomic::Ordering::Relaxed);
|
||||||
self.sealing_work.locked().reset();
|
self.sealing_work.lock().reset();
|
||||||
} else {
|
} else {
|
||||||
*self.next_allowed_reseal.locked() = Instant::now() + self.options.reseal_min_period;
|
*self.next_allowed_reseal.lock() = Instant::now() + self.options.reseal_min_period;
|
||||||
// --------------------------------------------------------------------------
|
// --------------------------------------------------------------------------
|
||||||
// | NOTE Code below requires transaction_queue and sealing_work locks. |
|
// | NOTE Code below requires transaction_queue and sealing_work locks. |
|
||||||
// | Make sure to release the locks before calling that method. |
|
// | Make sure to release the locks before calling that method. |
|
||||||
@ -743,14 +742,14 @@ impl MinerService for Miner {
|
|||||||
trace!(target: "miner", "map_sealing_work: entering");
|
trace!(target: "miner", "map_sealing_work: entering");
|
||||||
self.enable_and_prepare_sealing(chain);
|
self.enable_and_prepare_sealing(chain);
|
||||||
trace!(target: "miner", "map_sealing_work: sealing prepared");
|
trace!(target: "miner", "map_sealing_work: sealing prepared");
|
||||||
let mut sealing_work = self.sealing_work.locked();
|
let mut sealing_work = self.sealing_work.lock();
|
||||||
let ret = sealing_work.use_last_ref();
|
let ret = sealing_work.use_last_ref();
|
||||||
trace!(target: "miner", "map_sealing_work: leaving use_last_ref={:?}", ret.as_ref().map(|b| b.block().fields().header.hash()));
|
trace!(target: "miner", "map_sealing_work: leaving use_last_ref={:?}", ret.as_ref().map(|b| b.block().fields().header.hash()));
|
||||||
ret.map(f)
|
ret.map(f)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn submit_seal(&self, chain: &MiningBlockChainClient, pow_hash: H256, seal: Vec<Bytes>) -> Result<(), Error> {
|
fn submit_seal(&self, chain: &MiningBlockChainClient, pow_hash: H256, seal: Vec<Bytes>) -> Result<(), Error> {
|
||||||
let result = if let Some(b) = self.sealing_work.locked().get_used_if(if self.options.enable_resubmission { GetAction::Clone } else { GetAction::Take }, |b| &b.hash() == &pow_hash) {
|
let result = if let Some(b) = self.sealing_work.lock().get_used_if(if self.options.enable_resubmission { GetAction::Clone } else { GetAction::Take }, |b| &b.hash() == &pow_hash) {
|
||||||
b.lock().try_seal(self.engine(), seal).or_else(|_| {
|
b.lock().try_seal(self.engine(), seal).or_else(|_| {
|
||||||
warn!(target: "miner", "Mined solution rejected: Invalid.");
|
warn!(target: "miner", "Mined solution rejected: Invalid.");
|
||||||
Err(Error::PowInvalid)
|
Err(Error::PowInvalid)
|
||||||
@ -797,7 +796,7 @@ impl MinerService for Miner {
|
|||||||
.par_iter()
|
.par_iter()
|
||||||
.map(|h| fetch_transactions(chain, h));
|
.map(|h| fetch_transactions(chain, h));
|
||||||
out_of_chain.for_each(|txs| {
|
out_of_chain.for_each(|txs| {
|
||||||
let mut transaction_queue = self.transaction_queue.locked();
|
let mut transaction_queue = self.transaction_queue.lock();
|
||||||
let _ = self.add_transactions_to_queue(
|
let _ = self.add_transactions_to_queue(
|
||||||
chain, txs, TransactionOrigin::External, &mut transaction_queue
|
chain, txs, TransactionOrigin::External, &mut transaction_queue
|
||||||
);
|
);
|
||||||
@ -811,7 +810,7 @@ impl MinerService for Miner {
|
|||||||
.map(|h: &H256| fetch_transactions(chain, h));
|
.map(|h: &H256| fetch_transactions(chain, h));
|
||||||
|
|
||||||
in_chain.for_each(|mut txs| {
|
in_chain.for_each(|mut txs| {
|
||||||
let mut transaction_queue = self.transaction_queue.locked();
|
let mut transaction_queue = self.transaction_queue.lock();
|
||||||
|
|
||||||
let to_remove = txs.drain(..)
|
let to_remove = txs.drain(..)
|
||||||
.map(|tx| {
|
.map(|tx| {
|
||||||
|
@ -38,7 +38,7 @@
|
|||||||
//! assert_eq!(miner.status().transactions_in_pending_queue, 0);
|
//! assert_eq!(miner.status().transactions_in_pending_queue, 0);
|
||||||
//!
|
//!
|
||||||
//! // Check block for sealing
|
//! // Check block for sealing
|
||||||
//! //assert!(miner.sealing_block(client.deref()).locked().is_some());
|
//! //assert!(miner.sealing_block(client.deref()).lock().is_some());
|
||||||
//! }
|
//! }
|
||||||
//! ```
|
//! ```
|
||||||
|
|
||||||
|
@ -80,14 +80,17 @@ impl PriceInfo {
|
|||||||
//#[ignore]
|
//#[ignore]
|
||||||
#[test]
|
#[test]
|
||||||
fn should_get_price_info() {
|
fn should_get_price_info() {
|
||||||
use std::sync::{Condvar, Mutex, Arc};
|
use std::sync::Arc;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use util::log::init_log;
|
use util::log::init_log;
|
||||||
|
use util::{Condvar, Mutex};
|
||||||
|
|
||||||
init_log();
|
init_log();
|
||||||
let done = Arc::new((Mutex::new(PriceInfo { ethusd: 0f32 }), Condvar::new()));
|
let done = Arc::new((Mutex::new(PriceInfo { ethusd: 0f32 }), Condvar::new()));
|
||||||
let rdone = done.clone();
|
let rdone = done.clone();
|
||||||
PriceInfo::get(move |price| { let mut p = rdone.0.lock().unwrap(); *p = price; rdone.1.notify_one(); }).unwrap();
|
PriceInfo::get(move |price| { let mut p = rdone.0.lock(); *p = price; rdone.1.notify_one(); }).unwrap();
|
||||||
let p = done.1.wait_timeout(done.0.lock().unwrap(), Duration::from_millis(10000)).unwrap();
|
let mut p = done.0.lock();
|
||||||
assert!(!p.1.timed_out());
|
let t = done.1.wait_for(&mut p, Duration::from_millis(10000));
|
||||||
assert!(p.0.ethusd != 0f32);
|
assert!(!t.timed_out());
|
||||||
|
assert!(p.ethusd != 0f32);
|
||||||
}
|
}
|
@ -61,13 +61,13 @@ impl WorkPoster {
|
|||||||
pub fn notify(&self, pow_hash: H256, difficulty: U256, number: u64) {
|
pub fn notify(&self, pow_hash: H256, difficulty: U256, number: u64) {
|
||||||
// TODO: move this to engine
|
// TODO: move this to engine
|
||||||
let target = Ethash::difficulty_to_boundary(&difficulty);
|
let target = Ethash::difficulty_to_boundary(&difficulty);
|
||||||
let seed_hash = &self.seed_compute.locked().get_seedhash(number);
|
let seed_hash = &self.seed_compute.lock().get_seedhash(number);
|
||||||
let seed_hash = H256::from_slice(&seed_hash[..]);
|
let seed_hash = H256::from_slice(&seed_hash[..]);
|
||||||
let body = format!(
|
let body = format!(
|
||||||
r#"{{ "result": ["0x{}","0x{}","0x{}","0x{:x}"] }}"#,
|
r#"{{ "result": ["0x{}","0x{}","0x{}","0x{:x}"] }}"#,
|
||||||
pow_hash.hex(), seed_hash.hex(), target.hex(), number
|
pow_hash.hex(), seed_hash.hex(), target.hex(), number
|
||||||
);
|
);
|
||||||
let mut client = self.client.locked();
|
let mut client = self.client.lock();
|
||||||
for u in &self.urls {
|
for u in &self.urls {
|
||||||
if let Err(e) = client.request(u.clone(), PostHandler { body: body.clone() }) {
|
if let Err(e) = client.request(u.clone(), PostHandler { body: body.clone() }) {
|
||||||
warn!("Error sending HTTP notification to {} : {}, retrying", u, e);
|
warn!("Error sending HTTP notification to {} : {}, retrying", u, e);
|
||||||
|
@ -136,10 +136,10 @@ impl Spec {
|
|||||||
|
|
||||||
/// Return the state root for the genesis state, memoising accordingly.
|
/// Return the state root for the genesis state, memoising accordingly.
|
||||||
pub fn state_root(&self) -> H256 {
|
pub fn state_root(&self) -> H256 {
|
||||||
if self.state_root_memo.unwrapped_read().is_none() {
|
if self.state_root_memo.read().is_none() {
|
||||||
*self.state_root_memo.unwrapped_write() = Some(self.genesis_state.root());
|
*self.state_root_memo.write() = Some(self.genesis_state.root());
|
||||||
}
|
}
|
||||||
self.state_root_memo.unwrapped_read().as_ref().unwrap().clone()
|
self.state_root_memo.read().as_ref().unwrap().clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the known knodes of the network in enode format.
|
/// Get the known knodes of the network in enode format.
|
||||||
@ -209,12 +209,12 @@ impl Spec {
|
|||||||
/// Alter the value of the genesis state.
|
/// Alter the value of the genesis state.
|
||||||
pub fn set_genesis_state(&mut self, s: PodState) {
|
pub fn set_genesis_state(&mut self, s: PodState) {
|
||||||
self.genesis_state = s;
|
self.genesis_state = s;
|
||||||
*self.state_root_memo.unwrapped_write() = None;
|
*self.state_root_memo.write() = None;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns `false` if the memoized state root is invalid. `true` otherwise.
|
/// Returns `false` if the memoized state root is invalid. `true` otherwise.
|
||||||
pub fn is_state_root_valid(&self) -> bool {
|
pub fn is_state_root_valid(&self) -> bool {
|
||||||
self.state_root_memo.unwrapped_read().clone().map_or(true, |sr| sr == self.genesis_state.root())
|
self.state_root_memo.read().clone().map_or(true, |sr| sr == self.genesis_state.root())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Ensure that the given state DB has the trie nodes in for the genesis state.
|
/// Ensure that the given state DB has the trie nodes in for the genesis state.
|
||||||
|
@ -18,17 +18,18 @@
|
|||||||
use std::ptr;
|
use std::ptr;
|
||||||
use std::ops::{Deref, DerefMut};
|
use std::ops::{Deref, DerefMut};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::sync::{RwLock, Arc};
|
use std::sync::Arc;
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use bloomchain::{Number, Config as BloomConfig};
|
use bloomchain::{Number, Config as BloomConfig};
|
||||||
use bloomchain::group::{BloomGroupDatabase, BloomGroupChain, GroupPosition, BloomGroup};
|
use bloomchain::group::{BloomGroupDatabase, BloomGroupChain, GroupPosition, BloomGroup};
|
||||||
use util::{H256, H264, Database, DatabaseConfig, DBTransaction, RwLockable};
|
use util::{H256, H264, Database, DatabaseConfig, DBTransaction, RwLock};
|
||||||
use header::BlockNumber;
|
use header::BlockNumber;
|
||||||
use trace::{BlockTraces, LocalizedTrace, Config, Switch, Filter, Database as TraceDatabase, ImportRequest, DatabaseExtras, Error};
|
use trace::{BlockTraces, LocalizedTrace, Config, Switch, Filter, Database as TraceDatabase, ImportRequest, DatabaseExtras, Error};
|
||||||
use db::{Key, Writable, Readable, CacheUpdatePolicy};
|
use db::{Key, Writable, Readable, CacheUpdatePolicy};
|
||||||
use blooms;
|
use blooms;
|
||||||
use super::flat::{FlatTrace, FlatBlockTraces, FlatTransactionTraces};
|
use super::flat::{FlatTrace, FlatBlockTraces, FlatTransactionTraces};
|
||||||
|
|
||||||
|
|
||||||
const TRACE_DB_VER: &'static [u8] = b"1.0";
|
const TRACE_DB_VER: &'static [u8] = b"1.0";
|
||||||
|
|
||||||
#[derive(Debug, Copy, Clone)]
|
#[derive(Debug, Copy, Clone)]
|
||||||
@ -231,7 +232,7 @@ impl<T> TraceDatabase for TraceDB<T> where T: DatabaseExtras {
|
|||||||
|
|
||||||
// at first, let's insert new block traces
|
// at first, let's insert new block traces
|
||||||
{
|
{
|
||||||
let mut traces = self.traces.unwrapped_write();
|
let mut traces = self.traces.write();
|
||||||
// it's important to use overwrite here,
|
// it's important to use overwrite here,
|
||||||
// cause this value might be queried by hash later
|
// cause this value might be queried by hash later
|
||||||
batch.write_with_cache(traces.deref_mut(), request.block_hash, request.traces, CacheUpdatePolicy::Overwrite);
|
batch.write_with_cache(traces.deref_mut(), request.block_hash, request.traces, CacheUpdatePolicy::Overwrite);
|
||||||
@ -259,7 +260,7 @@ impl<T> TraceDatabase for TraceDB<T> where T: DatabaseExtras {
|
|||||||
.map(|p| (From::from(p.0), From::from(p.1)))
|
.map(|p| (From::from(p.0), From::from(p.1)))
|
||||||
.collect::<HashMap<TraceGroupPosition, blooms::BloomGroup>>();
|
.collect::<HashMap<TraceGroupPosition, blooms::BloomGroup>>();
|
||||||
|
|
||||||
let mut blooms = self.blooms.unwrapped_write();
|
let mut blooms = self.blooms.write();
|
||||||
batch.extend_with_cache(blooms.deref_mut(), blooms_to_insert, CacheUpdatePolicy::Remove);
|
batch.extend_with_cache(blooms.deref_mut(), blooms_to_insert, CacheUpdatePolicy::Remove);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -19,10 +19,9 @@ use self::ansi_term::Colour::{White, Yellow, Green, Cyan, Blue, Purple};
|
|||||||
use self::ansi_term::Style;
|
use self::ansi_term::Style;
|
||||||
|
|
||||||
use std::time::{Instant, Duration};
|
use std::time::{Instant, Duration};
|
||||||
use std::sync::RwLock;
|
|
||||||
use std::ops::{Deref, DerefMut};
|
use std::ops::{Deref, DerefMut};
|
||||||
use ethsync::SyncStatus;
|
use ethsync::SyncStatus;
|
||||||
use util::{Uint, RwLockable, NetworkConfiguration};
|
use util::{Uint, RwLock, NetworkConfiguration};
|
||||||
use ethcore::client::*;
|
use ethcore::client::*;
|
||||||
use number_prefix::{binary_prefix, Standalone, Prefixed};
|
use number_prefix::{binary_prefix, Standalone, Prefixed};
|
||||||
|
|
||||||
@ -75,20 +74,21 @@ impl Informant {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
#[cfg_attr(feature="dev", allow(match_bool))]
|
#[cfg_attr(feature="dev", allow(match_bool))]
|
||||||
pub fn tick(&self, client: &Client, maybe_status: Option<(SyncStatus, NetworkConfiguration)>) {
|
pub fn tick(&self, client: &Client, maybe_status: Option<(SyncStatus, NetworkConfiguration)>) {
|
||||||
let elapsed = self.last_tick.unwrapped_read().elapsed();
|
let elapsed = self.last_tick.read().elapsed();
|
||||||
if elapsed < Duration::from_secs(5) {
|
if elapsed < Duration::from_secs(5) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
*self.last_tick.unwrapped_write() = Instant::now();
|
*self.last_tick.write() = Instant::now();
|
||||||
|
|
||||||
let chain_info = client.chain_info();
|
let chain_info = client.chain_info();
|
||||||
let queue_info = client.queue_info();
|
let queue_info = client.queue_info();
|
||||||
let cache_info = client.blockchain_cache_info();
|
let cache_info = client.blockchain_cache_info();
|
||||||
|
|
||||||
let mut write_report = self.report.unwrapped_write();
|
let mut write_report = self.report.write();
|
||||||
let report = client.report();
|
let report = client.report();
|
||||||
|
|
||||||
let paint = |c: Style, t: String| match self.with_color {
|
let paint = |c: Style, t: String| match self.with_color {
|
||||||
@ -97,8 +97,8 @@ impl Informant {
|
|||||||
};
|
};
|
||||||
|
|
||||||
if let (_, _, &Some(ref last_report)) = (
|
if let (_, _, &Some(ref last_report)) = (
|
||||||
self.chain_info.unwrapped_read().deref(),
|
self.chain_info.read().deref(),
|
||||||
self.cache_info.unwrapped_read().deref(),
|
self.cache_info.read().deref(),
|
||||||
write_report.deref()
|
write_report.deref()
|
||||||
) {
|
) {
|
||||||
println!("{} {} {} blk/s {} tx/s {} Mgas/s {}{}+{} Qed {} db {} chain {} queue{}",
|
println!("{} {} {} blk/s {} tx/s {} Mgas/s {}{}+{} Qed {} db {} chain {} queue{}",
|
||||||
@ -133,8 +133,8 @@ impl Informant {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
*self.chain_info.unwrapped_write().deref_mut() = Some(chain_info);
|
*self.chain_info.write().deref_mut() = Some(chain_info);
|
||||||
*self.cache_info.unwrapped_write().deref_mut() = Some(cache_info);
|
*self.cache_info.write().deref_mut() = Some(cache_info);
|
||||||
*write_report.deref_mut() = Some(report);
|
*write_report.deref_mut() = Some(report);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -74,7 +74,7 @@ mod url;
|
|||||||
|
|
||||||
use std::io::{Write, Read, BufReader, BufRead};
|
use std::io::{Write, Read, BufReader, BufRead};
|
||||||
use std::ops::Deref;
|
use std::ops::Deref;
|
||||||
use std::sync::{Arc, Mutex, Condvar};
|
use std::sync::Arc;
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
use std::str::{FromStr, from_utf8};
|
use std::str::{FromStr, from_utf8};
|
||||||
@ -82,7 +82,7 @@ use std::thread::sleep;
|
|||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use rustc_serialize::hex::FromHex;
|
use rustc_serialize::hex::FromHex;
|
||||||
use ctrlc::CtrlC;
|
use ctrlc::CtrlC;
|
||||||
use util::{Lockable, H256, ToPretty, PayloadInfo, Bytes, Colour, Applyable, version, journaldb};
|
use util::{H256, ToPretty, PayloadInfo, Bytes, Colour, Applyable, version, journaldb};
|
||||||
use util::panics::{MayPanic, ForwardPanic, PanicHandler};
|
use util::panics::{MayPanic, ForwardPanic, PanicHandler};
|
||||||
use ethcore::client::{BlockID, BlockChainClient, ClientConfig, get_db_path, BlockImportError,
|
use ethcore::client::{BlockID, BlockChainClient, ClientConfig, get_db_path, BlockImportError,
|
||||||
ChainNotify, Mode};
|
ChainNotify, Mode};
|
||||||
@ -93,6 +93,7 @@ use ethsync::EthSync;
|
|||||||
use ethcore::miner::{Miner, MinerService, ExternalMiner};
|
use ethcore::miner::{Miner, MinerService, ExternalMiner};
|
||||||
use migration::migrate;
|
use migration::migrate;
|
||||||
use informant::Informant;
|
use informant::Informant;
|
||||||
|
use util::{Mutex, Condvar};
|
||||||
|
|
||||||
use die::*;
|
use die::*;
|
||||||
use cli::print_version;
|
use cli::print_version;
|
||||||
@ -593,7 +594,7 @@ fn wait_for_exit(
|
|||||||
|
|
||||||
// Wait for signal
|
// Wait for signal
|
||||||
let mutex = Mutex::new(());
|
let mutex = Mutex::new(());
|
||||||
let _ = exit.wait(mutex.locked()).unwrap();
|
let _ = exit.wait(&mut mutex.lock());
|
||||||
info!("Finishing work, please wait...");
|
info!("Finishing work, please wait...");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -41,7 +41,6 @@ use rpc::v1::tests::helpers::{TestSyncProvider, Config as SyncConfig, TestMinerS
|
|||||||
use rpc::v1::{Eth, EthClient, EthFilter, EthFilterClient};
|
use rpc::v1::{Eth, EthClient, EthFilter, EthFilterClient};
|
||||||
use util::panics::MayPanic;
|
use util::panics::MayPanic;
|
||||||
use util::hash::Address;
|
use util::hash::Address;
|
||||||
use util::Lockable;
|
|
||||||
|
|
||||||
const USAGE: &'static str = r#"
|
const USAGE: &'static str = r#"
|
||||||
Parity rpctest client.
|
Parity rpctest client.
|
||||||
@ -138,7 +137,7 @@ impl Configuration {
|
|||||||
panic_handler.on_panic(move |_reason| { e.notify_all(); });
|
panic_handler.on_panic(move |_reason| { e.notify_all(); });
|
||||||
|
|
||||||
let mutex = Mutex::new(());
|
let mutex = Mutex::new(());
|
||||||
let _ = exit.wait(mutex.locked()).unwrap();
|
let _ = exit.wait(mutex.lock()).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -16,10 +16,10 @@
|
|||||||
|
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::time::{Instant, Duration};
|
use std::time::{Instant, Duration};
|
||||||
use std::sync::{mpsc, Mutex, RwLock, Arc};
|
use std::sync::{mpsc, Arc};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use jsonrpc_core;
|
use jsonrpc_core;
|
||||||
use util::{U256, Lockable, RwLockable};
|
use util::{Mutex, RwLock, U256};
|
||||||
use v1::helpers::{TransactionRequest, TransactionConfirmation};
|
use v1::helpers::{TransactionRequest, TransactionConfirmation};
|
||||||
|
|
||||||
/// Result that can be returned from JSON RPC.
|
/// Result that can be returned from JSON RPC.
|
||||||
@ -110,7 +110,7 @@ pub struct ConfirmationPromise {
|
|||||||
impl ConfirmationToken {
|
impl ConfirmationToken {
|
||||||
/// Submit solution to all listeners
|
/// Submit solution to all listeners
|
||||||
fn resolve(&self, result: Option<RpcResult>) {
|
fn resolve(&self, result: Option<RpcResult>) {
|
||||||
let mut res = self.result.locked();
|
let mut res = self.result.lock();
|
||||||
*res = result.map_or(ConfirmationResult::Rejected, |h| ConfirmationResult::Confirmed(h));
|
*res = result.map_or(ConfirmationResult::Rejected, |h| ConfirmationResult::Confirmed(h));
|
||||||
// Notify listener
|
// Notify listener
|
||||||
self.handle.unpark();
|
self.handle.unpark();
|
||||||
@ -142,7 +142,7 @@ impl ConfirmationPromise {
|
|||||||
// Park thread (may wake up spuriously)
|
// Park thread (may wake up spuriously)
|
||||||
thread::park_timeout(deadline - now);
|
thread::park_timeout(deadline - now);
|
||||||
// Take confirmation result
|
// Take confirmation result
|
||||||
let res = self.result.locked();
|
let res = self.result.lock();
|
||||||
// Check the result
|
// Check the result
|
||||||
match *res {
|
match *res {
|
||||||
ConfirmationResult::Rejected => return None,
|
ConfirmationResult::Rejected => return None,
|
||||||
@ -183,7 +183,7 @@ impl ConfirmationsQueue {
|
|||||||
/// This method can be used only once (only single consumer of events can exist).
|
/// This method can be used only once (only single consumer of events can exist).
|
||||||
pub fn start_listening<F>(&self, listener: F) -> Result<(), QueueError>
|
pub fn start_listening<F>(&self, listener: F) -> Result<(), QueueError>
|
||||||
where F: Fn(QueueEvent) -> () {
|
where F: Fn(QueueEvent) -> () {
|
||||||
let recv = self.receiver.locked().take();
|
let recv = self.receiver.lock().take();
|
||||||
if let None = recv {
|
if let None = recv {
|
||||||
return Err(QueueError::AlreadyUsed);
|
return Err(QueueError::AlreadyUsed);
|
||||||
}
|
}
|
||||||
@ -208,13 +208,13 @@ impl ConfirmationsQueue {
|
|||||||
/// Notifies receiver about the event happening in this queue.
|
/// Notifies receiver about the event happening in this queue.
|
||||||
fn notify(&self, message: QueueEvent) {
|
fn notify(&self, message: QueueEvent) {
|
||||||
// We don't really care about the result
|
// We don't really care about the result
|
||||||
let _ = self.sender.locked().send(message);
|
let _ = self.sender.lock().send(message);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Removes transaction from this queue and notifies `ConfirmationPromise` holders about the result.
|
/// Removes transaction from this queue and notifies `ConfirmationPromise` holders about the result.
|
||||||
/// Notifies also a receiver about that event.
|
/// Notifies also a receiver about that event.
|
||||||
fn remove(&self, id: U256, result: Option<RpcResult>) -> Option<TransactionConfirmation> {
|
fn remove(&self, id: U256, result: Option<RpcResult>) -> Option<TransactionConfirmation> {
|
||||||
let token = self.queue.unwrapped_write().remove(&id);
|
let token = self.queue.write().remove(&id);
|
||||||
|
|
||||||
if let Some(token) = token {
|
if let Some(token) = token {
|
||||||
// notify receiver about the event
|
// notify receiver about the event
|
||||||
@ -241,13 +241,13 @@ impl SigningQueue for ConfirmationsQueue {
|
|||||||
fn add_request(&self, transaction: TransactionRequest) -> ConfirmationPromise {
|
fn add_request(&self, transaction: TransactionRequest) -> ConfirmationPromise {
|
||||||
// Increment id
|
// Increment id
|
||||||
let id = {
|
let id = {
|
||||||
let mut last_id = self.id.locked();
|
let mut last_id = self.id.lock();
|
||||||
*last_id = *last_id + U256::from(1);
|
*last_id = *last_id + U256::from(1);
|
||||||
*last_id
|
*last_id
|
||||||
};
|
};
|
||||||
// Add request to queue
|
// Add request to queue
|
||||||
let res = {
|
let res = {
|
||||||
let mut queue = self.queue.unwrapped_write();
|
let mut queue = self.queue.write();
|
||||||
queue.insert(id, ConfirmationToken {
|
queue.insert(id, ConfirmationToken {
|
||||||
result: Arc::new(Mutex::new(ConfirmationResult::Waiting)),
|
result: Arc::new(Mutex::new(ConfirmationResult::Waiting)),
|
||||||
handle: thread::current(),
|
handle: thread::current(),
|
||||||
@ -266,7 +266,7 @@ impl SigningQueue for ConfirmationsQueue {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn peek(&self, id: &U256) -> Option<TransactionConfirmation> {
|
fn peek(&self, id: &U256) -> Option<TransactionConfirmation> {
|
||||||
self.queue.unwrapped_read().get(id).map(|token| token.request.clone())
|
self.queue.read().get(id).map(|token| token.request.clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn request_rejected(&self, id: U256) -> Option<TransactionConfirmation> {
|
fn request_rejected(&self, id: U256) -> Option<TransactionConfirmation> {
|
||||||
@ -280,17 +280,17 @@ impl SigningQueue for ConfirmationsQueue {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn requests(&self) -> Vec<TransactionConfirmation> {
|
fn requests(&self) -> Vec<TransactionConfirmation> {
|
||||||
let queue = self.queue.unwrapped_read();
|
let queue = self.queue.read();
|
||||||
queue.values().map(|token| token.request.clone()).collect()
|
queue.values().map(|token| token.request.clone()).collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn len(&self) -> usize {
|
fn len(&self) -> usize {
|
||||||
let queue = self.queue.unwrapped_read();
|
let queue = self.queue.read();
|
||||||
queue.len()
|
queue.len()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn is_empty(&self) -> bool {
|
fn is_empty(&self) -> bool {
|
||||||
let queue = self.queue.unwrapped_read();
|
let queue = self.queue.read();
|
||||||
queue.is_empty()
|
queue.is_empty()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -300,8 +300,8 @@ impl SigningQueue for ConfirmationsQueue {
|
|||||||
mod test {
|
mod test {
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::sync::{Arc, Mutex};
|
use std::sync::Arc;
|
||||||
use util::{Address, U256, H256, Lockable};
|
use util::{Address, U256, H256, Mutex};
|
||||||
use v1::helpers::{SigningQueue, ConfirmationsQueue, QueueEvent, TransactionRequest};
|
use v1::helpers::{SigningQueue, ConfirmationsQueue, QueueEvent, TransactionRequest};
|
||||||
use v1::types::H256 as NH256;
|
use v1::types::H256 as NH256;
|
||||||
use jsonrpc_core::to_value;
|
use jsonrpc_core::to_value;
|
||||||
@ -354,7 +354,7 @@ mod test {
|
|||||||
let r = received.clone();
|
let r = received.clone();
|
||||||
let handle = thread::spawn(move || {
|
let handle = thread::spawn(move || {
|
||||||
q.start_listening(move |notification| {
|
q.start_listening(move |notification| {
|
||||||
let mut v = r.locked();
|
let mut v = r.lock();
|
||||||
*v = Some(notification);
|
*v = Some(notification);
|
||||||
}).expect("Should be closed nicely.")
|
}).expect("Should be closed nicely.")
|
||||||
});
|
});
|
||||||
@ -363,7 +363,7 @@ mod test {
|
|||||||
|
|
||||||
// then
|
// then
|
||||||
handle.join().expect("Thread should finish nicely");
|
handle.join().expect("Thread should finish nicely");
|
||||||
let r = received.locked().take();
|
let r = received.lock().take();
|
||||||
assert_eq!(r, Some(QueueEvent::NewRequest(U256::from(1))));
|
assert_eq!(r, Some(QueueEvent::NewRequest(U256::from(1))));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -20,7 +20,7 @@ extern crate ethash;
|
|||||||
|
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::time::{Instant, Duration};
|
use std::time::{Instant, Duration};
|
||||||
use std::sync::{Arc, Weak, Mutex};
|
use std::sync::{Arc, Weak};
|
||||||
use std::ops::Deref;
|
use std::ops::Deref;
|
||||||
use ethsync::{SyncProvider, SyncState};
|
use ethsync::{SyncProvider, SyncState};
|
||||||
use ethcore::miner::{MinerService, ExternalMinerService};
|
use ethcore::miner::{MinerService, ExternalMinerService};
|
||||||
@ -28,7 +28,7 @@ use jsonrpc_core::*;
|
|||||||
use util::numbers::*;
|
use util::numbers::*;
|
||||||
use util::sha3::*;
|
use util::sha3::*;
|
||||||
use util::rlp::{encode, decode, UntrustedRlp, View};
|
use util::rlp::{encode, decode, UntrustedRlp, View};
|
||||||
use util::Lockable;
|
use util::Mutex;
|
||||||
use ethcore::account_provider::AccountProvider;
|
use ethcore::account_provider::AccountProvider;
|
||||||
use ethcore::client::{MiningBlockChainClient, BlockID, TransactionID, UncleID};
|
use ethcore::client::{MiningBlockChainClient, BlockID, TransactionID, UncleID};
|
||||||
use ethcore::header::Header as BlockHeader;
|
use ethcore::header::Header as BlockHeader;
|
||||||
@ -562,7 +562,7 @@ impl<C, S, M, EM> Eth for EthClient<C, S, M, EM> where
|
|||||||
miner.map_sealing_work(client.deref(), |b| {
|
miner.map_sealing_work(client.deref(), |b| {
|
||||||
let pow_hash = b.hash();
|
let pow_hash = b.hash();
|
||||||
let target = Ethash::difficulty_to_boundary(b.block().header().difficulty());
|
let target = Ethash::difficulty_to_boundary(b.block().header().difficulty());
|
||||||
let seed_hash = self.seed_compute.locked().get_seedhash(b.block().header().number());
|
let seed_hash = self.seed_compute.lock().get_seedhash(b.block().header().number());
|
||||||
let block_number = RpcU256::from(b.block().header().number());
|
let block_number = RpcU256::from(b.block().header().number());
|
||||||
to_value(&(RpcH256::from(pow_hash), RpcH256::from(seed_hash), RpcH256::from(target), block_number))
|
to_value(&(RpcH256::from(pow_hash), RpcH256::from(seed_hash), RpcH256::from(target), block_number))
|
||||||
}).unwrap_or(Err(Error::internal_error())) // no work found.
|
}).unwrap_or(Err(Error::internal_error())) // no work found.
|
||||||
|
@ -17,19 +17,18 @@
|
|||||||
//! Eth Filter RPC implementation
|
//! Eth Filter RPC implementation
|
||||||
|
|
||||||
use std::ops::Deref;
|
use std::ops::Deref;
|
||||||
use std::sync::{Arc, Weak, Mutex};
|
use std::sync::{Arc, Weak};
|
||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
use jsonrpc_core::*;
|
use jsonrpc_core::*;
|
||||||
use util::Lockable;
|
|
||||||
use ethcore::miner::MinerService;
|
use ethcore::miner::MinerService;
|
||||||
use ethcore::filter::Filter as EthcoreFilter;
|
use ethcore::filter::Filter as EthcoreFilter;
|
||||||
use ethcore::client::{BlockChainClient, BlockID};
|
use ethcore::client::{BlockChainClient, BlockID};
|
||||||
|
use util::Mutex;
|
||||||
use v1::traits::EthFilter;
|
use v1::traits::EthFilter;
|
||||||
use v1::types::{BlockNumber, Index, Filter, Log, H256 as RpcH256, U256 as RpcU256};
|
use v1::types::{BlockNumber, Index, Filter, Log, H256 as RpcH256, U256 as RpcU256};
|
||||||
use v1::helpers::{PollFilter, PollManager};
|
use v1::helpers::{PollFilter, PollManager};
|
||||||
use v1::impls::eth::pending_logs;
|
use v1::impls::eth::pending_logs;
|
||||||
|
|
||||||
|
|
||||||
/// Eth filter rpc implementation.
|
/// Eth filter rpc implementation.
|
||||||
pub struct EthFilterClient<C, M> where
|
pub struct EthFilterClient<C, M> where
|
||||||
C: BlockChainClient,
|
C: BlockChainClient,
|
||||||
@ -68,7 +67,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
try!(self.active());
|
try!(self.active());
|
||||||
from_params::<(Filter,)>(params)
|
from_params::<(Filter,)>(params)
|
||||||
.and_then(|(filter,)| {
|
.and_then(|(filter,)| {
|
||||||
let mut polls = self.polls.locked();
|
let mut polls = self.polls.lock();
|
||||||
let block_number = take_weak!(self.client).chain_info().best_block_number;
|
let block_number = take_weak!(self.client).chain_info().best_block_number;
|
||||||
let id = polls.create_poll(PollFilter::Logs(block_number, Default::default(), filter));
|
let id = polls.create_poll(PollFilter::Logs(block_number, Default::default(), filter));
|
||||||
to_value(&RpcU256::from(id))
|
to_value(&RpcU256::from(id))
|
||||||
@ -79,7 +78,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
try!(self.active());
|
try!(self.active());
|
||||||
match params {
|
match params {
|
||||||
Params::None => {
|
Params::None => {
|
||||||
let mut polls = self.polls.locked();
|
let mut polls = self.polls.lock();
|
||||||
let id = polls.create_poll(PollFilter::Block(take_weak!(self.client).chain_info().best_block_number));
|
let id = polls.create_poll(PollFilter::Block(take_weak!(self.client).chain_info().best_block_number));
|
||||||
to_value(&RpcU256::from(id))
|
to_value(&RpcU256::from(id))
|
||||||
},
|
},
|
||||||
@ -91,7 +90,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
try!(self.active());
|
try!(self.active());
|
||||||
match params {
|
match params {
|
||||||
Params::None => {
|
Params::None => {
|
||||||
let mut polls = self.polls.locked();
|
let mut polls = self.polls.lock();
|
||||||
let pending_transactions = take_weak!(self.miner).pending_transactions_hashes();
|
let pending_transactions = take_weak!(self.miner).pending_transactions_hashes();
|
||||||
let id = polls.create_poll(PollFilter::PendingTransaction(pending_transactions));
|
let id = polls.create_poll(PollFilter::PendingTransaction(pending_transactions));
|
||||||
|
|
||||||
@ -106,7 +105,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
let client = take_weak!(self.client);
|
let client = take_weak!(self.client);
|
||||||
from_params::<(Index,)>(params)
|
from_params::<(Index,)>(params)
|
||||||
.and_then(|(index,)| {
|
.and_then(|(index,)| {
|
||||||
let mut polls = self.polls.locked();
|
let mut polls = self.polls.lock();
|
||||||
match polls.poll_mut(&index.value()) {
|
match polls.poll_mut(&index.value()) {
|
||||||
None => Ok(Value::Array(vec![] as Vec<Value>)),
|
None => Ok(Value::Array(vec![] as Vec<Value>)),
|
||||||
Some(filter) => match *filter {
|
Some(filter) => match *filter {
|
||||||
@ -196,7 +195,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
try!(self.active());
|
try!(self.active());
|
||||||
from_params::<(Index,)>(params)
|
from_params::<(Index,)>(params)
|
||||||
.and_then(|(index,)| {
|
.and_then(|(index,)| {
|
||||||
let mut polls = self.polls.locked();
|
let mut polls = self.polls.lock();
|
||||||
match polls.poll(&index.value()) {
|
match polls.poll(&index.value()) {
|
||||||
Some(&PollFilter::Logs(ref _block_number, ref _previous_log, ref filter)) => {
|
Some(&PollFilter::Logs(ref _block_number, ref _previous_log, ref filter)) => {
|
||||||
let include_pending = filter.to_block == Some(BlockNumber::Pending);
|
let include_pending = filter.to_block == Some(BlockNumber::Pending);
|
||||||
@ -222,7 +221,7 @@ impl<C, M> EthFilter for EthFilterClient<C, M> where
|
|||||||
try!(self.active());
|
try!(self.active());
|
||||||
from_params::<(Index,)>(params)
|
from_params::<(Index,)>(params)
|
||||||
.and_then(|(index,)| {
|
.and_then(|(index,)| {
|
||||||
self.polls.locked().remove_poll(&index.value());
|
self.polls.lock().remove_poll(&index.value());
|
||||||
to_value(&true)
|
to_value(&true)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -16,7 +16,7 @@
|
|||||||
|
|
||||||
//! Test implementation of miner service.
|
//! Test implementation of miner service.
|
||||||
|
|
||||||
use util::{Address, H256, Bytes, U256, FixedHash, Uint, Lockable, RwLockable};
|
use util::{Address, H256, Bytes, U256, FixedHash, Uint};
|
||||||
use util::standard::*;
|
use util::standard::*;
|
||||||
use ethcore::error::{Error, ExecutionError};
|
use ethcore::error::{Error, ExecutionError};
|
||||||
use ethcore::client::{MiningBlockChainClient, Executed, CallAnalytics};
|
use ethcore::client::{MiningBlockChainClient, Executed, CallAnalytics};
|
||||||
@ -76,68 +76,68 @@ impl MinerService for TestMinerService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn set_author(&self, author: Address) {
|
fn set_author(&self, author: Address) {
|
||||||
*self.author.unwrapped_write() = author;
|
*self.author.write() = author;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_extra_data(&self, extra_data: Bytes) {
|
fn set_extra_data(&self, extra_data: Bytes) {
|
||||||
*self.extra_data.unwrapped_write() = extra_data;
|
*self.extra_data.write() = extra_data;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the lower gas limit we wish to target when sealing a new block.
|
/// Set the lower gas limit we wish to target when sealing a new block.
|
||||||
fn set_gas_floor_target(&self, target: U256) {
|
fn set_gas_floor_target(&self, target: U256) {
|
||||||
self.gas_range_target.unwrapped_write().0 = target;
|
self.gas_range_target.write().0 = target;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the upper gas limit we wish to target when sealing a new block.
|
/// Set the upper gas limit we wish to target when sealing a new block.
|
||||||
fn set_gas_ceil_target(&self, target: U256) {
|
fn set_gas_ceil_target(&self, target: U256) {
|
||||||
self.gas_range_target.unwrapped_write().1 = target;
|
self.gas_range_target.write().1 = target;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_minimal_gas_price(&self, min_gas_price: U256) {
|
fn set_minimal_gas_price(&self, min_gas_price: U256) {
|
||||||
*self.min_gas_price.unwrapped_write() = min_gas_price;
|
*self.min_gas_price.write() = min_gas_price;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_transactions_limit(&self, limit: usize) {
|
fn set_transactions_limit(&self, limit: usize) {
|
||||||
*self.limit.unwrapped_write() = limit;
|
*self.limit.write() = limit;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_tx_gas_limit(&self, limit: U256) {
|
fn set_tx_gas_limit(&self, limit: U256) {
|
||||||
*self.tx_gas_limit.unwrapped_write() = limit;
|
*self.tx_gas_limit.write() = limit;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn transactions_limit(&self) -> usize {
|
fn transactions_limit(&self) -> usize {
|
||||||
*self.limit.unwrapped_read()
|
*self.limit.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn author(&self) -> Address {
|
fn author(&self) -> Address {
|
||||||
*self.author.unwrapped_read()
|
*self.author.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn minimal_gas_price(&self) -> U256 {
|
fn minimal_gas_price(&self) -> U256 {
|
||||||
*self.min_gas_price.unwrapped_read()
|
*self.min_gas_price.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn extra_data(&self) -> Bytes {
|
fn extra_data(&self) -> Bytes {
|
||||||
self.extra_data.unwrapped_read().clone()
|
self.extra_data.read().clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn gas_floor_target(&self) -> U256 {
|
fn gas_floor_target(&self) -> U256 {
|
||||||
self.gas_range_target.unwrapped_read().0
|
self.gas_range_target.read().0
|
||||||
}
|
}
|
||||||
|
|
||||||
fn gas_ceil_target(&self) -> U256 {
|
fn gas_ceil_target(&self) -> U256 {
|
||||||
self.gas_range_target.unwrapped_read().1
|
self.gas_range_target.read().1
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Imports transactions to transaction queue.
|
/// Imports transactions to transaction queue.
|
||||||
fn import_external_transactions(&self, _chain: &MiningBlockChainClient, transactions: Vec<SignedTransaction>) ->
|
fn import_external_transactions(&self, _chain: &MiningBlockChainClient, transactions: Vec<SignedTransaction>) ->
|
||||||
Vec<Result<TransactionImportResult, Error>> {
|
Vec<Result<TransactionImportResult, Error>> {
|
||||||
// lets assume that all txs are valid
|
// lets assume that all txs are valid
|
||||||
self.imported_transactions.locked().extend_from_slice(&transactions);
|
self.imported_transactions.lock().extend_from_slice(&transactions);
|
||||||
|
|
||||||
for sender in transactions.iter().filter_map(|t| t.sender().ok()) {
|
for sender in transactions.iter().filter_map(|t| t.sender().ok()) {
|
||||||
let nonce = self.last_nonce(&sender).expect("last_nonce must be populated in tests");
|
let nonce = self.last_nonce(&sender).expect("last_nonce must be populated in tests");
|
||||||
self.last_nonces.unwrapped_write().insert(sender, nonce + U256::from(1));
|
self.last_nonces.write().insert(sender, nonce + U256::from(1));
|
||||||
}
|
}
|
||||||
transactions
|
transactions
|
||||||
.iter()
|
.iter()
|
||||||
@ -152,11 +152,11 @@ impl MinerService for TestMinerService {
|
|||||||
// keep the pending nonces up to date
|
// keep the pending nonces up to date
|
||||||
if let Ok(ref sender) = transaction.sender() {
|
if let Ok(ref sender) = transaction.sender() {
|
||||||
let nonce = self.last_nonce(sender).unwrap_or(chain.latest_nonce(sender));
|
let nonce = self.last_nonce(sender).unwrap_or(chain.latest_nonce(sender));
|
||||||
self.last_nonces.unwrapped_write().insert(sender.clone(), nonce + U256::from(1));
|
self.last_nonces.write().insert(sender.clone(), nonce + U256::from(1));
|
||||||
}
|
}
|
||||||
|
|
||||||
// lets assume that all txs are valid
|
// lets assume that all txs are valid
|
||||||
self.imported_transactions.locked().push(transaction);
|
self.imported_transactions.lock().push(transaction);
|
||||||
|
|
||||||
Ok(TransactionImportResult::Current)
|
Ok(TransactionImportResult::Current)
|
||||||
}
|
}
|
||||||
@ -186,23 +186,23 @@ impl MinerService for TestMinerService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn transaction(&self, hash: &H256) -> Option<SignedTransaction> {
|
fn transaction(&self, hash: &H256) -> Option<SignedTransaction> {
|
||||||
self.pending_transactions.locked().get(hash).cloned()
|
self.pending_transactions.lock().get(hash).cloned()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn all_transactions(&self) -> Vec<SignedTransaction> {
|
fn all_transactions(&self) -> Vec<SignedTransaction> {
|
||||||
self.pending_transactions.locked().values().cloned().collect()
|
self.pending_transactions.lock().values().cloned().collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn pending_transactions(&self) -> Vec<SignedTransaction> {
|
fn pending_transactions(&self) -> Vec<SignedTransaction> {
|
||||||
self.pending_transactions.locked().values().cloned().collect()
|
self.pending_transactions.lock().values().cloned().collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn pending_receipts(&self) -> BTreeMap<H256, Receipt> {
|
fn pending_receipts(&self) -> BTreeMap<H256, Receipt> {
|
||||||
self.pending_receipts.locked().clone()
|
self.pending_receipts.lock().clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn last_nonce(&self, address: &Address) -> Option<U256> {
|
fn last_nonce(&self, address: &Address) -> Option<U256> {
|
||||||
self.last_nonces.unwrapped_read().get(address).cloned()
|
self.last_nonces.read().get(address).cloned()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Submit `seal` as a valid solution for the header of `pow_hash`.
|
/// Submit `seal` as a valid solution for the header of `pow_hash`.
|
||||||
@ -212,7 +212,7 @@ impl MinerService for TestMinerService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn balance(&self, _chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
fn balance(&self, _chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
||||||
self.latest_closed_block.locked().as_ref().map_or_else(U256::zero, |b| b.block().fields().state.balance(address).clone())
|
self.latest_closed_block.lock().as_ref().map_or_else(U256::zero, |b| b.block().fields().state.balance(address).clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn call(&self, _chain: &MiningBlockChainClient, _t: &SignedTransaction, _analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
fn call(&self, _chain: &MiningBlockChainClient, _t: &SignedTransaction, _analytics: CallAnalytics) -> Result<Executed, ExecutionError> {
|
||||||
@ -220,7 +220,7 @@ impl MinerService for TestMinerService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn storage_at(&self, _chain: &MiningBlockChainClient, address: &Address, position: &H256) -> H256 {
|
fn storage_at(&self, _chain: &MiningBlockChainClient, address: &Address, position: &H256) -> H256 {
|
||||||
self.latest_closed_block.locked().as_ref().map_or_else(H256::default, |b| b.block().fields().state.storage_at(address, position).clone())
|
self.latest_closed_block.lock().as_ref().map_or_else(H256::default, |b| b.block().fields().state.storage_at(address, position).clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn nonce(&self, _chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
fn nonce(&self, _chain: &MiningBlockChainClient, address: &Address) -> U256 {
|
||||||
@ -230,7 +230,7 @@ impl MinerService for TestMinerService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn code(&self, _chain: &MiningBlockChainClient, address: &Address) -> Option<Bytes> {
|
fn code(&self, _chain: &MiningBlockChainClient, address: &Address) -> Option<Bytes> {
|
||||||
self.latest_closed_block.locked().as_ref().map_or(None, |b| b.block().fields().state.code(address).clone())
|
self.latest_closed_block.lock().as_ref().map_or(None, |b| b.block().fields().state.code(address).clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -16,9 +16,8 @@
|
|||||||
|
|
||||||
//! Test implementation of SyncProvider.
|
//! Test implementation of SyncProvider.
|
||||||
|
|
||||||
use util::{U256, RwLockable};
|
use util::{RwLock, U256};
|
||||||
use ethsync::{SyncProvider, SyncStatus, SyncState};
|
use ethsync::{SyncProvider, SyncStatus, SyncState};
|
||||||
use std::sync::RwLock;
|
|
||||||
|
|
||||||
/// TestSyncProvider config.
|
/// TestSyncProvider config.
|
||||||
pub struct Config {
|
pub struct Config {
|
||||||
@ -57,7 +56,7 @@ impl TestSyncProvider {
|
|||||||
|
|
||||||
impl SyncProvider for TestSyncProvider {
|
impl SyncProvider for TestSyncProvider {
|
||||||
fn status(&self) -> SyncStatus {
|
fn status(&self) -> SyncStatus {
|
||||||
self.status.unwrapped_read().clone()
|
self.status.read().clone()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -16,11 +16,11 @@
|
|||||||
|
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::sync::{Arc, RwLock};
|
use std::sync::Arc;
|
||||||
use jsonrpc_core::IoHandler;
|
use jsonrpc_core::IoHandler;
|
||||||
use util::RwLockable;
|
|
||||||
use util::hash::{Address, H256, FixedHash};
|
use util::hash::{Address, H256, FixedHash};
|
||||||
use util::numbers::{Uint, U256};
|
use util::numbers::{Uint, U256};
|
||||||
|
use util::RwLock;
|
||||||
use ethcore::account_provider::AccountProvider;
|
use ethcore::account_provider::AccountProvider;
|
||||||
use ethcore::client::{TestBlockChainClient, EachBlockWith, Executed, TransactionID};
|
use ethcore::client::{TestBlockChainClient, EachBlockWith, Executed, TransactionID};
|
||||||
use ethcore::log_entry::{LocalizedLogEntry, LogEntry};
|
use ethcore::log_entry::{LocalizedLogEntry, LogEntry};
|
||||||
@ -104,13 +104,13 @@ fn rpc_eth_syncing() {
|
|||||||
assert_eq!(tester.io.handle_request(request), Some(false_res.to_owned()));
|
assert_eq!(tester.io.handle_request(request), Some(false_res.to_owned()));
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut status = tester.sync.status.unwrapped_write();
|
let mut status = tester.sync.status.write();
|
||||||
status.state = SyncState::Blocks;
|
status.state = SyncState::Blocks;
|
||||||
status.highest_block_number = Some(2500);
|
status.highest_block_number = Some(2500);
|
||||||
|
|
||||||
// "sync" to 1000 blocks.
|
// "sync" to 1000 blocks.
|
||||||
// causes TestBlockChainClient to return 1000 for its best block number.
|
// causes TestBlockChainClient to return 1000 for its best block number.
|
||||||
let mut blocks = tester.client.blocks.unwrapped_write();
|
let mut blocks = tester.client.blocks.write();
|
||||||
for i in 0..1000 {
|
for i in 0..1000 {
|
||||||
blocks.insert(H256::from(i), Vec::new());
|
blocks.insert(H256::from(i), Vec::new());
|
||||||
}
|
}
|
||||||
@ -121,7 +121,7 @@ fn rpc_eth_syncing() {
|
|||||||
|
|
||||||
{
|
{
|
||||||
// finish "syncing"
|
// finish "syncing"
|
||||||
let mut blocks = tester.client.blocks.unwrapped_write();
|
let mut blocks = tester.client.blocks.write();
|
||||||
for i in 0..1500 {
|
for i in 0..1500 {
|
||||||
blocks.insert(H256::from(i + 1000), Vec::new());
|
blocks.insert(H256::from(i + 1000), Vec::new());
|
||||||
}
|
}
|
||||||
@ -133,9 +133,9 @@ fn rpc_eth_syncing() {
|
|||||||
#[test]
|
#[test]
|
||||||
fn rpc_eth_hashrate() {
|
fn rpc_eth_hashrate() {
|
||||||
let tester = EthTester::default();
|
let tester = EthTester::default();
|
||||||
tester.hashrates.unwrapped_write().insert(H256::from(0), U256::from(0xfffa));
|
tester.hashrates.write().insert(H256::from(0), U256::from(0xfffa));
|
||||||
tester.hashrates.unwrapped_write().insert(H256::from(0), U256::from(0xfffb));
|
tester.hashrates.write().insert(H256::from(0), U256::from(0xfffb));
|
||||||
tester.hashrates.unwrapped_write().insert(H256::from(1), U256::from(0x1));
|
tester.hashrates.write().insert(H256::from(1), U256::from(0x1));
|
||||||
|
|
||||||
let request = r#"{"jsonrpc": "2.0", "method": "eth_hashrate", "params": [], "id": 1}"#;
|
let request = r#"{"jsonrpc": "2.0", "method": "eth_hashrate", "params": [], "id": 1}"#;
|
||||||
let response = r#"{"jsonrpc":"2.0","result":"0xfffc","id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","result":"0xfffc","id":1}"#;
|
||||||
@ -158,7 +158,7 @@ fn rpc_eth_submit_hashrate() {
|
|||||||
let response = r#"{"jsonrpc":"2.0","result":true,"id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","result":true,"id":1}"#;
|
||||||
|
|
||||||
assert_eq!(tester.io.handle_request(request), Some(response.to_owned()));
|
assert_eq!(tester.io.handle_request(request), Some(response.to_owned()));
|
||||||
assert_eq!(tester.hashrates.unwrapped_read().get(&H256::from("0x59daa26581d0acd1fce254fb7e85952f4c09d0915afd33d3886cd914bc7d283c")).cloned(),
|
assert_eq!(tester.hashrates.read().get(&H256::from("0x59daa26581d0acd1fce254fb7e85952f4c09d0915afd33d3886cd914bc7d283c")).cloned(),
|
||||||
Some(U256::from(0x500_000)));
|
Some(U256::from(0x500_000)));
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -215,7 +215,7 @@ fn rpc_eth_mining() {
|
|||||||
let response = r#"{"jsonrpc":"2.0","result":false,"id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","result":false,"id":1}"#;
|
||||||
assert_eq!(tester.io.handle_request(request), Some(response.to_owned()));
|
assert_eq!(tester.io.handle_request(request), Some(response.to_owned()));
|
||||||
|
|
||||||
tester.hashrates.unwrapped_write().insert(H256::from(1), U256::from(0x1));
|
tester.hashrates.write().insert(H256::from(1), U256::from(0x1));
|
||||||
|
|
||||||
let request = r#"{"jsonrpc": "2.0", "method": "eth_mining", "params": [], "id": 1}"#;
|
let request = r#"{"jsonrpc": "2.0", "method": "eth_mining", "params": [], "id": 1}"#;
|
||||||
let response = r#"{"jsonrpc":"2.0","result":true,"id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","result":true,"id":1}"#;
|
||||||
@ -364,7 +364,7 @@ fn rpc_eth_pending_transaction_by_hash() {
|
|||||||
let tester = EthTester::default();
|
let tester = EthTester::default();
|
||||||
{
|
{
|
||||||
let tx: SignedTransaction = decode(&FromHex::from_hex("f85f800182520894095e7baea6a6c7c4c2dfeb977efac326af552d870a801ba048b55bfa915ac795c431978d8a6a992b628d557da5ff759b307d495a36649353a0efffd310ac743f371de3b9f7f9cb56c0b28ad43601b4ab949f53faa07bd2c804").unwrap());
|
let tx: SignedTransaction = decode(&FromHex::from_hex("f85f800182520894095e7baea6a6c7c4c2dfeb977efac326af552d870a801ba048b55bfa915ac795c431978d8a6a992b628d557da5ff759b307d495a36649353a0efffd310ac743f371de3b9f7f9cb56c0b28ad43601b4ab949f53faa07bd2c804").unwrap());
|
||||||
tester.miner.pending_transactions.locked().insert(H256::zero(), tx);
|
tester.miner.pending_transactions.lock().insert(H256::zero(), tx);
|
||||||
}
|
}
|
||||||
|
|
||||||
let response = r#"{"jsonrpc":"2.0","result":{"blockHash":null,"blockNumber":null,"creates":null,"from":"0x0f65fe9276bc9a24ae7083ae28e2660ef72df99e","gas":"0x5208","gasPrice":"0x01","hash":"0x41df922fd0d4766fcc02e161f8295ec28522f329ae487f14d811e4b64c8d6e31","input":"0x","nonce":"0x00","to":"0x095e7baea6a6c7c4c2dfeb977efac326af552d87","transactionIndex":null,"value":"0x0a"},"id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","result":{"blockHash":null,"blockNumber":null,"creates":null,"from":"0x0f65fe9276bc9a24ae7083ae28e2660ef72df99e","gas":"0x5208","gasPrice":"0x01","hash":"0x41df922fd0d4766fcc02e161f8295ec28522f329ae487f14d811e4b64c8d6e31","input":"0x","nonce":"0x00","to":"0x095e7baea6a6c7c4c2dfeb977efac326af552d87","transactionIndex":null,"value":"0x0a"},"id":1}"#;
|
||||||
@ -591,7 +591,7 @@ fn rpc_eth_send_transaction() {
|
|||||||
|
|
||||||
assert_eq!(tester.io.handle_request(&request), Some(response));
|
assert_eq!(tester.io.handle_request(&request), Some(response));
|
||||||
|
|
||||||
tester.miner.last_nonces.unwrapped_write().insert(address.clone(), U256::zero());
|
tester.miner.last_nonces.write().insert(address.clone(), U256::zero());
|
||||||
|
|
||||||
let t = Transaction {
|
let t = Transaction {
|
||||||
nonce: U256::one(),
|
nonce: U256::one(),
|
||||||
@ -749,7 +749,7 @@ fn returns_error_if_can_mine_and_no_closed_block() {
|
|||||||
use ethsync::{SyncState};
|
use ethsync::{SyncState};
|
||||||
|
|
||||||
let eth_tester = EthTester::default();
|
let eth_tester = EthTester::default();
|
||||||
eth_tester.sync.status.unwrapped_write().state = SyncState::Idle;
|
eth_tester.sync.status.write().state = SyncState::Idle;
|
||||||
|
|
||||||
let request = r#"{"jsonrpc": "2.0", "method": "eth_getWork", "params": [], "id": 1}"#;
|
let request = r#"{"jsonrpc": "2.0", "method": "eth_getWork", "params": [], "id": 1}"#;
|
||||||
let response = r#"{"jsonrpc":"2.0","error":{"code":-32603,"message":"Internal error","data":null},"id":1}"#;
|
let response = r#"{"jsonrpc":"2.0","error":{"code":-32603,"message":"Internal error","data":null},"id":1}"#;
|
||||||
|
@ -88,7 +88,7 @@ fn should_add_transaction_to_queue() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn should_dispatch_transaction_if_account_is_unlocked() {
|
fn should_dispatch_transaction_if_account_is_unlock() {
|
||||||
// given
|
// given
|
||||||
let tester = eth_signing();
|
let tester = eth_signing();
|
||||||
let acc = tester.accounts.new_account("test").unwrap();
|
let acc = tester.accounts.new_account("test").unwrap();
|
||||||
|
@ -18,7 +18,6 @@ use std::sync::Arc;
|
|||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
use jsonrpc_core::IoHandler;
|
use jsonrpc_core::IoHandler;
|
||||||
use util::numbers::*;
|
use util::numbers::*;
|
||||||
use util::RwLockable;
|
|
||||||
use ethcore::account_provider::AccountProvider;
|
use ethcore::account_provider::AccountProvider;
|
||||||
use v1::{PersonalClient, Personal};
|
use v1::{PersonalClient, Personal};
|
||||||
use v1::tests::helpers::TestMinerService;
|
use v1::tests::helpers::TestMinerService;
|
||||||
@ -175,7 +174,7 @@ fn sign_and_send_transaction() {
|
|||||||
|
|
||||||
assert_eq!(tester.io.handle_request(request.as_ref()), Some(response));
|
assert_eq!(tester.io.handle_request(request.as_ref()), Some(response));
|
||||||
|
|
||||||
tester.miner.last_nonces.unwrapped_write().insert(address.clone(), U256::zero());
|
tester.miner.last_nonces.write().insert(address.clone(), U256::zero());
|
||||||
|
|
||||||
let t = Transaction {
|
let t = Transaction {
|
||||||
nonce: U256::one(),
|
nonce: U256::one(),
|
||||||
|
@ -18,7 +18,6 @@ use std::sync::Arc;
|
|||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
use jsonrpc_core::IoHandler;
|
use jsonrpc_core::IoHandler;
|
||||||
use util::numbers::*;
|
use util::numbers::*;
|
||||||
use util::Lockable;
|
|
||||||
use ethcore::account_provider::AccountProvider;
|
use ethcore::account_provider::AccountProvider;
|
||||||
use ethcore::client::TestBlockChainClient;
|
use ethcore::client::TestBlockChainClient;
|
||||||
use ethcore::transaction::{Transaction, Action};
|
use ethcore::transaction::{Transaction, Action};
|
||||||
@ -113,7 +112,7 @@ fn should_reject_transaction_from_queue_without_dispatching() {
|
|||||||
// then
|
// then
|
||||||
assert_eq!(tester.io.handle_request(&request), Some(response.to_owned()));
|
assert_eq!(tester.io.handle_request(&request), Some(response.to_owned()));
|
||||||
assert_eq!(tester.queue.requests().len(), 0);
|
assert_eq!(tester.queue.requests().len(), 0);
|
||||||
assert_eq!(tester.miner.imported_transactions.locked().len(), 0);
|
assert_eq!(tester.miner.imported_transactions.lock().len(), 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@ -182,6 +181,6 @@ fn should_confirm_transaction_and_dispatch() {
|
|||||||
// then
|
// then
|
||||||
assert_eq!(tester.io.handle_request(&request), Some(response.to_owned()));
|
assert_eq!(tester.io.handle_request(&request), Some(response.to_owned()));
|
||||||
assert_eq!(tester.queue.requests().len(), 0);
|
assert_eq!(tester.queue.requests().len(), 0);
|
||||||
assert_eq!(tester.miner.imported_transactions.locked().len(), 1);
|
assert_eq!(tester.miner.imported_transactions.lock().len(), 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1142,7 +1142,7 @@ impl ChainSync {
|
|||||||
|e| format!("Error sending nodes: {:?}", e)),
|
|e| format!("Error sending nodes: {:?}", e)),
|
||||||
|
|
||||||
_ => {
|
_ => {
|
||||||
sync.unwrapped_write().on_packet(io, peer, packet_id, data);
|
sync.write().on_packet(io, peer, packet_id, data);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -73,9 +73,9 @@ extern crate rand;
|
|||||||
extern crate heapsize;
|
extern crate heapsize;
|
||||||
|
|
||||||
use std::ops::*;
|
use std::ops::*;
|
||||||
use std::sync::*;
|
use std::sync::Arc;
|
||||||
use util::network::{NetworkProtocolHandler, NetworkService, NetworkContext, PeerId, NetworkConfiguration};
|
use util::network::{NetworkProtocolHandler, NetworkService, NetworkContext, PeerId, NetworkConfiguration};
|
||||||
use util::{TimerToken, U256, H256, RwLockable, UtilError};
|
use util::{TimerToken, U256, H256, RwLock, UtilError};
|
||||||
use ethcore::client::{Client, ChainNotify};
|
use ethcore::client::{Client, ChainNotify};
|
||||||
use io::NetSyncIo;
|
use io::NetSyncIo;
|
||||||
use chain::ChainSync;
|
use chain::ChainSync;
|
||||||
@ -140,7 +140,7 @@ impl EthSync {
|
|||||||
impl SyncProvider for EthSync {
|
impl SyncProvider for EthSync {
|
||||||
/// Get sync status
|
/// Get sync status
|
||||||
fn status(&self) -> SyncStatus {
|
fn status(&self) -> SyncStatus {
|
||||||
self.handler.sync.unwrapped_read().status()
|
self.handler.sync.read().status()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -161,16 +161,16 @@ impl NetworkProtocolHandler for SyncProtocolHandler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn connected(&self, io: &NetworkContext, peer: &PeerId) {
|
fn connected(&self, io: &NetworkContext, peer: &PeerId) {
|
||||||
self.sync.unwrapped_write().on_peer_connected(&mut NetSyncIo::new(io, self.chain.deref()), *peer);
|
self.sync.write().on_peer_connected(&mut NetSyncIo::new(io, self.chain.deref()), *peer);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn disconnected(&self, io: &NetworkContext, peer: &PeerId) {
|
fn disconnected(&self, io: &NetworkContext, peer: &PeerId) {
|
||||||
self.sync.unwrapped_write().on_peer_aborting(&mut NetSyncIo::new(io, self.chain.deref()), *peer);
|
self.sync.write().on_peer_aborting(&mut NetSyncIo::new(io, self.chain.deref()), *peer);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn timeout(&self, io: &NetworkContext, _timer: TimerToken) {
|
fn timeout(&self, io: &NetworkContext, _timer: TimerToken) {
|
||||||
self.sync.unwrapped_write().maintain_peers(&mut NetSyncIo::new(io, self.chain.deref()));
|
self.sync.write().maintain_peers(&mut NetSyncIo::new(io, self.chain.deref()));
|
||||||
self.sync.unwrapped_write().maintain_sync(&mut NetSyncIo::new(io, self.chain.deref()));
|
self.sync.write().maintain_sync(&mut NetSyncIo::new(io, self.chain.deref()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -184,7 +184,7 @@ impl ChainNotify for EthSync {
|
|||||||
{
|
{
|
||||||
self.network.with_context(ETH_PROTOCOL, |context| {
|
self.network.with_context(ETH_PROTOCOL, |context| {
|
||||||
let mut sync_io = NetSyncIo::new(context, self.handler.chain.deref());
|
let mut sync_io = NetSyncIo::new(context, self.handler.chain.deref());
|
||||||
self.handler.sync.unwrapped_write().chain_new_blocks(
|
self.handler.sync.write().chain_new_blocks(
|
||||||
&mut sync_io,
|
&mut sync_io,
|
||||||
&imported,
|
&imported,
|
||||||
&invalid,
|
&invalid,
|
||||||
@ -241,7 +241,7 @@ impl ManageNetwork for EthSync {
|
|||||||
fn stop_network(&self) {
|
fn stop_network(&self) {
|
||||||
self.network.with_context(ETH_PROTOCOL, |context| {
|
self.network.with_context(ETH_PROTOCOL, |context| {
|
||||||
let mut sync_io = NetSyncIo::new(context, self.handler.chain.deref());
|
let mut sync_io = NetSyncIo::new(context, self.handler.chain.deref());
|
||||||
self.handler.sync.unwrapped_write().abort(&mut sync_io);
|
self.handler.sync.write().abort(&mut sync_io);
|
||||||
});
|
});
|
||||||
self.stop();
|
self.stop();
|
||||||
}
|
}
|
||||||
|
@ -27,7 +27,7 @@ fn two_peers() {
|
|||||||
net.peer_mut(2).chain.add_blocks(1000, EachBlockWith::Uncle);
|
net.peer_mut(2).chain.add_blocks(1000, EachBlockWith::Uncle);
|
||||||
net.sync();
|
net.sync();
|
||||||
assert!(net.peer(0).chain.block(BlockID::Number(1000)).is_some());
|
assert!(net.peer(0).chain.block(BlockID::Number(1000)).is_some());
|
||||||
assert_eq!(net.peer(0).chain.blocks.unwrapped_read().deref(), net.peer(1).chain.blocks.unwrapped_read().deref());
|
assert_eq!(net.peer(0).chain.blocks.read().deref(), net.peer(1).chain.blocks.read().deref());
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@ -37,7 +37,7 @@ fn long_chain() {
|
|||||||
net.peer_mut(1).chain.add_blocks(50000, EachBlockWith::Nothing);
|
net.peer_mut(1).chain.add_blocks(50000, EachBlockWith::Nothing);
|
||||||
net.sync();
|
net.sync();
|
||||||
assert!(net.peer(0).chain.block(BlockID::Number(50000)).is_some());
|
assert!(net.peer(0).chain.block(BlockID::Number(50000)).is_some());
|
||||||
assert_eq!(net.peer(0).chain.blocks.unwrapped_read().deref(), net.peer(1).chain.blocks.unwrapped_read().deref());
|
assert_eq!(net.peer(0).chain.blocks.read().deref(), net.peer(1).chain.blocks.read().deref());
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@ -47,7 +47,7 @@ fn status_after_sync() {
|
|||||||
net.peer_mut(1).chain.add_blocks(1000, EachBlockWith::Uncle);
|
net.peer_mut(1).chain.add_blocks(1000, EachBlockWith::Uncle);
|
||||||
net.peer_mut(2).chain.add_blocks(1000, EachBlockWith::Uncle);
|
net.peer_mut(2).chain.add_blocks(1000, EachBlockWith::Uncle);
|
||||||
net.sync();
|
net.sync();
|
||||||
let status = net.peer(0).sync.unwrapped_read().status();
|
let status = net.peer(0).sync.read().status();
|
||||||
assert_eq!(status.state, SyncState::Idle);
|
assert_eq!(status.state, SyncState::Idle);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -71,7 +71,7 @@ fn empty_blocks() {
|
|||||||
}
|
}
|
||||||
net.sync();
|
net.sync();
|
||||||
assert!(net.peer(0).chain.block(BlockID::Number(1000)).is_some());
|
assert!(net.peer(0).chain.block(BlockID::Number(1000)).is_some());
|
||||||
assert_eq!(net.peer(0).chain.blocks.unwrapped_read().deref(), net.peer(1).chain.blocks.unwrapped_read().deref());
|
assert_eq!(net.peer(0).chain.blocks.read().deref(), net.peer(1).chain.blocks.read().deref());
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@ -87,12 +87,12 @@ fn forked() {
|
|||||||
net.peer_mut(1).chain.add_blocks(100, EachBlockWith::Uncle); //fork between 1 and 2
|
net.peer_mut(1).chain.add_blocks(100, EachBlockWith::Uncle); //fork between 1 and 2
|
||||||
net.peer_mut(2).chain.add_blocks(10, EachBlockWith::Nothing);
|
net.peer_mut(2).chain.add_blocks(10, EachBlockWith::Nothing);
|
||||||
// peer 1 has the best chain of 601 blocks
|
// peer 1 has the best chain of 601 blocks
|
||||||
let peer1_chain = net.peer(1).chain.numbers.unwrapped_read().clone();
|
let peer1_chain = net.peer(1).chain.numbers.read().clone();
|
||||||
net.sync();
|
net.sync();
|
||||||
assert_eq!(net.peer(0).chain.difficulty.unwrapped_read().deref(), net.peer(1).chain.difficulty.unwrapped_read().deref());
|
assert_eq!(net.peer(0).chain.difficulty.read().deref(), net.peer(1).chain.difficulty.read().deref());
|
||||||
assert_eq!(net.peer(0).chain.numbers.unwrapped_read().deref(), &peer1_chain);
|
assert_eq!(net.peer(0).chain.numbers.read().deref(), &peer1_chain);
|
||||||
assert_eq!(net.peer(1).chain.numbers.unwrapped_read().deref(), &peer1_chain);
|
assert_eq!(net.peer(1).chain.numbers.read().deref(), &peer1_chain);
|
||||||
assert_eq!(net.peer(2).chain.numbers.unwrapped_read().deref(), &peer1_chain);
|
assert_eq!(net.peer(2).chain.numbers.read().deref(), &peer1_chain);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@ -107,14 +107,14 @@ fn restart() {
|
|||||||
assert!(net.peer(0).chain.chain_info().best_block_number > 100);
|
assert!(net.peer(0).chain.chain_info().best_block_number > 100);
|
||||||
net.restart_peer(0);
|
net.restart_peer(0);
|
||||||
|
|
||||||
let status = net.peer(0).sync.unwrapped_read().status();
|
let status = net.peer(0).sync.read().status();
|
||||||
assert_eq!(status.state, SyncState::ChainHead);
|
assert_eq!(status.state, SyncState::ChainHead);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn status_empty() {
|
fn status_empty() {
|
||||||
let net = TestNet::new(2);
|
let net = TestNet::new(2);
|
||||||
assert_eq!(net.peer(0).sync.unwrapped_read().status().state, SyncState::Idle);
|
assert_eq!(net.peer(0).sync.read().status().state, SyncState::Idle);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
@ -118,7 +118,7 @@ impl TestNet {
|
|||||||
for client in 0..self.peers.len() {
|
for client in 0..self.peers.len() {
|
||||||
if peer != client {
|
if peer != client {
|
||||||
let mut p = self.peers.get_mut(peer).unwrap();
|
let mut p = self.peers.get_mut(peer).unwrap();
|
||||||
p.sync.unwrapped_write().on_peer_connected(&mut TestIo::new(&mut p.chain, &mut p.queue, Some(client as PeerId)), client as PeerId);
|
p.sync.write().on_peer_connected(&mut TestIo::new(&mut p.chain, &mut p.queue, Some(client as PeerId)), client as PeerId);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -133,18 +133,18 @@ impl TestNet {
|
|||||||
trace!("----------------");
|
trace!("----------------");
|
||||||
}
|
}
|
||||||
let mut p = self.peers.get_mut(peer).unwrap();
|
let mut p = self.peers.get_mut(peer).unwrap();
|
||||||
p.sync.unwrapped_write().maintain_sync(&mut TestIo::new(&mut p.chain, &mut p.queue, None));
|
p.sync.write().maintain_sync(&mut TestIo::new(&mut p.chain, &mut p.queue, None));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn sync_step_peer(&mut self, peer_num: usize) {
|
pub fn sync_step_peer(&mut self, peer_num: usize) {
|
||||||
let mut peer = self.peer_mut(peer_num);
|
let mut peer = self.peer_mut(peer_num);
|
||||||
peer.sync.unwrapped_write().maintain_sync(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None));
|
peer.sync.write().maintain_sync(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None));
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn restart_peer(&mut self, i: usize) {
|
pub fn restart_peer(&mut self, i: usize) {
|
||||||
let peer = self.peer_mut(i);
|
let peer = self.peer_mut(i);
|
||||||
peer.sync.unwrapped_write().restart(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None));
|
peer.sync.write().restart(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None));
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn sync(&mut self) -> u32 {
|
pub fn sync(&mut self) -> u32 {
|
||||||
@ -173,6 +173,6 @@ impl TestNet {
|
|||||||
|
|
||||||
pub fn trigger_chain_new_blocks(&mut self, peer_id: usize) {
|
pub fn trigger_chain_new_blocks(&mut self, peer_id: usize) {
|
||||||
let mut peer = self.peer_mut(peer_id);
|
let mut peer = self.peer_mut(peer_id);
|
||||||
peer.sync.unwrapped_write().chain_new_blocks(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None), &[], &[], &[], &[], &[]);
|
peer.sync.write().chain_new_blocks(&mut TestIo::new(&mut peer.chain, &mut peer.queue, None), &[], &[], &[], &[], &[]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -35,6 +35,7 @@ vergen = "0.1"
|
|||||||
target_info = "0.1"
|
target_info = "0.1"
|
||||||
bigint = { path = "bigint" }
|
bigint = { path = "bigint" }
|
||||||
chrono = "0.2"
|
chrono = "0.2"
|
||||||
|
parking_lot = "0.2.6"
|
||||||
using_queue = { path = "using_queue" }
|
using_queue = { path = "using_queue" }
|
||||||
table = { path = "table" }
|
table = { path = "table" }
|
||||||
ansi_term = "0.7"
|
ansi_term = "0.7"
|
||||||
|
@ -14,18 +14,19 @@
|
|||||||
// You should have received a copy of the GNU General Public License
|
// You should have received a copy of the GNU General Public License
|
||||||
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
use std::sync::*;
|
use std::sync::Arc;
|
||||||
use std::thread::{self, JoinHandle};
|
use std::thread::{self, JoinHandle};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use mio::*;
|
use mio::*;
|
||||||
use crossbeam::sync::chase_lev;
|
use crossbeam::sync::chase_lev;
|
||||||
use slab::Slab;
|
use slab::Slab;
|
||||||
use error::*;
|
use error::*;
|
||||||
use misc::*;
|
|
||||||
use io::{IoError, IoHandler};
|
use io::{IoError, IoHandler};
|
||||||
use io::worker::{Worker, Work, WorkType};
|
use io::worker::{Worker, Work, WorkType};
|
||||||
use panics::*;
|
use panics::*;
|
||||||
|
|
||||||
|
use parking_lot::{Condvar, RwLock, Mutex};
|
||||||
|
|
||||||
/// Timer ID
|
/// Timer ID
|
||||||
pub type TimerToken = usize;
|
pub type TimerToken = usize;
|
||||||
/// Timer ID
|
/// Timer ID
|
||||||
@ -228,7 +229,7 @@ impl<Message> Handler for IoManager<Message> where Message: Send + Clone + Sync
|
|||||||
let handler_index = token.as_usize() / TOKENS_PER_HANDLER;
|
let handler_index = token.as_usize() / TOKENS_PER_HANDLER;
|
||||||
let token_id = token.as_usize() % TOKENS_PER_HANDLER;
|
let token_id = token.as_usize() % TOKENS_PER_HANDLER;
|
||||||
if let Some(handler) = self.handlers.get(handler_index) {
|
if let Some(handler) = self.handlers.get(handler_index) {
|
||||||
if let Some(timer) = self.timers.unwrapped_read().get(&token.as_usize()) {
|
if let Some(timer) = self.timers.read().get(&token.as_usize()) {
|
||||||
event_loop.timeout_ms(token, timer.delay).expect("Error re-registering user timer");
|
event_loop.timeout_ms(token, timer.delay).expect("Error re-registering user timer");
|
||||||
self.worker_channel.push(Work { work_type: WorkType::Timeout, token: token_id, handler: handler.clone(), handler_id: handler_index });
|
self.worker_channel.push(Work { work_type: WorkType::Timeout, token: token_id, handler: handler.clone(), handler_id: handler_index });
|
||||||
self.work_ready.notify_all();
|
self.work_ready.notify_all();
|
||||||
@ -250,7 +251,7 @@ impl<Message> Handler for IoManager<Message> where Message: Send + Clone + Sync
|
|||||||
// TODO: flush event loop
|
// TODO: flush event loop
|
||||||
self.handlers.remove(handler_id);
|
self.handlers.remove(handler_id);
|
||||||
// unregister timers
|
// unregister timers
|
||||||
let mut timers = self.timers.unwrapped_write();
|
let mut timers = self.timers.write();
|
||||||
let to_remove: Vec<_> = timers.keys().cloned().filter(|timer_id| timer_id / TOKENS_PER_HANDLER == handler_id).collect();
|
let to_remove: Vec<_> = timers.keys().cloned().filter(|timer_id| timer_id / TOKENS_PER_HANDLER == handler_id).collect();
|
||||||
for timer_id in to_remove {
|
for timer_id in to_remove {
|
||||||
let timer = timers.remove(&timer_id).expect("to_remove only contains keys from timers; qed");
|
let timer = timers.remove(&timer_id).expect("to_remove only contains keys from timers; qed");
|
||||||
@ -260,11 +261,11 @@ impl<Message> Handler for IoManager<Message> where Message: Send + Clone + Sync
|
|||||||
IoMessage::AddTimer { handler_id, token, delay } => {
|
IoMessage::AddTimer { handler_id, token, delay } => {
|
||||||
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
||||||
let timeout = event_loop.timeout_ms(Token(timer_id), delay).expect("Error registering user timer");
|
let timeout = event_loop.timeout_ms(Token(timer_id), delay).expect("Error registering user timer");
|
||||||
self.timers.unwrapped_write().insert(timer_id, UserTimer { delay: delay, timeout: timeout });
|
self.timers.write().insert(timer_id, UserTimer { delay: delay, timeout: timeout });
|
||||||
},
|
},
|
||||||
IoMessage::RemoveTimer { handler_id, token } => {
|
IoMessage::RemoveTimer { handler_id, token } => {
|
||||||
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
||||||
if let Some(timer) = self.timers.unwrapped_write().remove(&timer_id) {
|
if let Some(timer) = self.timers.write().remove(&timer_id) {
|
||||||
event_loop.clear_timeout(timer.timeout);
|
event_loop.clear_timeout(timer.timeout);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@ -278,7 +279,7 @@ impl<Message> Handler for IoManager<Message> where Message: Send + Clone + Sync
|
|||||||
handler.deregister_stream(token, event_loop);
|
handler.deregister_stream(token, event_loop);
|
||||||
// unregister a timer associated with the token (if any)
|
// unregister a timer associated with the token (if any)
|
||||||
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
let timer_id = token + handler_id * TOKENS_PER_HANDLER;
|
||||||
if let Some(timer) = self.timers.unwrapped_write().remove(&timer_id) {
|
if let Some(timer) = self.timers.write().remove(&timer_id) {
|
||||||
event_loop.clear_timeout(timer.timeout);
|
event_loop.clear_timeout(timer.timeout);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -14,7 +14,7 @@
|
|||||||
// You should have received a copy of the GNU General Public License
|
// You should have received a copy of the GNU General Public License
|
||||||
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
use std::sync::*;
|
use std::sync::Arc;
|
||||||
use std::mem;
|
use std::mem;
|
||||||
use std::thread::{JoinHandle, self};
|
use std::thread::{JoinHandle, self};
|
||||||
use std::sync::atomic::{AtomicBool, Ordering as AtomicOrdering};
|
use std::sync::atomic::{AtomicBool, Ordering as AtomicOrdering};
|
||||||
@ -22,7 +22,8 @@ use crossbeam::sync::chase_lev;
|
|||||||
use io::service::{HandlerId, IoChannel, IoContext};
|
use io::service::{HandlerId, IoChannel, IoContext};
|
||||||
use io::{IoHandler};
|
use io::{IoHandler};
|
||||||
use panics::*;
|
use panics::*;
|
||||||
use misc::Lockable;
|
|
||||||
|
use parking_lot::{Condvar, Mutex};
|
||||||
|
|
||||||
pub enum WorkType<Message> {
|
pub enum WorkType<Message> {
|
||||||
Readable,
|
Readable,
|
||||||
@ -82,11 +83,11 @@ impl Worker {
|
|||||||
where Message: Send + Sync + Clone + 'static {
|
where Message: Send + Sync + Clone + 'static {
|
||||||
loop {
|
loop {
|
||||||
{
|
{
|
||||||
let lock = wait_mutex.locked();
|
let mut lock = wait_mutex.lock();
|
||||||
if deleting.load(AtomicOrdering::Acquire) {
|
if deleting.load(AtomicOrdering::Acquire) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
let _ = wait.wait(lock).unwrap();
|
let _ = wait.wait(&mut lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
if deleting.load(AtomicOrdering::Acquire) {
|
if deleting.load(AtomicOrdering::Acquire) {
|
||||||
|
@ -20,7 +20,6 @@ use common::*;
|
|||||||
use rlp::*;
|
use rlp::*;
|
||||||
use hashdb::*;
|
use hashdb::*;
|
||||||
use memorydb::*;
|
use memorydb::*;
|
||||||
use misc::RwLockable;
|
|
||||||
use super::{DB_PREFIX_LEN, LATEST_ERA_KEY, VERSION_KEY};
|
use super::{DB_PREFIX_LEN, LATEST_ERA_KEY, VERSION_KEY};
|
||||||
use super::traits::JournalDB;
|
use super::traits::JournalDB;
|
||||||
use kvdb::{Database, DBTransaction, DatabaseConfig};
|
use kvdb::{Database, DBTransaction, DatabaseConfig};
|
||||||
@ -226,7 +225,7 @@ impl EarlyMergeDB {
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
fn can_reconstruct_refs(&self) -> bool {
|
fn can_reconstruct_refs(&self) -> bool {
|
||||||
let (latest_era, reconstructed) = Self::read_refs(&self.backing);
|
let (latest_era, reconstructed) = Self::read_refs(&self.backing);
|
||||||
let refs = self.refs.as_ref().unwrap().unwrapped_write();
|
let refs = self.refs.as_ref().unwrap().write();
|
||||||
if *refs != reconstructed || latest_era != self.latest_era {
|
if *refs != reconstructed || latest_era != self.latest_era {
|
||||||
let clean_refs = refs.iter().filter_map(|(k, v)| if reconstructed.get(k) == Some(v) {None} else {Some((k.clone(), v.clone()))}).collect::<HashMap<_, _>>();
|
let clean_refs = refs.iter().filter_map(|(k, v)| if reconstructed.get(k) == Some(v) {None} else {Some((k.clone(), v.clone()))}).collect::<HashMap<_, _>>();
|
||||||
let clean_recon = reconstructed.into_iter().filter_map(|(k, v)| if refs.get(&k) == Some(&v) {None} else {Some((k.clone(), v.clone()))}).collect::<HashMap<_, _>>();
|
let clean_recon = reconstructed.into_iter().filter_map(|(k, v)| if refs.get(&k) == Some(&v) {None} else {Some((k.clone(), v.clone()))}).collect::<HashMap<_, _>>();
|
||||||
@ -334,7 +333,7 @@ impl JournalDB for EarlyMergeDB {
|
|||||||
|
|
||||||
fn mem_used(&self) -> usize {
|
fn mem_used(&self) -> usize {
|
||||||
self.overlay.mem_used() + match self.refs {
|
self.overlay.mem_used() + match self.refs {
|
||||||
Some(ref c) => c.unwrapped_read().heap_size_of_children(),
|
Some(ref c) => c.read().heap_size_of_children(),
|
||||||
None => 0
|
None => 0
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -390,7 +389,7 @@ impl JournalDB for EarlyMergeDB {
|
|||||||
//
|
//
|
||||||
|
|
||||||
// record new commit's details.
|
// record new commit's details.
|
||||||
let mut refs = self.refs.as_ref().unwrap().unwrapped_write();
|
let mut refs = self.refs.as_ref().unwrap().write();
|
||||||
let batch = DBTransaction::new();
|
let batch = DBTransaction::new();
|
||||||
let trace = false;
|
let trace = false;
|
||||||
{
|
{
|
||||||
|
@ -20,7 +20,6 @@ use common::*;
|
|||||||
use rlp::*;
|
use rlp::*;
|
||||||
use hashdb::*;
|
use hashdb::*;
|
||||||
use memorydb::*;
|
use memorydb::*;
|
||||||
use misc::RwLockable;
|
|
||||||
use super::{DB_PREFIX_LEN, LATEST_ERA_KEY, VERSION_KEY};
|
use super::{DB_PREFIX_LEN, LATEST_ERA_KEY, VERSION_KEY};
|
||||||
use kvdb::{Database, DBTransaction, DatabaseConfig};
|
use kvdb::{Database, DBTransaction, DatabaseConfig};
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
@ -137,7 +136,7 @@ impl OverlayRecentDB {
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
fn can_reconstruct_refs(&self) -> bool {
|
fn can_reconstruct_refs(&self) -> bool {
|
||||||
let reconstructed = Self::read_overlay(&self.backing);
|
let reconstructed = Self::read_overlay(&self.backing);
|
||||||
let journal_overlay = self.journal_overlay.unwrapped_read();
|
let journal_overlay = self.journal_overlay.read();
|
||||||
*journal_overlay == reconstructed
|
*journal_overlay == reconstructed
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -207,7 +206,7 @@ impl JournalDB for OverlayRecentDB {
|
|||||||
|
|
||||||
fn mem_used(&self) -> usize {
|
fn mem_used(&self) -> usize {
|
||||||
let mut mem = self.transaction_overlay.mem_used();
|
let mut mem = self.transaction_overlay.mem_used();
|
||||||
let overlay = self.journal_overlay.unwrapped_read();
|
let overlay = self.journal_overlay.read();
|
||||||
mem += overlay.backing_overlay.mem_used();
|
mem += overlay.backing_overlay.mem_used();
|
||||||
mem += overlay.journal.heap_size_of_children();
|
mem += overlay.journal.heap_size_of_children();
|
||||||
mem
|
mem
|
||||||
@ -217,17 +216,17 @@ impl JournalDB for OverlayRecentDB {
|
|||||||
self.backing.get(&LATEST_ERA_KEY).expect("Low level database error").is_none()
|
self.backing.get(&LATEST_ERA_KEY).expect("Low level database error").is_none()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn latest_era(&self) -> Option<u64> { self.journal_overlay.unwrapped_read().latest_era }
|
fn latest_era(&self) -> Option<u64> { self.journal_overlay.read().latest_era }
|
||||||
|
|
||||||
fn state(&self, key: &H256) -> Option<Bytes> {
|
fn state(&self, key: &H256) -> Option<Bytes> {
|
||||||
let v = self.journal_overlay.unwrapped_read().backing_overlay.get(&OverlayRecentDB::to_short_key(key)).map(|v| v.to_vec());
|
let v = self.journal_overlay.read().backing_overlay.get(&OverlayRecentDB::to_short_key(key)).map(|v| v.to_vec());
|
||||||
v.or_else(|| self.backing.get_by_prefix(&key[0..DB_PREFIX_LEN]).map(|b| b.to_vec()))
|
v.or_else(|| self.backing.get_by_prefix(&key[0..DB_PREFIX_LEN]).map(|b| b.to_vec()))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn commit(&mut self, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
|
fn commit(&mut self, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
|
||||||
// record new commit's details.
|
// record new commit's details.
|
||||||
trace!("commit: #{} ({}), end era: {:?}", now, id, end);
|
trace!("commit: #{} ({}), end era: {:?}", now, id, end);
|
||||||
let mut journal_overlay = self.journal_overlay.unwrapped_write();
|
let mut journal_overlay = self.journal_overlay.write();
|
||||||
let batch = DBTransaction::new();
|
let batch = DBTransaction::new();
|
||||||
{
|
{
|
||||||
let mut r = RlpStream::new_list(3);
|
let mut r = RlpStream::new_list(3);
|
||||||
@ -334,7 +333,7 @@ impl HashDB for OverlayRecentDB {
|
|||||||
match k {
|
match k {
|
||||||
Some(&(ref d, rc)) if rc > 0 => Some(d),
|
Some(&(ref d, rc)) if rc > 0 => Some(d),
|
||||||
_ => {
|
_ => {
|
||||||
let v = self.journal_overlay.unwrapped_read().backing_overlay.get(&OverlayRecentDB::to_short_key(key)).map(|v| v.to_vec());
|
let v = self.journal_overlay.read().backing_overlay.get(&OverlayRecentDB::to_short_key(key)).map(|v| v.to_vec());
|
||||||
match v {
|
match v {
|
||||||
Some(x) => {
|
Some(x) => {
|
||||||
Some(&self.transaction_overlay.denote(key, x).0)
|
Some(&self.transaction_overlay.denote(key, x).0)
|
||||||
|
@ -116,6 +116,7 @@ extern crate libc;
|
|||||||
extern crate target_info;
|
extern crate target_info;
|
||||||
extern crate bigint;
|
extern crate bigint;
|
||||||
extern crate chrono;
|
extern crate chrono;
|
||||||
|
extern crate parking_lot;
|
||||||
pub extern crate using_queue;
|
pub extern crate using_queue;
|
||||||
pub extern crate table;
|
pub extern crate table;
|
||||||
extern crate ansi_term;
|
extern crate ansi_term;
|
||||||
|
@ -18,14 +18,14 @@
|
|||||||
|
|
||||||
use std::env;
|
use std::env;
|
||||||
use std::borrow::Cow;
|
use std::borrow::Cow;
|
||||||
use rlog::{LogLevelFilter};
|
use rlog::LogLevelFilter;
|
||||||
use env_logger::LogBuilder;
|
use env_logger::LogBuilder;
|
||||||
use std::sync::{RwLock, RwLockReadGuard};
|
|
||||||
use std::sync::atomic::{Ordering, AtomicBool};
|
use std::sync::atomic::{Ordering, AtomicBool};
|
||||||
use arrayvec::ArrayVec;
|
use arrayvec::ArrayVec;
|
||||||
use misc::RwLockable;
|
|
||||||
pub use ansi_term::{Colour, Style};
|
pub use ansi_term::{Colour, Style};
|
||||||
|
|
||||||
|
use parking_lot::{RwLock, RwLockReadGuard};
|
||||||
|
|
||||||
lazy_static! {
|
lazy_static! {
|
||||||
static ref USE_COLOR: AtomicBool = AtomicBool::new(false);
|
static ref USE_COLOR: AtomicBool = AtomicBool::new(false);
|
||||||
}
|
}
|
||||||
@ -91,7 +91,7 @@ impl RotatingLogger {
|
|||||||
|
|
||||||
/// Append new log entry
|
/// Append new log entry
|
||||||
pub fn append(&self, log: String) {
|
pub fn append(&self, log: String) {
|
||||||
self.logs.unwrapped_write().insert(0, log);
|
self.logs.write().insert(0, log);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Return levels
|
/// Return levels
|
||||||
@ -101,7 +101,7 @@ impl RotatingLogger {
|
|||||||
|
|
||||||
/// Return logs
|
/// Return logs
|
||||||
pub fn logs(&self) -> RwLockReadGuard<ArrayVec<[String; LOG_SIZE]>> {
|
pub fn logs(&self) -> RwLockReadGuard<ArrayVec<[String; LOG_SIZE]>> {
|
||||||
self.logs.unwrapped_read()
|
self.logs.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -64,29 +64,4 @@ pub fn version_data() -> Bytes {
|
|||||||
s.append(&rustc_version());
|
s.append(&rustc_version());
|
||||||
s.append(&&Target::os()[0..2]);
|
s.append(&&Target::os()[0..2]);
|
||||||
s.out()
|
s.out()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Object can be locked directly into a `MutexGuard`.
|
|
||||||
pub trait Lockable<T> {
|
|
||||||
/// Lock object directly into a `MutexGuard`.
|
|
||||||
fn locked(&self) -> MutexGuard<T>;
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<T> Lockable<T> for Mutex<T> {
|
|
||||||
fn locked(&self) -> MutexGuard<T> { self.lock().unwrap() }
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Object can be read or write locked directly into a guard.
|
|
||||||
pub trait RwLockable<T> {
|
|
||||||
/// Read-lock object directly into a `ReadGuard`.
|
|
||||||
fn unwrapped_read(&self) -> RwLockReadGuard<T>;
|
|
||||||
|
|
||||||
/// Write-lock object directly into a `WriteGuard`.
|
|
||||||
fn unwrapped_write(&self) -> RwLockWriteGuard<T>;
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<T> RwLockable<T> for RwLock<T> {
|
|
||||||
fn unwrapped_read(&self) -> RwLockReadGuard<T> { self.read().unwrap() }
|
|
||||||
fn unwrapped_write(&self) -> RwLockWriteGuard<T> { self.write().unwrap() }
|
|
||||||
}
|
|
||||||
|
|
@ -96,13 +96,13 @@ impl<Socket: GenericSocket> GenericConnection<Socket> {
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
Ok(_) => return Ok(None),
|
Ok(_) => return Ok(None),
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
debug!(target:"network", "Read error {} ({})", self.token, e);
|
debug!(target:"network", "Read error {} ({})", self.token, e);
|
||||||
return Err(e)
|
return Err(e)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Add a packet to send queue.
|
/// Add a packet to send queue.
|
||||||
pub fn send<Message>(&mut self, io: &IoContext<Message>, data: Bytes) where Message: Send + Clone {
|
pub fn send<Message>(&mut self, io: &IoContext<Message>, data: Bytes) where Message: Send + Clone {
|
||||||
@ -490,7 +490,7 @@ pub fn test_encryption() {
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
use std::sync::*;
|
use std::sync::Arc;
|
||||||
use std::sync::atomic::AtomicBool;
|
use std::sync::atomic::AtomicBool;
|
||||||
use super::super::stats::*;
|
use super::super::stats::*;
|
||||||
use std::io::{Read, Write, Error, Cursor, ErrorKind};
|
use std::io::{Read, Write, Error, Cursor, ErrorKind};
|
||||||
|
@ -17,7 +17,7 @@
|
|||||||
use std::net::SocketAddr;
|
use std::net::SocketAddr;
|
||||||
use std::collections::{HashMap, HashSet};
|
use std::collections::{HashMap, HashSet};
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
use std::sync::*;
|
use std::sync::Arc;
|
||||||
use std::sync::atomic::{AtomicUsize, AtomicBool, Ordering as AtomicOrdering};
|
use std::sync::atomic::{AtomicUsize, AtomicBool, Ordering as AtomicOrdering};
|
||||||
use std::ops::*;
|
use std::ops::*;
|
||||||
use std::cmp::min;
|
use std::cmp::min;
|
||||||
@ -42,6 +42,7 @@ use network::error::{NetworkError, DisconnectReason};
|
|||||||
use network::discovery::{Discovery, TableUpdates, NodeEntry};
|
use network::discovery::{Discovery, TableUpdates, NodeEntry};
|
||||||
use network::ip_utils::{map_external_address, select_public_address};
|
use network::ip_utils::{map_external_address, select_public_address};
|
||||||
use path::restrict_permissions_owner;
|
use path::restrict_permissions_owner;
|
||||||
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
|
||||||
type Slab<T> = ::slab::Slab<T, usize>;
|
type Slab<T> = ::slab::Slab<T, usize>;
|
||||||
|
|
||||||
@ -201,7 +202,7 @@ impl<'s> NetworkContext<'s> {
|
|||||||
protocol: ProtocolId,
|
protocol: ProtocolId,
|
||||||
session: Option<SharedSession>, sessions: Arc<RwLock<Slab<SharedSession>>>,
|
session: Option<SharedSession>, sessions: Arc<RwLock<Slab<SharedSession>>>,
|
||||||
reserved_peers: &'s HashSet<NodeId>) -> NetworkContext<'s> {
|
reserved_peers: &'s HashSet<NodeId>) -> NetworkContext<'s> {
|
||||||
let id = session.as_ref().map(|s| s.locked().token());
|
let id = session.as_ref().map(|s| s.lock().token());
|
||||||
NetworkContext {
|
NetworkContext {
|
||||||
io: io,
|
io: io,
|
||||||
protocol: protocol,
|
protocol: protocol,
|
||||||
@ -215,7 +216,7 @@ impl<'s> NetworkContext<'s> {
|
|||||||
fn resolve_session(&self, peer: PeerId) -> Option<SharedSession> {
|
fn resolve_session(&self, peer: PeerId) -> Option<SharedSession> {
|
||||||
match self.session_id {
|
match self.session_id {
|
||||||
Some(id) if id == peer => self.session.clone(),
|
Some(id) if id == peer => self.session.clone(),
|
||||||
_ => self.sessions.unwrapped_read().get(peer).cloned(),
|
_ => self.sessions.read().get(peer).cloned(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -223,7 +224,7 @@ impl<'s> NetworkContext<'s> {
|
|||||||
pub fn send(&self, peer: PeerId, packet_id: PacketId, data: Vec<u8>) -> Result<(), UtilError> {
|
pub fn send(&self, peer: PeerId, packet_id: PacketId, data: Vec<u8>) -> Result<(), UtilError> {
|
||||||
let session = self.resolve_session(peer);
|
let session = self.resolve_session(peer);
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
try!(session.locked().send_packet(self.io, self.protocol, packet_id as u8, &data));
|
try!(session.lock().send_packet(self.io, self.protocol, packet_id as u8, &data));
|
||||||
} else {
|
} else {
|
||||||
trace!(target: "network", "Send: Peer no longer exist")
|
trace!(target: "network", "Send: Peer no longer exist")
|
||||||
}
|
}
|
||||||
@ -256,7 +257,7 @@ impl<'s> NetworkContext<'s> {
|
|||||||
|
|
||||||
/// Check if the session is still active.
|
/// Check if the session is still active.
|
||||||
pub fn is_expired(&self) -> bool {
|
pub fn is_expired(&self) -> bool {
|
||||||
self.session.as_ref().map_or(false, |s| s.locked().expired())
|
self.session.as_ref().map_or(false, |s| s.lock().expired())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Register a new IO timer. 'IoHandler::timeout' will be called with the token.
|
/// Register a new IO timer. 'IoHandler::timeout' will be called with the token.
|
||||||
@ -273,7 +274,7 @@ impl<'s> NetworkContext<'s> {
|
|||||||
pub fn peer_info(&self, peer: PeerId) -> String {
|
pub fn peer_info(&self, peer: PeerId) -> String {
|
||||||
let session = self.resolve_session(peer);
|
let session = self.resolve_session(peer);
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
return session.locked().info.client_version.clone()
|
return session.lock().info.client_version.clone()
|
||||||
}
|
}
|
||||||
"unknown".to_owned()
|
"unknown".to_owned()
|
||||||
}
|
}
|
||||||
@ -416,8 +417,8 @@ impl Host {
|
|||||||
Ok(n) => {
|
Ok(n) => {
|
||||||
let entry = NodeEntry { endpoint: n.endpoint.clone(), id: n.id.clone() };
|
let entry = NodeEntry { endpoint: n.endpoint.clone(), id: n.id.clone() };
|
||||||
|
|
||||||
self.nodes.unwrapped_write().add_node(n);
|
self.nodes.write().add_node(n);
|
||||||
if let Some(ref mut discovery) = *self.discovery.locked() {
|
if let Some(ref mut discovery) = *self.discovery.lock() {
|
||||||
discovery.add_node(entry);
|
discovery.add_node(entry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -428,9 +429,9 @@ impl Host {
|
|||||||
let n = try!(Node::from_str(id));
|
let n = try!(Node::from_str(id));
|
||||||
|
|
||||||
let entry = NodeEntry { endpoint: n.endpoint.clone(), id: n.id.clone() };
|
let entry = NodeEntry { endpoint: n.endpoint.clone(), id: n.id.clone() };
|
||||||
self.reserved_nodes.unwrapped_write().insert(n.id.clone());
|
self.reserved_nodes.write().insert(n.id.clone());
|
||||||
|
|
||||||
if let Some(ref mut discovery) = *self.discovery.locked() {
|
if let Some(ref mut discovery) = *self.discovery.lock() {
|
||||||
discovery.add_node(entry);
|
discovery.add_node(entry);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -438,17 +439,17 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn set_non_reserved_mode(&self, mode: NonReservedPeerMode, io: &IoContext<NetworkIoMessage>) {
|
pub fn set_non_reserved_mode(&self, mode: NonReservedPeerMode, io: &IoContext<NetworkIoMessage>) {
|
||||||
let mut info = self.info.unwrapped_write();
|
let mut info = self.info.write();
|
||||||
|
|
||||||
if info.config.non_reserved_mode != mode {
|
if info.config.non_reserved_mode != mode {
|
||||||
info.config.non_reserved_mode = mode.clone();
|
info.config.non_reserved_mode = mode.clone();
|
||||||
drop(info);
|
drop(info);
|
||||||
if let NonReservedPeerMode::Deny = mode {
|
if let NonReservedPeerMode::Deny = mode {
|
||||||
// disconnect all non-reserved peers here.
|
// disconnect all non-reserved peers here.
|
||||||
let reserved: HashSet<NodeId> = self.reserved_nodes.unwrapped_read().clone();
|
let reserved: HashSet<NodeId> = self.reserved_nodes.read().clone();
|
||||||
let mut to_kill = Vec::new();
|
let mut to_kill = Vec::new();
|
||||||
for e in self.sessions.unwrapped_write().iter_mut() {
|
for e in self.sessions.write().iter_mut() {
|
||||||
let mut s = e.locked();
|
let mut s = e.lock();
|
||||||
{
|
{
|
||||||
let id = s.id();
|
let id = s.id();
|
||||||
if id.is_some() && reserved.contains(id.unwrap()) {
|
if id.is_some() && reserved.contains(id.unwrap()) {
|
||||||
@ -469,7 +470,7 @@ impl Host {
|
|||||||
|
|
||||||
pub fn remove_reserved_node(&self, id: &str) -> Result<(), UtilError> {
|
pub fn remove_reserved_node(&self, id: &str) -> Result<(), UtilError> {
|
||||||
let n = try!(Node::from_str(id));
|
let n = try!(Node::from_str(id));
|
||||||
self.reserved_nodes.unwrapped_write().remove(&n.id);
|
self.reserved_nodes.write().remove(&n.id);
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@ -479,11 +480,11 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn external_url(&self) -> Option<String> {
|
pub fn external_url(&self) -> Option<String> {
|
||||||
self.info.unwrapped_read().public_endpoint.as_ref().map(|e| format!("{}", Node::new(self.info.unwrapped_read().id().clone(), e.clone())))
|
self.info.read().public_endpoint.as_ref().map(|e| format!("{}", Node::new(self.info.read().id().clone(), e.clone())))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn local_url(&self) -> String {
|
pub fn local_url(&self) -> String {
|
||||||
let r = format!("{}", Node::new(self.info.unwrapped_read().id().clone(), self.info.unwrapped_read().local_endpoint.clone()));
|
let r = format!("{}", Node::new(self.info.read().id().clone(), self.info.read().local_endpoint.clone()));
|
||||||
println!("{}", r);
|
println!("{}", r);
|
||||||
r
|
r
|
||||||
}
|
}
|
||||||
@ -491,8 +492,8 @@ impl Host {
|
|||||||
pub fn stop(&self, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
pub fn stop(&self, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
||||||
self.stopping.store(true, AtomicOrdering::Release);
|
self.stopping.store(true, AtomicOrdering::Release);
|
||||||
let mut to_kill = Vec::new();
|
let mut to_kill = Vec::new();
|
||||||
for e in self.sessions.unwrapped_write().iter_mut() {
|
for e in self.sessions.write().iter_mut() {
|
||||||
let mut s = e.locked();
|
let mut s = e.lock();
|
||||||
s.disconnect(io, DisconnectReason::ClientQuit);
|
s.disconnect(io, DisconnectReason::ClientQuit);
|
||||||
to_kill.push(s.token());
|
to_kill.push(s.token());
|
||||||
}
|
}
|
||||||
@ -505,16 +506,16 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn init_public_interface(&self, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
fn init_public_interface(&self, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
||||||
if self.info.unwrapped_read().public_endpoint.is_some() {
|
if self.info.read().public_endpoint.is_some() {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
let local_endpoint = self.info.unwrapped_read().local_endpoint.clone();
|
let local_endpoint = self.info.read().local_endpoint.clone();
|
||||||
let public_address = self.info.unwrapped_read().config.public_address.clone();
|
let public_address = self.info.read().config.public_address.clone();
|
||||||
let public_endpoint = match public_address {
|
let public_endpoint = match public_address {
|
||||||
None => {
|
None => {
|
||||||
let public_address = select_public_address(local_endpoint.address.port());
|
let public_address = select_public_address(local_endpoint.address.port());
|
||||||
let public_endpoint = NodeEndpoint { address: public_address, udp_port: local_endpoint.udp_port };
|
let public_endpoint = NodeEndpoint { address: public_address, udp_port: local_endpoint.udp_port };
|
||||||
if self.info.unwrapped_read().config.nat_enabled {
|
if self.info.read().config.nat_enabled {
|
||||||
match map_external_address(&local_endpoint) {
|
match map_external_address(&local_endpoint) {
|
||||||
Some(endpoint) => {
|
Some(endpoint) => {
|
||||||
info!("NAT mapped to external address {}", endpoint.address);
|
info!("NAT mapped to external address {}", endpoint.address);
|
||||||
@ -529,7 +530,7 @@ impl Host {
|
|||||||
Some(addr) => NodeEndpoint { address: addr, udp_port: local_endpoint.udp_port }
|
Some(addr) => NodeEndpoint { address: addr, udp_port: local_endpoint.udp_port }
|
||||||
};
|
};
|
||||||
|
|
||||||
self.info.unwrapped_write().public_endpoint = Some(public_endpoint.clone());
|
self.info.write().public_endpoint = Some(public_endpoint.clone());
|
||||||
|
|
||||||
if let Some(url) = self.external_url() {
|
if let Some(url) = self.external_url() {
|
||||||
io.message(NetworkIoMessage::NetworkStarted(url)).unwrap_or_else(|e| warn!("Error sending IO notification: {:?}", e));
|
io.message(NetworkIoMessage::NetworkStarted(url)).unwrap_or_else(|e| warn!("Error sending IO notification: {:?}", e));
|
||||||
@ -537,7 +538,7 @@ impl Host {
|
|||||||
|
|
||||||
// Initialize discovery.
|
// Initialize discovery.
|
||||||
let discovery = {
|
let discovery = {
|
||||||
let info = self.info.unwrapped_read();
|
let info = self.info.read();
|
||||||
if info.config.discovery_enabled && info.config.non_reserved_mode == NonReservedPeerMode::Accept {
|
if info.config.discovery_enabled && info.config.non_reserved_mode == NonReservedPeerMode::Accept {
|
||||||
let mut udp_addr = local_endpoint.address.clone();
|
let mut udp_addr = local_endpoint.address.clone();
|
||||||
udp_addr.set_port(local_endpoint.udp_port);
|
udp_addr.set_port(local_endpoint.udp_port);
|
||||||
@ -546,11 +547,11 @@ impl Host {
|
|||||||
};
|
};
|
||||||
|
|
||||||
if let Some(mut discovery) = discovery {
|
if let Some(mut discovery) = discovery {
|
||||||
discovery.init_node_list(self.nodes.unwrapped_read().unordered_entries());
|
discovery.init_node_list(self.nodes.read().unordered_entries());
|
||||||
for n in self.nodes.unwrapped_read().unordered_entries() {
|
for n in self.nodes.read().unordered_entries() {
|
||||||
discovery.add_node(n.clone());
|
discovery.add_node(n.clone());
|
||||||
}
|
}
|
||||||
*self.discovery.locked() = Some(discovery);
|
*self.discovery.lock() = Some(discovery);
|
||||||
io.register_stream(DISCOVERY).expect("Error registering UDP listener");
|
io.register_stream(DISCOVERY).expect("Error registering UDP listener");
|
||||||
io.register_timer(DISCOVERY_REFRESH, 7200).expect("Error registering discovery timer");
|
io.register_timer(DISCOVERY_REFRESH, 7200).expect("Error registering discovery timer");
|
||||||
io.register_timer(DISCOVERY_ROUND, 300).expect("Error registering discovery timer");
|
io.register_timer(DISCOVERY_ROUND, 300).expect("Error registering discovery timer");
|
||||||
@ -566,7 +567,7 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn have_session(&self, id: &NodeId) -> bool {
|
fn have_session(&self, id: &NodeId) -> bool {
|
||||||
self.sessions.unwrapped_read().iter().any(|e| e.locked().info.id == Some(id.clone()))
|
self.sessions.read().iter().any(|e| e.lock().info.id == Some(id.clone()))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn session_count(&self) -> usize {
|
fn session_count(&self) -> usize {
|
||||||
@ -574,17 +575,17 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn connecting_to(&self, id: &NodeId) -> bool {
|
fn connecting_to(&self, id: &NodeId) -> bool {
|
||||||
self.sessions.unwrapped_read().iter().any(|e| e.locked().id() == Some(id))
|
self.sessions.read().iter().any(|e| e.lock().id() == Some(id))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn handshake_count(&self) -> usize {
|
fn handshake_count(&self) -> usize {
|
||||||
self.sessions.unwrapped_read().count() - self.session_count()
|
self.sessions.read().count() - self.session_count()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn keep_alive(&self, io: &IoContext<NetworkIoMessage>) {
|
fn keep_alive(&self, io: &IoContext<NetworkIoMessage>) {
|
||||||
let mut to_kill = Vec::new();
|
let mut to_kill = Vec::new();
|
||||||
for e in self.sessions.unwrapped_write().iter_mut() {
|
for e in self.sessions.write().iter_mut() {
|
||||||
let mut s = e.locked();
|
let mut s = e.lock();
|
||||||
if !s.keep_alive(io) {
|
if !s.keep_alive(io) {
|
||||||
s.disconnect(io, DisconnectReason::PingTimeout);
|
s.disconnect(io, DisconnectReason::PingTimeout);
|
||||||
to_kill.push(s.token());
|
to_kill.push(s.token());
|
||||||
@ -598,7 +599,7 @@ impl Host {
|
|||||||
|
|
||||||
fn connect_peers(&self, io: &IoContext<NetworkIoMessage>) {
|
fn connect_peers(&self, io: &IoContext<NetworkIoMessage>) {
|
||||||
let (ideal_peers, mut pin) = {
|
let (ideal_peers, mut pin) = {
|
||||||
let info = self.info.unwrapped_read();
|
let info = self.info.read();
|
||||||
if info.capabilities.is_empty() {
|
if info.capabilities.is_empty() {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@ -608,7 +609,7 @@ impl Host {
|
|||||||
};
|
};
|
||||||
|
|
||||||
let session_count = self.session_count();
|
let session_count = self.session_count();
|
||||||
let reserved_nodes = self.reserved_nodes.unwrapped_read();
|
let reserved_nodes = self.reserved_nodes.read();
|
||||||
if session_count >= ideal_peers as usize + reserved_nodes.len() {
|
if session_count >= ideal_peers as usize + reserved_nodes.len() {
|
||||||
// check if all pinned nodes are connected.
|
// check if all pinned nodes are connected.
|
||||||
if reserved_nodes.iter().all(|n| self.have_session(n) && self.connecting_to(n)) {
|
if reserved_nodes.iter().all(|n| self.have_session(n) && self.connecting_to(n)) {
|
||||||
@ -629,7 +630,7 @@ impl Host {
|
|||||||
// iterate over all nodes, reserved ones coming first.
|
// iterate over all nodes, reserved ones coming first.
|
||||||
// if we are pinned to only reserved nodes, ignore all others.
|
// if we are pinned to only reserved nodes, ignore all others.
|
||||||
let nodes = reserved_nodes.iter().cloned().chain(if !pin {
|
let nodes = reserved_nodes.iter().cloned().chain(if !pin {
|
||||||
self.nodes.unwrapped_read().nodes()
|
self.nodes.read().nodes()
|
||||||
} else {
|
} else {
|
||||||
Vec::new()
|
Vec::new()
|
||||||
});
|
});
|
||||||
@ -657,7 +658,7 @@ impl Host {
|
|||||||
|
|
||||||
let socket = {
|
let socket = {
|
||||||
let address = {
|
let address = {
|
||||||
let mut nodes = self.nodes.unwrapped_write();
|
let mut nodes = self.nodes.write();
|
||||||
if let Some(node) = nodes.get_mut(id) {
|
if let Some(node) = nodes.get_mut(id) {
|
||||||
node.last_attempted = Some(::time::now());
|
node.last_attempted = Some(::time::now());
|
||||||
node.endpoint.address
|
node.endpoint.address
|
||||||
@ -682,11 +683,11 @@ impl Host {
|
|||||||
|
|
||||||
#[cfg_attr(feature="dev", allow(block_in_if_condition_stmt))]
|
#[cfg_attr(feature="dev", allow(block_in_if_condition_stmt))]
|
||||||
fn create_connection(&self, socket: TcpStream, id: Option<&NodeId>, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
fn create_connection(&self, socket: TcpStream, id: Option<&NodeId>, io: &IoContext<NetworkIoMessage>) -> Result<(), UtilError> {
|
||||||
let nonce = self.info.unwrapped_write().next_nonce();
|
let nonce = self.info.write().next_nonce();
|
||||||
let mut sessions = self.sessions.unwrapped_write();
|
let mut sessions = self.sessions.write();
|
||||||
|
|
||||||
let token = sessions.insert_with_opt(|token| {
|
let token = sessions.insert_with_opt(|token| {
|
||||||
match Session::new(io, socket, token, id, &nonce, self.stats.clone(), &self.info.unwrapped_read()) {
|
match Session::new(io, socket, token, id, &nonce, self.stats.clone(), &self.info.read()) {
|
||||||
Ok(s) => Some(Arc::new(Mutex::new(s))),
|
Ok(s) => Some(Arc::new(Mutex::new(s))),
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
debug!(target: "network", "Session create error: {:?}", e);
|
debug!(target: "network", "Session create error: {:?}", e);
|
||||||
@ -707,7 +708,7 @@ impl Host {
|
|||||||
fn accept(&self, io: &IoContext<NetworkIoMessage>) {
|
fn accept(&self, io: &IoContext<NetworkIoMessage>) {
|
||||||
trace!(target: "network", "Accepting incoming connection");
|
trace!(target: "network", "Accepting incoming connection");
|
||||||
loop {
|
loop {
|
||||||
let socket = match self.tcp_listener.locked().accept() {
|
let socket = match self.tcp_listener.lock().accept() {
|
||||||
Ok(None) => break,
|
Ok(None) => break,
|
||||||
Ok(Some((sock, _addr))) => sock,
|
Ok(Some((sock, _addr))) => sock,
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
@ -722,11 +723,11 @@ impl Host {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn session_writable(&self, token: StreamToken, io: &IoContext<NetworkIoMessage>) {
|
fn session_writable(&self, token: StreamToken, io: &IoContext<NetworkIoMessage>) {
|
||||||
let session = { self.sessions.unwrapped_read().get(token).cloned() };
|
let session = { self.sessions.read().get(token).cloned() };
|
||||||
|
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
let mut s = session.locked();
|
let mut s = session.lock();
|
||||||
if let Err(e) = s.writable(io, &self.info.unwrapped_read()) {
|
if let Err(e) = s.writable(io, &self.info.read()) {
|
||||||
trace!(target: "network", "Session write error: {}: {:?}", token, e);
|
trace!(target: "network", "Session write error: {}: {:?}", token, e);
|
||||||
}
|
}
|
||||||
if s.done() {
|
if s.done() {
|
||||||
@ -745,16 +746,16 @@ impl Host {
|
|||||||
let mut ready_data: Vec<ProtocolId> = Vec::new();
|
let mut ready_data: Vec<ProtocolId> = Vec::new();
|
||||||
let mut packet_data: Vec<(ProtocolId, PacketId, Vec<u8>)> = Vec::new();
|
let mut packet_data: Vec<(ProtocolId, PacketId, Vec<u8>)> = Vec::new();
|
||||||
let mut kill = false;
|
let mut kill = false;
|
||||||
let session = { self.sessions.unwrapped_read().get(token).cloned() };
|
let session = { self.sessions.read().get(token).cloned() };
|
||||||
if let Some(session) = session.clone() {
|
if let Some(session) = session.clone() {
|
||||||
let mut s = session.locked();
|
let mut s = session.lock();
|
||||||
loop {
|
loop {
|
||||||
match s.readable(io, &self.info.unwrapped_read()) {
|
match s.readable(io, &self.info.read()) {
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
trace!(target: "network", "Session read error: {}:{:?} ({:?}) {:?}", token, s.id(), s.remote_addr(), e);
|
trace!(target: "network", "Session read error: {}:{:?} ({:?}) {:?}", token, s.id(), s.remote_addr(), e);
|
||||||
if let UtilError::Network(NetworkError::Disconnect(DisconnectReason::IncompatibleProtocol)) = e {
|
if let UtilError::Network(NetworkError::Disconnect(DisconnectReason::IncompatibleProtocol)) = e {
|
||||||
if let Some(id) = s.id() {
|
if let Some(id) = s.id() {
|
||||||
self.nodes.unwrapped_write().mark_as_useless(id);
|
self.nodes.write().mark_as_useless(id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
kill = true;
|
kill = true;
|
||||||
@ -764,9 +765,9 @@ impl Host {
|
|||||||
self.num_sessions.fetch_add(1, AtomicOrdering::SeqCst);
|
self.num_sessions.fetch_add(1, AtomicOrdering::SeqCst);
|
||||||
if !s.info.originated {
|
if !s.info.originated {
|
||||||
let session_count = self.session_count();
|
let session_count = self.session_count();
|
||||||
let reserved_nodes = self.reserved_nodes.unwrapped_read();
|
let reserved_nodes = self.reserved_nodes.read();
|
||||||
let (ideal_peers, reserved_only) = {
|
let (ideal_peers, reserved_only) = {
|
||||||
let info = self.info.unwrapped_read();
|
let info = self.info.read();
|
||||||
(info.config.ideal_peers, info.config.non_reserved_mode == NonReservedPeerMode::Deny)
|
(info.config.ideal_peers, info.config.non_reserved_mode == NonReservedPeerMode::Deny)
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -781,14 +782,14 @@ impl Host {
|
|||||||
// Add it no node table
|
// Add it no node table
|
||||||
if let Ok(address) = s.remote_addr() {
|
if let Ok(address) = s.remote_addr() {
|
||||||
let entry = NodeEntry { id: s.id().unwrap().clone(), endpoint: NodeEndpoint { address: address, udp_port: address.port() } };
|
let entry = NodeEntry { id: s.id().unwrap().clone(), endpoint: NodeEndpoint { address: address, udp_port: address.port() } };
|
||||||
self.nodes.unwrapped_write().add_node(Node::new(entry.id.clone(), entry.endpoint.clone()));
|
self.nodes.write().add_node(Node::new(entry.id.clone(), entry.endpoint.clone()));
|
||||||
let mut discovery = self.discovery.locked();
|
let mut discovery = self.discovery.lock();
|
||||||
if let Some(ref mut discovery) = *discovery.deref_mut() {
|
if let Some(ref mut discovery) = *discovery.deref_mut() {
|
||||||
discovery.add_node(entry);
|
discovery.add_node(entry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for (p, _) in self.handlers.unwrapped_read().iter() {
|
for (p, _) in self.handlers.read().iter() {
|
||||||
if s.have_capability(p) {
|
if s.have_capability(p) {
|
||||||
ready_data.push(p);
|
ready_data.push(p);
|
||||||
}
|
}
|
||||||
@ -799,7 +800,7 @@ impl Host {
|
|||||||
protocol,
|
protocol,
|
||||||
packet_id,
|
packet_id,
|
||||||
}) => {
|
}) => {
|
||||||
match self.handlers.unwrapped_read().get(protocol) {
|
match self.handlers.read().get(protocol) {
|
||||||
None => { warn!(target: "network", "No handler found for protocol: {:?}", protocol) },
|
None => { warn!(target: "network", "No handler found for protocol: {:?}", protocol) },
|
||||||
Some(_) => packet_data.push((protocol, packet_id, data)),
|
Some(_) => packet_data.push((protocol, packet_id, data)),
|
||||||
}
|
}
|
||||||
@ -812,16 +813,16 @@ impl Host {
|
|||||||
if kill {
|
if kill {
|
||||||
self.kill_connection(token, io, true);
|
self.kill_connection(token, io, true);
|
||||||
}
|
}
|
||||||
let handlers = self.handlers.unwrapped_read();
|
let handlers = self.handlers.read();
|
||||||
for p in ready_data {
|
for p in ready_data {
|
||||||
let h = handlers.get(p).unwrap().clone();
|
let h = handlers.get(p).unwrap().clone();
|
||||||
self.stats.inc_sessions();
|
self.stats.inc_sessions();
|
||||||
let reserved = self.reserved_nodes.unwrapped_read();
|
let reserved = self.reserved_nodes.read();
|
||||||
h.connected(&NetworkContext::new(io, p, session.clone(), self.sessions.clone(), &reserved), &token);
|
h.connected(&NetworkContext::new(io, p, session.clone(), self.sessions.clone(), &reserved), &token);
|
||||||
}
|
}
|
||||||
for (p, packet_id, data) in packet_data {
|
for (p, packet_id, data) in packet_data {
|
||||||
let h = handlers.get(p).unwrap().clone();
|
let h = handlers.get(p).unwrap().clone();
|
||||||
let reserved = self.reserved_nodes.unwrapped_read();
|
let reserved = self.reserved_nodes.read();
|
||||||
h.read(&NetworkContext::new(io, p, session.clone(), self.sessions.clone(), &reserved), &token, packet_id, &data[1..]);
|
h.read(&NetworkContext::new(io, p, session.clone(), self.sessions.clone(), &reserved), &token, packet_id, &data[1..]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -837,14 +838,14 @@ impl Host {
|
|||||||
let mut deregister = false;
|
let mut deregister = false;
|
||||||
let mut expired_session = None;
|
let mut expired_session = None;
|
||||||
if let FIRST_SESSION ... LAST_SESSION = token {
|
if let FIRST_SESSION ... LAST_SESSION = token {
|
||||||
let sessions = self.sessions.unwrapped_write();
|
let sessions = self.sessions.write();
|
||||||
if let Some(session) = sessions.get(token).cloned() {
|
if let Some(session) = sessions.get(token).cloned() {
|
||||||
expired_session = Some(session.clone());
|
expired_session = Some(session.clone());
|
||||||
let mut s = session.locked();
|
let mut s = session.lock();
|
||||||
if !s.expired() {
|
if !s.expired() {
|
||||||
if s.is_ready() {
|
if s.is_ready() {
|
||||||
self.num_sessions.fetch_sub(1, AtomicOrdering::SeqCst);
|
self.num_sessions.fetch_sub(1, AtomicOrdering::SeqCst);
|
||||||
for (p, _) in self.handlers.unwrapped_read().iter() {
|
for (p, _) in self.handlers.read().iter() {
|
||||||
if s.have_capability(p) {
|
if s.have_capability(p) {
|
||||||
to_disconnect.push(p);
|
to_disconnect.push(p);
|
||||||
}
|
}
|
||||||
@ -858,12 +859,12 @@ impl Host {
|
|||||||
}
|
}
|
||||||
if let Some(id) = failure_id {
|
if let Some(id) = failure_id {
|
||||||
if remote {
|
if remote {
|
||||||
self.nodes.unwrapped_write().note_failure(&id);
|
self.nodes.write().note_failure(&id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for p in to_disconnect {
|
for p in to_disconnect {
|
||||||
let h = self.handlers.unwrapped_read().get(p).unwrap().clone();
|
let h = self.handlers.read().get(p).unwrap().clone();
|
||||||
let reserved = self.reserved_nodes.unwrapped_read();
|
let reserved = self.reserved_nodes.read();
|
||||||
h.disconnected(&NetworkContext::new(io, p, expired_session.clone(), self.sessions.clone(), &reserved), &token);
|
h.disconnected(&NetworkContext::new(io, p, expired_session.clone(), self.sessions.clone(), &reserved), &token);
|
||||||
}
|
}
|
||||||
if deregister {
|
if deregister {
|
||||||
@ -874,9 +875,9 @@ impl Host {
|
|||||||
fn update_nodes(&self, io: &IoContext<NetworkIoMessage>, node_changes: TableUpdates) {
|
fn update_nodes(&self, io: &IoContext<NetworkIoMessage>, node_changes: TableUpdates) {
|
||||||
let mut to_remove: Vec<PeerId> = Vec::new();
|
let mut to_remove: Vec<PeerId> = Vec::new();
|
||||||
{
|
{
|
||||||
let sessions = self.sessions.unwrapped_write();
|
let sessions = self.sessions.write();
|
||||||
for c in sessions.iter() {
|
for c in sessions.iter() {
|
||||||
let s = c.locked();
|
let s = c.lock();
|
||||||
if let Some(id) = s.id() {
|
if let Some(id) = s.id() {
|
||||||
if node_changes.removed.contains(id) {
|
if node_changes.removed.contains(id) {
|
||||||
to_remove.push(s.token());
|
to_remove.push(s.token());
|
||||||
@ -888,11 +889,11 @@ impl Host {
|
|||||||
trace!(target: "network", "Removed from node table: {}", i);
|
trace!(target: "network", "Removed from node table: {}", i);
|
||||||
self.kill_connection(i, io, false);
|
self.kill_connection(i, io, false);
|
||||||
}
|
}
|
||||||
self.nodes.unwrapped_write().update(node_changes);
|
self.nodes.write().update(node_changes);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn with_context<F>(&self, protocol: ProtocolId, io: &IoContext<NetworkIoMessage>, action: F) where F: Fn(&NetworkContext) {
|
pub fn with_context<F>(&self, protocol: ProtocolId, io: &IoContext<NetworkIoMessage>, action: F) where F: Fn(&NetworkContext) {
|
||||||
let reserved = { self.reserved_nodes.unwrapped_read() };
|
let reserved = { self.reserved_nodes.read() };
|
||||||
|
|
||||||
let context = NetworkContext::new(io, protocol, None, self.sessions.clone(), &reserved);
|
let context = NetworkContext::new(io, protocol, None, self.sessions.clone(), &reserved);
|
||||||
action(&context);
|
action(&context);
|
||||||
@ -922,7 +923,7 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
match stream {
|
match stream {
|
||||||
FIRST_SESSION ... LAST_SESSION => self.session_readable(stream, io),
|
FIRST_SESSION ... LAST_SESSION => self.session_readable(stream, io),
|
||||||
DISCOVERY => {
|
DISCOVERY => {
|
||||||
let node_changes = { self.discovery.locked().as_mut().unwrap().readable(io) };
|
let node_changes = { self.discovery.lock().as_mut().unwrap().readable(io) };
|
||||||
if let Some(node_changes) = node_changes {
|
if let Some(node_changes) = node_changes {
|
||||||
self.update_nodes(io, node_changes);
|
self.update_nodes(io, node_changes);
|
||||||
}
|
}
|
||||||
@ -939,7 +940,7 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
match stream {
|
match stream {
|
||||||
FIRST_SESSION ... LAST_SESSION => self.session_writable(stream, io),
|
FIRST_SESSION ... LAST_SESSION => self.session_writable(stream, io),
|
||||||
DISCOVERY => {
|
DISCOVERY => {
|
||||||
self.discovery.locked().as_mut().unwrap().writable(io);
|
self.discovery.lock().as_mut().unwrap().writable(io);
|
||||||
}
|
}
|
||||||
_ => panic!("Received unknown writable token"),
|
_ => panic!("Received unknown writable token"),
|
||||||
}
|
}
|
||||||
@ -953,11 +954,11 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
IDLE => self.maintain_network(io),
|
IDLE => self.maintain_network(io),
|
||||||
FIRST_SESSION ... LAST_SESSION => self.connection_timeout(token, io),
|
FIRST_SESSION ... LAST_SESSION => self.connection_timeout(token, io),
|
||||||
DISCOVERY_REFRESH => {
|
DISCOVERY_REFRESH => {
|
||||||
self.discovery.locked().as_mut().unwrap().refresh();
|
self.discovery.lock().as_mut().unwrap().refresh();
|
||||||
io.update_registration(DISCOVERY).unwrap_or_else(|e| debug!("Error updating discovery registration: {:?}", e));
|
io.update_registration(DISCOVERY).unwrap_or_else(|e| debug!("Error updating discovery registration: {:?}", e));
|
||||||
},
|
},
|
||||||
DISCOVERY_ROUND => {
|
DISCOVERY_ROUND => {
|
||||||
let node_changes = { self.discovery.locked().as_mut().unwrap().round() };
|
let node_changes = { self.discovery.lock().as_mut().unwrap().round() };
|
||||||
if let Some(node_changes) = node_changes {
|
if let Some(node_changes) = node_changes {
|
||||||
self.update_nodes(io, node_changes);
|
self.update_nodes(io, node_changes);
|
||||||
}
|
}
|
||||||
@ -965,13 +966,13 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
},
|
},
|
||||||
NODE_TABLE => {
|
NODE_TABLE => {
|
||||||
trace!(target: "network", "Refreshing node table");
|
trace!(target: "network", "Refreshing node table");
|
||||||
self.nodes.unwrapped_write().clear_useless();
|
self.nodes.write().clear_useless();
|
||||||
},
|
},
|
||||||
_ => match self.timers.unwrapped_read().get(&token).cloned() {
|
_ => match self.timers.read().get(&token).cloned() {
|
||||||
Some(timer) => match self.handlers.unwrapped_read().get(timer.protocol).cloned() {
|
Some(timer) => match self.handlers.read().get(timer.protocol).cloned() {
|
||||||
None => { warn!(target: "network", "No handler found for protocol: {:?}", timer.protocol) },
|
None => { warn!(target: "network", "No handler found for protocol: {:?}", timer.protocol) },
|
||||||
Some(h) => {
|
Some(h) => {
|
||||||
let reserved = self.reserved_nodes.unwrapped_read();
|
let reserved = self.reserved_nodes.read();
|
||||||
h.timeout(&NetworkContext::new(io, timer.protocol, None, self.sessions.clone(), &reserved), timer.token);
|
h.timeout(&NetworkContext::new(io, timer.protocol, None, self.sessions.clone(), &reserved), timer.token);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@ -991,10 +992,10 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
ref versions
|
ref versions
|
||||||
} => {
|
} => {
|
||||||
let h = handler.clone();
|
let h = handler.clone();
|
||||||
let reserved = self.reserved_nodes.unwrapped_read();
|
let reserved = self.reserved_nodes.read();
|
||||||
h.initialize(&NetworkContext::new(io, protocol, None, self.sessions.clone(), &reserved));
|
h.initialize(&NetworkContext::new(io, protocol, None, self.sessions.clone(), &reserved));
|
||||||
self.handlers.unwrapped_write().insert(protocol, h);
|
self.handlers.write().insert(protocol, h);
|
||||||
let mut info = self.info.unwrapped_write();
|
let mut info = self.info.write();
|
||||||
for v in versions {
|
for v in versions {
|
||||||
info.capabilities.push(CapabilityInfo { protocol: protocol, version: *v, packet_count:0 });
|
info.capabilities.push(CapabilityInfo { protocol: protocol, version: *v, packet_count:0 });
|
||||||
}
|
}
|
||||||
@ -1005,29 +1006,29 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
ref token,
|
ref token,
|
||||||
} => {
|
} => {
|
||||||
let handler_token = {
|
let handler_token = {
|
||||||
let mut timer_counter = self.timer_counter.unwrapped_write();
|
let mut timer_counter = self.timer_counter.write();
|
||||||
let counter = &mut *timer_counter;
|
let counter = &mut *timer_counter;
|
||||||
let handler_token = *counter;
|
let handler_token = *counter;
|
||||||
*counter += 1;
|
*counter += 1;
|
||||||
handler_token
|
handler_token
|
||||||
};
|
};
|
||||||
self.timers.unwrapped_write().insert(handler_token, ProtocolTimer { protocol: protocol, token: *token });
|
self.timers.write().insert(handler_token, ProtocolTimer { protocol: protocol, token: *token });
|
||||||
io.register_timer(handler_token, *delay).unwrap_or_else(|e| debug!("Error registering timer {}: {:?}", token, e));
|
io.register_timer(handler_token, *delay).unwrap_or_else(|e| debug!("Error registering timer {}: {:?}", token, e));
|
||||||
},
|
},
|
||||||
NetworkIoMessage::Disconnect(ref peer) => {
|
NetworkIoMessage::Disconnect(ref peer) => {
|
||||||
let session = { self.sessions.unwrapped_read().get(*peer).cloned() };
|
let session = { self.sessions.read().get(*peer).cloned() };
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
session.locked().disconnect(io, DisconnectReason::DisconnectRequested);
|
session.lock().disconnect(io, DisconnectReason::DisconnectRequested);
|
||||||
}
|
}
|
||||||
trace!(target: "network", "Disconnect requested {}", peer);
|
trace!(target: "network", "Disconnect requested {}", peer);
|
||||||
self.kill_connection(*peer, io, false);
|
self.kill_connection(*peer, io, false);
|
||||||
},
|
},
|
||||||
NetworkIoMessage::DisablePeer(ref peer) => {
|
NetworkIoMessage::DisablePeer(ref peer) => {
|
||||||
let session = { self.sessions.unwrapped_read().get(*peer).cloned() };
|
let session = { self.sessions.read().get(*peer).cloned() };
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
session.locked().disconnect(io, DisconnectReason::DisconnectRequested);
|
session.lock().disconnect(io, DisconnectReason::DisconnectRequested);
|
||||||
if let Some(id) = session.locked().id() {
|
if let Some(id) = session.lock().id() {
|
||||||
self.nodes.unwrapped_write().mark_as_useless(id)
|
self.nodes.write().mark_as_useless(id)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
trace!(target: "network", "Disabling peer {}", peer);
|
trace!(target: "network", "Disabling peer {}", peer);
|
||||||
@ -1042,13 +1043,13 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
fn register_stream(&self, stream: StreamToken, reg: Token, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
fn register_stream(&self, stream: StreamToken, reg: Token, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
||||||
match stream {
|
match stream {
|
||||||
FIRST_SESSION ... LAST_SESSION => {
|
FIRST_SESSION ... LAST_SESSION => {
|
||||||
let session = { self.sessions.unwrapped_read().get(stream).cloned() };
|
let session = { self.sessions.read().get(stream).cloned() };
|
||||||
if let Some(session) = session {
|
if let Some(session) = session {
|
||||||
session.locked().register_socket(reg, event_loop).expect("Error registering socket");
|
session.lock().register_socket(reg, event_loop).expect("Error registering socket");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
DISCOVERY => self.discovery.locked().as_ref().unwrap().register_socket(event_loop).expect("Error registering discovery socket"),
|
DISCOVERY => self.discovery.lock().as_ref().unwrap().register_socket(event_loop).expect("Error registering discovery socket"),
|
||||||
TCP_ACCEPT => event_loop.register(&*self.tcp_listener.locked(), Token(TCP_ACCEPT), EventSet::all(), PollOpt::edge()).expect("Error registering stream"),
|
TCP_ACCEPT => event_loop.register(&*self.tcp_listener.lock(), Token(TCP_ACCEPT), EventSet::all(), PollOpt::edge()).expect("Error registering stream"),
|
||||||
_ => warn!("Unexpected stream registration")
|
_ => warn!("Unexpected stream registration")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1056,9 +1057,9 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
fn deregister_stream(&self, stream: StreamToken, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
fn deregister_stream(&self, stream: StreamToken, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
||||||
match stream {
|
match stream {
|
||||||
FIRST_SESSION ... LAST_SESSION => {
|
FIRST_SESSION ... LAST_SESSION => {
|
||||||
let mut connections = self.sessions.unwrapped_write();
|
let mut connections = self.sessions.write();
|
||||||
if let Some(connection) = connections.get(stream).cloned() {
|
if let Some(connection) = connections.get(stream).cloned() {
|
||||||
connection.locked().deregister_socket(event_loop).expect("Error deregistering socket");
|
connection.lock().deregister_socket(event_loop).expect("Error deregistering socket");
|
||||||
connections.remove(stream);
|
connections.remove(stream);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1070,13 +1071,13 @@ impl IoHandler<NetworkIoMessage> for Host {
|
|||||||
fn update_stream(&self, stream: StreamToken, reg: Token, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
fn update_stream(&self, stream: StreamToken, reg: Token, event_loop: &mut EventLoop<IoManager<NetworkIoMessage>>) {
|
||||||
match stream {
|
match stream {
|
||||||
FIRST_SESSION ... LAST_SESSION => {
|
FIRST_SESSION ... LAST_SESSION => {
|
||||||
let connection = { self.sessions.unwrapped_read().get(stream).cloned() };
|
let connection = { self.sessions.read().get(stream).cloned() };
|
||||||
if let Some(connection) = connection {
|
if let Some(connection) = connection {
|
||||||
connection.locked().update_socket(reg, event_loop).expect("Error updating socket");
|
connection.lock().update_socket(reg, event_loop).expect("Error updating socket");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
DISCOVERY => self.discovery.locked().as_ref().unwrap().update_registration(event_loop).expect("Error reregistering discovery socket"),
|
DISCOVERY => self.discovery.lock().as_ref().unwrap().update_registration(event_loop).expect("Error reregistering discovery socket"),
|
||||||
TCP_ACCEPT => event_loop.reregister(&*self.tcp_listener.locked(), Token(TCP_ACCEPT), EventSet::all(), PollOpt::edge()).expect("Error reregistering stream"),
|
TCP_ACCEPT => event_loop.reregister(&*self.tcp_listener.lock(), Token(TCP_ACCEPT), EventSet::all(), PollOpt::edge()).expect("Error reregistering stream"),
|
||||||
_ => warn!("Unexpected stream update")
|
_ => warn!("Unexpected stream update")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -14,16 +14,18 @@
|
|||||||
// You should have received a copy of the GNU General Public License
|
// You should have received a copy of the GNU General Public License
|
||||||
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
use std::sync::*;
|
|
||||||
use error::*;
|
use error::*;
|
||||||
use panics::*;
|
use panics::*;
|
||||||
use misc::RwLockable;
|
|
||||||
use network::{NetworkProtocolHandler, NetworkConfiguration};
|
use network::{NetworkProtocolHandler, NetworkConfiguration};
|
||||||
use network::error::NetworkError;
|
use network::error::NetworkError;
|
||||||
use network::host::{Host, NetworkContext, NetworkIoMessage, ProtocolId};
|
use network::host::{Host, NetworkContext, NetworkIoMessage, ProtocolId};
|
||||||
use network::stats::NetworkStats;
|
use network::stats::NetworkStats;
|
||||||
use io::*;
|
use io::*;
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
|
||||||
/// IO Service with networking
|
/// IO Service with networking
|
||||||
/// `Message` defines a notification data type.
|
/// `Message` defines a notification data type.
|
||||||
pub struct NetworkService {
|
pub struct NetworkService {
|
||||||
@ -86,19 +88,19 @@ impl NetworkService {
|
|||||||
|
|
||||||
/// Returns external url if available.
|
/// Returns external url if available.
|
||||||
pub fn external_url(&self) -> Option<String> {
|
pub fn external_url(&self) -> Option<String> {
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
host.as_ref().and_then(|h| h.external_url())
|
host.as_ref().and_then(|h| h.external_url())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns external url if available.
|
/// Returns external url if available.
|
||||||
pub fn local_url(&self) -> Option<String> {
|
pub fn local_url(&self) -> Option<String> {
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
host.as_ref().map(|h| h.local_url())
|
host.as_ref().map(|h| h.local_url())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Start network IO
|
/// Start network IO
|
||||||
pub fn start(&self) -> Result<(), UtilError> {
|
pub fn start(&self) -> Result<(), UtilError> {
|
||||||
let mut host = self.host.unwrapped_write();
|
let mut host = self.host.write();
|
||||||
if host.is_none() {
|
if host.is_none() {
|
||||||
let h = Arc::new(try!(Host::new(self.config.clone(), self.stats.clone())));
|
let h = Arc::new(try!(Host::new(self.config.clone(), self.stats.clone())));
|
||||||
try!(self.io_service.register_handler(h.clone()));
|
try!(self.io_service.register_handler(h.clone()));
|
||||||
@ -109,7 +111,7 @@ impl NetworkService {
|
|||||||
|
|
||||||
/// Stop network IO
|
/// Stop network IO
|
||||||
pub fn stop(&self) -> Result<(), UtilError> {
|
pub fn stop(&self) -> Result<(), UtilError> {
|
||||||
let mut host = self.host.unwrapped_write();
|
let mut host = self.host.write();
|
||||||
if let Some(ref host) = *host {
|
if let Some(ref host) = *host {
|
||||||
let io = IoContext::new(self.io_service.channel(), 0); //TODO: take token id from host
|
let io = IoContext::new(self.io_service.channel(), 0); //TODO: take token id from host
|
||||||
try!(host.stop(&io));
|
try!(host.stop(&io));
|
||||||
@ -120,7 +122,7 @@ impl NetworkService {
|
|||||||
|
|
||||||
/// Try to add a reserved peer.
|
/// Try to add a reserved peer.
|
||||||
pub fn add_reserved_peer(&self, peer: &str) -> Result<(), UtilError> {
|
pub fn add_reserved_peer(&self, peer: &str) -> Result<(), UtilError> {
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
if let Some(ref host) = *host {
|
if let Some(ref host) = *host {
|
||||||
host.add_reserved_node(peer)
|
host.add_reserved_node(peer)
|
||||||
} else {
|
} else {
|
||||||
@ -130,7 +132,7 @@ impl NetworkService {
|
|||||||
|
|
||||||
/// Try to remove a reserved peer.
|
/// Try to remove a reserved peer.
|
||||||
pub fn remove_reserved_peer(&self, peer: &str) -> Result<(), UtilError> {
|
pub fn remove_reserved_peer(&self, peer: &str) -> Result<(), UtilError> {
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
if let Some(ref host) = *host {
|
if let Some(ref host) = *host {
|
||||||
host.remove_reserved_node(peer)
|
host.remove_reserved_node(peer)
|
||||||
} else {
|
} else {
|
||||||
@ -140,7 +142,7 @@ impl NetworkService {
|
|||||||
|
|
||||||
/// Set the non-reserved peer mode.
|
/// Set the non-reserved peer mode.
|
||||||
pub fn set_non_reserved_mode(&self, mode: ::network::NonReservedPeerMode) {
|
pub fn set_non_reserved_mode(&self, mode: ::network::NonReservedPeerMode) {
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
if let Some(ref host) = *host {
|
if let Some(ref host) = *host {
|
||||||
let io_ctxt = IoContext::new(self.io_service.channel(), 0);
|
let io_ctxt = IoContext::new(self.io_service.channel(), 0);
|
||||||
host.set_non_reserved_mode(mode, &io_ctxt);
|
host.set_non_reserved_mode(mode, &io_ctxt);
|
||||||
@ -150,7 +152,7 @@ impl NetworkService {
|
|||||||
/// Executes action in the network context
|
/// Executes action in the network context
|
||||||
pub fn with_context<F>(&self, protocol: ProtocolId, action: F) where F: Fn(&NetworkContext) {
|
pub fn with_context<F>(&self, protocol: ProtocolId, action: F) where F: Fn(&NetworkContext) {
|
||||||
let io = IoContext::new(self.io_service.channel(), 0);
|
let io = IoContext::new(self.io_service.channel(), 0);
|
||||||
let host = self.host.unwrapped_read();
|
let host = self.host.read();
|
||||||
if let Some(ref host) = host.as_ref() {
|
if let Some(ref host) = host.as_ref() {
|
||||||
host.with_context(protocol, &io, action);
|
host.with_context(protocol, &io, action);
|
||||||
};
|
};
|
||||||
|
@ -18,7 +18,6 @@ use std::sync::atomic::{AtomicBool, Ordering as AtomicOrdering};
|
|||||||
use std::thread;
|
use std::thread;
|
||||||
use std::time::*;
|
use std::time::*;
|
||||||
use common::*;
|
use common::*;
|
||||||
use misc::*;
|
|
||||||
use network::*;
|
use network::*;
|
||||||
use io::TimerToken;
|
use io::TimerToken;
|
||||||
use crypto::KeyPair;
|
use crypto::KeyPair;
|
||||||
@ -47,7 +46,7 @@ impl TestProtocol {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn got_packet(&self) -> bool {
|
pub fn got_packet(&self) -> bool {
|
||||||
self.packet.locked().deref()[..] == b"hello"[..]
|
self.packet.lock().deref()[..] == b"hello"[..]
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn got_timeout(&self) -> bool {
|
pub fn got_timeout(&self) -> bool {
|
||||||
@ -66,7 +65,7 @@ impl NetworkProtocolHandler for TestProtocol {
|
|||||||
|
|
||||||
fn read(&self, _io: &NetworkContext, _peer: &PeerId, packet_id: u8, data: &[u8]) {
|
fn read(&self, _io: &NetworkContext, _peer: &PeerId, packet_id: u8, data: &[u8]) {
|
||||||
assert_eq!(packet_id, 33);
|
assert_eq!(packet_id, 33);
|
||||||
self.packet.locked().extend(data);
|
self.packet.lock().extend(data);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn connected(&self, io: &NetworkContext, peer: &PeerId) {
|
fn connected(&self, io: &NetworkContext, peer: &PeerId) {
|
||||||
|
@ -18,9 +18,10 @@
|
|||||||
|
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::ops::DerefMut;
|
use std::ops::DerefMut;
|
||||||
use std::sync::{Arc, Mutex};
|
use std::sync::Arc;
|
||||||
use std::default::Default;
|
use std::default::Default;
|
||||||
use misc::Lockable;
|
|
||||||
|
use parking_lot::Mutex;
|
||||||
|
|
||||||
/// Thread-safe closure for handling possible panics
|
/// Thread-safe closure for handling possible panics
|
||||||
pub trait OnPanicListener: Send + Sync + 'static {
|
pub trait OnPanicListener: Send + Sync + 'static {
|
||||||
@ -89,7 +90,7 @@ impl PanicHandler {
|
|||||||
/// Notifies all listeners in case there is a panic.
|
/// Notifies all listeners in case there is a panic.
|
||||||
/// You should use `catch_panic` instead of calling this method explicitly.
|
/// You should use `catch_panic` instead of calling this method explicitly.
|
||||||
pub fn notify_all(&self, r: String) {
|
pub fn notify_all(&self, r: String) {
|
||||||
let mut listeners = self.listeners.locked();
|
let mut listeners = self.listeners.lock();
|
||||||
for listener in listeners.deref_mut() {
|
for listener in listeners.deref_mut() {
|
||||||
listener.call(&r);
|
listener.call(&r);
|
||||||
}
|
}
|
||||||
@ -98,7 +99,7 @@ impl PanicHandler {
|
|||||||
|
|
||||||
impl MayPanic for PanicHandler {
|
impl MayPanic for PanicHandler {
|
||||||
fn on_panic<F>(&self, closure: F) where F: OnPanicListener {
|
fn on_panic<F>(&self, closure: F) where F: OnPanicListener {
|
||||||
self.listeners.locked().push(Box::new(closure));
|
self.listeners.lock().push(Box::new(closure));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -119,50 +120,46 @@ impl<F> OnPanicListener for F
|
|||||||
#[test]
|
#[test]
|
||||||
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
||||||
fn should_notify_listeners_about_panic () {
|
fn should_notify_listeners_about_panic () {
|
||||||
use std::sync::RwLock;
|
use parking_lot::RwLock;
|
||||||
use misc::RwLockable;
|
|
||||||
// given
|
// given
|
||||||
let invocations = Arc::new(RwLock::new(vec![]));
|
let invocations = Arc::new(RwLock::new(vec![]));
|
||||||
let i = invocations.clone();
|
let i = invocations.clone();
|
||||||
let p = PanicHandler::new();
|
let p = PanicHandler::new();
|
||||||
p.on_panic(move |t| i.unwrapped_write().push(t));
|
p.on_panic(move |t| i.write().push(t));
|
||||||
|
|
||||||
// when
|
// when
|
||||||
p.catch_panic(|| panic!("Panic!")).unwrap_err();
|
p.catch_panic(|| panic!("Panic!")).unwrap_err();
|
||||||
|
|
||||||
// then
|
// then
|
||||||
assert!(invocations.unwrapped_read()[0] == "Panic!");
|
assert!(invocations.read()[0] == "Panic!");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
||||||
fn should_notify_listeners_about_panic_when_string_is_dynamic () {
|
fn should_notify_listeners_about_panic_when_string_is_dynamic () {
|
||||||
use std::sync::RwLock;
|
use parking_lot::RwLock;
|
||||||
use misc::RwLockable;
|
|
||||||
// given
|
// given
|
||||||
let invocations = Arc::new(RwLock::new(vec![]));
|
let invocations = Arc::new(RwLock::new(vec![]));
|
||||||
let i = invocations.clone();
|
let i = invocations.clone();
|
||||||
let p = PanicHandler::new();
|
let p = PanicHandler::new();
|
||||||
p.on_panic(move |t| i.unwrapped_write().push(t));
|
p.on_panic(move |t| i.write().push(t));
|
||||||
|
|
||||||
// when
|
// when
|
||||||
p.catch_panic(|| panic!("Panic: {}", 1)).unwrap_err();
|
p.catch_panic(|| panic!("Panic: {}", 1)).unwrap_err();
|
||||||
|
|
||||||
// then
|
// then
|
||||||
assert!(invocations.unwrapped_read()[0] == "Panic: 1");
|
assert!(invocations.read()[0] == "Panic: 1");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn should_notify_listeners_about_panic_in_other_thread () {
|
fn should_notify_listeners_about_panic_in_other_thread () {
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::sync::RwLock;
|
use parking_lot::RwLock;
|
||||||
use misc::RwLockable;
|
|
||||||
|
|
||||||
// given
|
// given
|
||||||
let invocations = Arc::new(RwLock::new(vec![]));
|
let invocations = Arc::new(RwLock::new(vec![]));
|
||||||
let i = invocations.clone();
|
let i = invocations.clone();
|
||||||
let p = PanicHandler::new();
|
let p = PanicHandler::new();
|
||||||
p.on_panic(move |t| i.unwrapped_write().push(t));
|
p.on_panic(move |t| i.write().push(t));
|
||||||
|
|
||||||
// when
|
// when
|
||||||
let t = thread::spawn(move ||
|
let t = thread::spawn(move ||
|
||||||
@ -171,20 +168,18 @@ fn should_notify_listeners_about_panic_in_other_thread () {
|
|||||||
t.join().unwrap_err();
|
t.join().unwrap_err();
|
||||||
|
|
||||||
// then
|
// then
|
||||||
assert!(invocations.unwrapped_read()[0] == "Panic!");
|
assert!(invocations.read()[0] == "Panic!");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
#[ignore] // panic forwarding doesnt work on the same thread in beta
|
||||||
fn should_forward_panics () {
|
fn should_forward_panics () {
|
||||||
use std::sync::RwLock;
|
use parking_lot::RwLock;
|
||||||
use misc::RwLockable;
|
|
||||||
|
|
||||||
// given
|
// given
|
||||||
let invocations = Arc::new(RwLock::new(vec![]));
|
let invocations = Arc::new(RwLock::new(vec![]));
|
||||||
let i = invocations.clone();
|
let i = invocations.clone();
|
||||||
let p = PanicHandler::new_in_arc();
|
let p = PanicHandler::new_in_arc();
|
||||||
p.on_panic(move |t| i.unwrapped_write().push(t));
|
p.on_panic(move |t| i.write().push(t));
|
||||||
|
|
||||||
let p2 = PanicHandler::new();
|
let p2 = PanicHandler::new();
|
||||||
p.forward_from(&p2);
|
p.forward_from(&p2);
|
||||||
@ -193,5 +188,5 @@ use std::sync::RwLock;
|
|||||||
p2.catch_panic(|| panic!("Panic!")).unwrap_err();
|
p2.catch_panic(|| panic!("Panic!")).unwrap_err();
|
||||||
|
|
||||||
// then
|
// then
|
||||||
assert!(invocations.unwrapped_read()[0] == "Panic!");
|
assert!(invocations.read()[0] == "Panic!");
|
||||||
}
|
}
|
||||||
|
@ -36,7 +36,7 @@ pub use std::error::Error as StdError;
|
|||||||
|
|
||||||
pub use std::ops::*;
|
pub use std::ops::*;
|
||||||
pub use std::cmp::*;
|
pub use std::cmp::*;
|
||||||
pub use std::sync::*;
|
pub use std::sync::Arc;
|
||||||
pub use std::cell::*;
|
pub use std::cell::*;
|
||||||
pub use std::collections::*;
|
pub use std::collections::*;
|
||||||
|
|
||||||
@ -46,3 +46,5 @@ pub use rustc_serialize::hex::{FromHex, FromHexError};
|
|||||||
|
|
||||||
pub use heapsize::HeapSizeOf;
|
pub use heapsize::HeapSizeOf;
|
||||||
pub use itertools::Itertools;
|
pub use itertools::Itertools;
|
||||||
|
|
||||||
|
pub use parking_lot::{Condvar, Mutex, MutexGuard, RwLock, RwLockReadGuard, RwLockWriteGuard};
|
Loading…
Reference in New Issue
Block a user