More performance optimizations (#1649)
* Use tree index for DB * Set uncles_hash, tx_root, receipts_root from verified block * Use Filth instead of a bool * Fix empty root check * Flush block queue properly * Expunge deref
This commit is contained in:
parent
459dcbcef1
commit
4e447ccc68
@ -36,7 +36,7 @@ pub struct Account {
|
|||||||
// Code cache of the account.
|
// Code cache of the account.
|
||||||
code_cache: Bytes,
|
code_cache: Bytes,
|
||||||
// Account is new or has been modified
|
// Account is new or has been modified
|
||||||
dirty: bool,
|
filth: Filth,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Account {
|
impl Account {
|
||||||
@ -50,7 +50,7 @@ impl Account {
|
|||||||
storage_overlay: RefCell::new(storage.into_iter().map(|(k, v)| (k, (Filth::Dirty, v))).collect()),
|
storage_overlay: RefCell::new(storage.into_iter().map(|(k, v)| (k, (Filth::Dirty, v))).collect()),
|
||||||
code_hash: Some(code.sha3()),
|
code_hash: Some(code.sha3()),
|
||||||
code_cache: code,
|
code_cache: code,
|
||||||
dirty: true,
|
filth: Filth::Dirty,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -63,7 +63,7 @@ impl Account {
|
|||||||
storage_overlay: RefCell::new(pod.storage.into_iter().map(|(k, v)| (k, (Filth::Dirty, v))).collect()),
|
storage_overlay: RefCell::new(pod.storage.into_iter().map(|(k, v)| (k, (Filth::Dirty, v))).collect()),
|
||||||
code_hash: Some(pod.code.sha3()),
|
code_hash: Some(pod.code.sha3()),
|
||||||
code_cache: pod.code,
|
code_cache: pod.code,
|
||||||
dirty: true,
|
filth: Filth::Dirty,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -76,7 +76,7 @@ impl Account {
|
|||||||
storage_overlay: RefCell::new(HashMap::new()),
|
storage_overlay: RefCell::new(HashMap::new()),
|
||||||
code_hash: Some(SHA3_EMPTY),
|
code_hash: Some(SHA3_EMPTY),
|
||||||
code_cache: vec![],
|
code_cache: vec![],
|
||||||
dirty: true,
|
filth: Filth::Dirty,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -90,7 +90,7 @@ impl Account {
|
|||||||
storage_overlay: RefCell::new(HashMap::new()),
|
storage_overlay: RefCell::new(HashMap::new()),
|
||||||
code_hash: Some(r.val_at(3)),
|
code_hash: Some(r.val_at(3)),
|
||||||
code_cache: vec![],
|
code_cache: vec![],
|
||||||
dirty: false,
|
filth: Filth::Clean,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -104,7 +104,7 @@ impl Account {
|
|||||||
storage_overlay: RefCell::new(HashMap::new()),
|
storage_overlay: RefCell::new(HashMap::new()),
|
||||||
code_hash: None,
|
code_hash: None,
|
||||||
code_cache: vec![],
|
code_cache: vec![],
|
||||||
dirty: true,
|
filth: Filth::Dirty,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -113,7 +113,7 @@ impl Account {
|
|||||||
pub fn init_code(&mut self, code: Bytes) {
|
pub fn init_code(&mut self, code: Bytes) {
|
||||||
assert!(self.code_hash.is_none());
|
assert!(self.code_hash.is_none());
|
||||||
self.code_cache = code;
|
self.code_cache = code;
|
||||||
self.dirty = true;
|
self.filth = Filth::Dirty;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Reset this account's code to the given code.
|
/// Reset this account's code to the given code.
|
||||||
@ -125,7 +125,7 @@ impl Account {
|
|||||||
/// Set (and cache) the contents of the trie's storage at `key` to `value`.
|
/// Set (and cache) the contents of the trie's storage at `key` to `value`.
|
||||||
pub fn set_storage(&mut self, key: H256, value: H256) {
|
pub fn set_storage(&mut self, key: H256, value: H256) {
|
||||||
self.storage_overlay.borrow_mut().insert(key, (Filth::Dirty, value));
|
self.storage_overlay.borrow_mut().insert(key, (Filth::Dirty, value));
|
||||||
self.dirty = true;
|
self.filth = Filth::Dirty;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get (and cache) the contents of the trie's storage at `key`.
|
/// Get (and cache) the contents of the trie's storage at `key`.
|
||||||
@ -183,7 +183,7 @@ impl Account {
|
|||||||
|
|
||||||
/// Is this a new or modified account?
|
/// Is this a new or modified account?
|
||||||
pub fn is_dirty(&self) -> bool {
|
pub fn is_dirty(&self) -> bool {
|
||||||
self.dirty
|
self.filth == Filth::Dirty
|
||||||
}
|
}
|
||||||
/// Provide a database to get `code_hash`. Should not be called if it is a contract without code.
|
/// Provide a database to get `code_hash`. Should not be called if it is a contract without code.
|
||||||
pub fn cache_code(&mut self, db: &AccountDB) -> bool {
|
pub fn cache_code(&mut self, db: &AccountDB) -> bool {
|
||||||
@ -216,13 +216,13 @@ impl Account {
|
|||||||
/// Increment the nonce of the account by one.
|
/// Increment the nonce of the account by one.
|
||||||
pub fn inc_nonce(&mut self) {
|
pub fn inc_nonce(&mut self) {
|
||||||
self.nonce = self.nonce + U256::from(1u8);
|
self.nonce = self.nonce + U256::from(1u8);
|
||||||
self.dirty = true;
|
self.filth = Filth::Dirty;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Increment the nonce of the account by one.
|
/// Increment the nonce of the account by one.
|
||||||
pub fn add_balance(&mut self, x: &U256) {
|
pub fn add_balance(&mut self, x: &U256) {
|
||||||
self.balance = self.balance + *x;
|
self.balance = self.balance + *x;
|
||||||
self.dirty = true;
|
self.filth = Filth::Dirty;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Increment the nonce of the account by one.
|
/// Increment the nonce of the account by one.
|
||||||
@ -230,7 +230,7 @@ impl Account {
|
|||||||
pub fn sub_balance(&mut self, x: &U256) {
|
pub fn sub_balance(&mut self, x: &U256) {
|
||||||
assert!(self.balance >= *x);
|
assert!(self.balance >= *x);
|
||||||
self.balance = self.balance - *x;
|
self.balance = self.balance - *x;
|
||||||
self.dirty = true;
|
self.filth = Filth::Dirty;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Commit the `storage_overlay` to the backing DB and update `storage_root`.
|
/// Commit the `storage_overlay` to the backing DB and update `storage_root`.
|
||||||
|
@ -275,6 +275,15 @@ impl<'x> OpenBlock<'x> {
|
|||||||
/// Alter the gas limit for the block.
|
/// Alter the gas limit for the block.
|
||||||
pub fn set_gas_used(&mut self, a: U256) { self.block.base.header.set_gas_used(a); }
|
pub fn set_gas_used(&mut self, a: U256) { self.block.base.header.set_gas_used(a); }
|
||||||
|
|
||||||
|
/// Alter the uncles hash the block.
|
||||||
|
pub fn set_uncles_hash(&mut self, h: H256) { self.block.base.header.set_uncles_hash(h); }
|
||||||
|
|
||||||
|
/// Alter transactions root for the block.
|
||||||
|
pub fn set_transactions_root(&mut self, h: H256) { self.block.base.header.set_transactions_root(h); }
|
||||||
|
|
||||||
|
/// Alter the receipts root for the block.
|
||||||
|
pub fn set_receipts_root(&mut self, h: H256) { self.block.base.header.set_receipts_root(h); }
|
||||||
|
|
||||||
/// Alter the extra_data for the block.
|
/// Alter the extra_data for the block.
|
||||||
pub fn set_extra_data(&mut self, extra_data: Bytes) -> Result<(), BlockError> {
|
pub fn set_extra_data(&mut self, extra_data: Bytes) -> Result<(), BlockError> {
|
||||||
if extra_data.len() > self.engine.maximum_extra_data_size() {
|
if extra_data.len() > self.engine.maximum_extra_data_size() {
|
||||||
@ -365,11 +374,17 @@ impl<'x> OpenBlock<'x> {
|
|||||||
let mut s = self;
|
let mut s = self;
|
||||||
|
|
||||||
s.engine.on_close_block(&mut s.block);
|
s.engine.on_close_block(&mut s.block);
|
||||||
|
if s.block.base.header.transactions_root.is_zero() || s.block.base.header.transactions_root == SHA3_NULL_RLP {
|
||||||
s.block.base.header.transactions_root = ordered_trie_root(s.block.base.transactions.iter().map(|ref e| e.rlp_bytes().to_vec()).collect());
|
s.block.base.header.transactions_root = ordered_trie_root(s.block.base.transactions.iter().map(|ref e| e.rlp_bytes().to_vec()).collect());
|
||||||
|
}
|
||||||
let uncle_bytes = s.block.base.uncles.iter().fold(RlpStream::new_list(s.block.base.uncles.len()), |mut s, u| {s.append_raw(&u.rlp(Seal::With), 1); s} ).out();
|
let uncle_bytes = s.block.base.uncles.iter().fold(RlpStream::new_list(s.block.base.uncles.len()), |mut s, u| {s.append_raw(&u.rlp(Seal::With), 1); s} ).out();
|
||||||
|
if s.block.base.header.uncles_hash.is_zero() {
|
||||||
s.block.base.header.uncles_hash = uncle_bytes.sha3();
|
s.block.base.header.uncles_hash = uncle_bytes.sha3();
|
||||||
s.block.base.header.state_root = s.block.state.root().clone();
|
}
|
||||||
|
if s.block.base.header.receipts_root.is_zero() || s.block.base.header.receipts_root == SHA3_NULL_RLP {
|
||||||
s.block.base.header.receipts_root = ordered_trie_root(s.block.receipts.iter().map(|ref r| r.rlp_bytes().to_vec()).collect());
|
s.block.base.header.receipts_root = ordered_trie_root(s.block.receipts.iter().map(|ref r| r.rlp_bytes().to_vec()).collect());
|
||||||
|
}
|
||||||
|
s.block.base.header.state_root = s.block.state.root().clone();
|
||||||
s.block.base.header.log_bloom = s.block.receipts.iter().fold(LogBloom::zero(), |mut b, r| {b = &b | &r.log_bloom; b}); //TODO: use |= operator
|
s.block.base.header.log_bloom = s.block.receipts.iter().fold(LogBloom::zero(), |mut b, r| {b = &b | &r.log_bloom; b}); //TODO: use |= operator
|
||||||
s.block.base.header.gas_used = s.block.receipts.last().map_or(U256::zero(), |r| r.gas_used);
|
s.block.base.header.gas_used = s.block.receipts.last().map_or(U256::zero(), |r| r.gas_used);
|
||||||
s.block.base.header.note_dirty();
|
s.block.base.header.note_dirty();
|
||||||
@ -500,6 +515,9 @@ pub fn enact(
|
|||||||
b.set_timestamp(header.timestamp());
|
b.set_timestamp(header.timestamp());
|
||||||
b.set_author(header.author().clone());
|
b.set_author(header.author().clone());
|
||||||
b.set_extra_data(header.extra_data().clone()).unwrap_or_else(|e| warn!("Couldn't set extradata: {}. Ignoring.", e));
|
b.set_extra_data(header.extra_data().clone()).unwrap_or_else(|e| warn!("Couldn't set extradata: {}. Ignoring.", e));
|
||||||
|
b.set_uncles_hash(header.uncles_hash().clone());
|
||||||
|
b.set_transactions_root(header.transactions_root().clone());
|
||||||
|
b.set_receipts_root(header.receipts_root().clone());
|
||||||
for t in transactions { try!(b.push_transaction(t.clone(), None)); }
|
for t in transactions { try!(b.push_transaction(t.clone(), None)); }
|
||||||
for u in uncles { try!(b.push_uncle(u.clone())); }
|
for u in uncles { try!(b.push_uncle(u.clone())); }
|
||||||
Ok(b.close_and_lock())
|
Ok(b.close_and_lock())
|
||||||
|
@ -445,8 +445,8 @@ impl BlockChain {
|
|||||||
let mut from_branch = vec![];
|
let mut from_branch = vec![];
|
||||||
let mut to_branch = vec![];
|
let mut to_branch = vec![];
|
||||||
|
|
||||||
let mut from_details = self.block_details(&from).expect(&format!("0. Expected to find details for block {:?}", from));
|
let mut from_details = self.block_details(&from).unwrap_or_else(|| panic!("0. Expected to find details for block {:?}", from));
|
||||||
let mut to_details = self.block_details(&to).expect(&format!("1. Expected to find details for block {:?}", to));
|
let mut to_details = self.block_details(&to).unwrap_or_else(|| panic!("1. Expected to find details for block {:?}", to));
|
||||||
let mut current_from = from;
|
let mut current_from = from;
|
||||||
let mut current_to = to;
|
let mut current_to = to;
|
||||||
|
|
||||||
@ -454,13 +454,13 @@ impl BlockChain {
|
|||||||
while from_details.number > to_details.number {
|
while from_details.number > to_details.number {
|
||||||
from_branch.push(current_from);
|
from_branch.push(current_from);
|
||||||
current_from = from_details.parent.clone();
|
current_from = from_details.parent.clone();
|
||||||
from_details = self.block_details(&from_details.parent).expect(&format!("2. Expected to find details for block {:?}", from_details.parent));
|
from_details = self.block_details(&from_details.parent).unwrap_or_else(|| panic!("2. Expected to find details for block {:?}", from_details.parent));
|
||||||
}
|
}
|
||||||
|
|
||||||
while to_details.number > from_details.number {
|
while to_details.number > from_details.number {
|
||||||
to_branch.push(current_to);
|
to_branch.push(current_to);
|
||||||
current_to = to_details.parent.clone();
|
current_to = to_details.parent.clone();
|
||||||
to_details = self.block_details(&to_details.parent).expect(&format!("3. Expected to find details for block {:?}", to_details.parent));
|
to_details = self.block_details(&to_details.parent).unwrap_or_else(|| panic!("3. Expected to find details for block {:?}", to_details.parent));
|
||||||
}
|
}
|
||||||
|
|
||||||
assert_eq!(from_details.number, to_details.number);
|
assert_eq!(from_details.number, to_details.number);
|
||||||
@ -469,11 +469,11 @@ impl BlockChain {
|
|||||||
while current_from != current_to {
|
while current_from != current_to {
|
||||||
from_branch.push(current_from);
|
from_branch.push(current_from);
|
||||||
current_from = from_details.parent.clone();
|
current_from = from_details.parent.clone();
|
||||||
from_details = self.block_details(&from_details.parent).expect(&format!("4. Expected to find details for block {:?}", from_details.parent));
|
from_details = self.block_details(&from_details.parent).unwrap_or_else(|| panic!("4. Expected to find details for block {:?}", from_details.parent));
|
||||||
|
|
||||||
to_branch.push(current_to);
|
to_branch.push(current_to);
|
||||||
current_to = to_details.parent.clone();
|
current_to = to_details.parent.clone();
|
||||||
to_details = self.block_details(&to_details.parent).expect(&format!("5. Expected to find details for block {:?}", from_details.parent));
|
to_details = self.block_details(&to_details.parent).unwrap_or_else(|| panic!("5. Expected to find details for block {:?}", from_details.parent));
|
||||||
}
|
}
|
||||||
|
|
||||||
let index = from_branch.len();
|
let index = from_branch.len();
|
||||||
@ -613,7 +613,7 @@ impl BlockChain {
|
|||||||
let hash = block.sha3();
|
let hash = block.sha3();
|
||||||
let number = header.number();
|
let number = header.number();
|
||||||
let parent_hash = header.parent_hash();
|
let parent_hash = header.parent_hash();
|
||||||
let parent_details = self.block_details(&parent_hash).expect(format!("Invalid parent hash: {:?}", parent_hash).as_ref());
|
let parent_details = self.block_details(&parent_hash).unwrap_or_else(|| panic!("Invalid parent hash: {:?}", parent_hash));
|
||||||
let total_difficulty = parent_details.total_difficulty + header.difficulty();
|
let total_difficulty = parent_details.total_difficulty + header.difficulty();
|
||||||
let is_new_best = total_difficulty > self.best_block_total_difficulty();
|
let is_new_best = total_difficulty > self.best_block_total_difficulty();
|
||||||
|
|
||||||
@ -682,7 +682,7 @@ impl BlockChain {
|
|||||||
let parent_hash = header.parent_hash();
|
let parent_hash = header.parent_hash();
|
||||||
|
|
||||||
// update parent
|
// update parent
|
||||||
let mut parent_details = self.block_details(&parent_hash).expect(format!("Invalid parent hash: {:?}", parent_hash).as_ref());
|
let mut parent_details = self.block_details(&parent_hash).unwrap_or_else(|| panic!("Invalid parent hash: {:?}", parent_hash));
|
||||||
parent_details.children.push(info.hash.clone());
|
parent_details.children.push(info.hash.clone());
|
||||||
|
|
||||||
// create current block details
|
// create current block details
|
||||||
|
@ -252,6 +252,9 @@ impl Client {
|
|||||||
/// Flush the block import queue.
|
/// Flush the block import queue.
|
||||||
pub fn flush_queue(&self) {
|
pub fn flush_queue(&self) {
|
||||||
self.block_queue.flush();
|
self.block_queue.flush();
|
||||||
|
while !self.block_queue.queue_info().is_empty() {
|
||||||
|
self.import_verified_blocks(&IoChannel::disconnected());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_last_hashes(&self, parent_hash: H256) -> LastHashes {
|
fn build_last_hashes(&self, parent_hash: H256) -> LastHashes {
|
||||||
|
@ -18,7 +18,7 @@
|
|||||||
|
|
||||||
use util::*;
|
use util::*;
|
||||||
use basic_types::*;
|
use basic_types::*;
|
||||||
use time::now_utc;
|
use time::get_time;
|
||||||
|
|
||||||
/// Type for Block number
|
/// Type for Block number
|
||||||
pub type BlockNumber = u64;
|
pub type BlockNumber = u64;
|
||||||
@ -137,6 +137,10 @@ impl Header {
|
|||||||
pub fn state_root(&self) -> &H256 { &self.state_root }
|
pub fn state_root(&self) -> &H256 { &self.state_root }
|
||||||
/// Get the receipts root field of the header.
|
/// Get the receipts root field of the header.
|
||||||
pub fn receipts_root(&self) -> &H256 { &self.receipts_root }
|
pub fn receipts_root(&self) -> &H256 { &self.receipts_root }
|
||||||
|
/// Get the transactions root field of the header.
|
||||||
|
pub fn transactions_root(&self) -> &H256 { &self.transactions_root }
|
||||||
|
/// Get the uncles hash field of the header.
|
||||||
|
pub fn uncles_hash(&self) -> &H256 { &self.uncles_hash }
|
||||||
/// Get the gas limit field of the header.
|
/// Get the gas limit field of the header.
|
||||||
pub fn gas_limit(&self) -> &U256 { &self.gas_limit }
|
pub fn gas_limit(&self) -> &U256 { &self.gas_limit }
|
||||||
|
|
||||||
@ -162,7 +166,7 @@ impl Header {
|
|||||||
/// Set the timestamp field of the header.
|
/// Set the timestamp field of the header.
|
||||||
pub fn set_timestamp(&mut self, a: u64) { self.timestamp = a; self.note_dirty(); }
|
pub fn set_timestamp(&mut self, a: u64) { self.timestamp = a; self.note_dirty(); }
|
||||||
/// Set the timestamp field of the header to the current time.
|
/// Set the timestamp field of the header to the current time.
|
||||||
pub fn set_timestamp_now(&mut self, but_later_than: u64) { self.timestamp = max(now_utc().to_timespec().sec as u64, but_later_than + 1); self.note_dirty(); }
|
pub fn set_timestamp_now(&mut self, but_later_than: u64) { self.timestamp = max(get_time().sec as u64, but_later_than + 1); self.note_dirty(); }
|
||||||
/// Set the number field of the header.
|
/// Set the number field of the header.
|
||||||
pub fn set_number(&mut self, a: BlockNumber) { self.number = a; self.note_dirty(); }
|
pub fn set_number(&mut self, a: BlockNumber) { self.number = a; self.note_dirty(); }
|
||||||
/// Set the author field of the header.
|
/// Set the author field of the header.
|
||||||
|
@ -447,7 +447,7 @@ fn execute_import(conf: Configuration, panic_handler: Arc<PanicHandler>) {
|
|||||||
Err(BlockImportError::Import(ImportError::AlreadyInChain)) => { trace!("Skipping block already in chain."); }
|
Err(BlockImportError::Import(ImportError::AlreadyInChain)) => { trace!("Skipping block already in chain."); }
|
||||||
Err(e) => die!("Cannot import block: {:?}", e)
|
Err(e) => die!("Cannot import block: {:?}", e)
|
||||||
}
|
}
|
||||||
informant.tick(client.deref(), None);
|
informant.tick(&*client, None);
|
||||||
};
|
};
|
||||||
|
|
||||||
match format {
|
match format {
|
||||||
@ -473,6 +473,10 @@ fn execute_import(conf: Configuration, panic_handler: Arc<PanicHandler>) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
while !client.queue_info().is_empty() {
|
||||||
|
sleep(Duration::from_secs(1));
|
||||||
|
informant.tick(&*client, None);
|
||||||
|
}
|
||||||
client.flush_queue();
|
client.flush_queue();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -49,8 +49,7 @@ pub struct ArchiveDB {
|
|||||||
impl ArchiveDB {
|
impl ArchiveDB {
|
||||||
/// Create a new instance from file
|
/// Create a new instance from file
|
||||||
pub fn new(path: &str, config: DatabaseConfig) -> ArchiveDB {
|
pub fn new(path: &str, config: DatabaseConfig) -> ArchiveDB {
|
||||||
let opts = config.prefix(DB_PREFIX_LEN);
|
let backing = Database::open(&config, path).unwrap_or_else(|e| {
|
||||||
let backing = Database::open(&opts, path).unwrap_or_else(|e| {
|
|
||||||
panic!("Error opening state db: {}", e);
|
panic!("Error opening state db: {}", e);
|
||||||
});
|
});
|
||||||
if !backing.is_empty() {
|
if !backing.is_empty() {
|
||||||
|
@ -74,8 +74,7 @@ const PADDING : [u8; 10] = [ 0u8; 10 ];
|
|||||||
impl EarlyMergeDB {
|
impl EarlyMergeDB {
|
||||||
/// Create a new instance from file
|
/// Create a new instance from file
|
||||||
pub fn new(path: &str, config: DatabaseConfig) -> EarlyMergeDB {
|
pub fn new(path: &str, config: DatabaseConfig) -> EarlyMergeDB {
|
||||||
let opts = config.prefix(DB_PREFIX_LEN);
|
let backing = Database::open(&config, path).unwrap_or_else(|e| {
|
||||||
let backing = Database::open(&opts, path).unwrap_or_else(|e| {
|
|
||||||
panic!("Error opening state db: {}", e);
|
panic!("Error opening state db: {}", e);
|
||||||
});
|
});
|
||||||
if !backing.is_empty() {
|
if !backing.is_empty() {
|
||||||
|
@ -104,8 +104,7 @@ impl OverlayRecentDB {
|
|||||||
|
|
||||||
/// Create a new instance from file
|
/// Create a new instance from file
|
||||||
pub fn from_prefs(path: &str, config: DatabaseConfig) -> OverlayRecentDB {
|
pub fn from_prefs(path: &str, config: DatabaseConfig) -> OverlayRecentDB {
|
||||||
let opts = config.prefix(DB_PREFIX_LEN);
|
let backing = Database::open(&config, path).unwrap_or_else(|e| {
|
||||||
let backing = Database::open(&opts, path).unwrap_or_else(|e| {
|
|
||||||
panic!("Error opening state db: {}", e);
|
panic!("Error opening state db: {}", e);
|
||||||
});
|
});
|
||||||
if !backing.is_empty() {
|
if !backing.is_empty() {
|
||||||
|
@ -47,8 +47,7 @@ const PADDING : [u8; 10] = [ 0u8; 10 ];
|
|||||||
impl RefCountedDB {
|
impl RefCountedDB {
|
||||||
/// Create a new instance given a `backing` database.
|
/// Create a new instance given a `backing` database.
|
||||||
pub fn new(path: &str, config: DatabaseConfig) -> RefCountedDB {
|
pub fn new(path: &str, config: DatabaseConfig) -> RefCountedDB {
|
||||||
let opts = config.prefix(DB_PREFIX_LEN);
|
let backing = Database::open(&config, path).unwrap_or_else(|e| {
|
||||||
let backing = Database::open(&opts, path).unwrap_or_else(|e| {
|
|
||||||
panic!("Error opening state db: {}", e);
|
panic!("Error opening state db: {}", e);
|
||||||
});
|
});
|
||||||
if !backing.is_empty() {
|
if !backing.is_empty() {
|
||||||
|
@ -18,7 +18,7 @@
|
|||||||
|
|
||||||
use std::default::Default;
|
use std::default::Default;
|
||||||
use rocksdb::{DB, Writable, WriteBatch, WriteOptions, IteratorMode, DBVector, DBIterator,
|
use rocksdb::{DB, Writable, WriteBatch, WriteOptions, IteratorMode, DBVector, DBIterator,
|
||||||
IndexType, Options, DBCompactionStyle, BlockBasedOptions, Direction, Cache};
|
Options, DBCompactionStyle, BlockBasedOptions, Direction, Cache};
|
||||||
|
|
||||||
const DB_BACKGROUND_FLUSHES: i32 = 2;
|
const DB_BACKGROUND_FLUSHES: i32 = 2;
|
||||||
const DB_BACKGROUND_COMPACTIONS: i32 = 2;
|
const DB_BACKGROUND_COMPACTIONS: i32 = 2;
|
||||||
@ -83,8 +83,6 @@ impl CompactionProfile {
|
|||||||
|
|
||||||
/// Database configuration
|
/// Database configuration
|
||||||
pub struct DatabaseConfig {
|
pub struct DatabaseConfig {
|
||||||
/// Optional prefix size in bytes. Allows lookup by partial key.
|
|
||||||
pub prefix_size: Option<usize>,
|
|
||||||
/// Max number of open files.
|
/// Max number of open files.
|
||||||
pub max_open_files: i32,
|
pub max_open_files: i32,
|
||||||
/// Cache-size
|
/// Cache-size
|
||||||
@ -98,7 +96,6 @@ impl DatabaseConfig {
|
|||||||
pub fn with_cache(cache_size: usize) -> DatabaseConfig {
|
pub fn with_cache(cache_size: usize) -> DatabaseConfig {
|
||||||
DatabaseConfig {
|
DatabaseConfig {
|
||||||
cache_size: Some(cache_size),
|
cache_size: Some(cache_size),
|
||||||
prefix_size: None,
|
|
||||||
max_open_files: 256,
|
max_open_files: 256,
|
||||||
compaction: CompactionProfile::default(),
|
compaction: CompactionProfile::default(),
|
||||||
}
|
}
|
||||||
@ -109,19 +106,12 @@ impl DatabaseConfig {
|
|||||||
self.compaction = profile;
|
self.compaction = profile;
|
||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Modify the prefix of the db
|
|
||||||
pub fn prefix(mut self, prefix_size: usize) -> Self {
|
|
||||||
self.prefix_size = Some(prefix_size);
|
|
||||||
self
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Default for DatabaseConfig {
|
impl Default for DatabaseConfig {
|
||||||
fn default() -> DatabaseConfig {
|
fn default() -> DatabaseConfig {
|
||||||
DatabaseConfig {
|
DatabaseConfig {
|
||||||
cache_size: None,
|
cache_size: None,
|
||||||
prefix_size: None,
|
|
||||||
max_open_files: 256,
|
max_open_files: 256,
|
||||||
compaction: CompactionProfile::default(),
|
compaction: CompactionProfile::default(),
|
||||||
}
|
}
|
||||||
@ -171,17 +161,9 @@ impl Database {
|
|||||||
opts.set_max_background_flushes(DB_BACKGROUND_FLUSHES);
|
opts.set_max_background_flushes(DB_BACKGROUND_FLUSHES);
|
||||||
opts.set_max_background_compactions(DB_BACKGROUND_COMPACTIONS);
|
opts.set_max_background_compactions(DB_BACKGROUND_COMPACTIONS);
|
||||||
|
|
||||||
if let Some(size) = config.prefix_size {
|
|
||||||
let mut block_opts = BlockBasedOptions::new();
|
|
||||||
block_opts.set_index_type(IndexType::HashSearch);
|
|
||||||
opts.set_block_based_table_factory(&block_opts);
|
|
||||||
opts.set_prefix_extractor_fixed_size(size);
|
|
||||||
if let Some(cache_size) = config.cache_size {
|
if let Some(cache_size) = config.cache_size {
|
||||||
block_opts.set_cache(Cache::new(cache_size * 1024 * 1024));
|
|
||||||
}
|
|
||||||
} else if let Some(cache_size) = config.cache_size {
|
|
||||||
let mut block_opts = BlockBasedOptions::new();
|
let mut block_opts = BlockBasedOptions::new();
|
||||||
// half goes to read cache
|
// all goes to read cache
|
||||||
block_opts.set_cache(Cache::new(cache_size * 1024 * 1024));
|
block_opts.set_cache(Cache::new(cache_size * 1024 * 1024));
|
||||||
opts.set_block_based_table_factory(&block_opts);
|
opts.set_block_based_table_factory(&block_opts);
|
||||||
}
|
}
|
||||||
@ -281,11 +263,9 @@ mod tests {
|
|||||||
assert!(db.get(&key1).unwrap().is_none());
|
assert!(db.get(&key1).unwrap().is_none());
|
||||||
assert_eq!(db.get(&key3).unwrap().unwrap().deref(), b"elephant");
|
assert_eq!(db.get(&key3).unwrap().unwrap().deref(), b"elephant");
|
||||||
|
|
||||||
if config.prefix_size.is_some() {
|
|
||||||
assert_eq!(db.get_by_prefix(&key3).unwrap().deref(), b"elephant");
|
assert_eq!(db.get_by_prefix(&key3).unwrap().deref(), b"elephant");
|
||||||
assert_eq!(db.get_by_prefix(&key2).unwrap().deref(), b"dog");
|
assert_eq!(db.get_by_prefix(&key2).unwrap().deref(), b"dog");
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn kvdb() {
|
fn kvdb() {
|
||||||
@ -293,9 +273,6 @@ mod tests {
|
|||||||
let smoke = Database::open_default(path.as_path().to_str().unwrap()).unwrap();
|
let smoke = Database::open_default(path.as_path().to_str().unwrap()).unwrap();
|
||||||
assert!(smoke.is_empty());
|
assert!(smoke.is_empty());
|
||||||
test_db(&DatabaseConfig::default());
|
test_db(&DatabaseConfig::default());
|
||||||
test_db(&DatabaseConfig::default().prefix(12));
|
|
||||||
test_db(&DatabaseConfig::default().prefix(22));
|
|
||||||
test_db(&DatabaseConfig::default().prefix(8));
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -197,7 +197,6 @@ impl Manager {
|
|||||||
let config = self.config.clone();
|
let config = self.config.clone();
|
||||||
let migrations = try!(self.migrations_from(version).ok_or(Error::MigrationImpossible));
|
let migrations = try!(self.migrations_from(version).ok_or(Error::MigrationImpossible));
|
||||||
let db_config = DatabaseConfig {
|
let db_config = DatabaseConfig {
|
||||||
prefix_size: None,
|
|
||||||
max_open_files: 64,
|
max_open_files: 64,
|
||||||
cache_size: None,
|
cache_size: None,
|
||||||
compaction: CompactionProfile::default(),
|
compaction: CompactionProfile::default(),
|
||||||
|
Loading…
Reference in New Issue
Block a user