remove internal locking from DBTransaction (#2003)

This commit is contained in:
Robert Habermeier 2016-08-25 16:43:56 +02:00 committed by Arkadiy Paronyan
parent b18407b9e3
commit 2aef81cf90
17 changed files with 121 additions and 119 deletions

View File

@ -380,7 +380,7 @@ impl BlockChain {
children: vec![] children: vec![]
}; };
let batch = DBTransaction::new(&db); let mut batch = DBTransaction::new(&db);
batch.put(db::COL_HEADERS, &hash, block.header_rlp().as_raw()); batch.put(db::COL_HEADERS, &hash, block.header_rlp().as_raw());
batch.put(db::COL_BODIES, &hash, &Self::block_to_body(genesis)); batch.put(db::COL_BODIES, &hash, &Self::block_to_body(genesis));
@ -419,7 +419,7 @@ impl BlockChain {
} }
} }
let batch = db.transaction(); let mut batch = db.transaction();
batch.put(db::COL_EXTRA, b"first", &hash); batch.put(db::COL_EXTRA, b"first", &hash);
db.write(batch).expect("Low level database error."); db.write(batch).expect("Low level database error.");
@ -451,7 +451,7 @@ impl BlockChain {
#[cfg(test)] #[cfg(test)]
fn rewind(&self) -> Option<H256> { fn rewind(&self) -> Option<H256> {
use db::Key; use db::Key;
let batch = self.db.transaction(); let mut batch =self.db.transaction();
// track back to the best block we have in the blocks database // track back to the best block we have in the blocks database
if let Some(best_block_hash) = self.db.get(db::COL_EXTRA, b"best").unwrap() { if let Some(best_block_hash) = self.db.get(db::COL_EXTRA, b"best").unwrap() {
let best_block_hash = H256::from_slice(&best_block_hash); let best_block_hash = H256::from_slice(&best_block_hash);
@ -604,7 +604,7 @@ impl BlockChain {
assert!(self.pending_best_block.read().is_none()); assert!(self.pending_best_block.read().is_none());
let batch = self.db.transaction(); let mut batch = self.db.transaction();
let block_rlp = UntrustedRlp::new(bytes); let block_rlp = UntrustedRlp::new(bytes);
let compressed_header = block_rlp.at(0).unwrap().compress(RlpType::Blocks); let compressed_header = block_rlp.at(0).unwrap().compress(RlpType::Blocks);
@ -625,7 +625,7 @@ impl BlockChain {
location: BlockLocation::CanonChain, location: BlockLocation::CanonChain,
}; };
self.prepare_update(&batch, ExtrasUpdate { self.prepare_update(&mut batch, ExtrasUpdate {
block_hashes: self.prepare_block_hashes_update(bytes, &info), block_hashes: self.prepare_block_hashes_update(bytes, &info),
block_details: self.prepare_block_details_update(bytes, &info), block_details: self.prepare_block_details_update(bytes, &info),
block_receipts: self.prepare_block_receipts_update(receipts, &info), block_receipts: self.prepare_block_receipts_update(receipts, &info),
@ -659,7 +659,7 @@ impl BlockChain {
let mut update = HashMap::new(); let mut update = HashMap::new();
update.insert(hash, block_details); update.insert(hash, block_details);
self.prepare_update(&batch, ExtrasUpdate { self.prepare_update(&mut batch, ExtrasUpdate {
block_hashes: self.prepare_block_hashes_update(bytes, &info), block_hashes: self.prepare_block_hashes_update(bytes, &info),
block_details: update, block_details: update,
block_receipts: self.prepare_block_receipts_update(receipts, &info), block_receipts: self.prepare_block_receipts_update(receipts, &info),
@ -682,7 +682,7 @@ impl BlockChain {
let mut parent_details = self.block_details(&block_hash) let mut parent_details = self.block_details(&block_hash)
.unwrap_or_else(|| panic!("Invalid block hash: {:?}", block_hash)); .unwrap_or_else(|| panic!("Invalid block hash: {:?}", block_hash));
let batch = self.db.transaction(); let mut batch = self.db.transaction();
parent_details.children.push(child_hash); parent_details.children.push(child_hash);
let mut update = HashMap::new(); let mut update = HashMap::new();
@ -701,7 +701,7 @@ impl BlockChain {
/// Inserts the block into backing cache database. /// Inserts the block into backing cache database.
/// Expects the block to be valid and already verified. /// Expects the block to be valid and already verified.
/// If the block is already known, does nothing. /// If the block is already known, does nothing.
pub fn insert_block(&self, batch: &DBTransaction, bytes: &[u8], receipts: Vec<Receipt>) -> ImportRoute { pub fn insert_block(&self, batch: &mut DBTransaction, bytes: &[u8], receipts: Vec<Receipt>) -> ImportRoute {
// create views onto rlp // create views onto rlp
let block = BlockView::new(bytes); let block = BlockView::new(bytes);
let header = block.header_view(); let header = block.header_view();
@ -782,7 +782,7 @@ impl BlockChain {
} }
/// Prepares extras update. /// Prepares extras update.
fn prepare_update(&self, batch: &DBTransaction, update: ExtrasUpdate, is_best: bool) { fn prepare_update(&self, batch: &mut DBTransaction, update: ExtrasUpdate, is_best: bool) {
{ {
let block_hashes: Vec<_> = update.block_details.keys().cloned().collect(); let block_hashes: Vec<_> = update.block_details.keys().cloned().collect();
@ -1147,8 +1147,8 @@ mod tests {
assert_eq!(bc.best_block_number(), 0); assert_eq!(bc.best_block_number(), 0);
// when // when
let batch = db.transaction(); let mut batch =db.transaction();
bc.insert_block(&batch, &first, vec![]); bc.insert_block(&mut batch, &first, vec![]);
assert_eq!(bc.best_block_number(), 0); assert_eq!(bc.best_block_number(), 0);
bc.commit(); bc.commit();
// NOTE no db.write here (we want to check if best block is cached) // NOTE no db.write here (we want to check if best block is cached)
@ -1177,8 +1177,8 @@ mod tests {
assert_eq!(bc.block_hash(1), None); assert_eq!(bc.block_hash(1), None);
assert_eq!(bc.block_details(&genesis_hash).unwrap().children, vec![]); assert_eq!(bc.block_details(&genesis_hash).unwrap().children, vec![]);
let batch = db.transaction(); let mut batch =db.transaction();
bc.insert_block(&batch, &first, vec![]); bc.insert_block(&mut batch, &first, vec![]);
db.write(batch).unwrap(); db.write(batch).unwrap();
bc.commit(); bc.commit();
@ -1203,11 +1203,11 @@ mod tests {
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let mut block_hashes = vec![genesis_hash.clone()]; let mut block_hashes = vec![genesis_hash.clone()];
let batch = db.transaction(); let mut batch =db.transaction();
for _ in 0..10 { for _ in 0..10 {
let block = canon_chain.generate(&mut finalizer).unwrap(); let block = canon_chain.generate(&mut finalizer).unwrap();
block_hashes.push(BlockView::new(&block).header_view().sha3()); block_hashes.push(BlockView::new(&block).header_view().sha3());
bc.insert_block(&batch, &block, vec![]); bc.insert_block(&mut batch, &block, vec![]);
bc.commit(); bc.commit();
} }
db.write(batch).unwrap(); db.write(batch).unwrap();
@ -1238,20 +1238,20 @@ mod tests {
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let batch = db.transaction(); let mut batch =db.transaction();
for b in &[&b1a, &b1b, &b2a, &b2b, &b3a, &b3b, &b4a, &b4b, &b5a, &b5b] { for b in &[&b1a, &b1b, &b2a, &b2b, &b3a, &b3b, &b4a, &b4b, &b5a, &b5b] {
bc.insert_block(&batch, b, vec![]); bc.insert_block(&mut batch, b, vec![]);
bc.commit(); bc.commit();
} }
bc.insert_block(&batch, &b1b, vec![]); bc.insert_block(&mut batch, &b1b, vec![]);
bc.insert_block(&batch, &b2a, vec![]); bc.insert_block(&mut batch, &b2a, vec![]);
bc.insert_block(&batch, &b2b, vec![]); bc.insert_block(&mut batch, &b2b, vec![]);
bc.insert_block(&batch, &b3a, vec![]); bc.insert_block(&mut batch, &b3a, vec![]);
bc.insert_block(&batch, &b3b, vec![]); bc.insert_block(&mut batch, &b3b, vec![]);
bc.insert_block(&batch, &b4a, vec![]); bc.insert_block(&mut batch, &b4a, vec![]);
bc.insert_block(&batch, &b4b, vec![]); bc.insert_block(&mut batch, &b4b, vec![]);
bc.insert_block(&batch, &b5a, vec![]); bc.insert_block(&mut batch, &b5a, vec![]);
bc.insert_block(&batch, &b5b, vec![]); bc.insert_block(&mut batch, &b5b, vec![]);
db.write(batch).unwrap(); db.write(batch).unwrap();
assert_eq!( assert_eq!(
@ -1286,17 +1286,17 @@ mod tests {
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let batch = db.transaction(); let mut batch =db.transaction();
let ir1 = bc.insert_block(&batch, &b1, vec![]); let ir1 = bc.insert_block(&mut batch, &b1, vec![]);
bc.commit(); bc.commit();
let ir2 = bc.insert_block(&batch, &b2, vec![]); let ir2 = bc.insert_block(&mut batch, &b2, vec![]);
bc.commit(); bc.commit();
let ir3b = bc.insert_block(&batch, &b3b, vec![]); let ir3b = bc.insert_block(&mut batch, &b3b, vec![]);
bc.commit(); bc.commit();
db.write(batch).unwrap(); db.write(batch).unwrap();
assert_eq!(bc.block_hash(3).unwrap(), b3b_hash); assert_eq!(bc.block_hash(3).unwrap(), b3b_hash);
let batch = db.transaction(); let mut batch =db.transaction();
let ir3a = bc.insert_block(&batch, &b3a, vec![]); let ir3a = bc.insert_block(&mut batch, &b3a, vec![]);
bc.commit(); bc.commit();
db.write(batch).unwrap(); db.write(batch).unwrap();
@ -1402,8 +1402,8 @@ mod tests {
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
assert_eq!(bc.best_block_hash(), genesis_hash); assert_eq!(bc.best_block_hash(), genesis_hash);
let batch = db.transaction(); let mut batch =db.transaction();
bc.insert_block(&batch, &first, vec![]); bc.insert_block(&mut batch, &first, vec![]);
db.write(batch).unwrap(); db.write(batch).unwrap();
bc.commit(); bc.commit();
assert_eq!(bc.best_block_hash(), first_hash); assert_eq!(bc.best_block_hash(), first_hash);
@ -1467,8 +1467,8 @@ mod tests {
let temp = RandomTempPath::new(); let temp = RandomTempPath::new();
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let batch = db.transaction(); let mut batch =db.transaction();
bc.insert_block(&batch, &b1, vec![]); bc.insert_block(&mut batch, &b1, vec![]);
db.write(batch).unwrap(); db.write(batch).unwrap();
bc.commit(); bc.commit();
@ -1480,8 +1480,8 @@ mod tests {
} }
fn insert_block(db: &Arc<Database>, bc: &BlockChain, bytes: &[u8], receipts: Vec<Receipt>) -> ImportRoute { fn insert_block(db: &Arc<Database>, bc: &BlockChain, bytes: &[u8], receipts: Vec<Receipt>) -> ImportRoute {
let batch = db.transaction(); let mut batch =db.transaction();
let res = bc.insert_block(&batch, bytes, receipts); let res = bc.insert_block(&mut batch, bytes, receipts);
db.write(batch).unwrap(); db.write(batch).unwrap();
bc.commit(); bc.commit();
res res
@ -1569,16 +1569,16 @@ mod tests {
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let uncle = canon_chain.fork(1).generate(&mut finalizer.fork()).unwrap(); let uncle = canon_chain.fork(1).generate(&mut finalizer.fork()).unwrap();
let batch = db.transaction(); let mut batch =db.transaction();
// create a longer fork // create a longer fork
for _ in 0..5 { for _ in 0..5 {
let canon_block = canon_chain.generate(&mut finalizer).unwrap(); let canon_block = canon_chain.generate(&mut finalizer).unwrap();
bc.insert_block(&batch, &canon_block, vec![]); bc.insert_block(&mut batch, &canon_block, vec![]);
bc.commit(); bc.commit();
} }
assert_eq!(bc.best_block_number(), 5); assert_eq!(bc.best_block_number(), 5);
bc.insert_block(&batch, &uncle, vec![]); bc.insert_block(&mut batch, &uncle, vec![]);
db.write(batch).unwrap(); db.write(batch).unwrap();
bc.commit(); bc.commit();
} }
@ -1604,10 +1604,10 @@ mod tests {
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(Config::default(), &genesis, db.clone()); let bc = BlockChain::new(Config::default(), &genesis, db.clone());
let batch = db.transaction(); let mut batch =db.transaction();
bc.insert_block(&batch, &first, vec![]); bc.insert_block(&mut batch, &first, vec![]);
bc.commit(); bc.commit();
bc.insert_block(&batch, &second, vec![]); bc.insert_block(&mut batch, &second, vec![]);
bc.commit(); bc.commit();
db.write(batch).unwrap(); db.write(batch).unwrap();

View File

@ -172,8 +172,8 @@ impl Client {
let mut state_db = journaldb::new(db.clone(), config.pruning, ::db::COL_STATE); let mut state_db = journaldb::new(db.clone(), config.pruning, ::db::COL_STATE);
if state_db.is_empty() && try!(spec.ensure_db_good(state_db.as_hashdb_mut())) { if state_db.is_empty() && try!(spec.ensure_db_good(state_db.as_hashdb_mut())) {
let batch = DBTransaction::new(&db); let mut batch = DBTransaction::new(&db);
try!(state_db.commit(&batch, 0, &spec.genesis_header().hash(), None)); try!(state_db.commit(&mut batch, 0, &spec.genesis_header().hash(), None));
try!(db.write(batch).map_err(ClientError::Database)); try!(db.write(batch).map_err(ClientError::Database));
} }
@ -431,14 +431,14 @@ impl Client {
//let traces = From::from(block.traces().clone().unwrap_or_else(Vec::new)); //let traces = From::from(block.traces().clone().unwrap_or_else(Vec::new));
let batch = DBTransaction::new(&self.db); let mut batch = DBTransaction::new(&self.db);
// CHECK! I *think* this is fine, even if the state_root is equal to another // CHECK! I *think* this is fine, even if the state_root is equal to another
// already-imported block of the same number. // already-imported block of the same number.
// TODO: Prove it with a test. // TODO: Prove it with a test.
block.drain().commit(&batch, number, hash, ancient).expect("DB commit failed."); block.drain().commit(&mut batch, number, hash, ancient).expect("DB commit failed.");
let route = self.chain.insert_block(&batch, block_data, receipts); let route = self.chain.insert_block(&mut batch, block_data, receipts);
self.tracedb.import(&batch, TraceImportRequest { self.tracedb.import(&mut batch, TraceImportRequest {
traces: traces.into(), traces: traces.into(),
block_hash: hash.clone(), block_hash: hash.clone(),
block_number: number, block_number: number,

View File

@ -83,10 +83,10 @@ pub trait Key<T> {
/// Should be used to write value into database. /// Should be used to write value into database.
pub trait Writable { pub trait Writable {
/// Writes the value into the database. /// Writes the value into the database.
fn write<T, R>(&self, col: Option<u32>, key: &Key<T, Target = R>, value: &T) where T: Encodable, R: Deref<Target = [u8]>; fn write<T, R>(&mut self, col: Option<u32>, key: &Key<T, Target = R>, value: &T) where T: Encodable, R: Deref<Target = [u8]>;
/// Writes the value into the database and updates the cache. /// Writes the value into the database and updates the cache.
fn write_with_cache<K, T, R>(&self, col: Option<u32>, cache: &mut Cache<K, T>, key: K, value: T, policy: CacheUpdatePolicy) where fn write_with_cache<K, T, R>(&mut self, col: Option<u32>, cache: &mut Cache<K, T>, key: K, value: T, policy: CacheUpdatePolicy) where
K: Key<T, Target = R> + Hash + Eq, K: Key<T, Target = R> + Hash + Eq,
T: Encodable, T: Encodable,
R: Deref<Target = [u8]> { R: Deref<Target = [u8]> {
@ -102,7 +102,7 @@ pub trait Writable {
} }
/// Writes the values into the database and updates the cache. /// Writes the values into the database and updates the cache.
fn extend_with_cache<K, T, R>(&self, col: Option<u32>, cache: &mut Cache<K, T>, values: HashMap<K, T>, policy: CacheUpdatePolicy) where fn extend_with_cache<K, T, R>(&mut self, col: Option<u32>, cache: &mut Cache<K, T>, values: HashMap<K, T>, policy: CacheUpdatePolicy) where
K: Key<T, Target = R> + Hash + Eq, K: Key<T, Target = R> + Hash + Eq,
T: Encodable, T: Encodable,
R: Deref<Target = [u8]> { R: Deref<Target = [u8]> {
@ -169,7 +169,7 @@ pub trait Readable {
} }
impl Writable for DBTransaction { impl Writable for DBTransaction {
fn write<T, R>(&self, col: Option<u32>, key: &Key<T, Target = R>, value: &T) where T: Encodable, R: Deref<Target = [u8]> { fn write<T, R>(&mut self, col: Option<u32>, key: &Key<T, Target = R>, value: &T) where T: Encodable, R: Deref<Target = [u8]> {
self.put(col, &key.key(), &encode(value)); self.put(col, &key.key(), &encode(value));
} }
} }

View File

@ -492,8 +492,8 @@ impl StateRebuilder {
} }
let backing = self.db.backing().clone(); let backing = self.db.backing().clone();
let batch = backing.transaction(); let mut batch = backing.transaction();
try!(self.db.inject(&batch)); try!(self.db.inject(&mut batch));
try!(backing.write(batch).map_err(::util::UtilError::SimpleString)); try!(backing.write(batch).map_err(::util::UtilError::SimpleString));
trace!(target: "snapshot", "current state root: {:?}", self.state_root); trace!(target: "snapshot", "current state root: {:?}", self.state_root);
Ok(()) Ok(())

View File

@ -43,14 +43,16 @@ fn chunk_and_restore(amount: u64) {
let bc = BlockChain::new(Default::default(), &genesis, old_db.clone()); let bc = BlockChain::new(Default::default(), &genesis, old_db.clone());
// build the blockchain. // build the blockchain.
let mut batch = old_db.transaction();
for _ in 0..amount { for _ in 0..amount {
let block = canon_chain.generate(&mut finalizer).unwrap(); let block = canon_chain.generate(&mut finalizer).unwrap();
let batch = old_db.transaction(); bc.insert_block(&mut batch, &block, vec![]);
bc.insert_block(&batch, &block, vec![]);
bc.commit(); bc.commit();
old_db.write(batch).unwrap();
} }
old_db.write(batch).unwrap();
let best_hash = bc.best_block_hash(); let best_hash = bc.best_block_hash();
// snapshot it. // snapshot it.

View File

@ -259,9 +259,9 @@ pub fn generate_dummy_blockchain(block_number: u32) -> GuardedTempResult<BlockCh
let db = new_db(temp.as_str()); let db = new_db(temp.as_str());
let bc = BlockChain::new(BlockChainConfig::default(), &create_unverifiable_block(0, H256::zero()), db.clone()); let bc = BlockChain::new(BlockChainConfig::default(), &create_unverifiable_block(0, H256::zero()), db.clone());
let batch = db.transaction(); let mut batch = db.transaction();
for block_order in 1..block_number { for block_order in 1..block_number {
bc.insert_block(&batch, &create_unverifiable_block(block_order, bc.best_block_hash()), vec![]); bc.insert_block(&mut batch, &create_unverifiable_block(block_order, bc.best_block_hash()), vec![]);
bc.commit(); bc.commit();
} }
db.write(batch).unwrap(); db.write(batch).unwrap();
@ -278,9 +278,9 @@ pub fn generate_dummy_blockchain_with_extra(block_number: u32) -> GuardedTempRes
let bc = BlockChain::new(BlockChainConfig::default(), &create_unverifiable_block(0, H256::zero()), db.clone()); let bc = BlockChain::new(BlockChainConfig::default(), &create_unverifiable_block(0, H256::zero()), db.clone());
let batch = db.transaction(); let mut batch = db.transaction();
for block_order in 1..block_number { for block_order in 1..block_number {
bc.insert_block(&batch, &create_unverifiable_block_with_extra(block_order, bc.best_block_hash(), None), vec![]); bc.insert_block(&mut batch, &create_unverifiable_block_with_extra(block_order, bc.best_block_hash(), None), vec![]);
bc.commit(); bc.commit();
} }
db.write(batch).unwrap(); db.write(batch).unwrap();

View File

@ -142,7 +142,7 @@ impl<T> TraceDB<T> where T: DatabaseExtras {
false => [0x0] false => [0x0]
}; };
let batch = DBTransaction::new(&tracesdb); let mut batch = DBTransaction::new(&tracesdb);
batch.put(db::COL_TRACE, b"enabled", &encoded_tracing); batch.put(db::COL_TRACE, b"enabled", &encoded_tracing);
batch.put(db::COL_TRACE, b"version", TRACE_DB_VER); batch.put(db::COL_TRACE, b"version", TRACE_DB_VER);
tracesdb.write(batch).unwrap(); tracesdb.write(batch).unwrap();
@ -261,7 +261,7 @@ impl<T> TraceDatabase for TraceDB<T> where T: DatabaseExtras {
/// Traces of import request's enacted blocks are expected to be already in database /// Traces of import request's enacted blocks are expected to be already in database
/// or to be the currently inserted trace. /// or to be the currently inserted trace.
fn import(&self, batch: &DBTransaction, request: ImportRequest) { fn import(&self, batch: &mut DBTransaction, request: ImportRequest) {
// valid (canon): retracted 0, enacted 1 => false, true, // valid (canon): retracted 0, enacted 1 => false, true,
// valid (branch): retracted 0, enacted 0 => false, false, // valid (branch): retracted 0, enacted 0 => false, false,
// valid (bbcc): retracted 1, enacted 1 => true, true, // valid (bbcc): retracted 1, enacted 1 => true, true,
@ -611,8 +611,8 @@ mod tests {
// import block 0 // import block 0
let request = create_simple_import_request(0, block_0.clone()); let request = create_simple_import_request(0, block_0.clone());
let batch = DBTransaction::new(&db); let mut batch = DBTransaction::new(&db);
tracedb.import(&batch, request); tracedb.import(&mut batch, request);
db.write(batch).unwrap(); db.write(batch).unwrap();
let filter = Filter { let filter = Filter {
@ -627,8 +627,8 @@ mod tests {
// import block 1 // import block 1
let request = create_simple_import_request(1, block_1.clone()); let request = create_simple_import_request(1, block_1.clone());
let batch = DBTransaction::new(&db); let mut batch = DBTransaction::new(&db);
tracedb.import(&batch, request); tracedb.import(&mut batch, request);
db.write(batch).unwrap(); db.write(batch).unwrap();
let filter = Filter { let filter = Filter {
@ -686,8 +686,8 @@ mod tests {
// import block 0 // import block 0
let request = create_simple_import_request(0, block_0.clone()); let request = create_simple_import_request(0, block_0.clone());
let batch = DBTransaction::new(&db); let mut batch = DBTransaction::new(&db);
tracedb.import(&batch, request); tracedb.import(&mut batch, request);
db.write(batch).unwrap(); db.write(batch).unwrap();
} }

View File

@ -121,7 +121,7 @@ pub trait Database {
fn tracing_enabled(&self) -> bool; fn tracing_enabled(&self) -> bool;
/// Imports new block traces. /// Imports new block traces.
fn import(&self, batch: &DBTransaction, request: ImportRequest); fn import(&self, batch: &mut DBTransaction, request: ImportRequest);
/// Returns localized trace at given position. /// Returns localized trace at given position.
fn trace(&self, block_number: BlockNumber, tx_position: usize, trace_position: Vec<usize>) -> Option<LocalizedTrace>; fn trace(&self, block_number: BlockNumber, tx_position: usize, trace_position: Vec<usize>) -> Option<LocalizedTrace>;

View File

@ -156,7 +156,7 @@ impl JournalDB for ArchiveDB {
self.latest_era.is_none() self.latest_era.is_none()
} }
fn commit(&mut self, batch: &DBTransaction, now: u64, _id: &H256, _end: Option<(u64, H256)>) -> Result<u32, UtilError> { fn commit(&mut self, batch: &mut DBTransaction, now: u64, _id: &H256, _end: Option<(u64, H256)>) -> Result<u32, UtilError> {
let mut inserts = 0usize; let mut inserts = 0usize;
let mut deletes = 0usize; let mut deletes = 0usize;
@ -185,7 +185,7 @@ impl JournalDB for ArchiveDB {
Ok((inserts + deletes) as u32) Ok((inserts + deletes) as u32)
} }
fn inject(&mut self, batch: &DBTransaction) -> Result<u32, UtilError> { fn inject(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError> {
let mut inserts = 0usize; let mut inserts = 0usize;
let mut deletes = 0usize; let mut deletes = 0usize;

View File

@ -101,13 +101,13 @@ impl EarlyMergeDB {
} }
// The next three are valid only as long as there is an insert operation of `key` in the journal. // The next three are valid only as long as there is an insert operation of `key` in the journal.
fn set_already_in(batch: &DBTransaction, col: Option<u32>, key: &H256) { batch.put(col, &Self::morph_key(key, 0), &[1u8]); } fn set_already_in(batch: &mut DBTransaction, col: Option<u32>, key: &H256) { batch.put(col, &Self::morph_key(key, 0), &[1u8]); }
fn reset_already_in(batch: &DBTransaction, col: Option<u32>, key: &H256) { batch.delete(col, &Self::morph_key(key, 0)); } fn reset_already_in(batch: &mut DBTransaction, col: Option<u32>, key: &H256) { batch.delete(col, &Self::morph_key(key, 0)); }
fn is_already_in(backing: &Database, col: Option<u32>, key: &H256) -> bool { fn is_already_in(backing: &Database, col: Option<u32>, key: &H256) -> bool {
backing.get(col, &Self::morph_key(key, 0)).expect("Low-level database error. Some issue with your hard disk?").is_some() backing.get(col, &Self::morph_key(key, 0)).expect("Low-level database error. Some issue with your hard disk?").is_some()
} }
fn insert_keys(inserts: &[(H256, Bytes)], backing: &Database, col: Option<u32>, refs: &mut HashMap<H256, RefInfo>, batch: &DBTransaction, trace: bool) { fn insert_keys(inserts: &[(H256, Bytes)], backing: &Database, col: Option<u32>, refs: &mut HashMap<H256, RefInfo>, batch: &mut DBTransaction, trace: bool) {
for &(ref h, ref d) in inserts { for &(ref h, ref d) in inserts {
if let Some(c) = refs.get_mut(h) { if let Some(c) = refs.get_mut(h) {
// already counting. increment. // already counting. increment.
@ -156,7 +156,7 @@ impl EarlyMergeDB {
trace!(target: "jdb.fine", "replay_keys: (end) refs={:?}", refs); trace!(target: "jdb.fine", "replay_keys: (end) refs={:?}", refs);
} }
fn remove_keys(deletes: &[H256], refs: &mut HashMap<H256, RefInfo>, batch: &DBTransaction, col: Option<u32>, from: RemoveFrom, trace: bool) { fn remove_keys(deletes: &[H256], refs: &mut HashMap<H256, RefInfo>, batch: &mut DBTransaction, col: Option<u32>, from: RemoveFrom, trace: bool) {
// with a remove on {queue_refs: 1, in_archive: true}, we have two options: // with a remove on {queue_refs: 1, in_archive: true}, we have two options:
// - convert to {queue_refs: 1, in_archive: false} (i.e. remove it from the conceptual archive) // - convert to {queue_refs: 1, in_archive: false} (i.e. remove it from the conceptual archive)
// - convert to {queue_refs: 0, in_archive: true} (i.e. remove it from the conceptual queue) // - convert to {queue_refs: 0, in_archive: true} (i.e. remove it from the conceptual queue)
@ -337,7 +337,7 @@ impl JournalDB for EarlyMergeDB {
} }
#[cfg_attr(feature="dev", allow(cyclomatic_complexity))] #[cfg_attr(feature="dev", allow(cyclomatic_complexity))]
fn commit(&mut self, batch: &DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> { fn commit(&mut self, batch: &mut DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
// journal format: // journal format:
// [era, 0] => [ id, [insert_0, ...], [remove_0, ...] ] // [era, 0] => [ id, [insert_0, ...], [remove_0, ...] ]
// [era, 1] => [ id, [insert_0, ...], [remove_0, ...] ] // [era, 1] => [ id, [insert_0, ...], [remove_0, ...] ]
@ -514,7 +514,7 @@ impl JournalDB for EarlyMergeDB {
Ok(0) Ok(0)
} }
fn inject(&mut self, batch: &DBTransaction) -> Result<u32, UtilError> { fn inject(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError> {
let mut ops = 0; let mut ops = 0;
for (key, (value, rc)) in self.overlay.drain() { for (key, (value, rc)) in self.overlay.drain() {
if rc != 0 { ops += 1 } if rc != 0 { ops += 1 }

View File

@ -222,7 +222,7 @@ impl JournalDB for OverlayRecentDB {
.or_else(|| self.backing.get_by_prefix(self.column, &key[0..DB_PREFIX_LEN]).map(|b| b.to_vec())) .or_else(|| self.backing.get_by_prefix(self.column, &key[0..DB_PREFIX_LEN]).map(|b| b.to_vec()))
} }
fn commit(&mut self, batch: &DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> { fn commit(&mut self, batch: &mut DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
// record new commit's details. // record new commit's details.
trace!("commit: #{} ({}), end era: {:?}", now, id, end); trace!("commit: #{} ({}), end era: {:?}", now, id, end);
let mut journal_overlay = self.journal_overlay.write(); let mut journal_overlay = self.journal_overlay.write();
@ -314,7 +314,7 @@ impl JournalDB for OverlayRecentDB {
self.journal_overlay.write().pending_overlay.clear(); self.journal_overlay.write().pending_overlay.clear();
} }
fn inject(&mut self, batch: &DBTransaction) -> Result<u32, UtilError> { fn inject(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError> {
let mut ops = 0; let mut ops = 0;
for (key, (value, rc)) in self.transaction_overlay.drain() { for (key, (value, rc)) in self.transaction_overlay.drain() {
if rc != 0 { ops += 1 } if rc != 0 { ops += 1 }

View File

@ -109,7 +109,7 @@ impl JournalDB for RefCountedDB {
self.backing.get_by_prefix(self.column, &id[0..DB_PREFIX_LEN]).map(|b| b.to_vec()) self.backing.get_by_prefix(self.column, &id[0..DB_PREFIX_LEN]).map(|b| b.to_vec())
} }
fn commit(&mut self, batch: &DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> { fn commit(&mut self, batch: &mut DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
// journal format: // journal format:
// [era, 0] => [ id, [insert_0, ...], [remove_0, ...] ] // [era, 0] => [ id, [insert_0, ...], [remove_0, ...] ]
// [era, 1] => [ id, [insert_0, ...], [remove_0, ...] ] // [era, 1] => [ id, [insert_0, ...], [remove_0, ...] ]
@ -182,11 +182,11 @@ impl JournalDB for RefCountedDB {
} }
} }
let r = try!(self.forward.commit_to_batch(&batch)); let r = try!(self.forward.commit_to_batch(batch));
Ok(r) Ok(r)
} }
fn inject(&mut self, batch: &DBTransaction) -> Result<u32, UtilError> { fn inject(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError> {
self.inserts.clear(); self.inserts.clear();
for remove in self.removes.drain(..) { for remove in self.removes.drain(..) {
self.forward.remove(&remove); self.forward.remove(&remove);

View File

@ -37,7 +37,7 @@ pub trait JournalDB: HashDB {
/// Commit all recent insert operations and canonical historical commits' removals from the /// Commit all recent insert operations and canonical historical commits' removals from the
/// old era to the backing database, reverting any non-canonical historical commit's inserts. /// old era to the backing database, reverting any non-canonical historical commit's inserts.
fn commit(&mut self, batch: &DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError>; fn commit(&mut self, batch: &mut DBTransaction, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError>;
/// Commit all queued insert and delete operations without affecting any journalling -- this requires that all insertions /// Commit all queued insert and delete operations without affecting any journalling -- this requires that all insertions
/// and deletions are indeed canonical and will likely lead to an invalid database if that assumption is violated. /// and deletions are indeed canonical and will likely lead to an invalid database if that assumption is violated.
@ -46,7 +46,7 @@ pub trait JournalDB: HashDB {
/// by any previous `commit` operations. Essentially, this means that `inject` can be used /// by any previous `commit` operations. Essentially, this means that `inject` can be used
/// either to restore a state to a fresh database, or to insert data which may only be journalled /// either to restore a state to a fresh database, or to insert data which may only be journalled
/// from this point onwards. /// from this point onwards.
fn inject(&mut self, batch: &DBTransaction) -> Result<u32, UtilError>; fn inject(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError>;
/// State data query /// State data query
fn state(&self, _id: &H256) -> Option<Bytes>; fn state(&self, _id: &H256) -> Option<Bytes>;
@ -67,8 +67,8 @@ pub trait JournalDB: HashDB {
/// Commit all changes in a single batch /// Commit all changes in a single batch
#[cfg(test)] #[cfg(test)]
fn commit_batch(&mut self, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> { fn commit_batch(&mut self, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
let batch = self.backing().transaction(); let mut batch = self.backing().transaction();
let res = try!(self.commit(&batch, now, id, end)); let res = try!(self.commit(&mut batch, now, id, end));
let result = self.backing().write(batch).map(|_| res).map_err(Into::into); let result = self.backing().write(batch).map(|_| res).map_err(Into::into);
self.flush(); self.flush();
result result
@ -77,8 +77,8 @@ pub trait JournalDB: HashDB {
/// Inject all changes in a single batch. /// Inject all changes in a single batch.
#[cfg(test)] #[cfg(test)]
fn inject_batch(&mut self) -> Result<u32, UtilError> { fn inject_batch(&mut self) -> Result<u32, UtilError> {
let batch = self.backing().transaction(); let mut batch = self.backing().transaction();
let res = try!(self.inject(&batch)); let res = try!(self.inject(&mut batch));
self.backing().write(batch).map(|_| res).map_err(Into::into) self.backing().write(batch).map(|_| res).map_err(Into::into)
} }
} }

View File

@ -28,7 +28,7 @@ const DB_BACKGROUND_COMPACTIONS: i32 = 2;
/// Write transaction. Batches a sequence of put/delete operations for efficiency. /// Write transaction. Batches a sequence of put/delete operations for efficiency.
pub struct DBTransaction { pub struct DBTransaction {
ops: Mutex<Vec<DBOp>>, ops: Vec<DBOp>,
} }
enum DBOp { enum DBOp {
@ -52,15 +52,15 @@ impl DBTransaction {
/// Create new transaction. /// Create new transaction.
pub fn new(_db: &Database) -> DBTransaction { pub fn new(_db: &Database) -> DBTransaction {
DBTransaction { DBTransaction {
ops: Mutex::new(Vec::with_capacity(256)), ops: Vec::with_capacity(256),
} }
} }
/// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write. /// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write.
pub fn put(&self, col: Option<u32>, key: &[u8], value: &[u8]) { pub fn put(&mut self, col: Option<u32>, key: &[u8], value: &[u8]) {
let mut ekey = ElasticArray32::new(); let mut ekey = ElasticArray32::new();
ekey.append_slice(key); ekey.append_slice(key);
self.ops.lock().push(DBOp::Insert { self.ops.push(DBOp::Insert {
col: col, col: col,
key: ekey, key: ekey,
value: value.to_vec(), value: value.to_vec(),
@ -68,10 +68,10 @@ impl DBTransaction {
} }
/// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write. /// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write.
pub fn put_vec(&self, col: Option<u32>, key: &[u8], value: Bytes) { pub fn put_vec(&mut self, col: Option<u32>, key: &[u8], value: Bytes) {
let mut ekey = ElasticArray32::new(); let mut ekey = ElasticArray32::new();
ekey.append_slice(key); ekey.append_slice(key);
self.ops.lock().push(DBOp::Insert { self.ops.push(DBOp::Insert {
col: col, col: col,
key: ekey, key: ekey,
value: value, value: value,
@ -80,10 +80,10 @@ impl DBTransaction {
/// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write. /// Insert a key-value pair in the transaction. Any existing value value will be overwritten upon write.
/// Value will be RLP-compressed on flush /// Value will be RLP-compressed on flush
pub fn put_compressed(&self, col: Option<u32>, key: &[u8], value: Bytes) { pub fn put_compressed(&mut self, col: Option<u32>, key: &[u8], value: Bytes) {
let mut ekey = ElasticArray32::new(); let mut ekey = ElasticArray32::new();
ekey.append_slice(key); ekey.append_slice(key);
self.ops.lock().push(DBOp::InsertCompressed { self.ops.push(DBOp::InsertCompressed {
col: col, col: col,
key: ekey, key: ekey,
value: value, value: value,
@ -91,10 +91,10 @@ impl DBTransaction {
} }
/// Delete value by key. /// Delete value by key.
pub fn delete(&self, col: Option<u32>, key: &[u8]) { pub fn delete(&mut self, col: Option<u32>, key: &[u8]) {
let mut ekey = ElasticArray32::new(); let mut ekey = ElasticArray32::new();
ekey.append_slice(key); ekey.append_slice(key);
self.ops.lock().push(DBOp::Delete { self.ops.push(DBOp::Delete {
col: col, col: col,
key: ekey, key: ekey,
}); });
@ -299,7 +299,7 @@ impl Database {
/// Commit transaction to database. /// Commit transaction to database.
pub fn write_buffered(&self, tr: DBTransaction) { pub fn write_buffered(&self, tr: DBTransaction) {
let mut overlay = self.overlay.write(); let mut overlay = self.overlay.write();
let ops = tr.ops.into_inner(); let ops = tr.ops;
for op in ops { for op in ops {
match op { match op {
DBOp::Insert { col, key, value } => { DBOp::Insert { col, key, value } => {
@ -359,7 +359,7 @@ impl Database {
/// Commit transaction to database. /// Commit transaction to database.
pub fn write(&self, tr: DBTransaction) -> Result<(), String> { pub fn write(&self, tr: DBTransaction) -> Result<(), String> {
let batch = WriteBatch::new(); let batch = WriteBatch::new();
let ops = tr.ops.into_inner(); let ops = tr.ops;
for op in ops { for op in ops {
match op { match op {
DBOp::Insert { col, key, value } => { DBOp::Insert { col, key, value } => {
@ -425,7 +425,7 @@ mod tests {
let key2 = H256::from_str("03c69be41d0b7e40352fc85be1cd65eb03d40ef8427a0ca4596b1ead9a00e9fc").unwrap(); let key2 = H256::from_str("03c69be41d0b7e40352fc85be1cd65eb03d40ef8427a0ca4596b1ead9a00e9fc").unwrap();
let key3 = H256::from_str("01c69be41d0b7e40352fc85be1cd65eb03d40ef8427a0ca4596b1ead9a00e9fc").unwrap(); let key3 = H256::from_str("01c69be41d0b7e40352fc85be1cd65eb03d40ef8427a0ca4596b1ead9a00e9fc").unwrap();
let batch = db.transaction(); let mut batch = db.transaction();
batch.put(None, &key1, b"cat"); batch.put(None, &key1, b"cat");
batch.put(None, &key2, b"dog"); batch.put(None, &key2, b"dog");
db.write(batch).unwrap(); db.write(batch).unwrap();
@ -439,17 +439,17 @@ mod tests {
assert_eq!(&*contents[1].0, &*key2); assert_eq!(&*contents[1].0, &*key2);
assert_eq!(&*contents[1].1, b"dog"); assert_eq!(&*contents[1].1, b"dog");
let batch = db.transaction(); let mut batch = db.transaction();
batch.delete(None, &key1); batch.delete(None, &key1);
db.write(batch).unwrap(); db.write(batch).unwrap();
assert!(db.get(None, &key1).unwrap().is_none()); assert!(db.get(None, &key1).unwrap().is_none());
let batch = db.transaction(); let mut batch = db.transaction();
batch.put(None, &key1, b"cat"); batch.put(None, &key1, b"cat");
db.write(batch).unwrap(); db.write(batch).unwrap();
let transaction = db.transaction(); let mut transaction = db.transaction();
transaction.put(None, &key3, b"elephant"); transaction.put(None, &key3, b"elephant");
transaction.delete(None, &key1); transaction.delete(None, &key1);
db.write(transaction).unwrap(); db.write(transaction).unwrap();
@ -459,7 +459,7 @@ mod tests {
assert_eq!(&*db.get_by_prefix(None, &key3).unwrap(), b"elephant"); assert_eq!(&*db.get_by_prefix(None, &key3).unwrap(), b"elephant");
assert_eq!(&*db.get_by_prefix(None, &key2).unwrap(), b"dog"); assert_eq!(&*db.get_by_prefix(None, &key2).unwrap(), b"dog");
let transaction = db.transaction(); let mut transaction = db.transaction();
transaction.put(None, &key1, b"horse"); transaction.put(None, &key1, b"horse");
transaction.delete(None, &key3); transaction.delete(None, &key3);
db.write_buffered(transaction); db.write_buffered(transaction);

View File

@ -72,7 +72,7 @@ impl Batch {
pub fn commit(&mut self, dest: &mut Database) -> Result<(), Error> { pub fn commit(&mut self, dest: &mut Database) -> Result<(), Error> {
if self.inner.is_empty() { return Ok(()) } if self.inner.is_empty() { return Ok(()) }
let transaction = DBTransaction::new(dest); let mut transaction = DBTransaction::new(dest);
for keypair in &self.inner { for keypair in &self.inner {
transaction.put(self.column, &keypair.0, &keypair.1); transaction.put(self.column, &keypair.0, &keypair.1);

View File

@ -35,7 +35,7 @@ fn db_path(path: &Path) -> PathBuf {
fn make_db(path: &Path, pairs: BTreeMap<Vec<u8>, Vec<u8>>) { fn make_db(path: &Path, pairs: BTreeMap<Vec<u8>, Vec<u8>>) {
let db = Database::open_default(path.to_str().unwrap()).expect("failed to open temp database"); let db = Database::open_default(path.to_str().unwrap()).expect("failed to open temp database");
{ {
let transaction = db.transaction(); let mut transaction = db.transaction();
for (k, v) in pairs { for (k, v) in pairs {
transaction.put(None, &k, &v); transaction.put(None, &k, &v);
} }

View File

@ -58,13 +58,13 @@ impl OverlayDB {
/// Commit all operations in a single batch. /// Commit all operations in a single batch.
#[cfg(test)] #[cfg(test)]
pub fn commit(&mut self) -> Result<u32, UtilError> { pub fn commit(&mut self) -> Result<u32, UtilError> {
let batch = self.backing.transaction(); let mut batch = self.backing.transaction();
let res = try!(self.commit_to_batch(&batch)); let res = try!(self.commit_to_batch(&mut batch));
self.backing.write(batch).map(|_| res).map_err(|e| e.into()) self.backing.write(batch).map(|_| res).map_err(|e| e.into())
} }
/// Commit all operations to given batch. /// Commit all operations to given batch.
pub fn commit_to_batch(&mut self, batch: &DBTransaction) -> Result<u32, UtilError> { pub fn commit_to_batch(&mut self, batch: &mut DBTransaction) -> Result<u32, UtilError> {
let mut ret = 0u32; let mut ret = 0u32;
let mut deletes = 0usize; let mut deletes = 0usize;
for i in self.overlay.drain().into_iter() { for i in self.overlay.drain().into_iter() {
@ -111,7 +111,7 @@ impl OverlayDB {
} }
/// Put the refs and value of the given key, possibly deleting it from the db. /// Put the refs and value of the given key, possibly deleting it from the db.
fn put_payload_in_batch(&self, batch: &DBTransaction, key: &H256, payload: (Bytes, u32)) -> bool { fn put_payload_in_batch(&self, batch: &mut DBTransaction, key: &H256, payload: (Bytes, u32)) -> bool {
if payload.1 > 0 { if payload.1 > 0 {
let mut s = RlpStream::new_list(2); let mut s = RlpStream::new_list(2);
s.append(&payload.1); s.append(&payload.1);
@ -195,8 +195,8 @@ impl HashDB for OverlayDB {
fn overlaydb_revert() { fn overlaydb_revert() {
let mut m = OverlayDB::new_temp(); let mut m = OverlayDB::new_temp();
let foo = m.insert(b"foo"); // insert foo. let foo = m.insert(b"foo"); // insert foo.
let batch = m.backing.transaction(); let mut batch = m.backing.transaction();
m.commit_to_batch(&batch).unwrap(); // commit - new operations begin here... m.commit_to_batch(&mut batch).unwrap(); // commit - new operations begin here...
m.backing.write(batch).unwrap(); m.backing.write(batch).unwrap();
let bar = m.insert(b"bar"); // insert bar. let bar = m.insert(b"bar"); // insert bar.
m.remove(&foo); // remove foo. m.remove(&foo); // remove foo.
@ -300,7 +300,7 @@ fn playpen() {
use std::fs; use std::fs;
{ {
let db = Database::open_default("/tmp/test").unwrap(); let db = Database::open_default("/tmp/test").unwrap();
let batch = db.transaction(); let mut batch = db.transaction();
batch.put(None, b"test", b"test2"); batch.put(None, b"test", b"test2");
db.write(batch).unwrap(); db.write(batch).unwrap();
match db.get(None, b"test") { match db.get(None, b"test") {
@ -308,7 +308,7 @@ fn playpen() {
Ok(None) => println!("No value for that key"), Ok(None) => println!("No value for that key"),
Err(..) => println!("Gah"), Err(..) => println!("Gah"),
} }
let batch = db.transaction(); let mut batch = db.transaction();
batch.delete(None, b"test"); batch.delete(None, b"test");
db.write(batch).unwrap(); db.write(batch).unwrap();
} }