Merge branch 'master' into updating_clippy
Conflicts: util/src/journaldb/archivedb.rs
This commit is contained in:
commit
d0f7555348
@ -13,17 +13,14 @@ pub struct AccountDB<'db> {
|
|||||||
|
|
||||||
#[inline]
|
#[inline]
|
||||||
fn combine_key<'a>(address: &'a H256, key: &'a H256) -> H256 {
|
fn combine_key<'a>(address: &'a H256, key: &'a H256) -> H256 {
|
||||||
let mut addr_hash = address.sha3();
|
address ^ key
|
||||||
// preserve 96 bits of original key for db lookup
|
|
||||||
addr_hash[0..12].clone_from_slice(&[0u8; 12]);
|
|
||||||
&addr_hash ^ key
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'db> AccountDB<'db> {
|
impl<'db> AccountDB<'db> {
|
||||||
pub fn new(db: &'db HashDB, address: &Address) -> AccountDB<'db> {
|
pub fn new(db: &'db HashDB, address: &Address) -> AccountDB<'db> {
|
||||||
AccountDB {
|
AccountDB {
|
||||||
db: db,
|
db: db,
|
||||||
address: x!(address.clone()),
|
address: x!(address),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -70,7 +67,7 @@ impl<'db> AccountDBMut<'db> {
|
|||||||
pub fn new(db: &'db mut HashDB, address: &Address) -> AccountDBMut<'db> {
|
pub fn new(db: &'db mut HashDB, address: &Address) -> AccountDBMut<'db> {
|
||||||
AccountDBMut {
|
AccountDBMut {
|
||||||
db: db,
|
db: db,
|
||||||
address: x!(address.clone()),
|
address: x!(address),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -116,7 +116,7 @@ pub struct Client<V = CanonVerifier> where V: Verifier {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const HISTORY: u64 = 1000;
|
const HISTORY: u64 = 1000;
|
||||||
const CLIENT_DB_VER_STR: &'static str = "5.1";
|
const CLIENT_DB_VER_STR: &'static str = "5.2";
|
||||||
|
|
||||||
impl Client<CanonVerifier> {
|
impl Client<CanonVerifier> {
|
||||||
/// Create a new client with given spec and DB path.
|
/// Create a new client with given spec and DB path.
|
||||||
|
@ -81,10 +81,9 @@ Protocol Options:
|
|||||||
--testnet Equivalent to --chain testnet (geth-compatible).
|
--testnet Equivalent to --chain testnet (geth-compatible).
|
||||||
--networkid INDEX Override the network identifier from the chain we are on.
|
--networkid INDEX Override the network identifier from the chain we are on.
|
||||||
--pruning METHOD Configure pruning of the state/storage trie. METHOD may be one of: archive,
|
--pruning METHOD Configure pruning of the state/storage trie. METHOD may be one of: archive,
|
||||||
light (experimental) [default: archive].
|
light (experimental), fast (experimental) [default: archive].
|
||||||
-d --datadir PATH Specify the database & configuration directory path [default: $HOME/.parity]
|
-d --datadir PATH Specify the database & configuration directory path [default: $HOME/.parity]
|
||||||
--db-path PATH Specify the database & configuration directory path [default: $HOME/.parity]
|
--db-path PATH Specify the database & configuration directory path [default: $HOME/.parity]
|
||||||
--pruning Client should prune the state/storage trie.
|
|
||||||
--keys-path PATH Specify the path for JSON key files to be found [default: $HOME/.web3/keys]
|
--keys-path PATH Specify the path for JSON key files to be found [default: $HOME/.web3/keys]
|
||||||
--identity NAME Specify your node's name.
|
--identity NAME Specify your node's name.
|
||||||
|
|
||||||
@ -428,7 +427,7 @@ impl Configuration {
|
|||||||
client_config.pruning = match self.args.flag_pruning.as_str() {
|
client_config.pruning = match self.args.flag_pruning.as_str() {
|
||||||
"" | "archive" => journaldb::Algorithm::Archive,
|
"" | "archive" => journaldb::Algorithm::Archive,
|
||||||
"pruned" => journaldb::Algorithm::EarlyMerge,
|
"pruned" => journaldb::Algorithm::EarlyMerge,
|
||||||
// "fast" => journaldb::Algorithm::OverlayRecent, // TODO: @arkpar uncomment this once option 2 is merged.
|
"fast" => journaldb::Algorithm::OverlayRecent,
|
||||||
// "slow" => journaldb::Algorithm::RefCounted, // TODO: @gavofyork uncomment this once ref-count algo is merged.
|
// "slow" => journaldb::Algorithm::RefCounted, // TODO: @gavofyork uncomment this once ref-count algo is merged.
|
||||||
_ => { die!("Invalid pruning method given."); }
|
_ => { die!("Invalid pruning method given."); }
|
||||||
};
|
};
|
||||||
|
@ -364,7 +364,7 @@ mod tests {
|
|||||||
fn reopen_fork() {
|
fn reopen_fork() {
|
||||||
let mut dir = ::std::env::temp_dir();
|
let mut dir = ::std::env::temp_dir();
|
||||||
dir.push(H32::random().hex());
|
dir.push(H32::random().hex());
|
||||||
let (foo, _bar, _baz) = {
|
let (foo, _, _) = {
|
||||||
let mut jdb = ArchiveDB::new(dir.to_str().unwrap());
|
let mut jdb = ArchiveDB::new(dir.to_str().unwrap());
|
||||||
// history is 1
|
// history is 1
|
||||||
let foo = jdb.insert(b"foo");
|
let foo = jdb.insert(b"foo");
|
||||||
|
@ -22,6 +22,7 @@ use common::*;
|
|||||||
pub mod traits;
|
pub mod traits;
|
||||||
mod archivedb;
|
mod archivedb;
|
||||||
mod optiononedb;
|
mod optiononedb;
|
||||||
|
mod overlay;
|
||||||
|
|
||||||
/// Export the JournalDB trait.
|
/// Export the JournalDB trait.
|
||||||
pub use self::traits::JournalDB;
|
pub use self::traits::JournalDB;
|
||||||
@ -73,6 +74,7 @@ pub fn new(path: &str, algorithm: Algorithm) -> Box<JournalDB> {
|
|||||||
match algorithm {
|
match algorithm {
|
||||||
Algorithm::Archive => Box::new(archivedb::ArchiveDB::new(path)),
|
Algorithm::Archive => Box::new(archivedb::ArchiveDB::new(path)),
|
||||||
Algorithm::EarlyMerge => Box::new(optiononedb::OptionOneDB::new(path)),
|
Algorithm::EarlyMerge => Box::new(optiononedb::OptionOneDB::new(path)),
|
||||||
|
Algorithm::OverlayRecent => Box::new(overlay::JournalOverlayDB::new(path)),
|
||||||
_ => unimplemented!(),
|
_ => unimplemented!(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
873
util/src/journaldb/overlay.rs
Normal file
873
util/src/journaldb/overlay.rs
Normal file
@ -0,0 +1,873 @@
|
|||||||
|
// Copyright 2015, 2016 Ethcore (UK) Ltd.
|
||||||
|
// This file is part of Parity.
|
||||||
|
|
||||||
|
// Parity is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
|
||||||
|
// Parity is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
//! JournalDB over in-memory overlay
|
||||||
|
|
||||||
|
use common::*;
|
||||||
|
use rlp::*;
|
||||||
|
use hashdb::*;
|
||||||
|
use memorydb::*;
|
||||||
|
use kvdb::{Database, DBTransaction, DatabaseConfig};
|
||||||
|
#[cfg(test)]
|
||||||
|
use std::env;
|
||||||
|
use super::JournalDB;
|
||||||
|
|
||||||
|
/// Implementation of the JournalDB trait for a disk-backed database with a memory overlay
|
||||||
|
/// and, possibly, latent-removal semantics.
|
||||||
|
///
|
||||||
|
/// Like OverlayDB, there is a memory overlay; `commit()` must be called in order to
|
||||||
|
/// write operations out to disk. Unlike OverlayDB, `remove()` operations do not take effect
|
||||||
|
/// immediately. Rather some age (based on a linear but arbitrary metric) must pass before
|
||||||
|
/// the removals actually take effect.
|
||||||
|
///
|
||||||
|
/// There are two memory overlays:
|
||||||
|
/// - Transaction overlay contains current transaction data. It is merged with with history
|
||||||
|
/// overlay on each `commit()`
|
||||||
|
/// - History overlay contains all data inserted during the history period. When the node
|
||||||
|
/// in the overlay becomes ancient it is written to disk on `commit()`
|
||||||
|
///
|
||||||
|
/// There is also a journal maintained in memory and on the disk as well which lists insertions
|
||||||
|
/// and removals for each commit during the history period. This is used to track
|
||||||
|
/// data nodes that go out of history scope and must be written to disk.
|
||||||
|
///
|
||||||
|
/// Commit workflow:
|
||||||
|
/// 1. Create a new journal record from the transaction overlay.
|
||||||
|
/// 2. Inseart each node from the transaction overlay into the History overlay increasing reference
|
||||||
|
/// count if it is already there. Note that the reference counting is managed by `MemoryDB`
|
||||||
|
/// 3. Clear the transaction overlay.
|
||||||
|
/// 4. For a canonical journal record that becomes ancient inserts its insertions into the disk DB
|
||||||
|
/// 5. For each journal record that goes out of the history scope (becomes ancient) remove its
|
||||||
|
/// insertions from the history overlay, decreasing the reference counter and removing entry if
|
||||||
|
/// if reaches zero.
|
||||||
|
/// 6. For a canonical journal record that becomes ancient delete its removals from the disk only if
|
||||||
|
/// the removed key is not present in the history overlay.
|
||||||
|
/// 7. Delete ancient record from memory and disk.
|
||||||
|
///
|
||||||
|
pub struct JournalOverlayDB {
|
||||||
|
transaction_overlay: MemoryDB,
|
||||||
|
backing: Arc<Database>,
|
||||||
|
journal_overlay: Arc<RwLock<JournalOverlay>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(PartialEq)]
|
||||||
|
struct JournalOverlay {
|
||||||
|
backing_overlay: MemoryDB,
|
||||||
|
journal: HashMap<u64, Vec<JournalEntry>>,
|
||||||
|
latest_era: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(PartialEq)]
|
||||||
|
struct JournalEntry {
|
||||||
|
id: H256,
|
||||||
|
insertions: Vec<H256>,
|
||||||
|
deletions: Vec<H256>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl HeapSizeOf for JournalEntry {
|
||||||
|
fn heap_size_of_children(&self) -> usize {
|
||||||
|
self.insertions.heap_size_of_children() + self.deletions.heap_size_of_children()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Clone for JournalOverlayDB {
|
||||||
|
fn clone(&self) -> JournalOverlayDB {
|
||||||
|
JournalOverlayDB {
|
||||||
|
transaction_overlay: MemoryDB::new(),
|
||||||
|
backing: self.backing.clone(),
|
||||||
|
journal_overlay: self.journal_overlay.clone(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// all keys must be at least 12 bytes
|
||||||
|
const LATEST_ERA_KEY : [u8; 12] = [ b'l', b'a', b's', b't', 0, 0, 0, 0, 0, 0, 0, 0 ];
|
||||||
|
const VERSION_KEY : [u8; 12] = [ b'j', b'v', b'e', b'r', 0, 0, 0, 0, 0, 0, 0, 0 ];
|
||||||
|
const DB_VERSION : u32 = 0x200 + 3;
|
||||||
|
const PADDING : [u8; 10] = [ 0u8; 10 ];
|
||||||
|
|
||||||
|
impl JournalOverlayDB {
|
||||||
|
/// Create a new instance from file
|
||||||
|
pub fn new(path: &str) -> JournalOverlayDB {
|
||||||
|
Self::from_prefs(path)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new instance from file
|
||||||
|
pub fn from_prefs(path: &str) -> JournalOverlayDB {
|
||||||
|
let opts = DatabaseConfig {
|
||||||
|
prefix_size: Some(12) //use 12 bytes as prefix, this must match account_db prefix
|
||||||
|
};
|
||||||
|
let backing = Database::open(&opts, path).unwrap_or_else(|e| {
|
||||||
|
panic!("Error opening state db: {}", e);
|
||||||
|
});
|
||||||
|
if !backing.is_empty() {
|
||||||
|
match backing.get(&VERSION_KEY).map(|d| d.map(|v| decode::<u32>(&v))) {
|
||||||
|
Ok(Some(DB_VERSION)) => {}
|
||||||
|
v => panic!("Incompatible DB version, expected {}, got {:?}", DB_VERSION, v)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
backing.put(&VERSION_KEY, &encode(&DB_VERSION)).expect("Error writing version to database");
|
||||||
|
}
|
||||||
|
|
||||||
|
let journal_overlay = Arc::new(RwLock::new(JournalOverlayDB::read_overlay(&backing)));
|
||||||
|
JournalOverlayDB {
|
||||||
|
transaction_overlay: MemoryDB::new(),
|
||||||
|
backing: Arc::new(backing),
|
||||||
|
journal_overlay: journal_overlay,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new instance with an anonymous temporary database.
|
||||||
|
#[cfg(test)]
|
||||||
|
pub fn new_temp() -> JournalOverlayDB {
|
||||||
|
let mut dir = env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
Self::new(dir.to_str().unwrap())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
fn can_reconstruct_refs(&self) -> bool {
|
||||||
|
let reconstructed = Self::read_overlay(&self.backing);
|
||||||
|
let journal_overlay = self.journal_overlay.read().unwrap();
|
||||||
|
*journal_overlay == reconstructed
|
||||||
|
}
|
||||||
|
|
||||||
|
fn payload(&self, key: &H256) -> Option<Bytes> {
|
||||||
|
self.backing.get(&key.bytes()).expect("Low-level database error. Some issue with your hard disk?").map(|v| v.to_vec())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn read_overlay(db: &Database) -> JournalOverlay {
|
||||||
|
let mut journal = HashMap::new();
|
||||||
|
let mut overlay = MemoryDB::new();
|
||||||
|
let mut count = 0;
|
||||||
|
let mut latest_era = 0;
|
||||||
|
if let Some(val) = db.get(&LATEST_ERA_KEY).expect("Low-level database error.") {
|
||||||
|
latest_era = decode::<u64>(&val);
|
||||||
|
let mut era = latest_era;
|
||||||
|
loop {
|
||||||
|
let mut index = 0usize;
|
||||||
|
while let Some(rlp_data) = db.get({
|
||||||
|
let mut r = RlpStream::new_list(3);
|
||||||
|
r.append(&era);
|
||||||
|
r.append(&index);
|
||||||
|
r.append(&&PADDING[..]);
|
||||||
|
&r.drain()
|
||||||
|
}).expect("Low-level database error.") {
|
||||||
|
trace!("read_overlay: era={}, index={}", era, index);
|
||||||
|
let rlp = Rlp::new(&rlp_data);
|
||||||
|
let id: H256 = rlp.val_at(0);
|
||||||
|
let insertions = rlp.at(1);
|
||||||
|
let deletions: Vec<H256> = rlp.val_at(2);
|
||||||
|
let mut inserted_keys = Vec::new();
|
||||||
|
for r in insertions.iter() {
|
||||||
|
let k: H256 = r.val_at(0);
|
||||||
|
let v: Bytes = r.val_at(1);
|
||||||
|
overlay.emplace(k.clone(), v);
|
||||||
|
inserted_keys.push(k);
|
||||||
|
count += 1;
|
||||||
|
}
|
||||||
|
journal.entry(era).or_insert_with(Vec::new).push(JournalEntry {
|
||||||
|
id: id,
|
||||||
|
insertions: inserted_keys,
|
||||||
|
deletions: deletions,
|
||||||
|
});
|
||||||
|
index += 1;
|
||||||
|
};
|
||||||
|
if index == 0 || era == 0 {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
era -= 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
trace!("Recovered {} overlay entries, {} journal entries", count, journal.len());
|
||||||
|
JournalOverlay { backing_overlay: overlay, journal: journal, latest_era: latest_era }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl JournalDB for JournalOverlayDB {
|
||||||
|
fn spawn(&self) -> Box<JournalDB> {
|
||||||
|
Box::new(self.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn mem_used(&self) -> usize {
|
||||||
|
let mut mem = self.transaction_overlay.mem_used();
|
||||||
|
let overlay = self.journal_overlay.read().unwrap();
|
||||||
|
mem += overlay.backing_overlay.mem_used();
|
||||||
|
mem += overlay.journal.heap_size_of_children();
|
||||||
|
mem
|
||||||
|
}
|
||||||
|
|
||||||
|
fn is_empty(&self) -> bool {
|
||||||
|
self.backing.get(&LATEST_ERA_KEY).expect("Low level database error").is_none()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn commit(&mut self, now: u64, id: &H256, end: Option<(u64, H256)>) -> Result<u32, UtilError> {
|
||||||
|
// record new commit's details.
|
||||||
|
trace!("commit: #{} ({}), end era: {:?}", now, id, end);
|
||||||
|
let mut journal_overlay = self.journal_overlay.write().unwrap();
|
||||||
|
let batch = DBTransaction::new();
|
||||||
|
{
|
||||||
|
let mut r = RlpStream::new_list(3);
|
||||||
|
let mut tx = self.transaction_overlay.drain();
|
||||||
|
let inserted_keys: Vec<_> = tx.iter().filter_map(|(k, &(_, c))| if c > 0 { Some(k.clone()) } else { None }).collect();
|
||||||
|
let removed_keys: Vec<_> = tx.iter().filter_map(|(k, &(_, c))| if c < 0 { Some(k.clone()) } else { None }).collect();
|
||||||
|
// Increase counter for each inserted key no matter if the block is canonical or not.
|
||||||
|
let insertions = tx.drain().filter_map(|(k, (v, c))| if c > 0 { Some((k, v)) } else { None });
|
||||||
|
r.append(id);
|
||||||
|
r.begin_list(inserted_keys.len());
|
||||||
|
for (k, v) in insertions {
|
||||||
|
r.begin_list(2);
|
||||||
|
r.append(&k);
|
||||||
|
r.append(&v);
|
||||||
|
journal_overlay.backing_overlay.emplace(k, v);
|
||||||
|
}
|
||||||
|
r.append(&removed_keys);
|
||||||
|
|
||||||
|
let mut k = RlpStream::new_list(3);
|
||||||
|
let index = journal_overlay.journal.get(&now).map(|j| j.len()).unwrap_or(0);
|
||||||
|
k.append(&now);
|
||||||
|
k.append(&index);
|
||||||
|
k.append(&&PADDING[..]);
|
||||||
|
try!(batch.put(&k.drain(), r.as_raw()));
|
||||||
|
if now >= journal_overlay.latest_era {
|
||||||
|
try!(batch.put(&LATEST_ERA_KEY, &encode(&now)));
|
||||||
|
journal_overlay.latest_era = now;
|
||||||
|
}
|
||||||
|
journal_overlay.journal.entry(now).or_insert_with(Vec::new).push(JournalEntry { id: id.clone(), insertions: inserted_keys, deletions: removed_keys });
|
||||||
|
}
|
||||||
|
|
||||||
|
let journal_overlay = journal_overlay.deref_mut();
|
||||||
|
// apply old commits' details
|
||||||
|
if let Some((end_era, canon_id)) = end {
|
||||||
|
if let Some(ref mut records) = journal_overlay.journal.get_mut(&end_era) {
|
||||||
|
let mut canon_insertions: Vec<(H256, Bytes)> = Vec::new();
|
||||||
|
let mut canon_deletions: Vec<H256> = Vec::new();
|
||||||
|
let mut overlay_deletions: Vec<H256> = Vec::new();
|
||||||
|
let mut index = 0usize;
|
||||||
|
for mut journal in records.drain(..) {
|
||||||
|
//delete the record from the db
|
||||||
|
let mut r = RlpStream::new_list(3);
|
||||||
|
r.append(&end_era);
|
||||||
|
r.append(&index);
|
||||||
|
r.append(&&PADDING[..]);
|
||||||
|
try!(batch.delete(&r.drain()));
|
||||||
|
trace!("commit: Delete journal for time #{}.{}: {}, (canon was {}): +{} -{} entries", end_era, index, journal.id, canon_id, journal.insertions.len(), journal.deletions.len());
|
||||||
|
{
|
||||||
|
if canon_id == journal.id {
|
||||||
|
for h in &journal.insertions {
|
||||||
|
if let Some(&(ref d, rc)) = journal_overlay.backing_overlay.raw(h) {
|
||||||
|
if rc > 0 {
|
||||||
|
canon_insertions.push((h.clone(), d.clone())); //TODO: optimize this to avoid data copy
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
canon_deletions = journal.deletions;
|
||||||
|
}
|
||||||
|
overlay_deletions.append(&mut journal.insertions);
|
||||||
|
}
|
||||||
|
index += 1;
|
||||||
|
}
|
||||||
|
// apply canon inserts first
|
||||||
|
for (k, v) in canon_insertions {
|
||||||
|
try!(batch.put(&k, &v));
|
||||||
|
}
|
||||||
|
// update the overlay
|
||||||
|
for k in overlay_deletions {
|
||||||
|
journal_overlay.backing_overlay.kill(&k);
|
||||||
|
}
|
||||||
|
// apply canon deletions
|
||||||
|
for k in canon_deletions {
|
||||||
|
if !journal_overlay.backing_overlay.exists(&k) {
|
||||||
|
try!(batch.delete(&k));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
journal_overlay.backing_overlay.purge();
|
||||||
|
}
|
||||||
|
journal_overlay.journal.remove(&end_era);
|
||||||
|
}
|
||||||
|
try!(self.backing.write(batch));
|
||||||
|
Ok(0)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
impl HashDB for JournalOverlayDB {
|
||||||
|
fn keys(&self) -> HashMap<H256, i32> {
|
||||||
|
let mut ret: HashMap<H256, i32> = HashMap::new();
|
||||||
|
for (key, _) in self.backing.iter() {
|
||||||
|
let h = H256::from_slice(key.deref());
|
||||||
|
ret.insert(h, 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
for (key, refs) in self.transaction_overlay.keys().into_iter() {
|
||||||
|
let refs = *ret.get(&key).unwrap_or(&0) + refs;
|
||||||
|
ret.insert(key, refs);
|
||||||
|
}
|
||||||
|
ret
|
||||||
|
}
|
||||||
|
|
||||||
|
fn lookup(&self, key: &H256) -> Option<&[u8]> {
|
||||||
|
let k = self.transaction_overlay.raw(key);
|
||||||
|
match k {
|
||||||
|
Some(&(ref d, rc)) if rc > 0 => Some(d),
|
||||||
|
_ => {
|
||||||
|
let v = self.journal_overlay.read().unwrap().backing_overlay.lookup(key).map(|v| v.to_vec());
|
||||||
|
match v {
|
||||||
|
Some(x) => {
|
||||||
|
Some(&self.transaction_overlay.denote(key, x).0)
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
if let Some(x) = self.payload(key) {
|
||||||
|
Some(&self.transaction_overlay.denote(key, x).0)
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn exists(&self, key: &H256) -> bool {
|
||||||
|
self.lookup(key).is_some()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn insert(&mut self, value: &[u8]) -> H256 {
|
||||||
|
self.transaction_overlay.insert(value)
|
||||||
|
}
|
||||||
|
fn emplace(&mut self, key: H256, value: Bytes) {
|
||||||
|
self.transaction_overlay.emplace(key, value);
|
||||||
|
}
|
||||||
|
fn kill(&mut self, key: &H256) {
|
||||||
|
self.transaction_overlay.kill(key);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use common::*;
|
||||||
|
use super::*;
|
||||||
|
use hashdb::*;
|
||||||
|
use log::init_log;
|
||||||
|
use journaldb::JournalDB;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn insert_same_in_fork() {
|
||||||
|
// history is 1
|
||||||
|
let mut jdb = JournalOverlayDB::new_temp();
|
||||||
|
|
||||||
|
let x = jdb.insert(b"X");
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(2, &b"2".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(3, &b"1002a".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(4, &b"1003a".sha3(), Some((2, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&x);
|
||||||
|
jdb.commit(3, &b"1002b".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
let x = jdb.insert(b"X");
|
||||||
|
jdb.commit(4, &b"1003b".sha3(), Some((2, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.commit(5, &b"1004a".sha3(), Some((3, b"1002a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(6, &b"1005a".sha3(), Some((4, b"1003a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
assert!(jdb.exists(&x));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn long_history() {
|
||||||
|
// history is 3
|
||||||
|
let mut jdb = JournalOverlayDB::new_temp();
|
||||||
|
let h = jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&h));
|
||||||
|
jdb.remove(&h);
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&h));
|
||||||
|
jdb.commit(2, &b"2".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&h));
|
||||||
|
jdb.commit(3, &b"3".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&h));
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(!jdb.exists(&h));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn complex() {
|
||||||
|
// history is 1
|
||||||
|
let mut jdb = JournalOverlayDB::new_temp();
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
let bar = jdb.insert(b"bar");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(jdb.exists(&bar));
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.remove(&bar);
|
||||||
|
let baz = jdb.insert(b"baz");
|
||||||
|
jdb.commit(1, &b"1".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(jdb.exists(&bar));
|
||||||
|
assert!(jdb.exists(&baz));
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.remove(&baz);
|
||||||
|
jdb.commit(2, &b"2".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(!jdb.exists(&bar));
|
||||||
|
assert!(jdb.exists(&baz));
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(3, &b"3".sha3(), Some((2, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(!jdb.exists(&bar));
|
||||||
|
assert!(!jdb.exists(&baz));
|
||||||
|
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((3, b"3".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(!jdb.exists(&foo));
|
||||||
|
assert!(!jdb.exists(&bar));
|
||||||
|
assert!(!jdb.exists(&baz));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn fork() {
|
||||||
|
// history is 1
|
||||||
|
let mut jdb = JournalOverlayDB::new_temp();
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
let bar = jdb.insert(b"bar");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(jdb.exists(&bar));
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
let baz = jdb.insert(b"baz");
|
||||||
|
jdb.commit(1, &b"1a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&bar);
|
||||||
|
jdb.commit(1, &b"1b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(jdb.exists(&bar));
|
||||||
|
assert!(jdb.exists(&baz));
|
||||||
|
|
||||||
|
jdb.commit(2, &b"2b".sha3(), Some((1, b"1b".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(!jdb.exists(&baz));
|
||||||
|
assert!(!jdb.exists(&bar));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn overwrite() {
|
||||||
|
// history is 1
|
||||||
|
let mut jdb = JournalOverlayDB::new_temp();
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(1, &b"1".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
jdb.commit(2, &b"2".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
jdb.commit(3, &b"2".sha3(), Some((0, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn fork_same_key_one() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1c".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
jdb.commit(2, &b"2a".sha3(), Some((1, b"1a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn fork_same_key_other() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1c".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
jdb.commit(2, &b"2b".sha3(), Some((1, b"1b".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn fork_ins_del_ins() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(2, &b"2a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(2, &b"2b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(3, &b"3a".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(3, &b"3b".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.commit(4, &b"4a".sha3(), Some((2, b"2a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.commit(5, &b"5a".sha3(), Some((3, b"3a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn reopen() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
let bar = H256::random();
|
||||||
|
|
||||||
|
let foo = {
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
// history is 1
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.emplace(bar.clone(), b"bar".to_vec());
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
foo
|
||||||
|
};
|
||||||
|
|
||||||
|
{
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(1, &b"1".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(jdb.exists(&bar));
|
||||||
|
jdb.commit(2, &b"2".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(!jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn insert_delete_insert_delete_insert_expunge() {
|
||||||
|
init_log();
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
|
||||||
|
// history is 4
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(2, &b"2".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(3, &b"3".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
// expunge foo
|
||||||
|
jdb.commit(5, &b"5".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn forked_insert_delete_insert_delete_insert_expunge() {
|
||||||
|
init_log();
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
|
||||||
|
// history is 4
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(1, &b"1a".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(1, &b"1b".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(2, &b"2a".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(2, &b"2b".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(3, &b"3a".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(3, &b"3b".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(4, &b"4a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(4, &b"4b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
// expunge foo
|
||||||
|
jdb.commit(5, &b"5".sha3(), Some((1, b"1a".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn broken_assert() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
// history is 1
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(1, &b"1".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
// foo is ancient history.
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(2, &b"2".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(3, &b"3".sha3(), Some((2, b"2".sha3()))).unwrap(); // BROKEN
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((3, b"3".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.commit(5, &b"5".sha3(), Some((4, b"4".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(!jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn reopen_test() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
// history is 4
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(2, &b"2".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(3, &b"3".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
// foo is ancient history.
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
let bar = jdb.insert(b"bar");
|
||||||
|
jdb.commit(5, &b"5".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.remove(&bar);
|
||||||
|
jdb.commit(6, &b"6".sha3(), Some((2, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.insert(b"bar");
|
||||||
|
jdb.commit(7, &b"7".sha3(), Some((3, b"3".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn reopen_remove_three() {
|
||||||
|
init_log();
|
||||||
|
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
|
||||||
|
let foo = b"foo".sha3();
|
||||||
|
|
||||||
|
{
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
// history is 1
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.commit(1, &b"1".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
// foo is ancient history.
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(2, &b"2".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
jdb.insert(b"foo");
|
||||||
|
jdb.commit(3, &b"3".sha3(), Some((1, b"1".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
// incantation to reopen the db
|
||||||
|
}; { let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
|
||||||
|
jdb.remove(&foo);
|
||||||
|
jdb.commit(4, &b"4".sha3(), Some((2, b"2".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
// incantation to reopen the db
|
||||||
|
}; { let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
|
||||||
|
jdb.commit(5, &b"5".sha3(), Some((3, b"3".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
|
||||||
|
// incantation to reopen the db
|
||||||
|
}; { let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
|
||||||
|
jdb.commit(6, &b"6".sha3(), Some((4, b"4".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(!jdb.exists(&foo));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn reopen_fork() {
|
||||||
|
let mut dir = ::std::env::temp_dir();
|
||||||
|
dir.push(H32::random().hex());
|
||||||
|
let (foo, bar, baz) = {
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
// history is 1
|
||||||
|
let foo = jdb.insert(b"foo");
|
||||||
|
let bar = jdb.insert(b"bar");
|
||||||
|
jdb.commit(0, &b"0".sha3(), None).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
jdb.remove(&foo);
|
||||||
|
let baz = jdb.insert(b"baz");
|
||||||
|
jdb.commit(1, &b"1a".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
|
||||||
|
jdb.remove(&bar);
|
||||||
|
jdb.commit(1, &b"1b".sha3(), Some((0, b"0".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
(foo, bar, baz)
|
||||||
|
};
|
||||||
|
|
||||||
|
{
|
||||||
|
let mut jdb = JournalOverlayDB::new(dir.to_str().unwrap());
|
||||||
|
jdb.commit(2, &b"2b".sha3(), Some((1, b"1b".sha3()))).unwrap();
|
||||||
|
assert!(jdb.can_reconstruct_refs());
|
||||||
|
assert!(jdb.exists(&foo));
|
||||||
|
assert!(!jdb.exists(&baz));
|
||||||
|
assert!(!jdb.exists(&bar));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@ -70,6 +70,7 @@ use std::default::Default;
|
|||||||
/// assert!(!m.exists(&k));
|
/// assert!(!m.exists(&k));
|
||||||
/// }
|
/// }
|
||||||
/// ```
|
/// ```
|
||||||
|
#[derive(PartialEq)]
|
||||||
pub struct MemoryDB {
|
pub struct MemoryDB {
|
||||||
data: HashMap<H256, (Bytes, i32)>,
|
data: HashMap<H256, (Bytes, i32)>,
|
||||||
static_null_rlp: (Bytes, i32),
|
static_null_rlp: (Bytes, i32),
|
||||||
|
Loading…
Reference in New Issue
Block a user