header_chain writes to database
This commit is contained in:
parent
fe52e969b6
commit
d013a13be6
@ -23,9 +23,9 @@
|
|||||||
//! This is separate from the `BlockChain` for two reasons:
|
//! This is separate from the `BlockChain` for two reasons:
|
||||||
//! - It stores only headers (and a pruned subset of them)
|
//! - It stores only headers (and a pruned subset of them)
|
||||||
//! - To allow for flexibility in the database layout once that's incorporated.
|
//! - To allow for flexibility in the database layout once that's incorporated.
|
||||||
// TODO: use DB instead of memory. DB Layout: just the contents of `candidates`/`headers`
|
|
||||||
|
|
||||||
use std::collections::{BTreeMap, HashMap};
|
use std::collections::{BTreeMap, HashMap};
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use cht;
|
use cht;
|
||||||
|
|
||||||
@ -34,7 +34,10 @@ use ethcore::error::BlockError;
|
|||||||
use ethcore::encoded;
|
use ethcore::encoded;
|
||||||
use ethcore::header::Header;
|
use ethcore::header::Header;
|
||||||
use ethcore::ids::BlockId;
|
use ethcore::ids::BlockId;
|
||||||
use util::{H256, U256, HeapSizeOf, Mutex, RwLock};
|
|
||||||
|
use rlp::{Encodable, Decodable, Decoder, DecoderError, RlpStream, View};
|
||||||
|
use util::{H256, U256, HeapSizeOf, RwLock};
|
||||||
|
use util::kvdb::{DBTransaction, KeyValueDB};
|
||||||
|
|
||||||
use smallvec::SmallVec;
|
use smallvec::SmallVec;
|
||||||
|
|
||||||
@ -43,6 +46,9 @@ use smallvec::SmallVec;
|
|||||||
/// relevant to any blocks we've got in memory.
|
/// relevant to any blocks we've got in memory.
|
||||||
const HISTORY: u64 = 2048;
|
const HISTORY: u64 = 2048;
|
||||||
|
|
||||||
|
/// The best block key. Maps to a `u64` best block number.
|
||||||
|
const BEST_KEY: &'static [u8] = &*b"best_block_key";
|
||||||
|
|
||||||
/// Information about a block.
|
/// Information about a block.
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub struct BlockDescriptor {
|
pub struct BlockDescriptor {
|
||||||
@ -75,22 +81,109 @@ impl HeapSizeOf for Entry {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Encodable for Entry {
|
||||||
|
fn rlp_append(&self, s: &mut RlpStream) {
|
||||||
|
s.begin_list(self.candidates.len());
|
||||||
|
|
||||||
|
for candidate in &self.candidates {
|
||||||
|
s.begin_list(3)
|
||||||
|
.append(&candidate.hash)
|
||||||
|
.append(&candidate.parent_hash)
|
||||||
|
.append(&candidate.total_difficulty);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Decodable for Entry {
|
||||||
|
fn decode<D: Decoder>(decoder: &D) -> Result<Self, DecoderError> {
|
||||||
|
let rlp = decoder.as_rlp();
|
||||||
|
|
||||||
|
let mut candidates = SmallVec::<[Candidate; 3]>::new();
|
||||||
|
|
||||||
|
for item in rlp.iter() {
|
||||||
|
candidates.push(Candidate {
|
||||||
|
hash: item.val_at(0)?,
|
||||||
|
parent_hash: item.val_at(1)?,
|
||||||
|
total_difficulty: item.val_at(2)?,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
if candidates.is_empty() { return Err(DecoderError::Custom("Empty candidates vector submitted.")) }
|
||||||
|
|
||||||
|
// rely on the invariant that the canonical entry is always first.
|
||||||
|
let canon_hash = candidates[0].hash;
|
||||||
|
Ok(Entry {
|
||||||
|
candidates: candidates,
|
||||||
|
canonical_hash: canon_hash,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn cht_key(number: u64) -> String {
|
||||||
|
format!("canonical_{}", number)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn era_key(number: u64) -> String {
|
||||||
|
format!("candidates_{}", number)
|
||||||
|
}
|
||||||
|
|
||||||
/// Header chain. See module docs for more details.
|
/// Header chain. See module docs for more details.
|
||||||
pub struct HeaderChain {
|
pub struct HeaderChain {
|
||||||
genesis_header: encoded::Header, // special-case the genesis.
|
genesis_header: encoded::Header, // special-case the genesis.
|
||||||
candidates: RwLock<BTreeMap<u64, Entry>>,
|
candidates: RwLock<BTreeMap<u64, Entry>>,
|
||||||
headers: RwLock<HashMap<H256, encoded::Header>>,
|
headers: RwLock<HashMap<H256, encoded::Header>>,
|
||||||
best_block: RwLock<BlockDescriptor>,
|
best_block: RwLock<BlockDescriptor>,
|
||||||
cht_roots: Mutex<Vec<H256>>,
|
db: Arc<KeyValueDB>,
|
||||||
|
col: Option<u32>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl HeaderChain {
|
impl HeaderChain {
|
||||||
/// Create a new header chain given this genesis block.
|
/// Create a new header chain given this genesis block and database to read from.
|
||||||
pub fn new(genesis: &[u8]) -> Self {
|
pub fn new(db: Arc<KeyValueDB>, col: Option<u32>, genesis: &[u8]) -> Result<Self, String> {
|
||||||
use ethcore::views::HeaderView;
|
use ethcore::views::HeaderView;
|
||||||
|
|
||||||
let g_view = HeaderView::new(genesis);
|
let chain = if let Some(best_number) = db.get(col, BEST_KEY)?.map(|x| ::rlp::decode(&x)) {
|
||||||
|
let mut cur_number = best_number;
|
||||||
|
let mut candidates = BTreeMap::new();
|
||||||
|
let mut headers = HashMap::new();
|
||||||
|
|
||||||
|
// load all era entries and referenced headers within them.
|
||||||
|
while let Some(entry) = db.get(col, era_key(cur_number).as_bytes())? {
|
||||||
|
let entry: Entry = ::rlp::decode(&entry);
|
||||||
|
for candidate in &entry.candidates {
|
||||||
|
match db.get(col, &*candidate.hash)? {
|
||||||
|
Some(hdr) => headers.insert(candidate.hash, encoded::Header::new(hdr.to_vec())),
|
||||||
|
None => return Err(format!("Database missing referenced header: {}", candidate.hash)),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
candidates.insert(cur_number, entry);
|
||||||
|
|
||||||
|
cur_number -= 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
// fill best block block descriptor.
|
||||||
|
if candidates.is_empty() { return Err(format!("Database corrupt: best block referenced but no data.")) }
|
||||||
|
let best_block = {
|
||||||
|
let era = candidates.get(&best_number)
|
||||||
|
.expect("candidates non-empty; filled in loop starting at best_number; qed");
|
||||||
|
let best = &era.candidates[0];
|
||||||
|
BlockDescriptor {
|
||||||
|
hash: best.hash,
|
||||||
|
number: best_number,
|
||||||
|
total_difficulty: best.total_difficulty,
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
HeaderChain {
|
||||||
|
genesis_header: encoded::Header::new(genesis.to_owned()),
|
||||||
|
best_block: RwLock::new(best_block),
|
||||||
|
candidates: RwLock::new(candidates),
|
||||||
|
headers: RwLock::new(headers),
|
||||||
|
db: db,
|
||||||
|
col: col,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
let g_view = HeaderView::new(genesis);
|
||||||
HeaderChain {
|
HeaderChain {
|
||||||
genesis_header: encoded::Header::new(genesis.to_owned()),
|
genesis_header: encoded::Header::new(genesis.to_owned()),
|
||||||
best_block: RwLock::new(BlockDescriptor {
|
best_block: RwLock::new(BlockDescriptor {
|
||||||
@ -100,14 +193,18 @@ impl HeaderChain {
|
|||||||
}),
|
}),
|
||||||
candidates: RwLock::new(BTreeMap::new()),
|
candidates: RwLock::new(BTreeMap::new()),
|
||||||
headers: RwLock::new(HashMap::new()),
|
headers: RwLock::new(HashMap::new()),
|
||||||
cht_roots: Mutex::new(Vec::new()),
|
db: db,
|
||||||
|
col: col,
|
||||||
}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(chain)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Insert a pre-verified header.
|
/// Insert a pre-verified header.
|
||||||
///
|
///
|
||||||
/// This blindly trusts that the data given to it is sensible.
|
/// This blindly trusts that the data given to it is sensible.
|
||||||
pub fn insert(&self, header: Header) -> Result<(), BlockError> {
|
pub fn insert(&self, transaction: &mut DBTransaction, header: Header) -> Result<(), BlockError> {
|
||||||
let hash = header.hash();
|
let hash = header.hash();
|
||||||
let number = header.number();
|
let number = header.number();
|
||||||
let parent_hash = *header.parent_hash();
|
let parent_hash = *header.parent_hash();
|
||||||
@ -129,15 +226,19 @@ impl HeaderChain {
|
|||||||
let total_difficulty = parent_td + *header.difficulty();
|
let total_difficulty = parent_td + *header.difficulty();
|
||||||
|
|
||||||
// insert headers and candidates entries.
|
// insert headers and candidates entries.
|
||||||
candidates.entry(number).or_insert_with(|| Entry { candidates: SmallVec::new(), canonical_hash: hash })
|
{
|
||||||
.candidates.push(Candidate {
|
let cur_era = candidates.entry(number)
|
||||||
|
.or_insert_with(|| Entry { candidates: SmallVec::new(), canonical_hash: hash });
|
||||||
|
cur_era.candidates.push(Candidate {
|
||||||
hash: hash,
|
hash: hash,
|
||||||
parent_hash: parent_hash,
|
parent_hash: parent_hash,
|
||||||
total_difficulty: total_difficulty,
|
total_difficulty: total_difficulty,
|
||||||
});
|
});
|
||||||
|
}
|
||||||
|
|
||||||
let raw = ::rlp::encode(&header).to_vec();
|
let raw = ::rlp::encode(&header);
|
||||||
self.headers.write().insert(hash, encoded::Header::new(raw));
|
transaction.put(self.col, &hash[..], &*raw);
|
||||||
|
self.headers.write().insert(hash, encoded::Header::new(raw.to_vec()));
|
||||||
|
|
||||||
// reorganize ancestors so canonical entries are first in their
|
// reorganize ancestors so canonical entries are first in their
|
||||||
// respective candidates vectors.
|
// respective candidates vectors.
|
||||||
@ -160,6 +261,10 @@ impl HeaderChain {
|
|||||||
// what about reorgs > cht::SIZE + HISTORY?
|
// what about reorgs > cht::SIZE + HISTORY?
|
||||||
// resetting to the last block of a given CHT should be possible.
|
// resetting to the last block of a given CHT should be possible.
|
||||||
canon_hash = entry.candidates[0].parent_hash;
|
canon_hash = entry.candidates[0].parent_hash;
|
||||||
|
|
||||||
|
// write altered era to disk.
|
||||||
|
let rlp_era = ::rlp::encode(&*entry);
|
||||||
|
transaction.put(self.col, era_key(height).as_bytes(), &rlp_era);
|
||||||
}
|
}
|
||||||
|
|
||||||
trace!(target: "chain", "New best block: ({}, {}), TD {}", number, hash, total_difficulty);
|
trace!(target: "chain", "New best block: ({}, {}), TD {}", number, hash, total_difficulty);
|
||||||
@ -168,13 +273,13 @@ impl HeaderChain {
|
|||||||
number: number,
|
number: number,
|
||||||
total_difficulty: total_difficulty,
|
total_difficulty: total_difficulty,
|
||||||
};
|
};
|
||||||
|
transaction.put(self.col, BEST_KEY, &*::rlp::encode(&number));
|
||||||
|
|
||||||
// produce next CHT root if it's time.
|
// produce next CHT root if it's time.
|
||||||
let earliest_era = *candidates.keys().next().expect("at least one era just created; qed");
|
let earliest_era = *candidates.keys().next().expect("at least one era just created; qed");
|
||||||
if earliest_era + HISTORY + cht::SIZE <= number {
|
if earliest_era + HISTORY + cht::SIZE <= number {
|
||||||
let cht_num = cht::block_to_cht_number(earliest_era)
|
let cht_num = cht::block_to_cht_number(earliest_era)
|
||||||
.expect("fails only for number == 0; genesis never imported; qed");
|
.expect("fails only for number == 0; genesis never imported; qed");
|
||||||
debug_assert_eq!(cht_num as usize, self.cht_roots.lock().len());
|
|
||||||
|
|
||||||
let mut headers = self.headers.write();
|
let mut headers = self.headers.write();
|
||||||
|
|
||||||
@ -186,10 +291,13 @@ impl HeaderChain {
|
|||||||
let iter = || {
|
let iter = || {
|
||||||
let era_entry = candidates.remove(&i)
|
let era_entry = candidates.remove(&i)
|
||||||
.expect("all eras are sequential with no gaps; qed");
|
.expect("all eras are sequential with no gaps; qed");
|
||||||
|
transaction.delete(self.col, era_key(i).as_bytes());
|
||||||
|
|
||||||
i += 1;
|
i += 1;
|
||||||
|
|
||||||
for ancient in &era_entry.candidates {
|
for ancient in &era_entry.candidates {
|
||||||
headers.remove(&ancient.hash);
|
headers.remove(&ancient.hash);
|
||||||
|
transaction.delete(self.col, &ancient.hash);
|
||||||
}
|
}
|
||||||
|
|
||||||
let canon = &era_entry.candidates[0];
|
let canon = &era_entry.candidates[0];
|
||||||
@ -199,9 +307,9 @@ impl HeaderChain {
|
|||||||
.expect("fails only when too few items; this is checked; qed")
|
.expect("fails only when too few items; this is checked; qed")
|
||||||
};
|
};
|
||||||
|
|
||||||
|
// write the CHT root to the database.
|
||||||
debug!(target: "chain", "Produced CHT {} root: {:?}", cht_num, cht_root);
|
debug!(target: "chain", "Produced CHT {} root: {:?}", cht_num, cht_root);
|
||||||
|
transaction.put(self.col, cht_key(cht_num).as_bytes(), &::rlp::encode(&cht_root));
|
||||||
self.cht_roots.lock().push(cht_root);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -257,7 +365,13 @@ impl HeaderChain {
|
|||||||
/// This is because it's assumed that the genesis hash is known,
|
/// This is because it's assumed that the genesis hash is known,
|
||||||
/// so including it within a CHT would be redundant.
|
/// so including it within a CHT would be redundant.
|
||||||
pub fn cht_root(&self, n: usize) -> Option<H256> {
|
pub fn cht_root(&self, n: usize) -> Option<H256> {
|
||||||
self.cht_roots.lock().get(n).map(|h| h.clone())
|
match self.db.get(self.col, cht_key(n as u64).as_bytes()) {
|
||||||
|
Ok(val) => val.map(|x| ::rlp::decode(&x)),
|
||||||
|
Err(e) => {
|
||||||
|
warn!(target: "chain", "Error reading from database: {}", e);
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the genesis hash.
|
/// Get the genesis hash.
|
||||||
@ -297,8 +411,7 @@ impl HeaderChain {
|
|||||||
impl HeapSizeOf for HeaderChain {
|
impl HeapSizeOf for HeaderChain {
|
||||||
fn heap_size_of_children(&self) -> usize {
|
fn heap_size_of_children(&self) -> usize {
|
||||||
self.candidates.read().heap_size_of_children() +
|
self.candidates.read().heap_size_of_children() +
|
||||||
self.headers.read().heap_size_of_children() +
|
self.headers.read().heap_size_of_children()
|
||||||
self.cht_roots.lock().heap_size_of_children()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -324,16 +437,23 @@ impl<'a> Iterator for AncestryIter<'a> {
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::HeaderChain;
|
use super::HeaderChain;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use ethcore::ids::BlockId;
|
use ethcore::ids::BlockId;
|
||||||
use ethcore::header::Header;
|
use ethcore::header::Header;
|
||||||
use ethcore::spec::Spec;
|
use ethcore::spec::Spec;
|
||||||
|
|
||||||
|
fn make_db() -> Arc<::util::KeyValueDB> {
|
||||||
|
Arc::new(::util::kvdb::in_memory(0))
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn basic_chain() {
|
fn basic_chain() {
|
||||||
let spec = Spec::new_test();
|
let spec = Spec::new_test();
|
||||||
let genesis_header = spec.genesis_header();
|
let genesis_header = spec.genesis_header();
|
||||||
|
let db = make_db();
|
||||||
|
|
||||||
let chain = HeaderChain::new(&::rlp::encode(&genesis_header));
|
let chain = HeaderChain::new(db.clone(), None, &::rlp::encode(&genesis_header)).unwrap();
|
||||||
|
|
||||||
let mut parent_hash = genesis_header.hash();
|
let mut parent_hash = genesis_header.hash();
|
||||||
let mut rolling_timestamp = genesis_header.timestamp();
|
let mut rolling_timestamp = genesis_header.timestamp();
|
||||||
@ -345,7 +465,9 @@ mod tests {
|
|||||||
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
||||||
parent_hash = header.hash();
|
parent_hash = header.hash();
|
||||||
|
|
||||||
chain.insert(header).unwrap();
|
let mut tx = db.transaction();
|
||||||
|
chain.insert(&mut tx, header).unwrap();
|
||||||
|
db.write(tx).unwrap();
|
||||||
|
|
||||||
rolling_timestamp += 10;
|
rolling_timestamp += 10;
|
||||||
}
|
}
|
||||||
@ -361,7 +483,8 @@ mod tests {
|
|||||||
let spec = Spec::new_test();
|
let spec = Spec::new_test();
|
||||||
let genesis_header = spec.genesis_header();
|
let genesis_header = spec.genesis_header();
|
||||||
|
|
||||||
let chain = HeaderChain::new(&::rlp::encode(&genesis_header));
|
let db = make_db();
|
||||||
|
let chain = HeaderChain::new(db.clone(), None, &::rlp::encode(&genesis_header)).unwrap();
|
||||||
|
|
||||||
let mut parent_hash = genesis_header.hash();
|
let mut parent_hash = genesis_header.hash();
|
||||||
let mut rolling_timestamp = genesis_header.timestamp();
|
let mut rolling_timestamp = genesis_header.timestamp();
|
||||||
@ -373,7 +496,9 @@ mod tests {
|
|||||||
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
||||||
parent_hash = header.hash();
|
parent_hash = header.hash();
|
||||||
|
|
||||||
chain.insert(header).unwrap();
|
let mut tx = db.transaction();
|
||||||
|
chain.insert(&mut tx, header).unwrap();
|
||||||
|
db.write(tx).unwrap();
|
||||||
|
|
||||||
rolling_timestamp += 10;
|
rolling_timestamp += 10;
|
||||||
}
|
}
|
||||||
@ -389,7 +514,9 @@ mod tests {
|
|||||||
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
||||||
parent_hash = header.hash();
|
parent_hash = header.hash();
|
||||||
|
|
||||||
chain.insert(header).unwrap();
|
let mut tx = db.transaction();
|
||||||
|
chain.insert(&mut tx, header).unwrap();
|
||||||
|
db.write(tx).unwrap();
|
||||||
|
|
||||||
rolling_timestamp += 10;
|
rolling_timestamp += 10;
|
||||||
}
|
}
|
||||||
@ -410,7 +537,9 @@ mod tests {
|
|||||||
header.set_difficulty(*genesis_header.difficulty() * (i * i).into());
|
header.set_difficulty(*genesis_header.difficulty() * (i * i).into());
|
||||||
parent_hash = header.hash();
|
parent_hash = header.hash();
|
||||||
|
|
||||||
chain.insert(header).unwrap();
|
let mut tx = db.transaction();
|
||||||
|
chain.insert(&mut tx, header).unwrap();
|
||||||
|
db.write(tx).unwrap();
|
||||||
|
|
||||||
rolling_timestamp += 11;
|
rolling_timestamp += 11;
|
||||||
}
|
}
|
||||||
@ -432,11 +561,46 @@ mod tests {
|
|||||||
fn earliest_is_latest() {
|
fn earliest_is_latest() {
|
||||||
let spec = Spec::new_test();
|
let spec = Spec::new_test();
|
||||||
let genesis_header = spec.genesis_header();
|
let genesis_header = spec.genesis_header();
|
||||||
|
let db = make_db();
|
||||||
|
|
||||||
let chain = HeaderChain::new(&::rlp::encode(&genesis_header));
|
let chain = HeaderChain::new(db.clone(), None, &::rlp::encode(&genesis_header)).unwrap();
|
||||||
|
|
||||||
assert!(chain.block_header(BlockId::Earliest).is_some());
|
assert!(chain.block_header(BlockId::Earliest).is_some());
|
||||||
assert!(chain.block_header(BlockId::Latest).is_some());
|
assert!(chain.block_header(BlockId::Latest).is_some());
|
||||||
assert!(chain.block_header(BlockId::Pending).is_some());
|
assert!(chain.block_header(BlockId::Pending).is_some());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn restore_from_db() {
|
||||||
|
let spec = Spec::new_test();
|
||||||
|
let genesis_header = spec.genesis_header();
|
||||||
|
let db = make_db();
|
||||||
|
|
||||||
|
{
|
||||||
|
let chain = HeaderChain::new(db.clone(), None, &::rlp::encode(&genesis_header)).unwrap();
|
||||||
|
let mut parent_hash = genesis_header.hash();
|
||||||
|
let mut rolling_timestamp = genesis_header.timestamp();
|
||||||
|
for i in 1..10000 {
|
||||||
|
let mut header = Header::new();
|
||||||
|
header.set_parent_hash(parent_hash);
|
||||||
|
header.set_number(i);
|
||||||
|
header.set_timestamp(rolling_timestamp);
|
||||||
|
header.set_difficulty(*genesis_header.difficulty() * i.into());
|
||||||
|
parent_hash = header.hash();
|
||||||
|
|
||||||
|
let mut tx = db.transaction();
|
||||||
|
chain.insert(&mut tx, header).unwrap();
|
||||||
|
db.write(tx).unwrap();
|
||||||
|
|
||||||
|
rolling_timestamp += 10;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let chain = HeaderChain::new(db.clone(), None, &::rlp::encode(&genesis_header)).unwrap();
|
||||||
|
assert!(chain.block_header(BlockId::Number(10)).is_none());
|
||||||
|
assert!(chain.block_header(BlockId::Number(9000)).is_some());
|
||||||
|
assert!(chain.cht_root(2).is_some());
|
||||||
|
assert!(chain.cht_root(3).is_none());
|
||||||
|
assert_eq!(chain.block_header(BlockId::Latest).unwrap().number(), 9999);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -111,10 +111,14 @@ pub struct Client {
|
|||||||
impl Client {
|
impl Client {
|
||||||
/// Create a new `Client`.
|
/// Create a new `Client`.
|
||||||
pub fn new(config: Config, spec: &Spec, io_channel: IoChannel<ClientIoMessage>) -> Self {
|
pub fn new(config: Config, spec: &Spec, io_channel: IoChannel<ClientIoMessage>) -> Self {
|
||||||
|
// TODO: use real DB.
|
||||||
|
let db = ::util::kvdb::in_memory(0);
|
||||||
|
let gh = ::rlp::encode(&spec.genesis_header());
|
||||||
|
|
||||||
Client {
|
Client {
|
||||||
queue: HeaderQueue::new(config.queue, spec.engine.clone(), io_channel, true),
|
queue: HeaderQueue::new(config.queue, spec.engine.clone(), io_channel, true),
|
||||||
engine: spec.engine.clone(),
|
engine: spec.engine.clone(),
|
||||||
chain: HeaderChain::new(&::rlp::encode(&spec.genesis_header())),
|
chain: HeaderChain::new(Arc::new(db), None, &gh).expect("new db every time"),
|
||||||
report: RwLock::new(ClientReport::default()),
|
report: RwLock::new(ClientReport::default()),
|
||||||
import_lock: Mutex::new(()),
|
import_lock: Mutex::new(()),
|
||||||
}
|
}
|
||||||
@ -201,7 +205,8 @@ impl Client {
|
|||||||
for verified_header in self.queue.drain(MAX) {
|
for verified_header in self.queue.drain(MAX) {
|
||||||
let (num, hash) = (verified_header.number(), verified_header.hash());
|
let (num, hash) = (verified_header.number(), verified_header.hash());
|
||||||
|
|
||||||
match self.chain.insert(verified_header) {
|
let mut tx = unimplemented!();
|
||||||
|
match self.chain.insert(&mut tx, verified_header) {
|
||||||
Ok(()) => {
|
Ok(()) => {
|
||||||
good.push(hash);
|
good.push(hash);
|
||||||
self.report.write().blocks_imported += 1;
|
self.report.write().blocks_imported += 1;
|
||||||
|
Loading…
Reference in New Issue
Block a user