Merge branch 'master' into td-strict-config
This commit is contained in:
commit
56311cf0b5
10
Cargo.lock
generated
10
Cargo.lock
generated
@ -204,6 +204,9 @@ dependencies = [
|
|||||||
name = "cc"
|
name = "cc"
|
||||||
version = "1.0.0"
|
version = "1.0.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
dependencies = [
|
||||||
|
"rayon 0.8.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cfg-if"
|
name = "cfg-if"
|
||||||
@ -1361,6 +1364,7 @@ dependencies = [
|
|||||||
"interleaved-ordered 0.1.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
"interleaved-ordered 0.1.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"kvdb 0.1.0",
|
"kvdb 0.1.0",
|
||||||
"log 0.3.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
"log 0.3.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"num_cpus 1.7.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"parking_lot 0.4.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
"parking_lot 0.4.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"regex 0.2.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
"regex 0.2.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"rlp 0.2.1",
|
"rlp 0.2.1",
|
||||||
@ -2654,7 +2658,7 @@ dependencies = [
|
|||||||
[[package]]
|
[[package]]
|
||||||
name = "rocksdb"
|
name = "rocksdb"
|
||||||
version = "0.4.5"
|
version = "0.4.5"
|
||||||
source = "git+https://github.com/paritytech/rust-rocksdb#8c4ad5411c141abc63d562d411053f7ebc1aa00c"
|
source = "git+https://github.com/paritytech/rust-rocksdb#166e14ed63cbd2e44b51267b8b98e4b89b0f236f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"libc 0.2.31 (registry+https://github.com/rust-lang/crates.io-index)",
|
"libc 0.2.31 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"local-encoding 0.2.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
"local-encoding 0.2.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
@ -2664,9 +2668,9 @@ dependencies = [
|
|||||||
[[package]]
|
[[package]]
|
||||||
name = "rocksdb-sys"
|
name = "rocksdb-sys"
|
||||||
version = "0.3.0"
|
version = "0.3.0"
|
||||||
source = "git+https://github.com/paritytech/rust-rocksdb#8c4ad5411c141abc63d562d411053f7ebc1aa00c"
|
source = "git+https://github.com/paritytech/rust-rocksdb#166e14ed63cbd2e44b51267b8b98e4b89b0f236f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"gcc 0.3.54 (registry+https://github.com/rust-lang/crates.io-index)",
|
"cc 1.0.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"libc 0.2.31 (registry+https://github.com/rust-lang/crates.io-index)",
|
"libc 0.2.31 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"snappy-sys 0.1.0 (git+https://github.com/paritytech/rust-snappy)",
|
"snappy-sys 0.1.0 (git+https://github.com/paritytech/rust-snappy)",
|
||||||
]
|
]
|
||||||
|
@ -64,11 +64,7 @@ impl<T: ChainDataFetcher> Service<T> {
|
|||||||
// initialize database.
|
// initialize database.
|
||||||
let mut db_config = DatabaseConfig::with_columns(db::NUM_COLUMNS);
|
let mut db_config = DatabaseConfig::with_columns(db::NUM_COLUMNS);
|
||||||
|
|
||||||
// give all rocksdb cache to the header chain column.
|
db_config.memory_budget = config.db_cache_size;
|
||||||
if let Some(size) = config.db_cache_size {
|
|
||||||
db_config.set_cache(db::COL_LIGHT_CHAIN, size);
|
|
||||||
}
|
|
||||||
|
|
||||||
db_config.compaction = config.db_compaction;
|
db_config.compaction = config.db_compaction;
|
||||||
db_config.wal = config.db_wal;
|
db_config.wal = config.db_wal;
|
||||||
|
|
||||||
|
@ -23,8 +23,6 @@ pub struct Config {
|
|||||||
pub pref_cache_size: usize,
|
pub pref_cache_size: usize,
|
||||||
/// Maximum cache size in bytes.
|
/// Maximum cache size in bytes.
|
||||||
pub max_cache_size: usize,
|
pub max_cache_size: usize,
|
||||||
/// Backing db cache_size
|
|
||||||
pub db_cache_size: Option<usize>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Default for Config {
|
impl Default for Config {
|
||||||
@ -32,8 +30,6 @@ impl Default for Config {
|
|||||||
Config {
|
Config {
|
||||||
pref_cache_size: 1 << 14,
|
pref_cache_size: 1 << 14,
|
||||||
max_cache_size: 1 << 20,
|
max_cache_size: 1 << 20,
|
||||||
db_cache_size: None,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -141,7 +141,7 @@ pub struct ClientConfig {
|
|||||||
pub pruning: journaldb::Algorithm,
|
pub pruning: journaldb::Algorithm,
|
||||||
/// The name of the client instance.
|
/// The name of the client instance.
|
||||||
pub name: String,
|
pub name: String,
|
||||||
/// RocksDB state column cache-size if not default
|
/// RocksDB column cache-size if not default
|
||||||
pub db_cache_size: Option<usize>,
|
pub db_cache_size: Option<usize>,
|
||||||
/// State db compaction profile
|
/// State db compaction profile
|
||||||
pub db_compaction: DatabaseCompactionProfile,
|
pub db_compaction: DatabaseCompactionProfile,
|
||||||
|
@ -79,12 +79,7 @@ impl ClientService {
|
|||||||
|
|
||||||
let mut db_config = DatabaseConfig::with_columns(::db::NUM_COLUMNS);
|
let mut db_config = DatabaseConfig::with_columns(::db::NUM_COLUMNS);
|
||||||
|
|
||||||
// give all rocksdb cache to state column; everything else has its
|
db_config.memory_budget = config.db_cache_size;
|
||||||
// own caches.
|
|
||||||
if let Some(size) = config.db_cache_size {
|
|
||||||
db_config.set_cache(::db::COL_STATE, size);
|
|
||||||
}
|
|
||||||
|
|
||||||
db_config.compaction = config.db_compaction.compaction_profile(client_path);
|
db_config.compaction = config.db_compaction.compaction_profile(client_path);
|
||||||
db_config.wal = config.db_wal;
|
db_config.wal = config.db_wal;
|
||||||
|
|
||||||
|
@ -33,7 +33,9 @@ import styles from './signerPending.css';
|
|||||||
|
|
||||||
@observer
|
@observer
|
||||||
class SignerPending extends Component {
|
class SignerPending extends Component {
|
||||||
static propTypes = {};
|
static propTypes = {
|
||||||
|
className: PropTypes.string
|
||||||
|
};
|
||||||
|
|
||||||
static contextTypes = {
|
static contextTypes = {
|
||||||
api: PropTypes.object.isRequired
|
api: PropTypes.object.isRequired
|
||||||
@ -91,11 +93,13 @@ class SignerPending extends Component {
|
|||||||
);
|
);
|
||||||
|
|
||||||
render () {
|
render () {
|
||||||
|
const { className } = this.props;
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Popup
|
<Popup
|
||||||
wide='very'
|
wide='very'
|
||||||
trigger={
|
trigger={
|
||||||
<div className={ [styles.signerPending].join(' ') }>
|
<div className={ [styles.signerPending, className].join(' ') }>
|
||||||
<Icon
|
<Icon
|
||||||
name={ this.store.pending.length > 0 ? 'bell' : 'bell outline' }
|
name={ this.store.pending.length > 0 ? 'bell' : 'bell outline' }
|
||||||
/>
|
/>
|
||||||
|
@ -52,16 +52,16 @@ $statusHeight: 2.75em;
|
|||||||
}
|
}
|
||||||
|
|
||||||
.plugins {
|
.plugins {
|
||||||
|
align-items: center;
|
||||||
|
display: flex;
|
||||||
flex-grow: 1;
|
flex-grow: 1;
|
||||||
text-align: right;
|
justify-content: flex-end;
|
||||||
vertical-align: middle;
|
|
||||||
|
|
||||||
> div,
|
> div,
|
||||||
> img,
|
> img,
|
||||||
> span {
|
> span {
|
||||||
display: inline-block !important;
|
align-self: center;
|
||||||
margin: 0 0 0 1rem;
|
margin: 0 0 0 1rem;
|
||||||
vertical-align: middle;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
.divider {
|
.divider {
|
||||||
@ -70,9 +70,29 @@ $statusHeight: 2.75em;
|
|||||||
opacity: 0.75;
|
opacity: 0.75;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.health,
|
||||||
|
.pending {
|
||||||
|
/* Same as default line-height, i.e. item same as text following */
|
||||||
|
height: 1.4285rem;
|
||||||
|
}
|
||||||
|
|
||||||
.health {
|
.health {
|
||||||
> span {
|
> span {
|
||||||
|
height: 100%;
|
||||||
|
/* re-instate the original margin, no negative offset */
|
||||||
|
margin-top: 0.2em;
|
||||||
width: auto;
|
width: auto;
|
||||||
|
|
||||||
|
> span {
|
||||||
|
/* default vertical alignment */
|
||||||
|
vertical-align: baseline;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
.pending {
|
||||||
|
> i {
|
||||||
|
height: 100%;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -69,7 +69,7 @@ function Status ({ className = '', upgradeStore }, { api }) {
|
|||||||
className={ styles.health }
|
className={ styles.health }
|
||||||
id='application.status.health'
|
id='application.status.health'
|
||||||
/>
|
/>
|
||||||
<SignerPending />
|
<SignerPending className={ styles.pending } />
|
||||||
|
|
||||||
<div className={ styles.divider } />
|
<div className={ styles.divider } />
|
||||||
<BlockNumber
|
<BlockNumber
|
||||||
|
@ -17,8 +17,10 @@
|
|||||||
use std::cmp::max;
|
use std::cmp::max;
|
||||||
|
|
||||||
const MIN_BC_CACHE_MB: u32 = 4;
|
const MIN_BC_CACHE_MB: u32 = 4;
|
||||||
const MIN_DB_CACHE_MB: u32 = 2;
|
const MIN_DB_CACHE_MB: u32 = 8;
|
||||||
const MIN_BLOCK_QUEUE_SIZE_LIMIT_MB: u32 = 16;
|
const MIN_BLOCK_QUEUE_SIZE_LIMIT_MB: u32 = 16;
|
||||||
|
const DEFAULT_DB_CACHE_SIZE: u32 = 128;
|
||||||
|
const DEFAULT_BC_CACHE_SIZE: u32 = 8;
|
||||||
const DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB: u32 = 40;
|
const DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB: u32 = 40;
|
||||||
const DEFAULT_TRACE_CACHE_SIZE: u32 = 20;
|
const DEFAULT_TRACE_CACHE_SIZE: u32 = 20;
|
||||||
const DEFAULT_STATE_CACHE_SIZE: u32 = 25;
|
const DEFAULT_STATE_CACHE_SIZE: u32 = 25;
|
||||||
@ -41,7 +43,11 @@ pub struct CacheConfig {
|
|||||||
|
|
||||||
impl Default for CacheConfig {
|
impl Default for CacheConfig {
|
||||||
fn default() -> Self {
|
fn default() -> Self {
|
||||||
CacheConfig::new(32, 8, DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB, DEFAULT_STATE_CACHE_SIZE)
|
CacheConfig::new(
|
||||||
|
DEFAULT_DB_CACHE_SIZE,
|
||||||
|
DEFAULT_BC_CACHE_SIZE,
|
||||||
|
DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB,
|
||||||
|
DEFAULT_STATE_CACHE_SIZE)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -68,14 +74,9 @@ impl CacheConfig {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Size of db cache for blockchain.
|
/// Size of db cache.
|
||||||
pub fn db_blockchain_cache_size(&self) -> u32 {
|
pub fn db_cache_size(&self) -> u32 {
|
||||||
max(MIN_DB_CACHE_MB, self.db / 4)
|
max(MIN_DB_CACHE_MB, self.db)
|
||||||
}
|
|
||||||
|
|
||||||
/// Size of db cache for state.
|
|
||||||
pub fn db_state_cache_size(&self) -> u32 {
|
|
||||||
max(MIN_DB_CACHE_MB, self.db * 3 / 4)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Size of block queue size limit
|
/// Size of block queue size limit
|
||||||
@ -122,13 +123,16 @@ mod tests {
|
|||||||
fn test_cache_config_db_cache_sizes() {
|
fn test_cache_config_db_cache_sizes() {
|
||||||
let config = CacheConfig::new_with_total_cache_size(400);
|
let config = CacheConfig::new_with_total_cache_size(400);
|
||||||
assert_eq!(config.db, 280);
|
assert_eq!(config.db, 280);
|
||||||
assert_eq!(config.db_blockchain_cache_size(), 70);
|
assert_eq!(config.db_cache_size(), 280);
|
||||||
assert_eq!(config.db_state_cache_size(), 210);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_cache_config_default() {
|
fn test_cache_config_default() {
|
||||||
assert_eq!(CacheConfig::default(),
|
assert_eq!(CacheConfig::default(),
|
||||||
CacheConfig::new(32, 8, super::DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB, super::DEFAULT_STATE_CACHE_SIZE));
|
CacheConfig::new(
|
||||||
|
super::DEFAULT_DB_CACHE_SIZE,
|
||||||
|
super::DEFAULT_BC_CACHE_SIZE,
|
||||||
|
super::DEFAULT_BLOCK_QUEUE_SIZE_LIMIT_MB,
|
||||||
|
super::DEFAULT_STATE_CACHE_SIZE));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -779,7 +779,7 @@ usage! {
|
|||||||
"--pruning-memory=[MB]",
|
"--pruning-memory=[MB]",
|
||||||
"The ideal amount of memory in megabytes to use to store recent states. As many states as possible will be kept within this limit, and at least --pruning-history states will always be kept.",
|
"The ideal amount of memory in megabytes to use to store recent states. As many states as possible will be kept within this limit, and at least --pruning-history states will always be kept.",
|
||||||
|
|
||||||
ARG arg_cache_size_db: (u32) = 32u32, or |c: &Config| otry!(c.footprint).cache_size_db.clone(),
|
ARG arg_cache_size_db: (u32) = 128u32, or |c: &Config| otry!(c.footprint).cache_size_db.clone(),
|
||||||
"--cache-size-db=[MB]",
|
"--cache-size-db=[MB]",
|
||||||
"Override database cache size.",
|
"Override database cache size.",
|
||||||
|
|
||||||
@ -1820,7 +1820,7 @@ mod tests {
|
|||||||
pruning_memory: None,
|
pruning_memory: None,
|
||||||
fast_and_loose: None,
|
fast_and_loose: None,
|
||||||
cache_size: None,
|
cache_size: None,
|
||||||
cache_size_db: Some(128),
|
cache_size_db: Some(256),
|
||||||
cache_size_blocks: Some(16),
|
cache_size_blocks: Some(16),
|
||||||
cache_size_queue: Some(100),
|
cache_size_queue: Some(100),
|
||||||
cache_size_state: Some(25),
|
cache_size_state: Some(25),
|
||||||
|
@ -63,7 +63,7 @@ tx_queue_gas = "off"
|
|||||||
tracing = "on"
|
tracing = "on"
|
||||||
pruning = "fast"
|
pruning = "fast"
|
||||||
pruning_history = 64
|
pruning_history = 64
|
||||||
cache_size_db = 128
|
cache_size_db = 256
|
||||||
cache_size_blocks = 16
|
cache_size_blocks = 16
|
||||||
cache_size_queue = 100
|
cache_size_queue = 100
|
||||||
cache_size_state = 25
|
cache_size_state = 25
|
||||||
|
@ -227,10 +227,8 @@ pub fn to_client_config(
|
|||||||
client_config.blockchain.max_cache_size = cache_config.blockchain() as usize * mb;
|
client_config.blockchain.max_cache_size = cache_config.blockchain() as usize * mb;
|
||||||
// in bytes
|
// in bytes
|
||||||
client_config.blockchain.pref_cache_size = cache_config.blockchain() as usize * 3 / 4 * mb;
|
client_config.blockchain.pref_cache_size = cache_config.blockchain() as usize * 3 / 4 * mb;
|
||||||
// db blockchain cache size, in megabytes
|
// db cache size, in megabytes
|
||||||
client_config.blockchain.db_cache_size = Some(cache_config.db_blockchain_cache_size() as usize);
|
client_config.db_cache_size = Some(cache_config.db_cache_size() as usize);
|
||||||
// db state cache size, in megabytes
|
|
||||||
client_config.db_cache_size = Some(cache_config.db_state_cache_size() as usize);
|
|
||||||
// db queue cache size, in bytes
|
// db queue cache size, in bytes
|
||||||
client_config.queue.max_mem_use = cache_config.queue() as usize * mb;
|
client_config.queue.max_mem_use = cache_config.queue() as usize * mb;
|
||||||
// in bytes
|
// in bytes
|
||||||
|
@ -168,7 +168,7 @@ fn consolidate_database(
|
|||||||
let config = default_migration_settings(compaction_profile);
|
let config = default_migration_settings(compaction_profile);
|
||||||
let mut db_config = DatabaseConfig {
|
let mut db_config = DatabaseConfig {
|
||||||
max_open_files: 64,
|
max_open_files: 64,
|
||||||
cache_sizes: Default::default(),
|
memory_budget: None,
|
||||||
compaction: config.compaction_profile,
|
compaction: config.compaction_profile,
|
||||||
columns: None,
|
columns: None,
|
||||||
wal: true,
|
wal: true,
|
||||||
|
@ -8,6 +8,7 @@ elastic-array = "0.9"
|
|||||||
ethcore-bigint = { path = "../bigint" }
|
ethcore-bigint = { path = "../bigint" }
|
||||||
kvdb = { path = "../kvdb" }
|
kvdb = { path = "../kvdb" }
|
||||||
log = "0.3"
|
log = "0.3"
|
||||||
|
num_cpus = "1.0"
|
||||||
parking_lot = "0.4"
|
parking_lot = "0.4"
|
||||||
regex = "0.2"
|
regex = "0.2"
|
||||||
rlp = { path = "../rlp" }
|
rlp = { path = "../rlp" }
|
||||||
|
@ -18,15 +18,17 @@
|
|||||||
extern crate log;
|
extern crate log;
|
||||||
|
|
||||||
extern crate elastic_array;
|
extern crate elastic_array;
|
||||||
|
extern crate interleaved_ordered;
|
||||||
|
extern crate num_cpus;
|
||||||
extern crate parking_lot;
|
extern crate parking_lot;
|
||||||
extern crate regex;
|
extern crate regex;
|
||||||
extern crate rocksdb;
|
extern crate rocksdb;
|
||||||
extern crate interleaved_ordered;
|
|
||||||
|
|
||||||
extern crate ethcore_bigint as bigint;
|
extern crate ethcore_bigint as bigint;
|
||||||
extern crate kvdb;
|
extern crate kvdb;
|
||||||
extern crate rlp;
|
extern crate rlp;
|
||||||
|
|
||||||
|
use std::cmp;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::marker::PhantomData;
|
use std::marker::PhantomData;
|
||||||
use std::path::{PathBuf, Path};
|
use std::path::{PathBuf, Path};
|
||||||
@ -35,7 +37,7 @@ use std::{mem, fs, io};
|
|||||||
use parking_lot::{Mutex, MutexGuard, RwLock};
|
use parking_lot::{Mutex, MutexGuard, RwLock};
|
||||||
use rocksdb::{
|
use rocksdb::{
|
||||||
DB, Writable, WriteBatch, WriteOptions, IteratorMode, DBIterator,
|
DB, Writable, WriteBatch, WriteOptions, IteratorMode, DBIterator,
|
||||||
Options, DBCompactionStyle, BlockBasedOptions, Direction, Cache, Column, ReadOptions
|
Options, BlockBasedOptions, Direction, Cache, Column, ReadOptions
|
||||||
};
|
};
|
||||||
use interleaved_ordered::{interleave_ordered, InterleaveOrdered};
|
use interleaved_ordered::{interleave_ordered, InterleaveOrdered};
|
||||||
|
|
||||||
@ -50,9 +52,7 @@ use std::process::Command;
|
|||||||
#[cfg(target_os = "linux")]
|
#[cfg(target_os = "linux")]
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
|
|
||||||
const DB_BACKGROUND_FLUSHES: i32 = 2;
|
const DB_DEFAULT_MEMORY_BUDGET_MB: usize = 128;
|
||||||
const DB_BACKGROUND_COMPACTIONS: i32 = 2;
|
|
||||||
const DB_WRITE_BUFFER_SIZE: usize = 2048 * 1000;
|
|
||||||
|
|
||||||
enum KeyState {
|
enum KeyState {
|
||||||
Insert(DBValue),
|
Insert(DBValue),
|
||||||
@ -65,8 +65,8 @@ enum KeyState {
|
|||||||
pub struct CompactionProfile {
|
pub struct CompactionProfile {
|
||||||
/// L0-L1 target file size
|
/// L0-L1 target file size
|
||||||
pub initial_file_size: u64,
|
pub initial_file_size: u64,
|
||||||
/// L2-LN target file size multiplier
|
/// block size
|
||||||
pub file_size_multiplier: i32,
|
pub block_size: usize,
|
||||||
/// rate limiter for background flushes and compactions, bytes/sec, if any
|
/// rate limiter for background flushes and compactions, bytes/sec, if any
|
||||||
pub write_rate_limit: Option<u64>,
|
pub write_rate_limit: Option<u64>,
|
||||||
}
|
}
|
||||||
@ -136,8 +136,8 @@ impl CompactionProfile {
|
|||||||
/// Default profile suitable for SSD storage
|
/// Default profile suitable for SSD storage
|
||||||
pub fn ssd() -> CompactionProfile {
|
pub fn ssd() -> CompactionProfile {
|
||||||
CompactionProfile {
|
CompactionProfile {
|
||||||
initial_file_size: 32 * 1024 * 1024,
|
initial_file_size: 64 * 1024 * 1024,
|
||||||
file_size_multiplier: 2,
|
block_size: 16 * 1024,
|
||||||
write_rate_limit: None,
|
write_rate_limit: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -145,9 +145,9 @@ impl CompactionProfile {
|
|||||||
/// Slow HDD compaction profile
|
/// Slow HDD compaction profile
|
||||||
pub fn hdd() -> CompactionProfile {
|
pub fn hdd() -> CompactionProfile {
|
||||||
CompactionProfile {
|
CompactionProfile {
|
||||||
initial_file_size: 192 * 1024 * 1024,
|
initial_file_size: 256 * 1024 * 1024,
|
||||||
file_size_multiplier: 1,
|
block_size: 64 * 1024,
|
||||||
write_rate_limit: Some(8 * 1024 * 1024),
|
write_rate_limit: Some(16 * 1024 * 1024),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -157,8 +157,8 @@ impl CompactionProfile {
|
|||||||
pub struct DatabaseConfig {
|
pub struct DatabaseConfig {
|
||||||
/// Max number of open files.
|
/// Max number of open files.
|
||||||
pub max_open_files: i32,
|
pub max_open_files: i32,
|
||||||
/// Cache sizes (in MiB) for specific columns.
|
/// Memory budget (in MiB) used for setting block cache size, write buffer size.
|
||||||
pub cache_sizes: HashMap<Option<u32>, usize>,
|
pub memory_budget: Option<usize>,
|
||||||
/// Compaction profile
|
/// Compaction profile
|
||||||
pub compaction: CompactionProfile,
|
pub compaction: CompactionProfile,
|
||||||
/// Set number of columns
|
/// Set number of columns
|
||||||
@ -176,17 +176,20 @@ impl DatabaseConfig {
|
|||||||
config
|
config
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set the column cache size in MiB.
|
pub fn memory_budget(&self) -> usize {
|
||||||
pub fn set_cache(&mut self, col: Option<u32>, size: usize) {
|
self.memory_budget.unwrap_or(DB_DEFAULT_MEMORY_BUDGET_MB) * 1024 * 1024
|
||||||
self.cache_sizes.insert(col, size);
|
}
|
||||||
|
|
||||||
|
pub fn memory_budget_per_col(&self) -> usize {
|
||||||
|
self.memory_budget() / self.columns.unwrap_or(1) as usize
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Default for DatabaseConfig {
|
impl Default for DatabaseConfig {
|
||||||
fn default() -> DatabaseConfig {
|
fn default() -> DatabaseConfig {
|
||||||
DatabaseConfig {
|
DatabaseConfig {
|
||||||
cache_sizes: HashMap::new(),
|
|
||||||
max_open_files: 512,
|
max_open_files: 512,
|
||||||
|
memory_budget: None,
|
||||||
compaction: CompactionProfile::default(),
|
compaction: CompactionProfile::default(),
|
||||||
columns: None,
|
columns: None,
|
||||||
wal: true,
|
wal: true,
|
||||||
@ -217,27 +220,24 @@ struct DBAndColumns {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// get column family configuration from database config.
|
// get column family configuration from database config.
|
||||||
fn col_config(col: u32, config: &DatabaseConfig) -> Options {
|
fn col_config(config: &DatabaseConfig, block_opts: &BlockBasedOptions) -> Result<Options> {
|
||||||
// default cache size for columns not specified.
|
|
||||||
const DEFAULT_CACHE: usize = 2;
|
|
||||||
|
|
||||||
let mut opts = Options::new();
|
let mut opts = Options::new();
|
||||||
opts.set_compaction_style(DBCompactionStyle::DBUniversalCompaction);
|
|
||||||
|
opts.set_parsed_options("level_compaction_dynamic_level_bytes=true")?;
|
||||||
|
|
||||||
|
opts.set_block_based_table_factory(block_opts);
|
||||||
|
|
||||||
|
opts.set_parsed_options(
|
||||||
|
&format!("block_based_table_factory={{{};{}}}",
|
||||||
|
"cache_index_and_filter_blocks=true",
|
||||||
|
"pin_l0_filter_and_index_blocks_in_cache=true"))?;
|
||||||
|
|
||||||
|
opts.optimize_level_style_compaction(config.memory_budget_per_col() as i32);
|
||||||
opts.set_target_file_size_base(config.compaction.initial_file_size);
|
opts.set_target_file_size_base(config.compaction.initial_file_size);
|
||||||
opts.set_target_file_size_multiplier(config.compaction.file_size_multiplier);
|
|
||||||
opts.set_db_write_buffer_size(DB_WRITE_BUFFER_SIZE);
|
|
||||||
|
|
||||||
let col_opt = config.columns.map(|_| col);
|
opts.set_parsed_options("compression_per_level=")?;
|
||||||
|
|
||||||
{
|
Ok(opts)
|
||||||
let cache_size = config.cache_sizes.get(&col_opt).cloned().unwrap_or(DEFAULT_CACHE);
|
|
||||||
let mut block_opts = BlockBasedOptions::new();
|
|
||||||
// all goes to read cache.
|
|
||||||
block_opts.set_cache(Cache::new(cache_size * 1024 * 1024));
|
|
||||||
opts.set_block_based_table_factory(&block_opts);
|
|
||||||
}
|
|
||||||
|
|
||||||
opts
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Key-Value database.
|
/// Key-Value database.
|
||||||
@ -246,6 +246,7 @@ pub struct Database {
|
|||||||
config: DatabaseConfig,
|
config: DatabaseConfig,
|
||||||
write_opts: WriteOptions,
|
write_opts: WriteOptions,
|
||||||
read_opts: ReadOptions,
|
read_opts: ReadOptions,
|
||||||
|
block_opts: BlockBasedOptions,
|
||||||
path: String,
|
path: String,
|
||||||
// Dirty values added with `write_buffered`. Cleaned on `flush`.
|
// Dirty values added with `write_buffered`. Cleaned on `flush`.
|
||||||
overlay: RwLock<Vec<HashMap<ElasticArray32<u8>, KeyState>>>,
|
overlay: RwLock<Vec<HashMap<ElasticArray32<u8>, KeyState>>>,
|
||||||
@ -265,31 +266,35 @@ impl Database {
|
|||||||
/// Open database file. Creates if it does not exist.
|
/// Open database file. Creates if it does not exist.
|
||||||
pub fn open(config: &DatabaseConfig, path: &str) -> Result<Database> {
|
pub fn open(config: &DatabaseConfig, path: &str) -> Result<Database> {
|
||||||
let mut opts = Options::new();
|
let mut opts = Options::new();
|
||||||
|
|
||||||
if let Some(rate_limit) = config.compaction.write_rate_limit {
|
if let Some(rate_limit) = config.compaction.write_rate_limit {
|
||||||
opts.set_parsed_options(&format!("rate_limiter_bytes_per_sec={}", rate_limit))?;
|
opts.set_parsed_options(&format!("rate_limiter_bytes_per_sec={}", rate_limit))?;
|
||||||
}
|
}
|
||||||
opts.set_parsed_options(&format!("max_total_wal_size={}", 64 * 1024 * 1024))?;
|
|
||||||
opts.set_parsed_options("verify_checksums_in_compaction=0")?;
|
|
||||||
opts.set_parsed_options("keep_log_file_num=1")?;
|
|
||||||
opts.set_max_open_files(config.max_open_files);
|
|
||||||
opts.create_if_missing(true);
|
|
||||||
opts.set_use_fsync(false);
|
opts.set_use_fsync(false);
|
||||||
opts.set_db_write_buffer_size(DB_WRITE_BUFFER_SIZE);
|
opts.create_if_missing(true);
|
||||||
|
opts.set_max_open_files(config.max_open_files);
|
||||||
|
opts.set_parsed_options("keep_log_file_num=1")?;
|
||||||
|
opts.set_parsed_options("bytes_per_sync=1048576")?;
|
||||||
|
opts.set_db_write_buffer_size(config.memory_budget_per_col() / 2);
|
||||||
|
opts.increase_parallelism(cmp::max(1, ::num_cpus::get() as i32 / 2));
|
||||||
|
|
||||||
opts.set_max_background_flushes(DB_BACKGROUND_FLUSHES);
|
let mut block_opts = BlockBasedOptions::new();
|
||||||
opts.set_max_background_compactions(DB_BACKGROUND_COMPACTIONS);
|
|
||||||
|
|
||||||
// compaction settings
|
{
|
||||||
opts.set_compaction_style(DBCompactionStyle::DBUniversalCompaction);
|
block_opts.set_block_size(config.compaction.block_size);
|
||||||
opts.set_target_file_size_base(config.compaction.initial_file_size);
|
let cache_size = cmp::max(8, config.memory_budget() / 3);
|
||||||
opts.set_target_file_size_multiplier(config.compaction.file_size_multiplier);
|
let cache = Cache::new(cache_size);
|
||||||
|
block_opts.set_cache(cache);
|
||||||
|
}
|
||||||
|
|
||||||
let mut cf_options = Vec::with_capacity(config.columns.unwrap_or(0) as usize);
|
let columns = config.columns.unwrap_or(0) as usize;
|
||||||
let cfnames: Vec<_> = (0..config.columns.unwrap_or(0)).map(|c| format!("col{}", c)).collect();
|
|
||||||
|
let mut cf_options = Vec::with_capacity(columns);
|
||||||
|
let cfnames: Vec<_> = (0..columns).map(|c| format!("col{}", c)).collect();
|
||||||
let cfnames: Vec<&str> = cfnames.iter().map(|n| n as &str).collect();
|
let cfnames: Vec<&str> = cfnames.iter().map(|n| n as &str).collect();
|
||||||
|
|
||||||
for col in 0 .. config.columns.unwrap_or(0) {
|
for _ in 0 .. config.columns.unwrap_or(0) {
|
||||||
cf_options.push(col_config(col, &config));
|
cf_options.push(col_config(&config, &block_opts)?);
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut write_opts = WriteOptions::new();
|
let mut write_opts = WriteOptions::new();
|
||||||
@ -348,6 +353,7 @@ impl Database {
|
|||||||
flushing_lock: Mutex::new((false)),
|
flushing_lock: Mutex::new((false)),
|
||||||
path: path.to_owned(),
|
path: path.to_owned(),
|
||||||
read_opts: read_opts,
|
read_opts: read_opts,
|
||||||
|
block_opts: block_opts,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -632,7 +638,7 @@ impl Database {
|
|||||||
Some(DBAndColumns { ref mut db, ref mut cfs }) => {
|
Some(DBAndColumns { ref mut db, ref mut cfs }) => {
|
||||||
let col = cfs.len() as u32;
|
let col = cfs.len() as u32;
|
||||||
let name = format!("col{}", col);
|
let name = format!("col{}", col);
|
||||||
cfs.push(db.create_cf(&name, &col_config(col, &self.config))?);
|
cfs.push(db.create_cf(&name, &col_config(&self.config, &self.block_opts)?)?);
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
},
|
||||||
None => Ok(()),
|
None => Ok(()),
|
||||||
|
@ -264,7 +264,7 @@ impl Manager {
|
|||||||
trace!(target: "migration", "Expecting database to contain {:?} columns", columns);
|
trace!(target: "migration", "Expecting database to contain {:?} columns", columns);
|
||||||
let mut db_config = DatabaseConfig {
|
let mut db_config = DatabaseConfig {
|
||||||
max_open_files: 64,
|
max_open_files: 64,
|
||||||
cache_sizes: Default::default(),
|
memory_budget: None,
|
||||||
compaction: config.compaction_profile,
|
compaction: config.compaction_profile,
|
||||||
columns: columns,
|
columns: columns,
|
||||||
wal: true,
|
wal: true,
|
||||||
|
Loading…
Reference in New Issue
Block a user