* Remove dead bootnodes, add new geth bootnodes
* More granular locking when fetching state
Finish GetDataNode requests early if queries take too long
* typo
* Use latest kvdb-rocksdb
* Cleanup
* Update ethcore/sync/src/chain/supplier.rs
Co-Authored-By: Andronik Ordian <write@reusable.software>
* Address review grumbles
* Fix compilation
* Address review grumbles
Co-authored-by: Andronik Ordian <write@reusable.software>
* engine error: remove faulty/unused `From`
* dependencies: bump `derive_more v0.99`
The primilary motivation behind this is to have hard errors on `From` conversions that are ignored by
duplicated `variants` with the same value type
* sccache "stop server" - > "show stats"
* remove testing and beta from update, cli, etc.
* Beta->Nigthly updater
* Beta->Nightly
* ->Nightly and fix
* updater ->Stable
* Testing->Nightly
* Update scripts/gitlab/test-linux.sh
Co-Authored-By: Denis Pisarev <denis.pisarev@parity.io>
* sccache "stop server" - > "show stats"
* remove testing and beta from update, cli, etc.
* Beta->Nigthly updater
* Beta->Nightly
* ->Nightly and fix
* updater ->Stable
* Testing->Nightly
* Update scripts/gitlab/test-linux.sh
Co-Authored-By: Denis Pisarev <denis.pisarev@parity.io>
* Update CHANGELOGs and version
* temporarily allow darwin and windows to be built on any branch
* fix check-benches job
* Revert "temporarily allow darwin and windows to be built on any branch"
This reverts commit 45c72f69e99cbe891f694e528a53eb3c3bd8f331.
* fix check-benches job
* Revert changing track from `nightly` to `stable`
* fix test: rpc_parity_upgrade_ready
* fix tests: rpc_parity_version_info, rpc_parity_releases_info
Co-authored-by: Denis Pisarev <denis.pisarev@parity.io>
Co-authored-by: s3krit <pugh@s3kr.it>
* Only use kvdb "column families"
This PR contains the changes necessary to use the `kvdb-*` crates from https://github.com/paritytech/parity-common/pull/278 (so a synchronized merge is required) which drops support for the old-style rocksdb "default" column to get a smaller and less complex API.
As it stands this PR is working correctly except for secret-store; we need to migrate it to use a new column family.
* Fix secretstore build
* Fix secretstore build: include ethkey when building with the "accounts" feature
* typos
* Restore state test commit
* Override all of parity-common from git
* Be precise about version requirement to migrate secretstore code
* Update ethcore/db/src/db.rs
Co-Authored-By: Niklas Adolfsson <niklasadolfsson1@gmail.com>
* Address review grumbles
* Review grumbles
* Cleanup
Co-authored-by: Niklas Adolfsson <niklasadolfsson1@gmail.com>
* Use upstream rocksdb
…by way of https://github.com/paritytech/parity-common/pull/257 by @ordian.
* Hint at how `parity db reset` works in the error message
* migration-rocksdb: fix build
* Cargo.toml: use git dependency instead of path
* update to latest kvdb-rocksdb
* fix tests
* saner default for light client
* rename open_db to open_db_light
* update to latest kvdb-rocksdb
* moar update to latest kvdb-rocksdb
* even moar update to latest kvdb-rocksdb
* use kvdb-rocksdb from crates.io
* Update parity/db/rocksdb/helpers.rs
* add docs to memory_budget division
* Comments and todos
Use `snapshot_sync` as logging target
* fix compilation
* More todos, more logs
* Fix picking snapshot peer: prefer the one with the highest block number
More docs, comments, todos
* Adjust WAIT_PEERS_TIMEOUT to be a multiple of MAINTAIN_SYNC_TIMER to try to fix snapshot startup problems
Docs, todos, comments
* Tabs
* Formatting
* Don't build new rlp::EMPTY_LIST_RLP instances
* Dial down debug logging
* Don't warn about missing hashes in the manifest: it's normal
Log client version on peer connect
* Cleanup
* Do not skip snapshots further away than 30k block from the highest block seen
Currently we look for peers that seed snapshots that are close to the highest block seen on the network (where "close" means withing 30k blocks). When a node starts up we wait for some time (5sec, increased here to 10sec) to let peers connect and if we have found a suitable peer to sync a snapshot from at the end of that delay, we start the download; if none is found and --warp-barrier is used we stall, otherwise we start a slow-sync.
When looking for a suitable snapshot, we use the highest block seen on the network to check if a peer has a snapshot that is within 30k blocks of that highest block number. This means that in a situation where all available snapshots are older than that, we will often fail to start a snapshot at all. What's worse is that the longer we delay starting a snapshot sync (to let more peers connect, in the hope of finding a good snapshot), the more likely we are to have seen a high block and thus the more likely we become to accept a snapshot.
This commit removes this comparison with the highest blocknumber criteria entirely and picks the best snapshot we find in 10sec.
* lockfile
* Add a `ChunkType::Dupe` variant so that we do not disconnect a peer if they happen to send us a duplicate chunk (just ignore the chunk and keep going)
Resolve some documentation todos, add more
* tweak log message
* Don't warp sync twice
Check if our own block is beyond the given warp barrier (can happen after we've completed a warp sync but are not quite yet synced up to the tip) and if so, don't sync.
More docs, resolve todos.
Dial down some `sync` debug level logging to trace
* Avoid iterating over all snapshot block/state hashes to find the next work item
Use a HashSet instead of a Vec and remove items from the set as chunks are processed. Calculate and store the total number of chunks in the `Snapshot` struct instead of counting pending chunks each time.
* Address review grumbles
* Log correct number of bytes written to disk
* Revert ChunkType::Dup change
* whitespace grumble
* Cleanup debugging code
* Fix docs
* Fix import and a typo
* Fix test impl
* Use `indexmap::IndexSet` to ensure chunk hashes are accessed in order
* Revert increased SNAPSHOT_MANIFEST_TIMEOUT: 5sec should be enough
* WIP. Typos and logging.
* Format todos
* Pause pruning while a snapshot is under way
Logs, docs and todos
* Allocate memory for the full chunk
* Name snapshotting threads
* Ensure `taking_snapshot` is set to false whenever and however `take_snapshot`returns
Rename `take_at` to `request_snapshot_at`
Cleanup
* Let "in_progress" deletion fail
Fix tests
* Just use an atomic
* Review grumbles
* Finish the sentence
* Resolve a few todos and clarify comments.
* Calculate progress rate since last update
* Lockfile
* Fix tests
* typo
* Reinstate default snapshotting frequency
Cut down on the logging noise
* address grumble
* Log memory use with `journal_size()` and explain why.
* Upgrade to jsonrpc v14
Contains https://github.com/paritytech/jsonrpc/pull/495 with good bugfixes to resource usage.
* Bump tokio & futures.
* Bump even further.
* Upgrade tokio to 0.1.22
* Partially revert "Bump tokio & futures."
This reverts commit 100907eb91907aa124d856d52374637256118e86.
* Rename RegistryInfo -> RegistryInfoDeprecated
* Add BlockId parameter to Registrar::get_address and RegistrarClient::call_contract
* Remove RegistrarClient::Call (use async for now); add RegistrarClient::get_address
* Remove Registrar type in favour of naked trait
* Use CallContract trait bound instead of separate call_contract method
* Make RegistrarClient::get_address and URLHint::resolve synchronous
* RegistrarClient::get_address: return check if address is zero
* Modify RegistryInfo::registry_address to take &str
* return Result from RegistryInfo::registry_address
* Replace RegistryInfo with RegistrarClient
- Modifed RegistrarClient::registrar_address to return Option
- Removed BlockChainClient::registrar_address
* Fix other build configs
* Fix unit test builds
* Remove local RegistrarClient type from run::execute_impl
* Remove registrar.json from ethcore
* Formatting/line breaks
* Update RegistrarClient docs, remove explicit lifetime
* Weak ref to ethcore client from hash fetch client
* Fix unit tests
* Update a few dependencies
Updates two dependencies: `kvdb-rocksdb` and `h2`. Brings in `parking_lot 0.9` which is unintended but possibly fine.
* Bump parking_lot to 0.9
Bump kvdb-memorydb to 0.2 (from git atm)
* New kvdb-memorydb is not breaking
* Remove [patch]
* inject_batch && commit_batch are no longer a part of journaldb
* get rid of redundant KeyedHashDB trait
* journaldb edition 2018
* journaldb trait moved to the lib.rs file
* making journaldb more idiomatic
* fix parity_bytes reexport
* rename parity-util-mem package in Cargo.toml file