582 lines
22 KiB
Rust
582 lines
22 KiB
Rust
// Copyright 2015-2017 Parity Technologies (UK) Ltd.
|
|
// This file is part of Parity.
|
|
|
|
// Parity is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Parity is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Parity. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
use std::path::PathBuf;
|
|
use std::collections::BTreeMap;
|
|
use serde_json;
|
|
use tiny_keccak::Keccak;
|
|
use bigint::hash::H256;
|
|
use ethkey::{Secret, Public};
|
|
use kvdb_rocksdb::{Database, DatabaseIterator};
|
|
use types::all::{Error, ServiceConfiguration, ServerKeyId, NodeId};
|
|
use serialization::{SerializablePublic, SerializableSecret, SerializableH256};
|
|
|
|
/// Key of version value.
|
|
const DB_META_KEY_VERSION: &'static [u8; 7] = b"version";
|
|
/// Current db version.
|
|
const CURRENT_VERSION: u8 = 2;
|
|
/// Current type of serialized key shares.
|
|
type CurrentSerializableDocumentKeyShare = SerializableDocumentKeyShareV2;
|
|
/// Current type of serialized key shares versions.
|
|
type CurrentSerializableDocumentKeyVersion = SerializableDocumentKeyShareVersionV2;
|
|
|
|
/// Encrypted key share, stored by key storage on the single key server.
|
|
#[derive(Debug, Clone, PartialEq)]
|
|
#[cfg_attr(test, derive(Default))]
|
|
pub struct DocumentKeyShare {
|
|
/// Author of the entry.
|
|
pub author: Public,
|
|
/// Decryption threshold (at least threshold + 1 nodes are required to decrypt data).
|
|
pub threshold: usize,
|
|
/// Server public key.
|
|
pub public: Public,
|
|
/// Common (shared) encryption point.
|
|
pub common_point: Option<Public>,
|
|
/// Encrypted point.
|
|
pub encrypted_point: Option<Public>,
|
|
/// Key share versions.
|
|
pub versions: Vec<DocumentKeyShareVersion>,
|
|
}
|
|
|
|
/// Versioned portion of document key share.
|
|
#[derive(Debug, Clone, PartialEq)]
|
|
pub struct DocumentKeyShareVersion {
|
|
/// Version hash (Keccak(time + id_numbers)).
|
|
pub hash: H256,
|
|
/// Nodes ids numbers.
|
|
pub id_numbers: BTreeMap<NodeId, Secret>,
|
|
/// Node secret share.
|
|
pub secret_share: Secret,
|
|
}
|
|
|
|
/// Document encryption keys storage
|
|
pub trait KeyStorage: Send + Sync {
|
|
/// Insert document encryption key
|
|
fn insert(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error>;
|
|
/// Update document encryption key
|
|
fn update(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error>;
|
|
/// Get document encryption key
|
|
fn get(&self, document: &ServerKeyId) -> Result<Option<DocumentKeyShare>, Error>;
|
|
/// Remove document encryption key
|
|
fn remove(&self, document: &ServerKeyId) -> Result<(), Error>;
|
|
/// Clears the database
|
|
fn clear(&self) -> Result<(), Error>;
|
|
/// Check if storage contains document encryption key
|
|
fn contains(&self, document: &ServerKeyId) -> bool;
|
|
/// Iterate through storage
|
|
fn iter<'a>(&'a self) -> Box<Iterator<Item=(ServerKeyId, DocumentKeyShare)> + 'a>;
|
|
}
|
|
|
|
/// Persistent document encryption keys storage
|
|
pub struct PersistentKeyStorage {
|
|
db: Database,
|
|
}
|
|
|
|
/// Persistent document encryption keys storage iterator
|
|
pub struct PersistentKeyStorageIterator<'a> {
|
|
iter: Option<DatabaseIterator<'a>>,
|
|
}
|
|
|
|
/// V0 of encrypted key share, as it is stored by key storage on the single key server.
|
|
#[derive(Serialize, Deserialize)]
|
|
struct SerializableDocumentKeyShareV0 {
|
|
/// Decryption threshold (at least threshold + 1 nodes are required to decrypt data).
|
|
pub threshold: usize,
|
|
/// Nodes ids numbers.
|
|
pub id_numbers: BTreeMap<SerializablePublic, SerializableSecret>,
|
|
/// Node secret share.
|
|
pub secret_share: SerializableSecret,
|
|
/// Common (shared) encryption point.
|
|
pub common_point: SerializablePublic,
|
|
/// Encrypted point.
|
|
pub encrypted_point: SerializablePublic,
|
|
}
|
|
|
|
/// V1 of encrypted key share, as it is stored by key storage on the single key server.
|
|
#[derive(Serialize, Deserialize)]
|
|
struct SerializableDocumentKeyShareV1 {
|
|
/// Authore of the entry.
|
|
pub author: SerializablePublic,
|
|
/// Decryption threshold (at least threshold + 1 nodes are required to decrypt data).
|
|
pub threshold: usize,
|
|
/// Nodes ids numbers.
|
|
pub id_numbers: BTreeMap<SerializablePublic, SerializableSecret>,
|
|
/// Node secret share.
|
|
pub secret_share: SerializableSecret,
|
|
/// Common (shared) encryption point.
|
|
pub common_point: Option<SerializablePublic>,
|
|
/// Encrypted point.
|
|
pub encrypted_point: Option<SerializablePublic>,
|
|
}
|
|
|
|
/// V2 of encrypted key share, as it is stored by key storage on the single key server.
|
|
#[derive(Serialize, Deserialize)]
|
|
struct SerializableDocumentKeyShareV2 {
|
|
/// Author of the entry.
|
|
pub author: SerializablePublic,
|
|
/// Decryption threshold (at least threshold + 1 nodes are required to decrypt data).
|
|
pub threshold: usize,
|
|
/// Server public.
|
|
pub public: SerializablePublic,
|
|
/// Common (shared) encryption point.
|
|
pub common_point: Option<SerializablePublic>,
|
|
/// Encrypted point.
|
|
pub encrypted_point: Option<SerializablePublic>,
|
|
/// Versions.
|
|
pub versions: Vec<SerializableDocumentKeyShareVersionV2>}
|
|
|
|
/// V2 of encrypted key share version, as it is stored by key storage on the single key server.
|
|
#[derive(Serialize, Deserialize)]
|
|
struct SerializableDocumentKeyShareVersionV2 {
|
|
/// Version hash.
|
|
pub hash: SerializableH256,
|
|
/// Nodes ids numbers.
|
|
pub id_numbers: BTreeMap<SerializablePublic, SerializableSecret>,
|
|
/// Node secret share.
|
|
pub secret_share: SerializableSecret,
|
|
}
|
|
|
|
impl PersistentKeyStorage {
|
|
/// Create new persistent document encryption keys storage
|
|
pub fn new(config: &ServiceConfiguration) -> Result<Self, Error> {
|
|
let mut db_path = PathBuf::from(&config.data_path);
|
|
db_path.push("db");
|
|
let db_path = db_path.to_str().ok_or_else(|| Error::Database("Invalid secretstore path".to_owned()))?;
|
|
|
|
let db = Database::open_default(&db_path)?;
|
|
let db = upgrade_db(db)?;
|
|
|
|
Ok(PersistentKeyStorage {
|
|
db: db,
|
|
})
|
|
}
|
|
}
|
|
|
|
fn upgrade_db(db: Database) -> Result<Database, Error> {
|
|
let version = db.get(None, DB_META_KEY_VERSION)?;
|
|
let version = version.and_then(|v| v.get(0).cloned()).unwrap_or(0);
|
|
match version {
|
|
0 => {
|
|
let mut batch = db.transaction();
|
|
batch.put(None, DB_META_KEY_VERSION, &[CURRENT_VERSION]);
|
|
for (db_key, db_value) in db.iter(None).into_iter().flat_map(|inner| inner).filter(|&(ref k, _)| **k != *DB_META_KEY_VERSION) {
|
|
let v0_key = serde_json::from_slice::<SerializableDocumentKeyShareV0>(&db_value).map_err(|e| Error::Database(e.to_string()))?;
|
|
let current_key = CurrentSerializableDocumentKeyShare {
|
|
// author is used in separate generation + encrypt sessions.
|
|
// in v0 there have been only simultaneous GenEnc sessions.
|
|
author: Public::default().into(), // added in v1
|
|
threshold: v0_key.threshold,
|
|
public: Public::default().into(), // addded in v2
|
|
common_point: Some(v0_key.common_point),
|
|
encrypted_point: Some(v0_key.encrypted_point),
|
|
versions: vec![CurrentSerializableDocumentKeyVersion {
|
|
hash: DocumentKeyShareVersion::data_hash(v0_key.id_numbers.iter().map(|(k, v)| (&***k, &****v))).into(),
|
|
id_numbers: v0_key.id_numbers,
|
|
secret_share: v0_key.secret_share,
|
|
}],
|
|
};
|
|
let db_value = serde_json::to_vec(¤t_key).map_err(|e| Error::Database(e.to_string()))?;
|
|
batch.put(None, &*db_key, &*db_value);
|
|
}
|
|
db.write(batch)?;
|
|
Ok(db)
|
|
},
|
|
1 => {
|
|
let mut batch = db.transaction();
|
|
batch.put(None, DB_META_KEY_VERSION, &[CURRENT_VERSION]);
|
|
for (db_key, db_value) in db.iter(None).into_iter().flat_map(|inner| inner).filter(|&(ref k, _)| **k != *DB_META_KEY_VERSION) {
|
|
let v1_key = serde_json::from_slice::<SerializableDocumentKeyShareV1>(&db_value).map_err(|e| Error::Database(e.to_string()))?;
|
|
let current_key = CurrentSerializableDocumentKeyShare {
|
|
author: v1_key.author, // added in v1
|
|
threshold: v1_key.threshold,
|
|
public: Public::default().into(), // addded in v2
|
|
common_point: v1_key.common_point,
|
|
encrypted_point: v1_key.encrypted_point,
|
|
versions: vec![CurrentSerializableDocumentKeyVersion {
|
|
hash: DocumentKeyShareVersion::data_hash(v1_key.id_numbers.iter().map(|(k, v)| (&***k, &****v))).into(),
|
|
id_numbers: v1_key.id_numbers,
|
|
secret_share: v1_key.secret_share,
|
|
}],
|
|
};
|
|
let db_value = serde_json::to_vec(¤t_key).map_err(|e| Error::Database(e.to_string()))?;
|
|
batch.put(None, &*db_key, &*db_value);
|
|
}
|
|
db.write(batch)?;
|
|
Ok(db)
|
|
}
|
|
2 => Ok(db),
|
|
_ => Err(Error::Database(format!("unsupported SecretStore database version: {}", version))),
|
|
}
|
|
}
|
|
|
|
impl KeyStorage for PersistentKeyStorage {
|
|
fn insert(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error> {
|
|
let key: CurrentSerializableDocumentKeyShare = key.into();
|
|
let key = serde_json::to_vec(&key).map_err(|e| Error::Database(e.to_string()))?;
|
|
let mut batch = self.db.transaction();
|
|
batch.put(None, &document, &key);
|
|
self.db.write(batch).map_err(Into::into)
|
|
}
|
|
|
|
fn update(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error> {
|
|
self.insert(document, key)
|
|
}
|
|
|
|
fn get(&self, document: &ServerKeyId) -> Result<Option<DocumentKeyShare>, Error> {
|
|
self.db.get(None, document)
|
|
.map_err(|e| Error::Database(e.to_string()))
|
|
.and_then(|key| match key {
|
|
None => Ok(None),
|
|
Some(key) => serde_json::from_slice::<CurrentSerializableDocumentKeyShare>(&key)
|
|
.map_err(|e| Error::Database(e.to_string()))
|
|
.map(Into::into)
|
|
.map(Some),
|
|
})
|
|
}
|
|
|
|
fn remove(&self, document: &ServerKeyId) -> Result<(), Error> {
|
|
let mut batch = self.db.transaction();
|
|
batch.delete(None, &document);
|
|
self.db.write(batch).map_err(Into::into)
|
|
}
|
|
|
|
fn clear(&self) -> Result<(), Error> {
|
|
let mut batch = self.db.transaction();
|
|
for (key, _) in self.iter() {
|
|
batch.delete(None, &key);
|
|
}
|
|
self.db.write(batch)
|
|
.map_err(|e| Error::Database(e.to_string()))
|
|
}
|
|
|
|
fn contains(&self, document: &ServerKeyId) -> bool {
|
|
self.db.get(None, document)
|
|
.map(|k| k.is_some())
|
|
.unwrap_or(false)
|
|
}
|
|
|
|
fn iter<'a>(&'a self) -> Box<Iterator<Item=(ServerKeyId, DocumentKeyShare)> + 'a> {
|
|
Box::new(PersistentKeyStorageIterator {
|
|
iter: self.db.iter(None),
|
|
})
|
|
}
|
|
}
|
|
|
|
impl<'a> Iterator for PersistentKeyStorageIterator<'a> {
|
|
type Item = (ServerKeyId, DocumentKeyShare);
|
|
|
|
fn next(&mut self) -> Option<(ServerKeyId, DocumentKeyShare)> {
|
|
self.iter.as_mut()
|
|
.and_then(|iter| iter.next()
|
|
.and_then(|(db_key, db_val)| serde_json::from_slice::<CurrentSerializableDocumentKeyShare>(&db_val)
|
|
.ok()
|
|
.map(|key| ((*db_key).into(), key.into()))))
|
|
}
|
|
}
|
|
|
|
impl DocumentKeyShare {
|
|
/// Get last version reference.
|
|
#[cfg(test)]
|
|
pub fn last_version(&self) -> Result<&DocumentKeyShareVersion, Error> {
|
|
self.versions.iter().rev()
|
|
.nth(0)
|
|
.ok_or_else(|| Error::Database("key version is not found".into()))
|
|
}
|
|
|
|
/// Get given version reference.
|
|
pub fn version(&self, version: &H256) -> Result<&DocumentKeyShareVersion, Error> {
|
|
self.versions.iter().rev()
|
|
.find(|v| &v.hash == version)
|
|
.ok_or_else(|| Error::Database("key version is not found".into()))
|
|
}
|
|
}
|
|
|
|
impl DocumentKeyShareVersion {
|
|
/// Create new version
|
|
pub fn new(id_numbers: BTreeMap<NodeId, Secret>, secret_share: Secret) -> Self {
|
|
DocumentKeyShareVersion {
|
|
hash: Self::data_hash(id_numbers.iter().map(|(k, v)| (&**k, &***v))),
|
|
id_numbers: id_numbers,
|
|
secret_share: secret_share,
|
|
}
|
|
}
|
|
|
|
|
|
/// Calculate hash of given version data.
|
|
pub fn data_hash<'a, I>(id_numbers: I) -> H256 where I: Iterator<Item=(&'a [u8], &'a [u8])> {
|
|
let mut nodes_keccak = Keccak::new_keccak256();
|
|
|
|
for (node, node_number) in id_numbers {
|
|
nodes_keccak.update(node);
|
|
nodes_keccak.update(node_number);
|
|
}
|
|
|
|
let mut nodes_keccak_value = [0u8; 32];
|
|
nodes_keccak.finalize(&mut nodes_keccak_value);
|
|
|
|
nodes_keccak_value.into()
|
|
}
|
|
}
|
|
|
|
impl From<DocumentKeyShare> for SerializableDocumentKeyShareV2 {
|
|
fn from(key: DocumentKeyShare) -> Self {
|
|
SerializableDocumentKeyShareV2 {
|
|
author: key.author.into(),
|
|
threshold: key.threshold,
|
|
public: key.public.into(),
|
|
common_point: key.common_point.map(Into::into),
|
|
encrypted_point: key.encrypted_point.map(Into::into),
|
|
versions: key.versions.into_iter().map(Into::into).collect(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<DocumentKeyShareVersion> for SerializableDocumentKeyShareVersionV2 {
|
|
fn from(version: DocumentKeyShareVersion) -> Self {
|
|
SerializableDocumentKeyShareVersionV2 {
|
|
hash: version.hash.into(),
|
|
id_numbers: version.id_numbers.into_iter().map(|(k, v)| (k.into(), v.into())).collect(),
|
|
secret_share: version.secret_share.into(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<SerializableDocumentKeyShareV2> for DocumentKeyShare {
|
|
fn from(key: SerializableDocumentKeyShareV2) -> Self {
|
|
DocumentKeyShare {
|
|
author: key.author.into(),
|
|
threshold: key.threshold,
|
|
public: key.public.into(),
|
|
common_point: key.common_point.map(Into::into),
|
|
encrypted_point: key.encrypted_point.map(Into::into),
|
|
versions: key.versions.into_iter()
|
|
.map(|v| DocumentKeyShareVersion {
|
|
hash: v.hash.into(),
|
|
id_numbers: v.id_numbers.into_iter().map(|(k, v)| (k.into(), v.into())).collect(),
|
|
secret_share: v.secret_share.into(),
|
|
})
|
|
.collect(),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
pub mod tests {
|
|
extern crate tempdir;
|
|
|
|
use std::collections::{BTreeMap, HashMap};
|
|
use parking_lot::RwLock;
|
|
use serde_json;
|
|
use self::tempdir::TempDir;
|
|
use ethkey::{Random, Generator, Public, Secret};
|
|
use kvdb_rocksdb::Database;
|
|
use types::all::{Error, NodeAddress, ServiceConfiguration, ClusterConfiguration, ServerKeyId};
|
|
use super::{DB_META_KEY_VERSION, CURRENT_VERSION, KeyStorage, PersistentKeyStorage, DocumentKeyShare,
|
|
DocumentKeyShareVersion, SerializableDocumentKeyShareV0, SerializableDocumentKeyShareV1,
|
|
CurrentSerializableDocumentKeyShare, upgrade_db};
|
|
|
|
/// In-memory document encryption keys storage
|
|
#[derive(Default)]
|
|
pub struct DummyKeyStorage {
|
|
keys: RwLock<HashMap<ServerKeyId, DocumentKeyShare>>,
|
|
}
|
|
|
|
impl KeyStorage for DummyKeyStorage {
|
|
fn insert(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error> {
|
|
self.keys.write().insert(document, key);
|
|
Ok(())
|
|
}
|
|
|
|
fn update(&self, document: ServerKeyId, key: DocumentKeyShare) -> Result<(), Error> {
|
|
self.keys.write().insert(document, key);
|
|
Ok(())
|
|
}
|
|
|
|
fn get(&self, document: &ServerKeyId) -> Result<Option<DocumentKeyShare>, Error> {
|
|
Ok(self.keys.read().get(document).cloned())
|
|
}
|
|
|
|
fn remove(&self, document: &ServerKeyId) -> Result<(), Error> {
|
|
self.keys.write().remove(document);
|
|
Ok(())
|
|
}
|
|
|
|
fn clear(&self) -> Result<(), Error> {
|
|
self.keys.write().clear();
|
|
Ok(())
|
|
}
|
|
|
|
fn contains(&self, document: &ServerKeyId) -> bool {
|
|
self.keys.read().contains_key(document)
|
|
}
|
|
|
|
fn iter<'a>(&'a self) -> Box<Iterator<Item=(ServerKeyId, DocumentKeyShare)> + 'a> {
|
|
Box::new(self.keys.read().clone().into_iter())
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn persistent_key_storage() {
|
|
let tempdir = TempDir::new("").unwrap();
|
|
let config = ServiceConfiguration {
|
|
listener_address: None,
|
|
service_contract_address: None,
|
|
acl_check_enabled: true,
|
|
data_path: tempdir.path().display().to_string(),
|
|
cluster_config: ClusterConfiguration {
|
|
threads: 1,
|
|
listener_address: NodeAddress {
|
|
address: "0.0.0.0".to_owned(),
|
|
port: 8083,
|
|
},
|
|
nodes: BTreeMap::new(),
|
|
allow_connecting_to_higher_nodes: false,
|
|
admin_public: None,
|
|
},
|
|
};
|
|
|
|
let key1 = ServerKeyId::from(1);
|
|
let value1 = DocumentKeyShare {
|
|
author: Public::default(),
|
|
threshold: 100,
|
|
public: Public::default(),
|
|
common_point: Some(Random.generate().unwrap().public().clone()),
|
|
encrypted_point: Some(Random.generate().unwrap().public().clone()),
|
|
versions: vec![DocumentKeyShareVersion {
|
|
hash: Default::default(),
|
|
id_numbers: vec![
|
|
(Random.generate().unwrap().public().clone(), Random.generate().unwrap().secret().clone())
|
|
].into_iter().collect(),
|
|
secret_share: Random.generate().unwrap().secret().clone(),
|
|
}],
|
|
};
|
|
let key2 = ServerKeyId::from(2);
|
|
let value2 = DocumentKeyShare {
|
|
author: Public::default(),
|
|
threshold: 200,
|
|
public: Public::default(),
|
|
common_point: Some(Random.generate().unwrap().public().clone()),
|
|
encrypted_point: Some(Random.generate().unwrap().public().clone()),
|
|
versions: vec![DocumentKeyShareVersion {
|
|
hash: Default::default(),
|
|
id_numbers: vec![
|
|
(Random.generate().unwrap().public().clone(), Random.generate().unwrap().secret().clone())
|
|
].into_iter().collect(),
|
|
secret_share: Random.generate().unwrap().secret().clone(),
|
|
}],
|
|
};
|
|
let key3 = ServerKeyId::from(3);
|
|
|
|
let key_storage = PersistentKeyStorage::new(&config).unwrap();
|
|
key_storage.insert(key1.clone(), value1.clone()).unwrap();
|
|
key_storage.insert(key2.clone(), value2.clone()).unwrap();
|
|
assert_eq!(key_storage.get(&key1), Ok(Some(value1.clone())));
|
|
assert_eq!(key_storage.get(&key2), Ok(Some(value2.clone())));
|
|
assert_eq!(key_storage.get(&key3), Ok(None));
|
|
drop(key_storage);
|
|
|
|
let key_storage = PersistentKeyStorage::new(&config).unwrap();
|
|
assert_eq!(key_storage.get(&key1), Ok(Some(value1)));
|
|
assert_eq!(key_storage.get(&key2), Ok(Some(value2)));
|
|
assert_eq!(key_storage.get(&key3), Ok(None));
|
|
}
|
|
|
|
#[test]
|
|
fn upgrade_db_from_0() {
|
|
let tempdir = TempDir::new("").unwrap();
|
|
let db = Database::open_default(&tempdir.path().display().to_string()).unwrap();
|
|
|
|
// prepare v0 database
|
|
{
|
|
let key = serde_json::to_vec(&SerializableDocumentKeyShareV0 {
|
|
threshold: 777,
|
|
id_numbers: vec![(
|
|
"b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".into(),
|
|
"281b6bf43cb86d0dc7b98e1b7def4a80f3ce16d28d2308f934f116767306f06c".parse::<Secret>().unwrap().into(),
|
|
)].into_iter().collect(),
|
|
secret_share: "00125d85a05e5e63e214cb60fe63f132eec8a103aa29266b7e6e6c5b7597230b".parse::<Secret>().unwrap().into(),
|
|
common_point: "99e82b163b062d55a64085bacfd407bb55f194ba5fb7a1af9c34b84435455520f1372e0e650a4f91aed0058cb823f62146ccb5599c8d13372c300dea866b69fc".into(),
|
|
encrypted_point: "7e05df9dd077ec21ed4bc45c9fe9e0a43d65fa4be540630de615ced5e95cf5c3003035eb713317237d7667feeeb64335525158f5f7411f67aca9645169ea554c".into(),
|
|
}).unwrap();
|
|
let mut batch = db.transaction();
|
|
batch.put(None, &[7], &key);
|
|
db.write(batch).unwrap();
|
|
}
|
|
|
|
// upgrade database
|
|
let db = upgrade_db(db).unwrap();
|
|
|
|
// check upgrade
|
|
assert_eq!(db.get(None, DB_META_KEY_VERSION).unwrap().unwrap()[0], CURRENT_VERSION);
|
|
let key = serde_json::from_slice::<CurrentSerializableDocumentKeyShare>(&db.get(None, &[7]).unwrap().map(|key| key.to_vec()).unwrap()).unwrap();
|
|
assert_eq!(Public::default(), key.author.clone().into());
|
|
assert_eq!(777, key.threshold);
|
|
assert_eq!(Some("99e82b163b062d55a64085bacfd407bb55f194ba5fb7a1af9c34b84435455520f1372e0e650a4f91aed0058cb823f62146ccb5599c8d13372c300dea866b69fc".parse::<Public>().unwrap()), key.common_point.clone().map(Into::into));
|
|
assert_eq!(Some("7e05df9dd077ec21ed4bc45c9fe9e0a43d65fa4be540630de615ced5e95cf5c3003035eb713317237d7667feeeb64335525158f5f7411f67aca9645169ea554c".parse::<Public>().unwrap()), key.encrypted_point.clone().map(Into::into));
|
|
|
|
assert_eq!(key.versions.len(), 1);
|
|
assert_eq!(vec![(
|
|
"b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".parse::<Public>().unwrap(),
|
|
"281b6bf43cb86d0dc7b98e1b7def4a80f3ce16d28d2308f934f116767306f06c".parse::<Secret>().unwrap(),
|
|
)], key.versions[0].id_numbers.clone().into_iter().map(|(k, v)| (k.into(), v.into())).collect::<Vec<(Public, Secret)>>());
|
|
assert_eq!("00125d85a05e5e63e214cb60fe63f132eec8a103aa29266b7e6e6c5b7597230b".parse::<Secret>().unwrap(), key.versions[0].secret_share.clone().into());
|
|
}
|
|
|
|
#[test]
|
|
fn upgrade_db_from_1() {
|
|
let tempdir = TempDir::new("").unwrap();
|
|
let db = Database::open_default(&tempdir.path().display().to_string()).unwrap();
|
|
|
|
// prepare v1 database
|
|
{
|
|
let key = serde_json::to_vec(&SerializableDocumentKeyShareV1 {
|
|
author: "b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".into(),
|
|
threshold: 777,
|
|
id_numbers: vec![(
|
|
"b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".into(),
|
|
"281b6bf43cb86d0dc7b98e1b7def4a80f3ce16d28d2308f934f116767306f06c".parse::<Secret>().unwrap().into(),
|
|
)].into_iter().collect(),
|
|
secret_share: "00125d85a05e5e63e214cb60fe63f132eec8a103aa29266b7e6e6c5b7597230b".parse::<Secret>().unwrap().into(),
|
|
common_point: Some("99e82b163b062d55a64085bacfd407bb55f194ba5fb7a1af9c34b84435455520f1372e0e650a4f91aed0058cb823f62146ccb5599c8d13372c300dea866b69fc".into()),
|
|
encrypted_point: Some("7e05df9dd077ec21ed4bc45c9fe9e0a43d65fa4be540630de615ced5e95cf5c3003035eb713317237d7667feeeb64335525158f5f7411f67aca9645169ea554c".into()),
|
|
}).unwrap();
|
|
let mut batch = db.transaction();
|
|
batch.put(None, DB_META_KEY_VERSION, &[1]);
|
|
batch.put(None, &[7], &key);
|
|
db.write(batch).unwrap();
|
|
}
|
|
|
|
// upgrade database
|
|
let db = upgrade_db(db).unwrap();
|
|
|
|
// check upgrade
|
|
assert_eq!(db.get(None, DB_META_KEY_VERSION).unwrap().unwrap()[0], CURRENT_VERSION);
|
|
let key = serde_json::from_slice::<CurrentSerializableDocumentKeyShare>(&db.get(None, &[7]).unwrap().map(|key| key.to_vec()).unwrap()).unwrap();
|
|
assert_eq!(777, key.threshold);
|
|
assert_eq!(Some("99e82b163b062d55a64085bacfd407bb55f194ba5fb7a1af9c34b84435455520f1372e0e650a4f91aed0058cb823f62146ccb5599c8d13372c300dea866b69fc".parse::<Public>().unwrap()), key.common_point.clone().map(Into::into));
|
|
assert_eq!(Some("7e05df9dd077ec21ed4bc45c9fe9e0a43d65fa4be540630de615ced5e95cf5c3003035eb713317237d7667feeeb64335525158f5f7411f67aca9645169ea554c".parse::<Public>().unwrap()), key.encrypted_point.clone().map(Into::into));
|
|
assert_eq!(key.author, "b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".into());
|
|
|
|
assert_eq!(key.versions.len(), 1);
|
|
assert_eq!(vec![(
|
|
"b486d3840218837b035c66196ecb15e6b067ca20101e11bd5e626288ab6806ecc70b8307012626bd512bad1559112d11d21025cef48cc7a1d2f3976da08f36c8".parse::<Public>().unwrap(),
|
|
"281b6bf43cb86d0dc7b98e1b7def4a80f3ce16d28d2308f934f116767306f06c".parse::<Secret>().unwrap(),
|
|
)], key.versions[0].id_numbers.clone().into_iter().map(|(k, v)| (k.into(), v.into())).collect::<Vec<(Public, Secret)>>());
|
|
|
|
assert_eq!("00125d85a05e5e63e214cb60fe63f132eec8a103aa29266b7e6e6c5b7597230b".parse::<Secret>().unwrap(), key.versions[0].secret_share.clone().into());
|
|
}
|
|
}
|