Compare commits
13 Commits
contract-m
...
tmp-simple
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
35e00b1dab
|
||
|
4cabc22c34
|
|||
|
0fac7246f0
|
|||
|
6dd775051c
|
|||
|
ee37a23ac5
|
|||
|
f53e9a7d2a
|
|||
|
fad9b62539
|
|||
|
a44e2734a4
|
|||
|
52059fe302
|
|||
|
bbba6f48e2
|
|||
|
d71e87198e
|
|||
|
71e864695b
|
|||
|
071cbcc6fb
|
4
.gitignore
vendored
4
.gitignore
vendored
@@ -4,7 +4,3 @@ service-configs/*
|
|||||||
__pycache__
|
__pycache__
|
||||||
*.pyc
|
*.pyc
|
||||||
*.o
|
*.o
|
||||||
gmon.out
|
|
||||||
*.egg-info
|
|
||||||
dist/
|
|
||||||
build/
|
|
||||||
|
|||||||
@@ -6,4 +6,3 @@ HOST=localhost
|
|||||||
PORT=5432
|
PORT=5432
|
||||||
ENGINE=postgresql
|
ENGINE=postgresql
|
||||||
DRIVER=psycopg2
|
DRIVER=psycopg2
|
||||||
DEBUG=
|
|
||||||
|
|||||||
@@ -6,4 +6,3 @@ HOST=localhost
|
|||||||
PORT=5432
|
PORT=5432
|
||||||
ENGINE=sqlite
|
ENGINE=sqlite
|
||||||
DRIVER=pysqlite
|
DRIVER=pysqlite
|
||||||
DEBUG=
|
|
||||||
|
|||||||
@@ -2,14 +2,9 @@
|
|||||||
import logging
|
import logging
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .list import (
|
from .list import list_transactions_mined
|
||||||
list_transactions_mined,
|
from .list import list_transactions_account_mined
|
||||||
list_transactions_account_mined,
|
from .list import add_transaction
|
||||||
add_transaction,
|
|
||||||
tag_transaction,
|
|
||||||
add_tag,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
logg = logging.getLogger()
|
logg = logging.getLogger()
|
||||||
|
|
||||||
|
|||||||
@@ -2,9 +2,8 @@
|
|||||||
import logging
|
import logging
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
from cic_cache.db.models.base import SessionBase
|
from cic_cache.db.models.base import SessionBase
|
||||||
from sqlalchemy import text
|
|
||||||
|
|
||||||
logg = logging.getLogger()
|
logg = logging.getLogger()
|
||||||
|
|
||||||
@@ -51,8 +50,7 @@ def list_transactions_account_mined(
|
|||||||
|
|
||||||
|
|
||||||
def add_transaction(
|
def add_transaction(
|
||||||
session,
|
session, tx_hash,
|
||||||
tx_hash,
|
|
||||||
block_number,
|
block_number,
|
||||||
tx_index,
|
tx_index,
|
||||||
sender,
|
sender,
|
||||||
@@ -64,33 +62,6 @@ def add_transaction(
|
|||||||
success,
|
success,
|
||||||
timestamp,
|
timestamp,
|
||||||
):
|
):
|
||||||
"""Adds a single transaction to the cache persistent storage. Sensible interpretation of all fields is the responsibility of the caller.
|
|
||||||
|
|
||||||
:param session: Persistent storage session object
|
|
||||||
:type session: SQLAlchemy session
|
|
||||||
:param tx_hash: Transaction hash
|
|
||||||
:type tx_hash: str, 0x-hex
|
|
||||||
:param block_number: Block number
|
|
||||||
:type block_number: int
|
|
||||||
:param tx_index: Transaction index in block
|
|
||||||
:type tx_index: int
|
|
||||||
:param sender: Ethereum address of effective sender
|
|
||||||
:type sender: str, 0x-hex
|
|
||||||
:param receiver: Ethereum address of effective recipient
|
|
||||||
:type receiver: str, 0x-hex
|
|
||||||
:param source_token: Ethereum address of token used by sender
|
|
||||||
:type source_token: str, 0x-hex
|
|
||||||
:param destination_token: Ethereum address of token received by recipient
|
|
||||||
:type destination_token: str, 0x-hex
|
|
||||||
:param from_value: Source token value spent in transaction
|
|
||||||
:type from_value: int
|
|
||||||
:param to_value: Destination token value received in transaction
|
|
||||||
:type to_value: int
|
|
||||||
:param success: True if code execution on network was successful
|
|
||||||
:type success: bool
|
|
||||||
:param date_block: Block timestamp
|
|
||||||
:type date_block: datetime
|
|
||||||
"""
|
|
||||||
date_block = datetime.datetime.fromtimestamp(timestamp)
|
date_block = datetime.datetime.fromtimestamp(timestamp)
|
||||||
s = "INSERT INTO tx (tx_hash, block_number, tx_index, sender, recipient, source_token, destination_token, from_value, to_value, success, date_block) VALUES ('{}', {}, {}, '{}', '{}', '{}', '{}', {}, {}, {}, '{}')".format(
|
s = "INSERT INTO tx (tx_hash, block_number, tx_index, sender, recipient, source_token, destination_token, from_value, to_value, success, date_block) VALUES ('{}', {}, {}, '{}', '{}', '{}', '{}', {}, {}, {}, '{}')".format(
|
||||||
tx_hash,
|
tx_hash,
|
||||||
@@ -106,74 +77,3 @@ def add_transaction(
|
|||||||
date_block,
|
date_block,
|
||||||
)
|
)
|
||||||
session.execute(s)
|
session.execute(s)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def tag_transaction(
|
|
||||||
session,
|
|
||||||
tx_hash,
|
|
||||||
name,
|
|
||||||
domain=None,
|
|
||||||
):
|
|
||||||
"""Tag a single transaction with a single tag.
|
|
||||||
|
|
||||||
Tag must already exist in storage.
|
|
||||||
|
|
||||||
:param session: Persistent storage session object
|
|
||||||
:type session: SQLAlchemy session
|
|
||||||
:param tx_hash: Transaction hash
|
|
||||||
:type tx_hash: str, 0x-hex
|
|
||||||
:param name: Tag value
|
|
||||||
:type name: str
|
|
||||||
:param domain: Tag domain
|
|
||||||
:type domain: str
|
|
||||||
:raises ValueError: Unknown tag or transaction hash
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
s = text("SELECT id from tx where tx_hash = :a")
|
|
||||||
r = session.execute(s, {'a': tx_hash}).fetchall()
|
|
||||||
tx_id = r[0].values()[0]
|
|
||||||
|
|
||||||
if tx_id == None:
|
|
||||||
raise ValueError('unknown tx hash {}'.format(tx_hash))
|
|
||||||
|
|
||||||
#s = text("SELECT id from tag where value = :a and domain = :b")
|
|
||||||
if domain == None:
|
|
||||||
s = text("SELECT id from tag where value = :a")
|
|
||||||
else:
|
|
||||||
s = text("SELECT id from tag where value = :a and domain = :b")
|
|
||||||
r = session.execute(s, {'a': name, 'b': domain}).fetchall()
|
|
||||||
tag_id = r[0].values()[0]
|
|
||||||
|
|
||||||
logg.debug('type {} {}'.format(type(tag_id), type(tx_id)))
|
|
||||||
|
|
||||||
if tag_id == None:
|
|
||||||
raise ValueError('unknown tag name {} domain {}'.format(name, domain))
|
|
||||||
|
|
||||||
s = text("INSERT INTO tag_tx_link (tag_id, tx_id) VALUES (:a, :b)")
|
|
||||||
r = session.execute(s, {'a': int(tag_id), 'b': int(tx_id)})
|
|
||||||
|
|
||||||
|
|
||||||
def add_tag(
|
|
||||||
session,
|
|
||||||
name,
|
|
||||||
domain=None,
|
|
||||||
):
|
|
||||||
"""Add a single tag to storage.
|
|
||||||
|
|
||||||
:param session: Persistent storage session object
|
|
||||||
:type session: SQLAlchemy session
|
|
||||||
:param name: Tag value
|
|
||||||
:type name: str
|
|
||||||
:param domain: Tag domain
|
|
||||||
:type domain: str
|
|
||||||
:raises sqlalchemy.exc.IntegrityError: Tag already exists
|
|
||||||
"""
|
|
||||||
|
|
||||||
s = None
|
|
||||||
if domain == None:
|
|
||||||
s = text("INSERT INTO tag (value) VALUES (:b)")
|
|
||||||
else:
|
|
||||||
s = text("INSERT INTO tag (domain, value) VALUES (:a, :b)")
|
|
||||||
session.execute(s, {'a': domain, 'b': name})
|
|
||||||
|
|||||||
@@ -1,38 +0,0 @@
|
|||||||
"""Transaction tags
|
|
||||||
|
|
||||||
Revision ID: aaf2bdce7d6e
|
|
||||||
Revises: 6604de4203e2
|
|
||||||
Create Date: 2021-05-01 09:20:20.775082
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = 'aaf2bdce7d6e'
|
|
||||||
down_revision = '6604de4203e2'
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.create_table(
|
|
||||||
'tag',
|
|
||||||
sa.Column('id', sa.Integer, primary_key=True),
|
|
||||||
sa.Column('domain', sa.String(), nullable=True),
|
|
||||||
sa.Column('value', sa.String(), nullable=False),
|
|
||||||
)
|
|
||||||
op.create_index('idx_tag_domain_value', 'tag', ['domain', 'value'], unique=True)
|
|
||||||
|
|
||||||
op.create_table(
|
|
||||||
'tag_tx_link',
|
|
||||||
sa.Column('id', sa.Integer, primary_key=True),
|
|
||||||
sa.Column('tag_id', sa.Integer, sa.ForeignKey('tag.id'), nullable=False),
|
|
||||||
sa.Column('tx_id', sa.Integer, sa.ForeignKey('tx.id'), nullable=False),
|
|
||||||
)
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_table('tag_tx_link')
|
|
||||||
op.drop_index('idx_tag_domain_value')
|
|
||||||
op.drop_table('tag')
|
|
||||||
@@ -1,27 +1,2 @@
|
|||||||
class TagSyncFilter:
|
class SyncFilter:
|
||||||
"""Holds tag name and domain for an implementing filter.
|
pass
|
||||||
|
|
||||||
:param name: Tag value
|
|
||||||
:type name: str
|
|
||||||
:param domain: Tag domain
|
|
||||||
:type domain: str
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, name, domain=None):
|
|
||||||
self.tag_name = name
|
|
||||||
self.tag_domain = domain
|
|
||||||
|
|
||||||
|
|
||||||
def tag(self):
|
|
||||||
"""Return tag value/domain.
|
|
||||||
|
|
||||||
:rtype: Tuple
|
|
||||||
:returns: tag value/domain.
|
|
||||||
"""
|
|
||||||
return (self.tag_name, self.tag_domain)
|
|
||||||
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
if self.tag_domain == None:
|
|
||||||
return self.tag_name
|
|
||||||
return '{}.{}'.format(self.tag_domain, self.tag_name)
|
|
||||||
|
|||||||
@@ -15,16 +15,15 @@ from cic_eth_registry.error import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .base import TagSyncFilter
|
from .base import SyncFilter
|
||||||
from cic_cache import db as cic_cache_db
|
from cic_cache import db as cic_cache_db
|
||||||
|
|
||||||
logg = logging.getLogger().getChild(__name__)
|
logg = logging.getLogger().getChild(__name__)
|
||||||
|
|
||||||
|
|
||||||
class ERC20TransferFilter(TagSyncFilter):
|
class ERC20TransferFilter(SyncFilter):
|
||||||
|
|
||||||
def __init__(self, chain_spec):
|
def __init__(self, chain_spec):
|
||||||
super(ERC20TransferFilter, self).__init__('transfer', domain='erc20')
|
|
||||||
self.chain_spec = chain_spec
|
self.chain_spec = chain_spec
|
||||||
|
|
||||||
|
|
||||||
@@ -47,9 +46,6 @@ class ERC20TransferFilter(TagSyncFilter):
|
|||||||
except RequestMismatchException:
|
except RequestMismatchException:
|
||||||
logg.debug('erc20 match but not a transfer, skipping')
|
logg.debug('erc20 match but not a transfer, skipping')
|
||||||
return False
|
return False
|
||||||
except ValueError:
|
|
||||||
logg.debug('erc20 match but bogus data, skipping')
|
|
||||||
return False
|
|
||||||
|
|
||||||
token_sender = tx.outputs[0]
|
token_sender = tx.outputs[0]
|
||||||
token_recipient = transfer_data[0]
|
token_recipient = transfer_data[0]
|
||||||
@@ -71,13 +67,7 @@ class ERC20TransferFilter(TagSyncFilter):
|
|||||||
tx.status == Status.SUCCESS,
|
tx.status == Status.SUCCESS,
|
||||||
block.timestamp,
|
block.timestamp,
|
||||||
)
|
)
|
||||||
db_session.flush()
|
#db_session.flush()
|
||||||
cic_cache_db.tag_transaction(
|
|
||||||
db_session,
|
|
||||||
tx.hash,
|
|
||||||
self.tag_name,
|
|
||||||
domain=self.tag_domain,
|
|
||||||
)
|
|
||||||
db_session.commit()
|
db_session.commit()
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|||||||
@@ -7,10 +7,9 @@ import argparse
|
|||||||
import sys
|
import sys
|
||||||
import re
|
import re
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import confini
|
import confini
|
||||||
import celery
|
import celery
|
||||||
import sqlalchemy
|
|
||||||
import rlp
|
import rlp
|
||||||
import cic_base.config
|
import cic_base.config
|
||||||
import cic_base.log
|
import cic_base.log
|
||||||
@@ -35,10 +34,7 @@ from chainsyncer.driver import (
|
|||||||
from chainsyncer.db.models.base import SessionBase
|
from chainsyncer.db.models.base import SessionBase
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_cache.db import (
|
from cic_cache.db import dsn_from_config
|
||||||
dsn_from_config,
|
|
||||||
add_tag,
|
|
||||||
)
|
|
||||||
from cic_cache.runnable.daemons.filters import (
|
from cic_cache.runnable.daemons.filters import (
|
||||||
ERC20TransferFilter,
|
ERC20TransferFilter,
|
||||||
)
|
)
|
||||||
@@ -63,17 +59,6 @@ chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
|||||||
cic_base.rpc.setup(chain_spec, config.get('ETH_PROVIDER'))
|
cic_base.rpc.setup(chain_spec, config.get('ETH_PROVIDER'))
|
||||||
|
|
||||||
|
|
||||||
def register_filter_tags(filters, session):
|
|
||||||
for f in filters:
|
|
||||||
tag = f.tag()
|
|
||||||
try:
|
|
||||||
add_tag(session, tag[0], domain=tag[1])
|
|
||||||
session.commit()
|
|
||||||
logg.info('added tag name "{}" domain "{}"'.format(tag[0], tag[1]))
|
|
||||||
except sqlalchemy.exc.IntegrityError:
|
|
||||||
logg.debug('already have tag name "{}" domain "{}"'.format(tag[0], tag[1]))
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# Connect to blockchain with chainlib
|
# Connect to blockchain with chainlib
|
||||||
rpc = RPCConnection.connect(chain_spec, 'default')
|
rpc = RPCConnection.connect(chain_spec, 'default')
|
||||||
@@ -113,19 +98,10 @@ def main():
|
|||||||
|
|
||||||
erc20_transfer_filter = ERC20TransferFilter(chain_spec)
|
erc20_transfer_filter = ERC20TransferFilter(chain_spec)
|
||||||
|
|
||||||
filters = [
|
|
||||||
erc20_transfer_filter,
|
|
||||||
]
|
|
||||||
|
|
||||||
session = SessionBase.create_session()
|
|
||||||
register_filter_tags(filters, session)
|
|
||||||
session.close()
|
|
||||||
|
|
||||||
i = 0
|
i = 0
|
||||||
for syncer in syncers:
|
for syncer in syncers:
|
||||||
logg.debug('running syncer index {}'.format(i))
|
logg.debug('running syncer index {}'.format(i))
|
||||||
for f in filters:
|
syncer.add_filter(erc20_transfer_filter)
|
||||||
syncer.add_filter(f)
|
|
||||||
r = syncer.loop(int(config.get('SYNCER_LOOP_INTERVAL')), rpc)
|
r = syncer.loop(int(config.get('SYNCER_LOOP_INTERVAL')), rpc)
|
||||||
sys.stderr.write("sync {} done at block {}\n".format(syncer, r))
|
sys.stderr.write("sync {} done at block {}\n".format(syncer, r))
|
||||||
|
|
||||||
|
|||||||
@@ -6,4 +6,4 @@ HOST=localhost
|
|||||||
PORT=5432
|
PORT=5432
|
||||||
ENGINE=postgresql
|
ENGINE=postgresql
|
||||||
DRIVER=psycopg2
|
DRIVER=psycopg2
|
||||||
DEBUG=0
|
DEBUG=
|
||||||
|
|||||||
@@ -1,4 +1,2 @@
|
|||||||
[cic]
|
[cic]
|
||||||
registry_address =
|
registry_address =
|
||||||
chain_spec =
|
|
||||||
trust_address =
|
|
||||||
|
|||||||
@@ -6,4 +6,4 @@ HOST=localhost
|
|||||||
PORT=5432
|
PORT=5432
|
||||||
ENGINE=sqlite
|
ENGINE=sqlite
|
||||||
DRIVER=pysqlite
|
DRIVER=pysqlite
|
||||||
DEBUG=1
|
DEBUG=
|
||||||
|
|||||||
@@ -43,6 +43,10 @@ COPY cic-cache/config/ /usr/local/etc/cic-cache/
|
|||||||
RUN git clone https://github.com/vishnubob/wait-for-it.git /usr/local/bin/wait-for-it/
|
RUN git clone https://github.com/vishnubob/wait-for-it.git /usr/local/bin/wait-for-it/
|
||||||
COPY cic-cache/cic_cache/db/migrations/ /usr/local/share/cic-cache/alembic/
|
COPY cic-cache/cic_cache/db/migrations/ /usr/local/share/cic-cache/alembic/
|
||||||
|
|
||||||
|
RUN git clone https://gitlab.com/grassrootseconomics/cic-contracts.git && \
|
||||||
|
mkdir -p /usr/local/share/cic/solidity && \
|
||||||
|
cp -R cic-contracts/abis /usr/local/share/cic/solidity/abi
|
||||||
|
|
||||||
COPY cic-cache/docker/start_tracker.sh ./start_tracker.sh
|
COPY cic-cache/docker/start_tracker.sh ./start_tracker.sh
|
||||||
COPY cic-cache/docker/db.sh ./db.sh
|
COPY cic-cache/docker/db.sh ./db.sh
|
||||||
RUN chmod 755 ./*.sh
|
RUN chmod 755 ./*.sh
|
||||||
|
|||||||
@@ -2,9 +2,4 @@
|
|||||||
|
|
||||||
. ./db.sh
|
. ./db.sh
|
||||||
|
|
||||||
if [ $? -ne "0" ]; then
|
|
||||||
>&2 echo db migrate fail
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
/usr/local/bin/cic-cache-trackerd $@
|
/usr/local/bin/cic-cache-trackerd $@
|
||||||
|
|||||||
@@ -1,9 +1,9 @@
|
|||||||
cic-base~=0.1.2b6
|
cic-base~=0.1.2a77
|
||||||
alembic==1.4.2
|
alembic==1.4.2
|
||||||
confini~=0.3.6rc3
|
confini~=0.3.6rc3
|
||||||
uwsgi==2.0.19.1
|
uwsgi==2.0.19.1
|
||||||
moolb~=0.1.0
|
moolb~=0.1.0
|
||||||
cic-eth-registry~=0.5.5a1
|
cic-eth-registry~=0.5.4a16
|
||||||
SQLAlchemy==1.3.20
|
SQLAlchemy==1.3.20
|
||||||
semver==2.13.0
|
semver==2.13.0
|
||||||
psycopg2==2.8.6
|
psycopg2==2.8.6
|
||||||
|
|||||||
@@ -4,10 +4,3 @@ pytest-mock==3.3.1
|
|||||||
pysqlite3==0.4.3
|
pysqlite3==0.4.3
|
||||||
sqlparse==0.4.1
|
sqlparse==0.4.1
|
||||||
pytest-celery==0.0.0a1
|
pytest-celery==0.0.0a1
|
||||||
eth_tester==0.5.0b3
|
|
||||||
py-evm==0.3.0a20
|
|
||||||
web3==5.12.2
|
|
||||||
cic-eth-registry~=0.5.5a3
|
|
||||||
giftable-erc20-token~=0.0.8a10
|
|
||||||
eth-address-index~=0.1.1a10
|
|
||||||
sarafu-faucet~=0.0.3a1
|
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
@@ -84,7 +84,3 @@ def txs(
|
|||||||
|
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
return [
|
|
||||||
tx_hash_first,
|
|
||||||
tx_hash_second,
|
|
||||||
]
|
|
||||||
|
|||||||
@@ -1,3 +0,0 @@
|
|||||||
from chainlib.eth.pytest import *
|
|
||||||
from cic_eth_registry.pytest.fixtures_tokens import *
|
|
||||||
|
|
||||||
@@ -1,69 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import os
|
|
||||||
import datetime
|
|
||||||
import logging
|
|
||||||
import json
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
import pytest
|
|
||||||
from sqlalchemy import text
|
|
||||||
from chainlib.eth.tx import Tx
|
|
||||||
from chainlib.eth.block import Block
|
|
||||||
from chainlib.chain import ChainSpec
|
|
||||||
from hexathon import (
|
|
||||||
strip_0x,
|
|
||||||
add_0x,
|
|
||||||
)
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from cic_cache.db import add_tag
|
|
||||||
from cic_cache.runnable.daemons.filters.erc20 import ERC20TransferFilter
|
|
||||||
|
|
||||||
logg = logging.getLogger()
|
|
||||||
|
|
||||||
|
|
||||||
def test_cache(
|
|
||||||
eth_rpc,
|
|
||||||
foo_token,
|
|
||||||
init_database,
|
|
||||||
list_defaults,
|
|
||||||
list_actors,
|
|
||||||
tags,
|
|
||||||
):
|
|
||||||
|
|
||||||
chain_spec = ChainSpec('foo', 'bar', 42, 'baz')
|
|
||||||
|
|
||||||
fltr = ERC20TransferFilter(chain_spec)
|
|
||||||
|
|
||||||
add_tag(init_database, fltr.tag_name, domain=fltr.tag_domain)
|
|
||||||
|
|
||||||
data = 'a9059cbb'
|
|
||||||
data += strip_0x(list_actors['alice'])
|
|
||||||
data += '1000'.ljust(64, '0')
|
|
||||||
|
|
||||||
block = Block({
|
|
||||||
'hash': os.urandom(32).hex(),
|
|
||||||
'number': 42,
|
|
||||||
'timestamp': datetime.datetime.utcnow().timestamp(),
|
|
||||||
'transactions': [],
|
|
||||||
})
|
|
||||||
|
|
||||||
tx = Tx({
|
|
||||||
'to': foo_token,
|
|
||||||
'from': list_actors['bob'],
|
|
||||||
'data': data,
|
|
||||||
'value': 0,
|
|
||||||
'hash': os.urandom(32).hex(),
|
|
||||||
'nonce': 13,
|
|
||||||
'gasPrice': 10000000,
|
|
||||||
'gas': 123456,
|
|
||||||
})
|
|
||||||
block.txs.append(tx)
|
|
||||||
tx.block = block
|
|
||||||
|
|
||||||
r = fltr.filter(eth_rpc, block, tx, db_session=init_database)
|
|
||||||
assert r
|
|
||||||
|
|
||||||
s = text("SELECT x.tx_hash FROM tag a INNER JOIN tag_tx_link l ON l.tag_id = a.id INNER JOIN tx x ON x.id = l.tx_id WHERE a.domain = :a AND a.value = :b")
|
|
||||||
r = init_database.execute(s, {'a': fltr.tag_domain, 'b': fltr.tag_name}).fetchone()
|
|
||||||
assert r[0] == tx.hash
|
|
||||||
@@ -2,7 +2,7 @@
|
|||||||
import os
|
import os
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import pytest
|
import pytest
|
||||||
import confini
|
import confini
|
||||||
|
|
||||||
@@ -13,7 +13,7 @@ logg = logging.getLogger(__file__)
|
|||||||
|
|
||||||
@pytest.fixture(scope='session')
|
@pytest.fixture(scope='session')
|
||||||
def load_config():
|
def load_config():
|
||||||
config_dir = os.path.join(root_dir, 'config/test')
|
config_dir = os.path.join(root_dir, '.config/test')
|
||||||
conf = confini.Config(config_dir, 'CICTEST')
|
conf = confini.Config(config_dir, 'CICTEST')
|
||||||
conf.process()
|
conf.process()
|
||||||
logg.debug('config {}'.format(conf))
|
logg.debug('config {}'.format(conf))
|
||||||
|
|||||||
@@ -3,16 +3,13 @@ import os
|
|||||||
import logging
|
import logging
|
||||||
import re
|
import re
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import pytest
|
import pytest
|
||||||
import sqlparse
|
import sqlparse
|
||||||
import alembic
|
|
||||||
from alembic.config import Config as AlembicConfig
|
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_cache.db.models.base import SessionBase
|
from cic_cache.db.models.base import SessionBase
|
||||||
from cic_cache.db import dsn_from_config
|
from cic_cache.db import dsn_from_config
|
||||||
from cic_cache.db import add_tag
|
|
||||||
|
|
||||||
logg = logging.getLogger(__file__)
|
logg = logging.getLogger(__file__)
|
||||||
|
|
||||||
@@ -29,10 +26,11 @@ def database_engine(
|
|||||||
except FileNotFoundError:
|
except FileNotFoundError:
|
||||||
pass
|
pass
|
||||||
dsn = dsn_from_config(load_config)
|
dsn = dsn_from_config(load_config)
|
||||||
SessionBase.connect(dsn, debug=load_config.true('DATABASE_DEBUG'))
|
SessionBase.connect(dsn)
|
||||||
return dsn
|
return dsn
|
||||||
|
|
||||||
|
|
||||||
|
# TODO: use alembic instead to migrate db, here we have to keep separate schema than migration script in script/migrate.py
|
||||||
@pytest.fixture(scope='function')
|
@pytest.fixture(scope='function')
|
||||||
def init_database(
|
def init_database(
|
||||||
load_config,
|
load_config,
|
||||||
@@ -40,23 +38,52 @@ def init_database(
|
|||||||
):
|
):
|
||||||
|
|
||||||
rootdir = os.path.dirname(os.path.dirname(__file__))
|
rootdir = os.path.dirname(os.path.dirname(__file__))
|
||||||
dbdir = os.path.join(rootdir, 'cic_cache', 'db')
|
schemadir = os.path.join(rootdir, 'db', load_config.get('DATABASE_DRIVER'))
|
||||||
migrationsdir = os.path.join(dbdir, 'migrations', load_config.get('DATABASE_ENGINE'))
|
|
||||||
if not os.path.isdir(migrationsdir):
|
if load_config.get('DATABASE_ENGINE') == 'sqlite':
|
||||||
migrationsdir = os.path.join(dbdir, 'migrations', 'default')
|
rconn = SessionBase.engine.raw_connection()
|
||||||
logg.info('using migrations directory {}'.format(migrationsdir))
|
f = open(os.path.join(schemadir, 'db.sql'))
|
||||||
|
s = f.read()
|
||||||
|
f.close()
|
||||||
|
rconn.executescript(s)
|
||||||
|
|
||||||
|
else:
|
||||||
|
rconn = SessionBase.engine.raw_connection()
|
||||||
|
rcursor = rconn.cursor()
|
||||||
|
|
||||||
|
#rcursor.execute('DROP FUNCTION IF EXISTS public.transaction_list')
|
||||||
|
#rcursor.execute('DROP FUNCTION IF EXISTS public.balances')
|
||||||
|
|
||||||
|
f = open(os.path.join(schemadir, 'db.sql'))
|
||||||
|
s = f.read()
|
||||||
|
f.close()
|
||||||
|
r = re.compile(r'^[A-Z]', re.MULTILINE)
|
||||||
|
for l in sqlparse.parse(s):
|
||||||
|
strl = str(l)
|
||||||
|
# we need to check for empty query lines, as sqlparse doesn't do that on its own (and psycopg complains when it gets them)
|
||||||
|
if not re.search(r, strl):
|
||||||
|
logg.warning('skipping parsed query line {}'.format(strl))
|
||||||
|
continue
|
||||||
|
rcursor.execute(strl)
|
||||||
|
rconn.commit()
|
||||||
|
|
||||||
|
rcursor.execute('SET search_path TO public')
|
||||||
|
|
||||||
|
# this doesn't work when run separately, no idea why
|
||||||
|
# functions have been manually added to original schema from cic-eth
|
||||||
|
# f = open(os.path.join(schemadir, 'proc_transaction_list.sql'))
|
||||||
|
# s = f.read()
|
||||||
|
# f.close()
|
||||||
|
# rcursor.execute(s)
|
||||||
|
#
|
||||||
|
# f = open(os.path.join(schemadir, 'proc_balances.sql'))
|
||||||
|
# s = f.read()
|
||||||
|
# f.close()
|
||||||
|
# rcursor.execute(s)
|
||||||
|
|
||||||
|
rcursor.close()
|
||||||
|
|
||||||
session = SessionBase.create_session()
|
session = SessionBase.create_session()
|
||||||
|
|
||||||
ac = AlembicConfig(os.path.join(migrationsdir, 'alembic.ini'))
|
|
||||||
ac.set_main_option('sqlalchemy.url', database_engine)
|
|
||||||
ac.set_main_option('script_location', migrationsdir)
|
|
||||||
|
|
||||||
alembic.command.downgrade(ac, 'base')
|
|
||||||
alembic.command.upgrade(ac, 'head')
|
|
||||||
|
|
||||||
session.commit()
|
|
||||||
|
|
||||||
yield session
|
yield session
|
||||||
session.commit()
|
session.commit()
|
||||||
session.close()
|
session.close()
|
||||||
@@ -89,14 +116,3 @@ def list_defaults(
|
|||||||
return {
|
return {
|
||||||
'block': 420000,
|
'block': 420000,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope='function')
|
|
||||||
def tags(
|
|
||||||
init_database,
|
|
||||||
):
|
|
||||||
|
|
||||||
add_tag(init_database, 'foo')
|
|
||||||
add_tag(init_database, 'baz', domain='bar')
|
|
||||||
add_tag(init_database, 'xyzzy', domain='bar')
|
|
||||||
init_database.commit()
|
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import datetime
|
|||||||
import logging
|
import logging
|
||||||
import json
|
import json
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
|
|||||||
@@ -1,37 +0,0 @@
|
|||||||
import os
|
|
||||||
import datetime
|
|
||||||
import logging
|
|
||||||
import json
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
import pytest
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from cic_cache.db import tag_transaction
|
|
||||||
|
|
||||||
logg = logging.getLogger()
|
|
||||||
|
|
||||||
|
|
||||||
def test_cache(
|
|
||||||
init_database,
|
|
||||||
list_defaults,
|
|
||||||
list_actors,
|
|
||||||
list_tokens,
|
|
||||||
txs,
|
|
||||||
tags,
|
|
||||||
):
|
|
||||||
|
|
||||||
tag_transaction(init_database, txs[0], 'foo')
|
|
||||||
tag_transaction(init_database, txs[0], 'baz', domain='bar')
|
|
||||||
tag_transaction(init_database, txs[1], 'xyzzy', domain='bar')
|
|
||||||
|
|
||||||
r = init_database.execute("SELECT x.tx_hash FROM tag a INNER JOIN tag_tx_link l ON l.tag_id = a.id INNER JOIN tx x ON x.id = l.tx_id WHERE a.value = 'foo'").fetchall()
|
|
||||||
assert r[0][0] == txs[0]
|
|
||||||
|
|
||||||
|
|
||||||
r = init_database.execute("SELECT x.tx_hash FROM tag a INNER JOIN tag_tx_link l ON l.tag_id = a.id INNER JOIN tx x ON x.id = l.tx_id WHERE a.domain = 'bar' AND a.value = 'baz'").fetchall()
|
|
||||||
assert r[0][0] == txs[0]
|
|
||||||
|
|
||||||
|
|
||||||
r = init_database.execute("SELECT x.tx_hash FROM tag a INNER JOIN tag_tx_link l ON l.tag_id = a.id INNER JOIN tx x ON x.id = l.tx_id WHERE a.domain = 'bar' AND a.value = 'xyzzy'").fetchall()
|
|
||||||
assert r[0][0] == txs[1]
|
|
||||||
@@ -2,7 +2,7 @@
|
|||||||
import datetime
|
import datetime
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import celery
|
import celery
|
||||||
from chainlib.eth.constant import ZERO_ADDRESS
|
from chainlib.eth.constant import ZERO_ADDRESS
|
||||||
from chainlib.chain import ChainSpec
|
from chainlib.chain import ChainSpec
|
||||||
@@ -32,9 +32,7 @@ def lock(chained_input, chain_spec_dict, address=ZERO_ADDRESS, flags=LockEnum.AL
|
|||||||
:returns: New lock state for address
|
:returns: New lock state for address
|
||||||
:rtype: number
|
:rtype: number
|
||||||
"""
|
"""
|
||||||
chain_str = '::'
|
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
||||||
if chain_spec_dict != None:
|
|
||||||
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
|
||||||
r = Lock.set(chain_str, flags, address=address, tx_hash=tx_hash)
|
r = Lock.set(chain_str, flags, address=address, tx_hash=tx_hash)
|
||||||
logg.debug('Locked {} for {}, flag now {}'.format(flags, address, r))
|
logg.debug('Locked {} for {}, flag now {}'.format(flags, address, r))
|
||||||
return chained_input
|
return chained_input
|
||||||
@@ -53,9 +51,7 @@ def unlock(chained_input, chain_spec_dict, address=ZERO_ADDRESS, flags=LockEnum.
|
|||||||
:returns: New lock state for address
|
:returns: New lock state for address
|
||||||
:rtype: number
|
:rtype: number
|
||||||
"""
|
"""
|
||||||
chain_str = '::'
|
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
||||||
if chain_spec_dict != None:
|
|
||||||
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
|
||||||
r = Lock.reset(chain_str, flags, address=address)
|
r = Lock.reset(chain_str, flags, address=address)
|
||||||
logg.debug('Unlocked {} for {}, flag now {}'.format(flags, address, r))
|
logg.debug('Unlocked {} for {}, flag now {}'.format(flags, address, r))
|
||||||
return chained_input
|
return chained_input
|
||||||
@@ -131,9 +127,7 @@ def unlock_queue(chained_input, chain_spec_dict, address=ZERO_ADDRESS):
|
|||||||
|
|
||||||
@celery_app.task(base=CriticalSQLAlchemyTask)
|
@celery_app.task(base=CriticalSQLAlchemyTask)
|
||||||
def check_lock(chained_input, chain_spec_dict, lock_flags, address=None):
|
def check_lock(chained_input, chain_spec_dict, lock_flags, address=None):
|
||||||
chain_str = '::'
|
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
||||||
if chain_spec_dict != None:
|
|
||||||
chain_str = str(ChainSpec.from_dict(chain_spec_dict))
|
|
||||||
session = SessionBase.create_session()
|
session = SessionBase.create_session()
|
||||||
r = Lock.check(chain_str, lock_flags, address=ZERO_ADDRESS, session=session)
|
r = Lock.check(chain_str, lock_flags, address=ZERO_ADDRESS, session=session)
|
||||||
if address != None:
|
if address != None:
|
||||||
@@ -145,9 +139,3 @@ def check_lock(chained_input, chain_spec_dict, lock_flags, address=None):
|
|||||||
session.flush()
|
session.flush()
|
||||||
session.close()
|
session.close()
|
||||||
return chained_input
|
return chained_input
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task()
|
|
||||||
def shutdown(message):
|
|
||||||
logg.critical('shutdown called: {}'.format(message))
|
|
||||||
celery_app.control.shutdown() #broadcast('shutdown')
|
|
||||||
|
|||||||
@@ -1,19 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
import celery
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from cic_eth.task import BaseTask
|
|
||||||
|
|
||||||
celery_app = celery.current_app
|
|
||||||
logg = logging.getLogger()
|
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task(bind=True, base=BaseTask)
|
|
||||||
def default_token(self):
|
|
||||||
return {
|
|
||||||
'symbol': self.default_token_symbol,
|
|
||||||
'address': self.default_token_address,
|
|
||||||
}
|
|
||||||
@@ -60,29 +60,6 @@ class AdminApi:
|
|||||||
self.call_address = call_address
|
self.call_address = call_address
|
||||||
|
|
||||||
|
|
||||||
def proxy_do(self, chain_spec, o):
|
|
||||||
s_proxy = celery.signature(
|
|
||||||
'cic_eth.task.rpc_proxy',
|
|
||||||
[
|
|
||||||
chain_spec.asdict(),
|
|
||||||
o,
|
|
||||||
'default',
|
|
||||||
],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
return s_proxy.apply_async()
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def registry(self):
|
|
||||||
s_registry = celery.signature(
|
|
||||||
'cic_eth.task.registry',
|
|
||||||
[],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
return s_registry.apply_async()
|
|
||||||
|
|
||||||
|
|
||||||
def unlock(self, chain_spec, address, flags=None):
|
def unlock(self, chain_spec, address, flags=None):
|
||||||
s_unlock = celery.signature(
|
s_unlock = celery.signature(
|
||||||
'cic_eth.admin.ctrl.unlock',
|
'cic_eth.admin.ctrl.unlock',
|
||||||
@@ -169,6 +146,7 @@ class AdminApi:
|
|||||||
|
|
||||||
# TODO: This check should most likely be in resend task itself
|
# TODO: This check should most likely be in resend task itself
|
||||||
tx_dict = s_get_tx_cache.apply_async().get()
|
tx_dict = s_get_tx_cache.apply_async().get()
|
||||||
|
#if tx_dict['status'] in [StatusEnum.REVERTED, StatusEnum.SUCCESS, StatusEnum.CANCELLED, StatusEnum.OBSOLETED]:
|
||||||
if not is_alive(getattr(StatusEnum, tx_dict['status']).value):
|
if not is_alive(getattr(StatusEnum, tx_dict['status']).value):
|
||||||
raise TxStateChangeError('Cannot resend mined or obsoleted transaction'.format(txold_hash_hex))
|
raise TxStateChangeError('Cannot resend mined or obsoleted transaction'.format(txold_hash_hex))
|
||||||
|
|
||||||
@@ -248,6 +226,9 @@ class AdminApi:
|
|||||||
break
|
break
|
||||||
last_nonce = nonce_otx
|
last_nonce = nonce_otx
|
||||||
|
|
||||||
|
#nonce_cache = Nonce.get(address)
|
||||||
|
#nonce_w3 = self.w3.eth.getTransactionCount(address, 'pending')
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'nonce': {
|
'nonce': {
|
||||||
#'network': nonce_cache,
|
#'network': nonce_cache,
|
||||||
@@ -291,6 +272,20 @@ class AdminApi:
|
|||||||
return s_nonce.apply_async()
|
return s_nonce.apply_async()
|
||||||
|
|
||||||
|
|
||||||
|
# # TODO: this is a stub, complete all checks
|
||||||
|
# def ready(self):
|
||||||
|
# """Checks whether all required initializations have been performed.
|
||||||
|
#
|
||||||
|
# :raises cic_eth.error.InitializationError: At least one setting pre-requisite has not been met.
|
||||||
|
# :raises KeyError: An address provided for initialization is not known by the keystore.
|
||||||
|
# """
|
||||||
|
# addr = AccountRole.get_address('ETH_GAS_PROVIDER_ADDRESS')
|
||||||
|
# if addr == ZERO_ADDRESS:
|
||||||
|
# raise InitializationError('missing account ETH_GAS_PROVIDER_ADDRESS')
|
||||||
|
#
|
||||||
|
# self.w3.eth.sign(addr, text='666f6f')
|
||||||
|
|
||||||
|
|
||||||
def account(self, chain_spec, address, include_sender=True, include_recipient=True, renderer=None, w=sys.stdout):
|
def account(self, chain_spec, address, include_sender=True, include_recipient=True, renderer=None, w=sys.stdout):
|
||||||
"""Lists locally originated transactions for the given Ethereum address.
|
"""Lists locally originated transactions for the given Ethereum address.
|
||||||
|
|
||||||
@@ -353,7 +348,6 @@ class AdminApi:
|
|||||||
|
|
||||||
|
|
||||||
# TODO: Add exception upon non-existent tx aswell as invalid tx data to docstring
|
# TODO: Add exception upon non-existent tx aswell as invalid tx data to docstring
|
||||||
# TODO: This method is WAY too long
|
|
||||||
def tx(self, chain_spec, tx_hash=None, tx_raw=None, registry=None, renderer=None, w=sys.stdout):
|
def tx(self, chain_spec, tx_hash=None, tx_raw=None, registry=None, renderer=None, w=sys.stdout):
|
||||||
"""Output local and network details about a given transaction with local origin.
|
"""Output local and network details about a given transaction with local origin.
|
||||||
|
|
||||||
@@ -376,6 +370,7 @@ class AdminApi:
|
|||||||
|
|
||||||
if tx_raw != None:
|
if tx_raw != None:
|
||||||
tx_hash = add_0x(keccak256_hex_to_hex(tx_raw))
|
tx_hash = add_0x(keccak256_hex_to_hex(tx_raw))
|
||||||
|
#tx_hash = self.w3.keccak(hexstr=tx_raw).hex()
|
||||||
|
|
||||||
s = celery.signature(
|
s = celery.signature(
|
||||||
'cic_eth.queue.query.get_tx_cache',
|
'cic_eth.queue.query.get_tx_cache',
|
||||||
@@ -391,78 +386,38 @@ class AdminApi:
|
|||||||
|
|
||||||
source_token = None
|
source_token = None
|
||||||
if tx['source_token'] != ZERO_ADDRESS:
|
if tx['source_token'] != ZERO_ADDRESS:
|
||||||
if registry != None:
|
try:
|
||||||
try:
|
source_token = registry.by_address(tx['source_token'])
|
||||||
source_token = registry.by_address(tx['source_token'])
|
#source_token = CICRegistry.get_address(chain_spec, tx['source_token']).contract
|
||||||
except UnknownContractError:
|
except UnknownContractError:
|
||||||
logg.warning('unknown source token contract {} (direct)'.format(tx['source_token']))
|
#source_token_contract = self.w3.eth.contract(abi=CICRegistry.abi('ERC20'), address=tx['source_token'])
|
||||||
else:
|
#source_token = CICRegistry.add_token(chain_spec, source_token_contract)
|
||||||
s = celery.signature(
|
logg.warning('unknown source token contract {}'.format(tx['source_token']))
|
||||||
'cic_eth.task.registry_address_lookup',
|
|
||||||
[
|
|
||||||
chain_spec.asdict(),
|
|
||||||
tx['source_token'],
|
|
||||||
],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
t = s.apply_async()
|
|
||||||
source_token = t.get()
|
|
||||||
if source_token == None:
|
|
||||||
logg.warning('unknown source token contract {} (task pool)'.format(tx['source_token']))
|
|
||||||
|
|
||||||
|
|
||||||
destination_token = None
|
destination_token = None
|
||||||
if tx['destination_token'] != ZERO_ADDRESS:
|
if tx['source_token'] != ZERO_ADDRESS:
|
||||||
if registry != None:
|
try:
|
||||||
try:
|
#destination_token = CICRegistry.get_address(chain_spec, tx['destination_token'])
|
||||||
destination_token = registry.by_address(tx['destination_token'])
|
destination_token = registry.by_address(tx['destination_token'])
|
||||||
except UnknownContractError:
|
except UnknownContractError:
|
||||||
logg.warning('unknown destination token contract {}'.format(tx['destination_token']))
|
#destination_token_contract = self.w3.eth.contract(abi=CICRegistry.abi('ERC20'), address=tx['source_token'])
|
||||||
else:
|
#destination_token = CICRegistry.add_token(chain_spec, destination_token_contract)
|
||||||
s = celery.signature(
|
logg.warning('unknown destination token contract {}'.format(tx['destination_token']))
|
||||||
'cic_eth.task.registry_address_lookup',
|
|
||||||
[
|
|
||||||
chain_spec.asdict(),
|
|
||||||
tx['destination_token'],
|
|
||||||
],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
t = s.apply_async()
|
|
||||||
destination_token = t.get()
|
|
||||||
if destination_token == None:
|
|
||||||
logg.warning('unknown destination token contract {} (task pool)'.format(tx['destination_token']))
|
|
||||||
|
|
||||||
|
|
||||||
tx['sender_description'] = 'Custodial account'
|
tx['sender_description'] = 'Custodial account'
|
||||||
tx['recipient_description'] = 'Custodial account'
|
tx['recipient_description'] = 'Custodial account'
|
||||||
|
|
||||||
o = code(tx['sender'])
|
o = code(tx['sender'])
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
if len(strip_0x(r, allow_empty=True)) > 0:
|
if len(strip_0x(r, allow_empty=True)) > 0:
|
||||||
if registry != None:
|
try:
|
||||||
try:
|
#sender_contract = CICRegistry.get_address(chain_spec, tx['sender'])
|
||||||
sender_contract = registry.by_address(tx['sender'], sender_address=self.call_address)
|
sender_contract = registry.by_address(tx['sender'], sender_address=self.call_address)
|
||||||
tx['sender_description'] = 'Contract at {}'.format(tx['sender'])
|
tx['sender_description'] = 'Contract at {}'.format(tx['sender']) #sender_contract)
|
||||||
except UnknownContractError:
|
except UnknownContractError:
|
||||||
tx['sender_description'] = 'Unknown contract'
|
tx['sender_description'] = 'Unknown contract'
|
||||||
except KeyError as e:
|
except KeyError as e:
|
||||||
tx['sender_description'] = 'Unknown contract'
|
tx['sender_description'] = 'Unknown contract'
|
||||||
else:
|
|
||||||
s = celery.signature(
|
|
||||||
'cic_eth.task.registry_address_lookup',
|
|
||||||
[
|
|
||||||
chain_spec.asdict(),
|
|
||||||
tx['sender'],
|
|
||||||
],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
t = s.apply_async()
|
|
||||||
tx['sender_description'] = t.get()
|
|
||||||
if tx['sender_description'] == None:
|
|
||||||
tx['sender_description'] = 'Unknown contract'
|
|
||||||
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
s = celery.signature(
|
s = celery.signature(
|
||||||
'cic_eth.eth.account.have',
|
'cic_eth.eth.account.have',
|
||||||
@@ -491,31 +446,16 @@ class AdminApi:
|
|||||||
tx['sender_description'] = role
|
tx['sender_description'] = role
|
||||||
|
|
||||||
o = code(tx['recipient'])
|
o = code(tx['recipient'])
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
if len(strip_0x(r, allow_empty=True)) > 0:
|
if len(strip_0x(r, allow_empty=True)) > 0:
|
||||||
if registry != None:
|
try:
|
||||||
try:
|
#recipient_contract = CICRegistry.by_address(tx['recipient'])
|
||||||
recipient_contract = registry.by_address(tx['recipient'])
|
recipient_contract = registry.by_address(tx['recipient'])
|
||||||
tx['recipient_description'] = 'Contract at {}'.format(tx['recipient'])
|
tx['recipient_description'] = 'Contract at {}'.format(tx['recipient']) #recipient_contract)
|
||||||
except UnknownContractError as e:
|
except UnknownContractError as e:
|
||||||
tx['recipient_description'] = 'Unknown contract'
|
tx['recipient_description'] = 'Unknown contract'
|
||||||
except KeyError as e:
|
except KeyError as e:
|
||||||
tx['recipient_description'] = 'Unknown contract'
|
tx['recipient_description'] = 'Unknown contract'
|
||||||
else:
|
|
||||||
s = celery.signature(
|
|
||||||
'cic_eth.task.registry_address_lookup',
|
|
||||||
[
|
|
||||||
chain_spec.asdict(),
|
|
||||||
tx['recipient'],
|
|
||||||
],
|
|
||||||
queue=self.queue
|
|
||||||
)
|
|
||||||
t = s.apply_async()
|
|
||||||
tx['recipient_description'] = t.get()
|
|
||||||
if tx['recipient_description'] == None:
|
|
||||||
tx['recipient_description'] = 'Unknown contract'
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
s = celery.signature(
|
s = celery.signature(
|
||||||
'cic_eth.eth.account.have',
|
'cic_eth.eth.account.have',
|
||||||
@@ -557,8 +497,7 @@ class AdminApi:
|
|||||||
r = None
|
r = None
|
||||||
try:
|
try:
|
||||||
o = transaction(tx_hash)
|
o = transaction(tx_hash)
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
if r != None:
|
if r != None:
|
||||||
tx['network_status'] = 'Mempool'
|
tx['network_status'] = 'Mempool'
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
@@ -567,8 +506,7 @@ class AdminApi:
|
|||||||
if r != None:
|
if r != None:
|
||||||
try:
|
try:
|
||||||
o = receipt(tx_hash)
|
o = receipt(tx_hash)
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
logg.debug('h {} o {}'.format(tx_hash, o))
|
logg.debug('h {} o {}'.format(tx_hash, o))
|
||||||
if int(strip_0x(r['status'])) == 1:
|
if int(strip_0x(r['status'])) == 1:
|
||||||
tx['network_status'] = 'Confirmed'
|
tx['network_status'] = 'Confirmed'
|
||||||
@@ -583,13 +521,11 @@ class AdminApi:
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
o = balance(tx['sender'])
|
o = balance(tx['sender'])
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
tx['sender_gas_balance'] = r
|
tx['sender_gas_balance'] = r
|
||||||
|
|
||||||
o = balance(tx['recipient'])
|
o = balance(tx['recipient'])
|
||||||
t = self.proxy_do(chain_spec, o)
|
r = self.rpc.do(o)
|
||||||
r = t.get()
|
|
||||||
tx['recipient_gas_balance'] = r
|
tx['recipient_gas_balance'] = r
|
||||||
|
|
||||||
tx_unpacked = unpack(bytes.fromhex(strip_0x(tx['signed_tx'])), chain_spec)
|
tx_unpacked = unpack(bytes.fromhex(strip_0x(tx['signed_tx'])), chain_spec)
|
||||||
|
|||||||
@@ -62,18 +62,6 @@ class Api:
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def default_token(self):
|
|
||||||
s_token = celery.signature(
|
|
||||||
'cic_eth.admin.token.default_token',
|
|
||||||
[],
|
|
||||||
queue=self.queue,
|
|
||||||
)
|
|
||||||
if self.callback_param != None:
|
|
||||||
s_token.link(self.callback_success)
|
|
||||||
|
|
||||||
return s_token.apply_async()
|
|
||||||
|
|
||||||
|
|
||||||
def convert_transfer(self, from_address, to_address, target_return, minimum_return, from_token_symbol, to_token_symbol):
|
def convert_transfer(self, from_address, to_address, target_return, minimum_return, from_token_symbol, to_token_symbol):
|
||||||
"""Executes a chain of celery tasks that performs conversion between two ERC20 tokens, and transfers to a specified receipient after convert has completed.
|
"""Executes a chain of celery tasks that performs conversion between two ERC20 tokens, and transfers to a specified receipient after convert has completed.
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
from cic_eth.db.models.base import SessionBase
|
|
||||||
|
|
||||||
|
|
||||||
def health(*args, **kwargs):
|
|
||||||
session = SessionBase.create_session()
|
|
||||||
session.execute('SELECT count(*) from alembic_version')
|
|
||||||
session.close()
|
|
||||||
return True
|
|
||||||
@@ -1,48 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
from chainlib.connection import RPCConnection
|
|
||||||
from chainlib.chain import ChainSpec
|
|
||||||
from chainlib.eth.gas import balance
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from cic_eth.db.models.role import AccountRole
|
|
||||||
from cic_eth.db.models.base import SessionBase
|
|
||||||
from cic_eth.db.enum import LockEnum
|
|
||||||
from cic_eth.error import LockedError
|
|
||||||
from cic_eth.admin.ctrl import check_lock
|
|
||||||
|
|
||||||
logg = logging.getLogger().getChild(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def health(*args, **kwargs):
|
|
||||||
|
|
||||||
session = SessionBase.create_session()
|
|
||||||
|
|
||||||
config = kwargs['config']
|
|
||||||
chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
|
||||||
logg.debug('check gas balance of gas gifter for chain {}'.format(chain_spec))
|
|
||||||
|
|
||||||
try:
|
|
||||||
check_lock(None, None, LockEnum.INIT)
|
|
||||||
except LockedError:
|
|
||||||
logg.warning('INIT lock is set, skipping GAS GIFTER balance check.')
|
|
||||||
return True
|
|
||||||
|
|
||||||
gas_provider = AccountRole.get_address('GAS_GIFTER', session=session)
|
|
||||||
session.close()
|
|
||||||
|
|
||||||
rpc = RPCConnection.connect(chain_spec, 'default')
|
|
||||||
o = balance(gas_provider)
|
|
||||||
r = rpc.do(o)
|
|
||||||
try:
|
|
||||||
r = int(r, 16)
|
|
||||||
except TypeError:
|
|
||||||
r = int(r)
|
|
||||||
gas_min = int(config.get('ETH_GAS_GIFTER_MINIMUM_BALANCE'))
|
|
||||||
if r < gas_min:
|
|
||||||
logg.error('EEK! gas gifter has balance {}, below minimum {}'.format(r, gas_min))
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
@@ -1,18 +0,0 @@
|
|||||||
# external imports
|
|
||||||
import redis
|
|
||||||
import os
|
|
||||||
|
|
||||||
|
|
||||||
def health(*args, **kwargs):
|
|
||||||
r = redis.Redis(
|
|
||||||
host=kwargs['config'].get('REDIS_HOST'),
|
|
||||||
port=kwargs['config'].get('REDIS_PORT'),
|
|
||||||
db=kwargs['config'].get('REDIS_DB'),
|
|
||||||
)
|
|
||||||
try:
|
|
||||||
r.set(kwargs['unit'], os.getpid())
|
|
||||||
except redis.connection.ConnectionError:
|
|
||||||
return False
|
|
||||||
except redis.connection.ResponseError:
|
|
||||||
return False
|
|
||||||
return True
|
|
||||||
@@ -1,37 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import time
|
|
||||||
import logging
|
|
||||||
from urllib.error import URLError
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
from chainlib.connection import RPCConnection
|
|
||||||
from chainlib.eth.constant import ZERO_ADDRESS
|
|
||||||
from chainlib.eth.sign import sign_message
|
|
||||||
from chainlib.error import JSONRPCException
|
|
||||||
|
|
||||||
logg = logging.getLogger().getChild(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def health(*args, **kwargs):
|
|
||||||
blocked = True
|
|
||||||
max_attempts = 5
|
|
||||||
conn = RPCConnection.connect(kwargs['config'].get('CIC_CHAIN_SPEC'), tag='signer')
|
|
||||||
for i in range(max_attempts):
|
|
||||||
idx = i + 1
|
|
||||||
logg.debug('attempt signer connection check {}/{}'.format(idx, max_attempts))
|
|
||||||
try:
|
|
||||||
conn.do(sign_message(ZERO_ADDRESS, '0x2a'))
|
|
||||||
except FileNotFoundError:
|
|
||||||
pass
|
|
||||||
except ConnectionError:
|
|
||||||
pass
|
|
||||||
except URLError:
|
|
||||||
pass
|
|
||||||
except JSONRPCException:
|
|
||||||
logg.debug('signer connection succeeded')
|
|
||||||
return True
|
|
||||||
|
|
||||||
if idx < max_attempts:
|
|
||||||
time.sleep(0.5)
|
|
||||||
|
|
||||||
return False
|
|
||||||
@@ -74,11 +74,10 @@ class LockEnum(enum.IntEnum):
|
|||||||
QUEUE: Disable queueing new or modified transactions
|
QUEUE: Disable queueing new or modified transactions
|
||||||
"""
|
"""
|
||||||
STICKY=1
|
STICKY=1
|
||||||
INIT=2
|
CREATE=2
|
||||||
CREATE=4
|
SEND=4
|
||||||
SEND=8
|
QUEUE=8
|
||||||
QUEUE=16
|
QUERY=16
|
||||||
QUERY=32
|
|
||||||
ALL=int(0xfffffffffffffffe)
|
ALL=int(0xfffffffffffffffe)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -5,11 +5,8 @@ Revises: 1f1b3b641d08
|
|||||||
Create Date: 2021-04-02 18:41:20.864265
|
Create Date: 2021-04-02 18:41:20.864265
|
||||||
|
|
||||||
"""
|
"""
|
||||||
import datetime
|
|
||||||
from alembic import op
|
from alembic import op
|
||||||
import sqlalchemy as sa
|
import sqlalchemy as sa
|
||||||
from chainlib.eth.constant import ZERO_ADDRESS
|
|
||||||
from cic_eth.db.enum import LockEnum
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
# revision identifiers, used by Alembic.
|
||||||
@@ -26,11 +23,10 @@ def upgrade():
|
|||||||
sa.Column("address", sa.String(42), nullable=True),
|
sa.Column("address", sa.String(42), nullable=True),
|
||||||
sa.Column('blockchain', sa.String),
|
sa.Column('blockchain', sa.String),
|
||||||
sa.Column("flags", sa.BIGINT(), nullable=False, default=0),
|
sa.Column("flags", sa.BIGINT(), nullable=False, default=0),
|
||||||
sa.Column("date_created", sa.DateTime, nullable=False, default=datetime.datetime.utcnow),
|
sa.Column("date_created", sa.DateTime, nullable=False),
|
||||||
sa.Column("otx_id", sa.Integer, sa.ForeignKey('otx.id'), nullable=True),
|
sa.Column("otx_id", sa.Integer, sa.ForeignKey('otx.id'), nullable=True),
|
||||||
)
|
)
|
||||||
op.create_index('idx_chain_address', 'lock', ['blockchain', 'address'], unique=True)
|
op.create_index('idx_chain_address', 'lock', ['blockchain', 'address'], unique=True)
|
||||||
op.execute("INSERT INTO lock (address, date_created, blockchain, flags) VALUES('{}', '{}', '::', {})".format(ZERO_ADDRESS, datetime.datetime.utcnow(), LockEnum.INIT | LockEnum.SEND | LockEnum.QUEUE))
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
def downgrade():
|
||||||
|
|||||||
@@ -48,8 +48,6 @@ class RoleMissingError(Exception):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class IntegrityError(Exception):
|
class IntegrityError(Exception):
|
||||||
"""Exception raised to signal irregularities with deduplication and ordering of tasks
|
"""Exception raised to signal irregularities with deduplication and ordering of tasks
|
||||||
|
|
||||||
@@ -64,19 +62,15 @@ class LockedError(Exception):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class SeppukuError(Exception):
|
class SignerError(Exception):
|
||||||
"""Exception base class for all errors that should cause system shutdown
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
|
|
||||||
class SignerError(SeppukuError):
|
|
||||||
"""Exception raised when signer is unavailable or generates an error
|
"""Exception raised when signer is unavailable or generates an error
|
||||||
|
|
||||||
"""
|
"""
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class RoleAgencyError(SeppukuError):
|
class EthError(Exception):
|
||||||
"""Exception raise when a role cannot perform its function. This is a critical exception
|
"""Exception raised when unspecified error from evm node is encountered
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
pass
|
||||||
|
|||||||
@@ -4,10 +4,10 @@ import logging
|
|||||||
# external imports
|
# external imports
|
||||||
import celery
|
import celery
|
||||||
from erc20_single_shot_faucet import SingleShotFaucet as Faucet
|
from erc20_single_shot_faucet import SingleShotFaucet as Faucet
|
||||||
|
from chainlib.eth.constant import ZERO_ADDRESS
|
||||||
from hexathon import (
|
from hexathon import (
|
||||||
strip_0x,
|
strip_0x,
|
||||||
)
|
)
|
||||||
from chainlib.eth.constant import ZERO_ADDRESS
|
|
||||||
from chainlib.connection import RPCConnection
|
from chainlib.connection import RPCConnection
|
||||||
from chainlib.eth.sign import (
|
from chainlib.eth.sign import (
|
||||||
new_account,
|
new_account,
|
||||||
@@ -19,7 +19,6 @@ from chainlib.eth.tx import (
|
|||||||
unpack,
|
unpack,
|
||||||
)
|
)
|
||||||
from chainlib.chain import ChainSpec
|
from chainlib.chain import ChainSpec
|
||||||
from chainlib.error import JSONRPCException
|
|
||||||
from eth_accounts_index import AccountRegistry
|
from eth_accounts_index import AccountRegistry
|
||||||
from sarafu_faucet import MinterFaucet as Faucet
|
from sarafu_faucet import MinterFaucet as Faucet
|
||||||
from chainqueue.db.models.tx import TxCache
|
from chainqueue.db.models.tx import TxCache
|
||||||
@@ -71,18 +70,11 @@ def create(self, password, chain_spec_dict):
|
|||||||
a = None
|
a = None
|
||||||
conn = RPCConnection.connect(chain_spec, 'signer')
|
conn = RPCConnection.connect(chain_spec, 'signer')
|
||||||
o = new_account()
|
o = new_account()
|
||||||
try:
|
a = conn.do(o)
|
||||||
a = conn.do(o)
|
|
||||||
except ConnectionError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
conn.disconnect()
|
conn.disconnect()
|
||||||
|
|
||||||
# TODO: It seems infeasible that a can be None in any case, verify
|
|
||||||
if a == None:
|
if a == None:
|
||||||
raise SignerError('create account')
|
raise SignerError('create account')
|
||||||
|
|
||||||
logg.debug('created account {}'.format(a))
|
logg.debug('created account {}'.format(a))
|
||||||
|
|
||||||
# Initialize nonce provider record for account
|
# Initialize nonce provider record for account
|
||||||
@@ -227,22 +219,21 @@ def have(self, account, chain_spec_dict):
|
|||||||
"""
|
"""
|
||||||
chain_spec = ChainSpec.from_dict(chain_spec_dict)
|
chain_spec = ChainSpec.from_dict(chain_spec_dict)
|
||||||
o = sign_message(account, '0x2a')
|
o = sign_message(account, '0x2a')
|
||||||
conn = RPCConnection.connect(chain_spec, 'signer')
|
try:
|
||||||
|
conn = RPCConnection.connect(chain_spec, 'signer')
|
||||||
|
except Exception as e:
|
||||||
|
logg.debug('cannot sign with {}: {}'.format(account, e))
|
||||||
|
return None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
conn.do(o)
|
conn.do(o)
|
||||||
except ConnectionError as e:
|
conn.disconnect()
|
||||||
raise SignerError(e)
|
return account
|
||||||
except FileNotFoundError as e:
|
except Exception as e:
|
||||||
raise SignerError(e)
|
|
||||||
except JSONRPCException as e:
|
|
||||||
logg.debug('cannot sign with {}: {}'.format(account, e))
|
logg.debug('cannot sign with {}: {}'.format(account, e))
|
||||||
conn.disconnect()
|
conn.disconnect()
|
||||||
return None
|
return None
|
||||||
|
|
||||||
conn.disconnect()
|
|
||||||
return account
|
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task(bind=True, base=CriticalSQLAlchemyTask)
|
@celery_app.task(bind=True, base=CriticalSQLAlchemyTask)
|
||||||
def set_role(self, tag, address, chain_spec_dict):
|
def set_role(self, tag, address, chain_spec_dict):
|
||||||
|
|||||||
@@ -108,13 +108,7 @@ def transfer(self, tokens, holder_address, receiver_address, value, chain_spec_d
|
|||||||
nonce_oracle = CustodialTaskNonceOracle(holder_address, self.request.root_id, session=session)
|
nonce_oracle = CustodialTaskNonceOracle(holder_address, self.request.root_id, session=session)
|
||||||
gas_oracle = self.create_gas_oracle(rpc, MaxGasOracle.gas)
|
gas_oracle = self.create_gas_oracle(rpc, MaxGasOracle.gas)
|
||||||
c = ERC20(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle, nonce_oracle=nonce_oracle)
|
c = ERC20(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle, nonce_oracle=nonce_oracle)
|
||||||
try:
|
(tx_hash_hex, tx_signed_raw_hex) = c.transfer(t['address'], holder_address, receiver_address, value, tx_format=TxFormat.RLP_SIGNED)
|
||||||
(tx_hash_hex, tx_signed_raw_hex) = c.transfer(t['address'], holder_address, receiver_address, value, tx_format=TxFormat.RLP_SIGNED)
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
except ConnectionError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
|
|
||||||
|
|
||||||
rpc_signer.disconnect()
|
rpc_signer.disconnect()
|
||||||
rpc.disconnect()
|
rpc.disconnect()
|
||||||
@@ -177,12 +171,7 @@ def approve(self, tokens, holder_address, spender_address, value, chain_spec_dic
|
|||||||
nonce_oracle = CustodialTaskNonceOracle(holder_address, self.request.root_id, session=session)
|
nonce_oracle = CustodialTaskNonceOracle(holder_address, self.request.root_id, session=session)
|
||||||
gas_oracle = self.create_gas_oracle(rpc, MaxGasOracle.gas)
|
gas_oracle = self.create_gas_oracle(rpc, MaxGasOracle.gas)
|
||||||
c = ERC20(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle, nonce_oracle=nonce_oracle)
|
c = ERC20(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle, nonce_oracle=nonce_oracle)
|
||||||
try:
|
(tx_hash_hex, tx_signed_raw_hex) = c.approve(t['address'], holder_address, spender_address, value, tx_format=TxFormat.RLP_SIGNED)
|
||||||
(tx_hash_hex, tx_signed_raw_hex) = c.approve(t['address'], holder_address, spender_address, value, tx_format=TxFormat.RLP_SIGNED)
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
except ConnectionError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
|
|
||||||
rpc_signer.disconnect()
|
rpc_signer.disconnect()
|
||||||
rpc.disconnect()
|
rpc.disconnect()
|
||||||
|
|||||||
@@ -328,12 +328,7 @@ def refill_gas(self, recipient_address, chain_spec_dict):
|
|||||||
|
|
||||||
# build and add transaction
|
# build and add transaction
|
||||||
logg.debug('tx send gas amount {} from provider {} to {}'.format(refill_amount, gas_provider, recipient_address))
|
logg.debug('tx send gas amount {} from provider {} to {}'.format(refill_amount, gas_provider, recipient_address))
|
||||||
try:
|
(tx_hash_hex, tx_signed_raw_hex) = c.create(gas_provider, recipient_address, refill_amount, tx_format=TxFormat.RLP_SIGNED)
|
||||||
(tx_hash_hex, tx_signed_raw_hex) = c.create(gas_provider, recipient_address, refill_amount, tx_format=TxFormat.RLP_SIGNED)
|
|
||||||
except ConnectionError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
logg.debug('adding queue refill gas tx {}'.format(tx_hash_hex))
|
logg.debug('adding queue refill gas tx {}'.format(tx_hash_hex))
|
||||||
cache_task = 'cic_eth.eth.gas.cache_gas_data'
|
cache_task = 'cic_eth.eth.gas.cache_gas_data'
|
||||||
register_tx(tx_hash_hex, tx_signed_raw_hex, chain_spec, queue, cache_task=cache_task, session=session)
|
register_tx(tx_hash_hex, tx_signed_raw_hex, chain_spec, queue, cache_task=cache_task, session=session)
|
||||||
@@ -409,12 +404,7 @@ def resend_with_higher_gas(self, txold_hash_hex, chain_spec_dict, gas=None, defa
|
|||||||
c = TxFactory(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle)
|
c = TxFactory(chain_spec, signer=rpc_signer, gas_oracle=gas_oracle)
|
||||||
logg.debug('change gas price from old {} to new {} for tx {}'.format(tx['gasPrice'], new_gas_price, tx))
|
logg.debug('change gas price from old {} to new {} for tx {}'.format(tx['gasPrice'], new_gas_price, tx))
|
||||||
tx['gasPrice'] = new_gas_price
|
tx['gasPrice'] = new_gas_price
|
||||||
try:
|
(tx_hash_hex, tx_signed_raw_hex) = c.build_raw(tx)
|
||||||
(tx_hash_hex, tx_signed_raw_hex) = c.build_raw(tx)
|
|
||||||
except ConnectionError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
raise SignerError(e)
|
|
||||||
queue_create(
|
queue_create(
|
||||||
chain_spec,
|
chain_spec,
|
||||||
tx['nonce'],
|
tx['nonce'],
|
||||||
|
|||||||
@@ -114,7 +114,7 @@ def list_tx_by_bloom(self, bloomspec, address, chain_spec_dict):
|
|||||||
|
|
||||||
# TODO: pass through registry to validate declarator entry of token
|
# TODO: pass through registry to validate declarator entry of token
|
||||||
#token = registry.by_address(tx['to'], sender_address=self.call_address)
|
#token = registry.by_address(tx['to'], sender_address=self.call_address)
|
||||||
token = ERC20Token(chain_spec, rpc, tx['to'])
|
token = ERC20Token(rpc, tx['to'])
|
||||||
token_symbol = token.symbol
|
token_symbol = token.symbol
|
||||||
token_decimals = token.decimals
|
token_decimals = token.decimals
|
||||||
times = tx_times(tx['hash'], chain_spec)
|
times = tx_times(tx['hash'], chain_spec)
|
||||||
|
|||||||
@@ -12,7 +12,6 @@ from chainqueue.error import NotLocalTxError
|
|||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_eth.task import CriticalSQLAlchemyAndWeb3Task
|
from cic_eth.task import CriticalSQLAlchemyAndWeb3Task
|
||||||
from cic_eth.db.models.base import SessionBase
|
|
||||||
|
|
||||||
celery_app = celery.current_app
|
celery_app = celery.current_app
|
||||||
|
|
||||||
|
|||||||
@@ -29,5 +29,5 @@ def connect(rpc, chain_spec, registry_address):
|
|||||||
CICRegistry.address = registry_address
|
CICRegistry.address = registry_address
|
||||||
registry = CICRegistry(chain_spec, rpc)
|
registry = CICRegistry(chain_spec, rpc)
|
||||||
registry_address = registry.by_name('ContractRegistry')
|
registry_address = registry.by_name('ContractRegistry')
|
||||||
|
|
||||||
return registry
|
return registry
|
||||||
|
|
||||||
|
|||||||
@@ -23,6 +23,7 @@ default_config_dir = os.environ.get('CONFINI_DIR', '/usr/local/etc/cic')
|
|||||||
|
|
||||||
argparser = argparse.ArgumentParser()
|
argparser = argparse.ArgumentParser()
|
||||||
argparser.add_argument('-p', '--provider', dest='p', default='http://localhost:8545', type=str, help='Web3 provider url (http only)')
|
argparser.add_argument('-p', '--provider', dest='p', default='http://localhost:8545', type=str, help='Web3 provider url (http only)')
|
||||||
|
argparser.add_argument('-r', '--registry-address', type=str, help='CIC registry address')
|
||||||
argparser.add_argument('-f', '--format', dest='f', default=default_format, type=str, help='Output format')
|
argparser.add_argument('-f', '--format', dest='f', default=default_format, type=str, help='Output format')
|
||||||
argparser.add_argument('-c', type=str, default=default_config_dir, help='config root to use')
|
argparser.add_argument('-c', type=str, default=default_config_dir, help='config root to use')
|
||||||
argparser.add_argument('-i', '--chain-spec', dest='i', type=str, help='chain spec')
|
argparser.add_argument('-i', '--chain-spec', dest='i', type=str, help='chain spec')
|
||||||
@@ -58,7 +59,6 @@ args_override = {
|
|||||||
'CIC_CHAIN_SPEC': getattr(args, 'i'),
|
'CIC_CHAIN_SPEC': getattr(args, 'i'),
|
||||||
}
|
}
|
||||||
# override args
|
# override args
|
||||||
config.dict_override(args_override, 'cli')
|
|
||||||
config.censor('PASSWORD', 'DATABASE')
|
config.censor('PASSWORD', 'DATABASE')
|
||||||
config.censor('PASSWORD', 'SSL')
|
config.censor('PASSWORD', 'SSL')
|
||||||
logg.debug('config loaded from {}:\n{}'.format(config_dir, config))
|
logg.debug('config loaded from {}:\n{}'.format(config_dir, config))
|
||||||
@@ -67,9 +67,7 @@ celery_app = celery.Celery(broker=config.get('CELERY_BROKER_URL'), backend=confi
|
|||||||
|
|
||||||
queue = args.q
|
queue = args.q
|
||||||
|
|
||||||
chain_spec = None
|
chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
||||||
if config.get('CIC_CHAIN_SPEC') != None and config.get('CIC_CHAIN_SPEC') != '::':
|
|
||||||
chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
|
||||||
admin_api = AdminApi(None)
|
admin_api = AdminApi(None)
|
||||||
|
|
||||||
|
|
||||||
@@ -84,9 +82,6 @@ def lock_names_to_flag(s):
|
|||||||
|
|
||||||
# TODO: move each command to submodule
|
# TODO: move each command to submodule
|
||||||
def main():
|
def main():
|
||||||
chain_spec_dict = None
|
|
||||||
if chain_spec != None:
|
|
||||||
chain_spec_dict = chain_spec.asdict()
|
|
||||||
if args.command == 'unlock':
|
if args.command == 'unlock':
|
||||||
flags = lock_names_to_flag(args.flags)
|
flags = lock_names_to_flag(args.flags)
|
||||||
if not is_checksum_address(args.address):
|
if not is_checksum_address(args.address):
|
||||||
@@ -96,7 +91,7 @@ def main():
|
|||||||
'cic_eth.admin.ctrl.unlock',
|
'cic_eth.admin.ctrl.unlock',
|
||||||
[
|
[
|
||||||
None,
|
None,
|
||||||
chain_spec_dict,
|
chain_spec.asdict(),
|
||||||
args.address,
|
args.address,
|
||||||
flags,
|
flags,
|
||||||
],
|
],
|
||||||
@@ -115,7 +110,7 @@ def main():
|
|||||||
'cic_eth.admin.ctrl.lock',
|
'cic_eth.admin.ctrl.lock',
|
||||||
[
|
[
|
||||||
None,
|
None,
|
||||||
chain_spec_dict,
|
chain_spec.asdict(),
|
||||||
args.address,
|
args.address,
|
||||||
flags,
|
flags,
|
||||||
],
|
],
|
||||||
|
|||||||
@@ -11,19 +11,10 @@ import websocket
|
|||||||
# external imports
|
# external imports
|
||||||
import celery
|
import celery
|
||||||
import confini
|
import confini
|
||||||
from chainlib.connection import (
|
from chainlib.connection import RPCConnection
|
||||||
RPCConnection,
|
from chainlib.eth.connection import EthUnixSignerConnection
|
||||||
ConnType,
|
|
||||||
)
|
|
||||||
from chainlib.eth.connection import (
|
|
||||||
EthUnixSignerConnection,
|
|
||||||
EthHTTPSignerConnection,
|
|
||||||
)
|
|
||||||
from chainlib.chain import ChainSpec
|
from chainlib.chain import ChainSpec
|
||||||
from chainqueue.db.models.otx import Otx
|
from chainqueue.db.models.otx import Otx
|
||||||
from cic_eth_registry.error import UnknownContractError
|
|
||||||
import liveness.linux
|
|
||||||
|
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_eth.eth import (
|
from cic_eth.eth import (
|
||||||
@@ -36,7 +27,6 @@ from cic_eth.eth import (
|
|||||||
from cic_eth.admin import (
|
from cic_eth.admin import (
|
||||||
debug,
|
debug,
|
||||||
ctrl,
|
ctrl,
|
||||||
token
|
|
||||||
)
|
)
|
||||||
from cic_eth.queue import (
|
from cic_eth.queue import (
|
||||||
query,
|
query,
|
||||||
@@ -61,8 +51,6 @@ from cic_eth.registry import (
|
|||||||
connect_declarator,
|
connect_declarator,
|
||||||
connect_token_registry,
|
connect_token_registry,
|
||||||
)
|
)
|
||||||
from cic_eth.task import BaseTask
|
|
||||||
|
|
||||||
|
|
||||||
logging.basicConfig(level=logging.WARNING)
|
logging.basicConfig(level=logging.WARNING)
|
||||||
logg = logging.getLogger()
|
logg = logging.getLogger()
|
||||||
@@ -74,7 +62,6 @@ argparser.add_argument('-p', '--provider', dest='p', type=str, help='rpc provide
|
|||||||
argparser.add_argument('-c', type=str, default=config_dir, help='config file')
|
argparser.add_argument('-c', type=str, default=config_dir, help='config file')
|
||||||
argparser.add_argument('-q', type=str, default='cic-eth', help='queue name for worker tasks')
|
argparser.add_argument('-q', type=str, default='cic-eth', help='queue name for worker tasks')
|
||||||
argparser.add_argument('-r', type=str, help='CIC registry address')
|
argparser.add_argument('-r', type=str, help='CIC registry address')
|
||||||
argparser.add_argument('--default-token-symbol', dest='default_token_symbol', type=str, help='Symbol of default token to use')
|
|
||||||
argparser.add_argument('--abi-dir', dest='abi_dir', type=str, help='Directory containing bytecode and abi')
|
argparser.add_argument('--abi-dir', dest='abi_dir', type=str, help='Directory containing bytecode and abi')
|
||||||
argparser.add_argument('--trace-queue-status', default=None, dest='trace_queue_status', action='store_true', help='set to perist all queue entry status changes to storage')
|
argparser.add_argument('--trace-queue-status', default=None, dest='trace_queue_status', action='store_true', help='set to perist all queue entry status changes to storage')
|
||||||
argparser.add_argument('-i', '--chain-spec', dest='i', type=str, help='chain spec')
|
argparser.add_argument('-i', '--chain-spec', dest='i', type=str, help='chain spec')
|
||||||
@@ -94,7 +81,6 @@ config.process()
|
|||||||
args_override = {
|
args_override = {
|
||||||
'CIC_CHAIN_SPEC': getattr(args, 'i'),
|
'CIC_CHAIN_SPEC': getattr(args, 'i'),
|
||||||
'CIC_REGISTRY_ADDRESS': getattr(args, 'r'),
|
'CIC_REGISTRY_ADDRESS': getattr(args, 'r'),
|
||||||
'CIC_DEFAULT_TOKEN_SYMBOL': getattr(args, 'default_token_symbol'),
|
|
||||||
'ETH_PROVIDER': getattr(args, 'p'),
|
'ETH_PROVIDER': getattr(args, 'p'),
|
||||||
'TASKS_TRACE_QUEUE_STATUS': getattr(args, 'trace_queue_status'),
|
'TASKS_TRACE_QUEUE_STATUS': getattr(args, 'trace_queue_status'),
|
||||||
}
|
}
|
||||||
@@ -104,15 +90,14 @@ config.censor('PASSWORD', 'DATABASE')
|
|||||||
config.censor('PASSWORD', 'SSL')
|
config.censor('PASSWORD', 'SSL')
|
||||||
logg.debug('config loaded from {}:\n{}'.format(args.c, config))
|
logg.debug('config loaded from {}:\n{}'.format(args.c, config))
|
||||||
|
|
||||||
health_modules = config.get('CIC_HEALTH_MODULES', [])
|
|
||||||
if len(health_modules) != 0:
|
|
||||||
health_modules = health_modules.split(',')
|
|
||||||
logg.debug('health mods {}'.format(health_modules))
|
|
||||||
|
|
||||||
# connect to database
|
# connect to database
|
||||||
dsn = dsn_from_config(config)
|
dsn = dsn_from_config(config)
|
||||||
SessionBase.connect(dsn, pool_size=int(config.get('DATABASE_POOL_SIZE')), debug=config.true('DATABASE_DEBUG'))
|
SessionBase.connect(dsn, pool_size=int(config.get('DATABASE_POOL_SIZE')), debug=config.true('DATABASE_DEBUG'))
|
||||||
|
|
||||||
|
# verify database connection with minimal sanity query
|
||||||
|
session = SessionBase.create_session()
|
||||||
|
session.execute('select version_num from alembic_version')
|
||||||
|
session.close()
|
||||||
|
|
||||||
# set up celery
|
# set up celery
|
||||||
current_app = celery.Celery(__name__)
|
current_app = celery.Celery(__name__)
|
||||||
@@ -149,18 +134,11 @@ else:
|
|||||||
})
|
})
|
||||||
|
|
||||||
chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
||||||
RPCConnection.register_constructor(ConnType.UNIX, EthUnixSignerConnection, 'signer')
|
|
||||||
RPCConnection.register_constructor(ConnType.HTTP, EthHTTPSignerConnection, 'signer')
|
|
||||||
RPCConnection.register_constructor(ConnType.HTTP_SSL, EthHTTPSignerConnection, 'signer')
|
|
||||||
RPCConnection.register_location(config.get('ETH_PROVIDER'), chain_spec, 'default')
|
RPCConnection.register_location(config.get('ETH_PROVIDER'), chain_spec, 'default')
|
||||||
RPCConnection.register_location(config.get('SIGNER_SOCKET_PATH'), chain_spec, 'signer')
|
RPCConnection.register_location(config.get('SIGNER_SOCKET_PATH'), chain_spec, 'signer', constructor=EthUnixSignerConnection)
|
||||||
|
|
||||||
Otx.tracing = config.true('TASKS_TRACE_QUEUE_STATUS')
|
Otx.tracing = config.true('TASKS_TRACE_QUEUE_STATUS')
|
||||||
|
|
||||||
#import cic_eth.checks.gas
|
|
||||||
#if not cic_eth.checks.gas.health(config=config):
|
|
||||||
# raise RuntimeError()
|
|
||||||
liveness.linux.load(health_modules, rundir=config.get('CIC_RUN_DIR'), config=config, unit='cic-eth-tasker')
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
argv = ['worker']
|
argv = ['worker']
|
||||||
@@ -184,11 +162,7 @@ def main():
|
|||||||
|
|
||||||
rpc = RPCConnection.connect(chain_spec, 'default')
|
rpc = RPCConnection.connect(chain_spec, 'default')
|
||||||
|
|
||||||
try:
|
connect_registry(rpc, chain_spec, config.get('CIC_REGISTRY_ADDRESS'))
|
||||||
registry = connect_registry(rpc, chain_spec, config.get('CIC_REGISTRY_ADDRESS'))
|
|
||||||
except UnknownContractError as e:
|
|
||||||
logg.exception('Registry contract connection failed for {}: {}'.format(config.get('CIC_REGISTRY_ADDRESS'), e))
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
||||||
if trusted_addresses_src == None:
|
if trusted_addresses_src == None:
|
||||||
@@ -197,18 +171,10 @@ def main():
|
|||||||
trusted_addresses = trusted_addresses_src.split(',')
|
trusted_addresses = trusted_addresses_src.split(',')
|
||||||
for address in trusted_addresses:
|
for address in trusted_addresses:
|
||||||
logg.info('using trusted address {}'.format(address))
|
logg.info('using trusted address {}'.format(address))
|
||||||
|
|
||||||
connect_declarator(rpc, chain_spec, trusted_addresses)
|
connect_declarator(rpc, chain_spec, trusted_addresses)
|
||||||
connect_token_registry(rpc, chain_spec)
|
connect_token_registry(rpc, chain_spec)
|
||||||
|
|
||||||
BaseTask.default_token_symbol = config.get('CIC_DEFAULT_TOKEN_SYMBOL')
|
|
||||||
BaseTask.default_token_address = registry.by_name(BaseTask.default_token_symbol)
|
|
||||||
BaseTask.run_dir = config.get('CIC_RUN_DIR')
|
|
||||||
logg.info('default token set to {} {}'.format(BaseTask.default_token_symbol, BaseTask.default_token_address))
|
|
||||||
|
|
||||||
liveness.linux.set(rundir=config.get('CIC_RUN_DIR'))
|
|
||||||
current_app.worker_main(argv)
|
current_app.worker_main(argv)
|
||||||
liveness.linux.reset(rundir=config.get('CIC_RUN_DIR'))
|
|
||||||
|
|
||||||
|
|
||||||
@celery.signals.eventlet_pool_postshutdown.connect
|
@celery.signals.eventlet_pool_postshutdown.connect
|
||||||
|
|||||||
@@ -1,65 +0,0 @@
|
|||||||
#!python3
|
|
||||||
|
|
||||||
# SPDX-License-Identifier: GPL-3.0-or-later
|
|
||||||
|
|
||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
import argparse
|
|
||||||
import os
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
import confini
|
|
||||||
import celery
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from cic_eth.api import Api
|
|
||||||
|
|
||||||
logging.basicConfig(level=logging.WARNING)
|
|
||||||
logg = logging.getLogger()
|
|
||||||
|
|
||||||
default_format = 'terminal'
|
|
||||||
default_config_dir = os.environ.get('CONFINI_DIR', '/usr/local/etc/cic')
|
|
||||||
|
|
||||||
|
|
||||||
argparser = argparse.ArgumentParser()
|
|
||||||
argparser.add_argument('-i', '--chain-spec', dest='i', type=str, help='chain spec')
|
|
||||||
argparser.add_argument('-c', type=str, default=default_config_dir, help='config root to use')
|
|
||||||
argparser.add_argument('-q', type=str, default='cic-eth', help='celery queue to submit transaction tasks to')
|
|
||||||
argparser.add_argument('--env-prefix', default=os.environ.get('CONFINI_ENV_PREFIX'), dest='env_prefix', type=str, help='environment prefix for variables to overwrite configuration')
|
|
||||||
argparser.add_argument('-v', action='store_true', help='Be verbose')
|
|
||||||
argparser.add_argument('-vv', help='be more verbose', action='store_true')
|
|
||||||
args = argparser.parse_args()
|
|
||||||
|
|
||||||
if args.v == True:
|
|
||||||
logging.getLogger().setLevel(logging.INFO)
|
|
||||||
elif args.vv == True:
|
|
||||||
logging.getLogger().setLevel(logging.DEBUG)
|
|
||||||
|
|
||||||
config_dir = os.path.join(args.c)
|
|
||||||
os.makedirs(config_dir, 0o777, True)
|
|
||||||
config = confini.Config(config_dir, args.env_prefix)
|
|
||||||
config.process()
|
|
||||||
args_override = {
|
|
||||||
'CIC_CHAIN_SPEC': getattr(args, 'i'),
|
|
||||||
}
|
|
||||||
config.dict_override(args_override, 'cli args')
|
|
||||||
config.censor('PASSWORD', 'DATABASE')
|
|
||||||
config.censor('PASSWORD', 'SSL')
|
|
||||||
logg.debug('config loaded from {}:\n{}'.format(config_dir, config))
|
|
||||||
|
|
||||||
|
|
||||||
celery_app = celery.Celery(broker=config.get('CELERY_BROKER_URL'), backend=config.get('CELERY_RESULT_URL'))
|
|
||||||
|
|
||||||
queue = args.q
|
|
||||||
|
|
||||||
api = Api(config.get('CIC_CHAIN_SPEC'), queue=queue)
|
|
||||||
|
|
||||||
def main():
|
|
||||||
t = api.default_token()
|
|
||||||
token_info = t.get()
|
|
||||||
print('Default token symbol: {}'.format(token_info['symbol']))
|
|
||||||
print('Default token address: {}'.format(token_info['address']))
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
|
||||||
main()
|
|
||||||
@@ -85,6 +85,9 @@ def main():
|
|||||||
callback_queue=args.q,
|
callback_queue=args.q,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
#register = not args.no_register
|
||||||
|
#logg.debug('register {}'.format(register))
|
||||||
|
#t = api.create_account(register=register)
|
||||||
t = api.transfer(config.get('_SENDER'), config.get('_RECIPIENT'), config.get('_VALUE'), config.get('_SYMBOL'))
|
t = api.transfer(config.get('_SENDER'), config.get('_RECIPIENT'), config.get('_VALUE'), config.get('_SYMBOL'))
|
||||||
|
|
||||||
ps.get_message()
|
ps.get_message()
|
||||||
|
|||||||
@@ -81,14 +81,10 @@ chain_spec = ChainSpec.from_chain_str(config.get('CIC_CHAIN_SPEC'))
|
|||||||
|
|
||||||
rpc = EthHTTPConnection(args.p)
|
rpc = EthHTTPConnection(args.p)
|
||||||
|
|
||||||
#registry_address = config.get('CIC_REGISTRY_ADDRESS')
|
registry_address = config.get('CIC_REGISTRY_ADDRESS')
|
||||||
|
|
||||||
admin_api = AdminApi(rpc)
|
admin_api = AdminApi(rpc)
|
||||||
|
|
||||||
t = admin_api.registry()
|
|
||||||
registry_address = t.get()
|
|
||||||
logg.info('got registry address from task pool: {}'.format(registry_address))
|
|
||||||
|
|
||||||
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
||||||
if trusted_addresses_src == None:
|
if trusted_addresses_src == None:
|
||||||
logg.critical('At least one trusted address must be declared in CIC_TRUST_ADDRESS')
|
logg.critical('At least one trusted address must be declared in CIC_TRUST_ADDRESS')
|
||||||
@@ -155,16 +151,14 @@ def main():
|
|||||||
txs = []
|
txs = []
|
||||||
renderer = render_tx
|
renderer = render_tx
|
||||||
if len(config.get('_QUERY')) > 66:
|
if len(config.get('_QUERY')) > 66:
|
||||||
#registry = connect_registry(rpc, chain_spec, registry_address)
|
registry = connect_registry(rpc, chain_spec, registry_address)
|
||||||
#admin_api.tx(chain_spec, tx_raw=config.get('_QUERY'), registry=registry, renderer=renderer)
|
admin_api.tx(chain_spec, tx_raw=config.get('_QUERY'), registry=registry, renderer=renderer)
|
||||||
admin_api.tx(chain_spec, tx_raw=config.get('_QUERY'), renderer=renderer)
|
|
||||||
elif len(config.get('_QUERY')) > 42:
|
elif len(config.get('_QUERY')) > 42:
|
||||||
#registry = connect_registry(rpc, chain_spec, registry_address)
|
registry = connect_registry(rpc, chain_spec, registry_address)
|
||||||
#admin_api.tx(chain_spec, tx_hash=config.get('_QUERY'), registry=registry, renderer=renderer)
|
admin_api.tx(chain_spec, tx_hash=config.get('_QUERY'), registry=registry, renderer=renderer)
|
||||||
admin_api.tx(chain_spec, tx_hash=config.get('_QUERY'), renderer=renderer)
|
|
||||||
|
|
||||||
elif len(config.get('_QUERY')) == 42:
|
elif len(config.get('_QUERY')) == 42:
|
||||||
#registry = connect_registry(rpc, chain_spec, registry_address)
|
registry = connect_registry(rpc, chain_spec, registry_address)
|
||||||
txs = admin_api.account(chain_spec, config.get('_QUERY'), include_recipient=False, renderer=render_account)
|
txs = admin_api.account(chain_spec, config.get('_QUERY'), include_recipient=False, renderer=render_account)
|
||||||
renderer = render_account
|
renderer = render_account
|
||||||
elif len(config.get('_QUERY')) >= 4 and config.get('_QUERY')[:4] == 'lock':
|
elif len(config.get('_QUERY')) >= 4 and config.get('_QUERY')[:4] == 'lock':
|
||||||
|
|||||||
@@ -7,20 +7,18 @@ import uuid
|
|||||||
# external imports
|
# external imports
|
||||||
import celery
|
import celery
|
||||||
import sqlalchemy
|
import sqlalchemy
|
||||||
from chainlib.chain import ChainSpec
|
|
||||||
from chainlib.connection import RPCConnection
|
|
||||||
from chainlib.eth.constant import ZERO_ADDRESS
|
from chainlib.eth.constant import ZERO_ADDRESS
|
||||||
from chainlib.eth.nonce import RPCNonceOracle
|
from chainlib.eth.nonce import RPCNonceOracle
|
||||||
from chainlib.eth.gas import RPCGasOracle
|
from chainlib.eth.gas import RPCGasOracle
|
||||||
from cic_eth_registry import CICRegistry
|
|
||||||
from cic_eth_registry.error import UnknownContractError
|
|
||||||
import liveness.linux
|
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_eth.error import SeppukuError
|
from cic_eth.error import (
|
||||||
|
SignerError,
|
||||||
|
EthError,
|
||||||
|
)
|
||||||
from cic_eth.db.models.base import SessionBase
|
from cic_eth.db.models.base import SessionBase
|
||||||
|
|
||||||
logg = logging.getLogger().getChild(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
celery_app = celery.current_app
|
celery_app = celery.current_app
|
||||||
|
|
||||||
@@ -31,9 +29,6 @@ class BaseTask(celery.Task):
|
|||||||
call_address = ZERO_ADDRESS
|
call_address = ZERO_ADDRESS
|
||||||
create_nonce_oracle = RPCNonceOracle
|
create_nonce_oracle = RPCNonceOracle
|
||||||
create_gas_oracle = RPCGasOracle
|
create_gas_oracle = RPCGasOracle
|
||||||
default_token_address = None
|
|
||||||
default_token_symbol = None
|
|
||||||
run_dir = '/run'
|
|
||||||
|
|
||||||
def create_session(self):
|
def create_session(self):
|
||||||
return BaseTask.session_func()
|
return BaseTask.session_func()
|
||||||
@@ -43,19 +38,6 @@ class BaseTask(celery.Task):
|
|||||||
logg.debug('task {} root uuid {}'.format(self.__class__.__name__, self.request.root_id))
|
logg.debug('task {} root uuid {}'.format(self.__class__.__name__, self.request.root_id))
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
def on_failure(self, exc, task_id, args, kwargs, einfo):
|
|
||||||
if isinstance(exc, SeppukuError):
|
|
||||||
liveness.linux.reset(rundir=self.run_dir)
|
|
||||||
logg.critical(einfo)
|
|
||||||
msg = 'received critical exception {}, calling shutdown'.format(str(exc))
|
|
||||||
s = celery.signature(
|
|
||||||
'cic_eth.admin.ctrl.shutdown',
|
|
||||||
[msg],
|
|
||||||
queue=self.request.delivery_info.get('routing_key'),
|
|
||||||
)
|
|
||||||
s.apply_async()
|
|
||||||
|
|
||||||
|
|
||||||
class CriticalTask(BaseTask):
|
class CriticalTask(BaseTask):
|
||||||
retry_jitter = True
|
retry_jitter = True
|
||||||
@@ -85,6 +67,7 @@ class CriticalSQLAlchemyAndWeb3Task(CriticalTask):
|
|||||||
sqlalchemy.exc.TimeoutError,
|
sqlalchemy.exc.TimeoutError,
|
||||||
requests.exceptions.ConnectionError,
|
requests.exceptions.ConnectionError,
|
||||||
sqlalchemy.exc.ResourceClosedError,
|
sqlalchemy.exc.ResourceClosedError,
|
||||||
|
EthError,
|
||||||
)
|
)
|
||||||
safe_gas_threshold_amount = 2000000000 * 60000 * 3
|
safe_gas_threshold_amount = 2000000000 * 60000 * 3
|
||||||
safe_gas_refill_amount = safe_gas_threshold_amount * 5
|
safe_gas_refill_amount = safe_gas_threshold_amount * 5
|
||||||
@@ -95,45 +78,19 @@ class CriticalSQLAlchemyAndSignerTask(CriticalTask):
|
|||||||
sqlalchemy.exc.DatabaseError,
|
sqlalchemy.exc.DatabaseError,
|
||||||
sqlalchemy.exc.TimeoutError,
|
sqlalchemy.exc.TimeoutError,
|
||||||
sqlalchemy.exc.ResourceClosedError,
|
sqlalchemy.exc.ResourceClosedError,
|
||||||
|
SignerError,
|
||||||
)
|
)
|
||||||
|
|
||||||
class CriticalWeb3AndSignerTask(CriticalTask):
|
class CriticalWeb3AndSignerTask(CriticalTask):
|
||||||
autoretry_for = (
|
autoretry_for = (
|
||||||
requests.exceptions.ConnectionError,
|
requests.exceptions.ConnectionError,
|
||||||
|
SignerError,
|
||||||
)
|
)
|
||||||
safe_gas_threshold_amount = 2000000000 * 60000 * 3
|
safe_gas_threshold_amount = 2000000000 * 60000 * 3
|
||||||
safe_gas_refill_amount = safe_gas_threshold_amount * 5
|
safe_gas_refill_amount = safe_gas_threshold_amount * 5
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task()
|
@celery_app.task(bind=True, base=BaseTask)
|
||||||
def check_health(self):
|
def hello(self):
|
||||||
pass
|
time.sleep(0.1)
|
||||||
|
return id(SessionBase.create_session)
|
||||||
|
|
||||||
# TODO: registry / rpc methods should perhaps be moved to better named module
|
|
||||||
@celery_app.task()
|
|
||||||
def registry():
|
|
||||||
return CICRegistry.address
|
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task()
|
|
||||||
def registry_address_lookup(chain_spec_dict, address, connection_tag='default'):
|
|
||||||
chain_spec = ChainSpec.from_dict(chain_spec_dict)
|
|
||||||
conn = RPCConnection.connect(chain_spec, tag=connection_tag)
|
|
||||||
registry = CICRegistry(chain_spec, conn)
|
|
||||||
return registry.by_address(address)
|
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task(throws=(UnknownContractError,))
|
|
||||||
def registry_name_lookup(chain_spec_dict, name, connection_tag='default'):
|
|
||||||
chain_spec = ChainSpec.from_dict(chain_spec_dict)
|
|
||||||
conn = RPCConnection.connect(chain_spec, tag=connection_tag)
|
|
||||||
registry = CICRegistry(chain_spec, conn)
|
|
||||||
return registry.by_name(name)
|
|
||||||
|
|
||||||
|
|
||||||
@celery_app.task()
|
|
||||||
def rpc_proxy(chain_spec_dict, o, connection_tag='default'):
|
|
||||||
chain_spec = ChainSpec.from_dict(chain_spec_dict)
|
|
||||||
conn = RPCConnection.connect(chain_spec, tag=connection_tag)
|
|
||||||
return conn.do(o)
|
|
||||||
|
|||||||
@@ -10,7 +10,7 @@ version = (
|
|||||||
0,
|
0,
|
||||||
11,
|
11,
|
||||||
0,
|
0,
|
||||||
'beta.11',
|
'beta.6',
|
||||||
)
|
)
|
||||||
|
|
||||||
version_object = semver.VersionInfo(
|
version_object = semver.VersionInfo(
|
||||||
|
|||||||
@@ -3,6 +3,3 @@ registry_address =
|
|||||||
chain_spec = evm:bloxberg:8996
|
chain_spec = evm:bloxberg:8996
|
||||||
tx_retry_delay =
|
tx_retry_delay =
|
||||||
trust_address =
|
trust_address =
|
||||||
default_token_symbol = GFT
|
|
||||||
health_modules = cic_eth.check.db,cic_eth.check.redis,cic_eth.check.signer,cic_eth.check.gas
|
|
||||||
run_dir = /run
|
|
||||||
|
|||||||
@@ -3,6 +3,3 @@ registry_address =
|
|||||||
chain_spec = evm:bloxberg:8996
|
chain_spec = evm:bloxberg:8996
|
||||||
trust_address = 0xEb3907eCad74a0013c259D5874AE7f22DcBcC95C
|
trust_address = 0xEb3907eCad74a0013c259D5874AE7f22DcBcC95C
|
||||||
tx_retry_delay = 20
|
tx_retry_delay = 20
|
||||||
default_token_symbol = GFT
|
|
||||||
health_modules = cic_eth.check.db,cic_eth.check.redis,cic_eth.check.signer,cic_eth.check.gas
|
|
||||||
run_dir = /run
|
|
||||||
|
|||||||
@@ -1,3 +1,8 @@
|
|||||||
[eth]
|
[eth]
|
||||||
|
#ws_provider = ws://localhost:8546
|
||||||
|
#ttp_provider = http://localhost:8545
|
||||||
provider = http://localhost:63545
|
provider = http://localhost:63545
|
||||||
gas_gifter_minimum_balance = 10000000000000000000000
|
gas_provider_address =
|
||||||
|
#chain_id =
|
||||||
|
abi_dir = /home/lash/src/ext/cic/grassrootseconomics/cic-contracts/abis
|
||||||
|
account_accounts_index_writer =
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
[signer]
|
[signer]
|
||||||
socket_path = ipc:///tmp/crypto-dev-signer/jsonrpc.ipc
|
socket_path = /tmp/crypto-dev-signer/jsonrpc.ipc
|
||||||
secret = deedbeef
|
secret = deedbeef
|
||||||
database_name = signer_test
|
database_name = signer_test
|
||||||
dev_keys_path =
|
dev_keys_path =
|
||||||
|
|||||||
@@ -1,3 +1,8 @@
|
|||||||
[eth]
|
[eth]
|
||||||
|
#ws_provider = ws://localhost:8546
|
||||||
|
#ttp_provider = http://localhost:8545
|
||||||
provider = http://localhost:8545
|
provider = http://localhost:8545
|
||||||
gas_gifter_minimum_balance = 10000000000000000000000
|
gas_provider_address =
|
||||||
|
#chain_id =
|
||||||
|
abi_dir = /usr/local/share/cic/solidity/abi
|
||||||
|
account_accounts_index_writer =
|
||||||
|
|||||||
@@ -50,4 +50,6 @@ COPY cic-eth/config/ /usr/local/etc/cic-eth/
|
|||||||
COPY cic-eth/cic_eth/db/migrations/ /usr/local/share/cic-eth/alembic/
|
COPY cic-eth/cic_eth/db/migrations/ /usr/local/share/cic-eth/alembic/
|
||||||
COPY cic-eth/crypto_dev_signer_config/ /usr/local/etc/crypto-dev-signer/
|
COPY cic-eth/crypto_dev_signer_config/ /usr/local/etc/crypto-dev-signer/
|
||||||
|
|
||||||
COPY util/liveness/health.sh /usr/local/bin/health.sh
|
RUN git clone https://gitlab.com/grassrootseconomics/cic-contracts.git && \
|
||||||
|
mkdir -p /usr/local/share/cic/solidity && \
|
||||||
|
cp -R cic-contracts/abis /usr/local/share/cic/solidity/abi
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
cic-base==0.1.2b5
|
cic-base~=0.1.2a76
|
||||||
celery==4.4.7
|
celery==4.4.7
|
||||||
crypto-dev-signer~=0.4.14b3
|
crypto-dev-signer~=0.4.14b2
|
||||||
confini~=0.3.6rc3
|
confini~=0.3.6rc3
|
||||||
cic-eth-registry~=0.5.4a16
|
cic-eth-registry~=0.5.4a16
|
||||||
#cic-bancor~=0.0.6
|
#cic-bancor~=0.0.6
|
||||||
@@ -15,10 +15,10 @@ semver==2.13.0
|
|||||||
websocket-client==0.57.0
|
websocket-client==0.57.0
|
||||||
moolb~=0.1.1b2
|
moolb~=0.1.1b2
|
||||||
eth-address-index~=0.1.1a9
|
eth-address-index~=0.1.1a9
|
||||||
chainlib~=0.0.2a20
|
chainlib~=0.0.2a13
|
||||||
hexathon~=0.0.1a7
|
hexathon~=0.0.1a7
|
||||||
chainsyncer[sql]~=0.0.2a2
|
chainsyncer[sql]~=0.0.2a2
|
||||||
chainqueue~=0.0.2a2
|
chainqueue~=0.0.1a7
|
||||||
pysha3==1.0.2
|
pysha3==1.0.2
|
||||||
coincurve==15.0.0
|
coincurve==15.0.0
|
||||||
sarafu-faucet==0.0.2a28
|
sarafu-faucet==0.0.2a28
|
||||||
|
|||||||
@@ -38,7 +38,6 @@ packages =
|
|||||||
cic_eth.runnable.daemons.filters
|
cic_eth.runnable.daemons.filters
|
||||||
cic_eth.callbacks
|
cic_eth.callbacks
|
||||||
cic_eth.sync
|
cic_eth.sync
|
||||||
cic_eth.check
|
|
||||||
scripts =
|
scripts =
|
||||||
./scripts/migrate.py
|
./scripts/migrate.py
|
||||||
|
|
||||||
@@ -53,7 +52,6 @@ console_scripts =
|
|||||||
cic-eth-create = cic_eth.runnable.create:main
|
cic-eth-create = cic_eth.runnable.create:main
|
||||||
cic-eth-inspect = cic_eth.runnable.view:main
|
cic-eth-inspect = cic_eth.runnable.view:main
|
||||||
cic-eth-ctl = cic_eth.runnable.ctrl:main
|
cic-eth-ctl = cic_eth.runnable.ctrl:main
|
||||||
cic-eth-info = cic_eth.runnable.info:main
|
|
||||||
# TODO: Merge this with ctl when subcmds sorted to submodules
|
# TODO: Merge this with ctl when subcmds sorted to submodules
|
||||||
cic-eth-tag = cic_eth.runnable.tag:main
|
cic-eth-tag = cic_eth.runnable.tag:main
|
||||||
cic-eth-resend = cic_eth.runnable.resend:main
|
cic-eth-resend = cic_eth.runnable.resend:main
|
||||||
|
|||||||
@@ -3,12 +3,8 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
# external imports
|
|
||||||
from chainlib.eth.erc20 import ERC20
|
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from cic_eth.api import Api
|
from cic_eth.api import Api
|
||||||
from cic_eth.task import BaseTask
|
|
||||||
|
|
||||||
script_dir = os.path.dirname(os.path.realpath(__file__))
|
script_dir = os.path.dirname(os.path.realpath(__file__))
|
||||||
root_dir = os.path.dirname(script_dir)
|
root_dir = os.path.dirname(script_dir)
|
||||||
@@ -32,26 +28,3 @@ def api(
|
|||||||
):
|
):
|
||||||
chain_str = str(default_chain_spec)
|
chain_str = str(default_chain_spec)
|
||||||
return Api(chain_str, queue=None, callback_param='foo')
|
return Api(chain_str, queue=None, callback_param='foo')
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope='function')
|
|
||||||
def foo_token_symbol(
|
|
||||||
default_chain_spec,
|
|
||||||
foo_token,
|
|
||||||
eth_rpc,
|
|
||||||
contract_roles,
|
|
||||||
):
|
|
||||||
|
|
||||||
c = ERC20(default_chain_spec)
|
|
||||||
o = c.symbol(foo_token, sender_address=contract_roles['CONTRACT_DEPLOYER'])
|
|
||||||
r = eth_rpc.do(o)
|
|
||||||
return c.parse_symbol(r)
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope='function')
|
|
||||||
def default_token(
|
|
||||||
foo_token,
|
|
||||||
foo_token_symbol,
|
|
||||||
):
|
|
||||||
BaseTask.default_token_symbol = foo_token_symbol
|
|
||||||
BaseTask.default_token_address = foo_token
|
|
||||||
|
|||||||
@@ -210,11 +210,9 @@ def test_callback_filter(
|
|||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.results = {}
|
self.results = {}
|
||||||
self.queue = 'test'
|
|
||||||
|
|
||||||
def call_back(self, transfer_type, result):
|
def call_back(self, transfer_type, result):
|
||||||
self.results[transfer_type] = result
|
self.results[transfer_type] = result
|
||||||
return self
|
|
||||||
|
|
||||||
mock = CallbackMock()
|
mock = CallbackMock()
|
||||||
fltr.call_back = mock.call_back
|
fltr.call_back = mock.call_back
|
||||||
|
|||||||
@@ -65,7 +65,6 @@ def test_tx(
|
|||||||
tx_hash_hex_orig = tx_hash_hex
|
tx_hash_hex_orig = tx_hash_hex
|
||||||
|
|
||||||
gas_oracle = OverrideGasOracle(price=1100000000, limit=21000)
|
gas_oracle = OverrideGasOracle(price=1100000000, limit=21000)
|
||||||
c = Gas(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
|
||||||
(tx_hash_hex, tx_signed_raw_hex) = c.create(agent_roles['ALICE'], agent_roles['BOB'], 100 * (10 ** 6), tx_format=TxFormat.RLP_SIGNED)
|
(tx_hash_hex, tx_signed_raw_hex) = c.create(agent_roles['ALICE'], agent_roles['BOB'], 100 * (10 ** 6), tx_format=TxFormat.RLP_SIGNED)
|
||||||
queue_create(
|
queue_create(
|
||||||
default_chain_spec,
|
default_chain_spec,
|
||||||
|
|||||||
@@ -34,7 +34,6 @@ def celery_includes():
|
|||||||
'cic_eth.admin.ctrl',
|
'cic_eth.admin.ctrl',
|
||||||
'cic_eth.admin.nonce',
|
'cic_eth.admin.nonce',
|
||||||
'cic_eth.admin.debug',
|
'cic_eth.admin.debug',
|
||||||
'cic_eth.admin.token',
|
|
||||||
'cic_eth.eth.account',
|
'cic_eth.eth.account',
|
||||||
'cic_eth.callbacks.noop',
|
'cic_eth.callbacks.noop',
|
||||||
'cic_eth.callbacks.http',
|
'cic_eth.callbacks.http',
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
import os
|
import os
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
# external imports
|
# third-party imports
|
||||||
import pytest
|
import pytest
|
||||||
import alembic
|
import alembic
|
||||||
from alembic.config import Config as AlembicConfig
|
from alembic.config import Config as AlembicConfig
|
||||||
@@ -53,9 +53,6 @@ def init_database(
|
|||||||
alembic.command.downgrade(ac, 'base')
|
alembic.command.downgrade(ac, 'base')
|
||||||
alembic.command.upgrade(ac, 'head')
|
alembic.command.upgrade(ac, 'head')
|
||||||
|
|
||||||
session.execute('DELETE FROM lock')
|
|
||||||
session.commit()
|
|
||||||
|
|
||||||
yield session
|
yield session
|
||||||
session.commit()
|
session.commit()
|
||||||
session.close()
|
session.close()
|
||||||
|
|||||||
@@ -273,7 +273,7 @@ def test_tx(
|
|||||||
eth_signer,
|
eth_signer,
|
||||||
agent_roles,
|
agent_roles,
|
||||||
contract_roles,
|
contract_roles,
|
||||||
celery_session_worker,
|
celery_worker,
|
||||||
):
|
):
|
||||||
|
|
||||||
nonce_oracle = RPCNonceOracle(agent_roles['ALICE'], eth_rpc)
|
nonce_oracle = RPCNonceOracle(agent_roles['ALICE'], eth_rpc)
|
||||||
|
|||||||
@@ -35,7 +35,7 @@ def test_list_tx(
|
|||||||
foo_token,
|
foo_token,
|
||||||
register_tokens,
|
register_tokens,
|
||||||
init_eth_tester,
|
init_eth_tester,
|
||||||
celery_session_worker,
|
celery_worker,
|
||||||
):
|
):
|
||||||
|
|
||||||
tx_hashes = []
|
tx_hashes = []
|
||||||
|
|||||||
@@ -1,21 +0,0 @@
|
|||||||
# external imports
|
|
||||||
import celery
|
|
||||||
|
|
||||||
|
|
||||||
def test_default_token(
|
|
||||||
default_token,
|
|
||||||
celery_session_worker,
|
|
||||||
foo_token,
|
|
||||||
foo_token_symbol,
|
|
||||||
):
|
|
||||||
|
|
||||||
s = celery.signature(
|
|
||||||
'cic_eth.admin.token.default_token',
|
|
||||||
[],
|
|
||||||
queue=None,
|
|
||||||
)
|
|
||||||
t = s.apply_async()
|
|
||||||
r = t.get()
|
|
||||||
|
|
||||||
assert r['address'] == foo_token
|
|
||||||
assert r['symbol'] == foo_token_symbol
|
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
[pgp]
|
[pgp]
|
||||||
exports_dir = /root/pgp
|
exports_dir = pgp
|
||||||
privatekey_file = privatekeys.asc
|
privatekey_file = privatekeys.asc
|
||||||
passphrase = merman
|
passphrase = merman
|
||||||
publickey_trusted_file = publickeys.asc
|
publickey_trusted_file = publickeys.asc
|
||||||
|
|||||||
1
apps/cic-meta/.gitignore
vendored
1
apps/cic-meta/.gitignore
vendored
@@ -3,3 +3,4 @@ dist
|
|||||||
dist-web
|
dist-web
|
||||||
dist-server
|
dist-server
|
||||||
scratch
|
scratch
|
||||||
|
tests
|
||||||
|
|||||||
@@ -2,21 +2,17 @@ FROM node:15.3.0-alpine3.10
|
|||||||
|
|
||||||
WORKDIR /tmp/src/cic-meta
|
WORKDIR /tmp/src/cic-meta
|
||||||
|
|
||||||
RUN apk add --no-cache postgresql bash
|
|
||||||
|
|
||||||
COPY cic-meta/package.json \
|
COPY cic-meta/package.json \
|
||||||
./
|
./
|
||||||
|
|
||||||
COPY cic-meta/src/ src/
|
COPY cic-meta/src/ src/
|
||||||
COPY cic-meta/tests/ tests/
|
COPY cic-meta/tests/ tests/
|
||||||
COPY cic-meta/scripts/ scripts/
|
COPY cic-meta/scripts/ scripts/
|
||||||
|
#COPY docker/*.sh /root/
|
||||||
|
|
||||||
RUN npm install
|
|
||||||
|
|
||||||
# see exports_dir gpg.ini
|
|
||||||
COPY cic-meta/tests/*.asc /root/pgp/
|
|
||||||
RUN alias tsc=node_modules/typescript/bin/tsc
|
RUN alias tsc=node_modules/typescript/bin/tsc
|
||||||
|
|
||||||
|
RUN npm install
|
||||||
|
|
||||||
COPY cic-meta/.config/ /usr/local/etc/cic-meta/
|
COPY cic-meta/.config/ /usr/local/etc/cic-meta/
|
||||||
# COPY cic-meta/scripts/server/initdb/server.postgres.sql /usr/local/share/cic-meta/sql/server.sql
|
# COPY cic-meta/scripts/server/initdb/server.postgres.sql /usr/local/share/cic-meta/sql/server.sql
|
||||||
@@ -24,9 +20,8 @@ COPY cic-meta/.config/ /usr/local/etc/cic-meta/
|
|||||||
COPY cic-meta/docker/db.sh ./db.sh
|
COPY cic-meta/docker/db.sh ./db.sh
|
||||||
RUN chmod 755 ./db.sh
|
RUN chmod 755 ./db.sh
|
||||||
|
|
||||||
#RUN alias ts-node=/tmp/src/cic-meta/node_modules/ts-node/dist/bin.js
|
RUN alias ts-node=/tmp/src/cic-meta/node_modules/ts-node/dist/bin.js
|
||||||
#ENTRYPOINT [ "./node_modules/ts-node/dist/bin.js", "./scripts/server/server.ts" ]
|
ENTRYPOINT [ "./node_modules/ts-node/dist/bin.js", "./scripts/server/server.ts" ]
|
||||||
|
|
||||||
COPY cic-meta/docker/start_server.sh ./start_server.sh
|
# COPY cic-meta/docker/start_server.sh ./start_server.sh
|
||||||
RUN chmod 755 ./start_server.sh
|
# RUN chmod 755 ./start_server.sh
|
||||||
ENTRYPOINT ["sh", "./start_server.sh"]
|
|
||||||
|
|||||||
@@ -1,6 +1,3 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
set -e
|
|
||||||
|
|
||||||
PGPASSWORD=$DATABASE_PASSWORD psql -v ON_ERROR_STOP=1 -U $DATABASE_USER -h $DATABASE_HOST -p $DATABASE_PORT -d $DATABASE_NAME -f $SCHEMA_SQL_PATH
|
|
||||||
|
|
||||||
|
|
||||||
|
PGPASSWORD=$DATABASE_PASSWORD psql -U $DATABASE_USER -h $DATABASE_HOST -p $DATABASE_PORT -d $DATABASE_NAME /usr/local/share/cic-meta/sql/server.sql
|
||||||
|
|||||||
@@ -1,9 +1,3 @@
|
|||||||
#!/bin/bash
|
|
||||||
set -euo pipefail
|
|
||||||
|
|
||||||
# db migration
|
|
||||||
sh ./db.sh
|
sh ./db.sh
|
||||||
|
|
||||||
# /usr/local/bin/node /usr/local/bin/cic-meta-server $@
|
/usr/local/bin/node /usr/local/bin/cic-meta-server $@
|
||||||
# ./node_modules/ts-node/dist/bin.js", "./scripts/server/server.ts $@
|
|
||||||
npm run start "$@"
|
|
||||||
|
|||||||
101
apps/cic-meta/package-lock.json
generated
101
apps/cic-meta/package-lock.json
generated
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "cic-client-meta",
|
"name": "cic-client-meta",
|
||||||
"version": "0.0.7-alpha.7",
|
"version": "0.0.7-alpha.2",
|
||||||
"lockfileVersion": 1,
|
"lockfileVersion": 1,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
@@ -852,75 +852,6 @@
|
|||||||
"printj": "~1.1.0"
|
"printj": "~1.1.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"crdt-meta": {
|
|
||||||
"version": "0.0.8",
|
|
||||||
"resolved": "https://registry.npmjs.org/crdt-meta/-/crdt-meta-0.0.8.tgz",
|
|
||||||
"integrity": "sha512-CS0sS0L2QWthz7vmu6vzl3p4kcpJ+IKILBJ4tbgN4A3iNG8wnBeuDIv/z3KFFQjcfuP4QAh6E9LywKUTxtDc3g==",
|
|
||||||
"requires": {
|
|
||||||
"automerge": "^0.14.2",
|
|
||||||
"ini": "^1.3.8",
|
|
||||||
"openpgp": "^4.10.8",
|
|
||||||
"pg": "^8.5.1",
|
|
||||||
"sqlite3": "^5.0.2"
|
|
||||||
},
|
|
||||||
"dependencies": {
|
|
||||||
"automerge": {
|
|
||||||
"version": "0.14.2",
|
|
||||||
"resolved": "https://registry.npmjs.org/automerge/-/automerge-0.14.2.tgz",
|
|
||||||
"integrity": "sha512-shiwuJHCbNRI23WZyIECLV4Ovf3WiAFJ7P9BH4l5gON1In/UUbjcSJKRygtIirObw2UQumeYxp3F2XBdSvQHnA==",
|
|
||||||
"requires": {
|
|
||||||
"immutable": "^3.8.2",
|
|
||||||
"transit-immutable-js": "^0.7.0",
|
|
||||||
"transit-js": "^0.8.861",
|
|
||||||
"uuid": "^3.4.0"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node-addon-api": {
|
|
||||||
"version": "3.1.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/node-addon-api/-/node-addon-api-3.1.0.tgz",
|
|
||||||
"integrity": "sha512-flmrDNB06LIl5lywUz7YlNGZH/5p0M7W28k8hzd9Lshtdh1wshD2Y+U4h9LD6KObOy1f+fEVdgprPrEymjM5uw=="
|
|
||||||
},
|
|
||||||
"pg": {
|
|
||||||
"version": "8.6.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/pg/-/pg-8.6.0.tgz",
|
|
||||||
"integrity": "sha512-qNS9u61lqljTDFvmk/N66EeGq3n6Ujzj0FFyNMGQr6XuEv4tgNTXvJQTfJdcvGit5p5/DWPu+wj920hAJFI+QQ==",
|
|
||||||
"requires": {
|
|
||||||
"buffer-writer": "2.0.0",
|
|
||||||
"packet-reader": "1.0.0",
|
|
||||||
"pg-connection-string": "^2.5.0",
|
|
||||||
"pg-pool": "^3.3.0",
|
|
||||||
"pg-protocol": "^1.5.0",
|
|
||||||
"pg-types": "^2.1.0",
|
|
||||||
"pgpass": "1.x"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"pg-connection-string": {
|
|
||||||
"version": "2.5.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/pg-connection-string/-/pg-connection-string-2.5.0.tgz",
|
|
||||||
"integrity": "sha512-r5o/V/ORTA6TmUnyWZR9nCj1klXCO2CEKNRlVuJptZe85QuhFayC7WeMic7ndayT5IRIR0S0xFxFi2ousartlQ=="
|
|
||||||
},
|
|
||||||
"pg-pool": {
|
|
||||||
"version": "3.3.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/pg-pool/-/pg-pool-3.3.0.tgz",
|
|
||||||
"integrity": "sha512-0O5huCql8/D6PIRFAlmccjphLYWC+JIzvUhSzXSpGaf+tjTZc4nn+Lr7mLXBbFJfvwbP0ywDv73EiaBsxn7zdg=="
|
|
||||||
},
|
|
||||||
"pg-protocol": {
|
|
||||||
"version": "1.5.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/pg-protocol/-/pg-protocol-1.5.0.tgz",
|
|
||||||
"integrity": "sha512-muRttij7H8TqRNu/DxrAJQITO4Ac7RmX3Klyr/9mJEOBeIpgnF8f9jAfRz5d3XwQZl5qBjF9gLsUtMPJE0vezQ=="
|
|
||||||
},
|
|
||||||
"sqlite3": {
|
|
||||||
"version": "5.0.2",
|
|
||||||
"resolved": "https://registry.npmjs.org/sqlite3/-/sqlite3-5.0.2.tgz",
|
|
||||||
"integrity": "sha512-1SdTNo+BVU211Xj1csWa8lV6KM0CtucDwRyA0VHl91wEH1Mgh7RxUpI4rVvG7OhHrzCSGaVyW5g8vKvlrk9DJA==",
|
|
||||||
"requires": {
|
|
||||||
"node-addon-api": "^3.0.0",
|
|
||||||
"node-gyp": "3.x",
|
|
||||||
"node-pre-gyp": "^0.11.0"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"create-hash": {
|
"create-hash": {
|
||||||
"version": "1.2.0",
|
"version": "1.2.0",
|
||||||
"resolved": "https://registry.npmjs.org/create-hash/-/create-hash-1.2.0.tgz",
|
"resolved": "https://registry.npmjs.org/create-hash/-/create-hash-1.2.0.tgz",
|
||||||
@@ -1035,17 +966,17 @@
|
|||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
"elliptic": {
|
"elliptic": {
|
||||||
"version": "6.5.4",
|
"version": "6.5.3",
|
||||||
"resolved": "https://registry.npmjs.org/elliptic/-/elliptic-6.5.4.tgz",
|
"resolved": "https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz",
|
||||||
"integrity": "sha512-iLhC6ULemrljPZb+QutR5TQGB+pdW6KGD5RSegS+8sorOZT+rdQFbsQFJgvN3eRqNALqJer4oQ16YvJHlU8hzQ==",
|
"integrity": "sha512-IMqzv5wNQf+E6aHeIqATs0tOLeOTwj1QKbRcS3jBbYkl5oLAserA8yJTT7/VyHUYG91PRmPyeQDObKLPpeS4dw==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"bn.js": "^4.11.9",
|
"bn.js": "^4.4.0",
|
||||||
"brorand": "^1.1.0",
|
"brorand": "^1.0.1",
|
||||||
"hash.js": "^1.0.0",
|
"hash.js": "^1.0.0",
|
||||||
"hmac-drbg": "^1.0.1",
|
"hmac-drbg": "^1.0.0",
|
||||||
"inherits": "^2.0.4",
|
"inherits": "^2.0.1",
|
||||||
"minimalistic-assert": "^1.0.1",
|
"minimalistic-assert": "^1.0.0",
|
||||||
"minimalistic-crypto-utils": "^1.0.1"
|
"minimalistic-crypto-utils": "^1.0.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"emoji-regex": {
|
"emoji-regex": {
|
||||||
@@ -1558,9 +1489,9 @@
|
|||||||
"integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ=="
|
"integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ=="
|
||||||
},
|
},
|
||||||
"ini": {
|
"ini": {
|
||||||
"version": "1.3.8",
|
"version": "1.3.5",
|
||||||
"resolved": "https://registry.npmjs.org/ini/-/ini-1.3.8.tgz",
|
"resolved": "https://registry.npmjs.org/ini/-/ini-1.3.5.tgz",
|
||||||
"integrity": "sha512-JV/yugV2uzW5iMRSiZAyDtQd+nxtUnjeLt0acNdw98kKLrvuRVyB80tsREOE7yvGVgalhZ6RNXCmEHkUKBKxew=="
|
"integrity": "sha512-RZY5huIKCMRWDUqZlEi72f/lmXKMvuszcMBduliQ3nnWbx9X/ZBQO7DijMEYS9EhHBb2qacRUMtC7svLwe0lcw=="
|
||||||
},
|
},
|
||||||
"interpret": {
|
"interpret": {
|
||||||
"version": "2.2.0",
|
"version": "2.2.0",
|
||||||
@@ -2026,9 +1957,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"y18n": {
|
"y18n": {
|
||||||
"version": "4.0.3",
|
"version": "4.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/y18n/-/y18n-4.0.3.tgz",
|
"resolved": "https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz",
|
||||||
"integrity": "sha512-JKhqTOwSrqNA1NY5lSztJ1GrBiUodLMmIZuLiDaMRJ+itFd+ABVE8XBjOvIWL+rSqNDC74LCSFmlb/U4UZ4hJQ==",
|
"integrity": "sha512-r9S/ZyXu/Xu9q1tYlpsLIsa3EeLXXk0VwlxqTcFRfg9EhMW+17kbt9G0NrgCmhGb5vT2hyhJZLfDGx+7+5Uj/w==",
|
||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
"yargs": {
|
"yargs": {
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "cic-client-meta",
|
"name": "cic-client-meta",
|
||||||
"version": "0.0.7-alpha.8",
|
"version": "0.0.7-alpha.6",
|
||||||
"description": "Signed CRDT metadata graphs for the CIC network",
|
"description": "Signed CRDT metadata graphs for the CIC network",
|
||||||
"main": "dist/index.js",
|
"main": "dist/index.js",
|
||||||
"types": "dist/index.d.ts",
|
"types": "dist/index.d.ts",
|
||||||
@@ -10,15 +10,13 @@
|
|||||||
"build-server": "tsc -d --outDir dist-server scripts/server/*.ts",
|
"build-server": "tsc -d --outDir dist-server scripts/server/*.ts",
|
||||||
"pack": "node_modules/typescript/bin/tsc -d --outDir dist && webpack",
|
"pack": "node_modules/typescript/bin/tsc -d --outDir dist && webpack",
|
||||||
"clean": "rm -rf dist",
|
"clean": "rm -rf dist",
|
||||||
"prepare": "npm run build && npm run build-server",
|
"prepare": "npm run build && npm run build-server"
|
||||||
"start": "./node_modules/ts-node/dist/bin.js ./scripts/server/server.ts"
|
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@ethereumjs/tx": "^3.0.0-beta.1",
|
"@ethereumjs/tx": "^3.0.0-beta.1",
|
||||||
"automerge": "^0.14.1",
|
"automerge": "^0.14.1",
|
||||||
"crdt-meta": "0.0.8",
|
|
||||||
"ethereumjs-wallet": "^1.0.1",
|
"ethereumjs-wallet": "^1.0.1",
|
||||||
"ini": "^1.3.8",
|
"ini": "^1.3.5",
|
||||||
"openpgp": "^4.10.8",
|
"openpgp": "^4.10.8",
|
||||||
"pg": "^8.4.2",
|
"pg": "^8.4.2",
|
||||||
"sqlite3": "^5.0.0",
|
"sqlite3": "^5.0.0",
|
||||||
@@ -42,6 +40,6 @@
|
|||||||
],
|
],
|
||||||
"license": "GPL-3.0-or-later",
|
"license": "GPL-3.0-or-later",
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=14.16.1"
|
"node": "~14.16.1"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { Config } from 'crdt-meta';
|
const config = require('./src/config');
|
||||||
const fs = require('fs');
|
const fs = require('fs');
|
||||||
|
|
||||||
if (process.argv[2] === undefined) {
|
if (process.argv[2] === undefined) {
|
||||||
@@ -15,6 +15,6 @@ try {
|
|||||||
process.exit(1);
|
process.exit(1);
|
||||||
}
|
}
|
||||||
|
|
||||||
const c = new Config(process.argv[2], process.env['CONFINI_ENV_PREFIX']);
|
const c = new config.Config(process.argv[2], process.env['CONFINI_ENV_PREFIX']);
|
||||||
c.process();
|
c.process();
|
||||||
process.stdout.write(c.toString());
|
process.stdout.write(c.toString());
|
||||||
|
|||||||
15
apps/cic-meta/scripts/initdb/postgresql.sh
Executable file
15
apps/cic-meta/scripts/initdb/postgresql.sh
Executable file
@@ -0,0 +1,15 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
psql -v ON_ERROR_STOP=1 --username grassroots --dbname cic_meta <<-EOSQL
|
||||||
|
create table if not exists store (
|
||||||
|
id serial primary key not null,
|
||||||
|
owner_fingerprint text not null,
|
||||||
|
hash char(64) not null unique,
|
||||||
|
content text not null
|
||||||
|
);
|
||||||
|
|
||||||
|
create index if not exists idx_fp on store ((lower(owner_fingerprint)));
|
||||||
|
EOSQL
|
||||||
|
|
||||||
|
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
create table if not exists store (
|
create table if not exists cic_meta.store (
|
||||||
id serial primary key not null,
|
id serial primary key not null,
|
||||||
owner_fingerprint text not null,
|
owner_fingerprint text not null,
|
||||||
hash char(64) not null unique,
|
hash char(64) not null unique,
|
||||||
|
|||||||
@@ -1,7 +1,8 @@
|
|||||||
import * as Automerge from 'automerge';
|
import * as Automerge from 'automerge';
|
||||||
import * as pgp from 'openpgp';
|
import * as pgp from 'openpgp';
|
||||||
|
import * as pg from 'pg';
|
||||||
|
|
||||||
import { Envelope, Syncable } from 'crdt-meta';
|
import { Envelope, Syncable } from '../../src/sync';
|
||||||
|
|
||||||
|
|
||||||
function handleNoMergeGet(db, digest, keystore) {
|
function handleNoMergeGet(db, digest, keystore) {
|
||||||
|
|||||||
@@ -1,11 +1,15 @@
|
|||||||
import * as http from 'http';
|
import * as http from 'http';
|
||||||
import * as fs from 'fs';
|
import * as fs from 'fs';
|
||||||
import * as path from 'path';
|
import * as path from 'path';
|
||||||
|
import * as pgp from 'openpgp';
|
||||||
|
|
||||||
import * as handlers from './handlers';
|
import * as handlers from './handlers';
|
||||||
import { PGPKeyStore, PGPSigner, Config, SqliteAdapter, PostgresAdapter } from 'crdt-meta';
|
import { Envelope, Syncable } from '../../src/sync';
|
||||||
|
import { PGPKeyStore, PGPSigner } from '../../src/auth';
|
||||||
|
|
||||||
import { standardArgs } from './args';
|
import { standardArgs } from './args';
|
||||||
|
import { Config } from '../../src/config';
|
||||||
|
import { SqliteAdapter, PostgresAdapter } from '../../src/db';
|
||||||
|
|
||||||
let configPath = '/usr/local/etc/cic-meta';
|
let configPath = '/usr/local/etc/cic-meta';
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Syncable, Addressable, mergeKey } from 'crdt-meta';
|
import { ArgPair, Syncable } from '../sync';
|
||||||
|
import { Addressable, mergeKey } from '../digest';
|
||||||
|
|
||||||
class Phone extends Syncable implements Addressable {
|
class Phone extends Syncable implements Addressable {
|
||||||
|
|
||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Syncable, Addressable, toAddressKey } from 'crdt-meta';
|
import { ArgPair, Syncable } from '../sync';
|
||||||
|
import { Addressable, addressToBytes, bytesToHex, toAddressKey } from '../digest';
|
||||||
|
|
||||||
const keySalt = new TextEncoder().encode(':cic.person');
|
const keySalt = new TextEncoder().encode(':cic.person');
|
||||||
class User extends Syncable implements Addressable {
|
class User extends Syncable implements Addressable {
|
||||||
191
apps/cic-meta/src/auth.ts
Normal file
191
apps/cic-meta/src/auth.ts
Normal file
@@ -0,0 +1,191 @@
|
|||||||
|
import * as pgp from 'openpgp';
|
||||||
|
import * as crypto from 'crypto';
|
||||||
|
|
||||||
|
interface Signable {
|
||||||
|
digest():string;
|
||||||
|
}
|
||||||
|
|
||||||
|
type KeyGetter = () => any;
|
||||||
|
|
||||||
|
type Signature = {
|
||||||
|
engine:string
|
||||||
|
algo:string
|
||||||
|
data:string
|
||||||
|
digest:string
|
||||||
|
}
|
||||||
|
|
||||||
|
interface Signer {
|
||||||
|
prepare(Signable):boolean;
|
||||||
|
onsign(Signature):void;
|
||||||
|
onverify(boolean):void;
|
||||||
|
sign(digest:string):void
|
||||||
|
verify(digest:string, signature:Signature):void
|
||||||
|
fingerprint():string
|
||||||
|
}
|
||||||
|
|
||||||
|
interface Authoritative {
|
||||||
|
}
|
||||||
|
|
||||||
|
interface KeyStore {
|
||||||
|
getPrivateKey: KeyGetter
|
||||||
|
getFingerprint: () => string
|
||||||
|
getTrustedKeys: () => Array<any>
|
||||||
|
getTrustedActiveKeys: () => Array<any>
|
||||||
|
getEncryptKeys: () => Array<any>
|
||||||
|
}
|
||||||
|
|
||||||
|
class PGPKeyStore implements KeyStore {
|
||||||
|
|
||||||
|
fingerprint: string
|
||||||
|
pk: any
|
||||||
|
|
||||||
|
pubk = {
|
||||||
|
active: [],
|
||||||
|
trusted: [],
|
||||||
|
encrypt: [],
|
||||||
|
}
|
||||||
|
loads = 0x00;
|
||||||
|
loadsTarget = 0x0f;
|
||||||
|
onload: (k:KeyStore) => void;
|
||||||
|
|
||||||
|
constructor(passphrase:string, pkArmor:string, pubkActiveArmor:string, pubkTrustedArmor:string, pubkEncryptArmor:string, onload = (ks:KeyStore) => {}) {
|
||||||
|
this._readKey(pkArmor, undefined, 1, passphrase);
|
||||||
|
this._readKey(pubkActiveArmor, 'active', 2);
|
||||||
|
this._readKey(pubkTrustedArmor, 'trusted', 4);
|
||||||
|
this._readKey(pubkEncryptArmor, 'encrypt', 8);
|
||||||
|
this.onload = onload;
|
||||||
|
}
|
||||||
|
|
||||||
|
private _readKey(a:string, x:any, n:number, pass?:string) {
|
||||||
|
pgp.key.readArmored(a).then((k) => {
|
||||||
|
if (pass !== undefined) {
|
||||||
|
this.pk = k.keys[0];
|
||||||
|
this.pk.decrypt(pass).then(() => {
|
||||||
|
this.fingerprint = this.pk.getFingerprint();
|
||||||
|
console.log('private key (sign)', this.fingerprint);
|
||||||
|
this._registerLoad(n);
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
this.pubk[x] = k.keys;
|
||||||
|
k.keys.forEach((pubk) => {
|
||||||
|
console.log('public key (' + x + ')', pubk.getFingerprint());
|
||||||
|
});
|
||||||
|
this._registerLoad(n);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
private _registerLoad(b:number) {
|
||||||
|
this.loads |= b;
|
||||||
|
if (this.loads == this.loadsTarget) {
|
||||||
|
this.onload(this);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public getTrustedKeys(): Array<any> {
|
||||||
|
return this.pubk['trusted'];
|
||||||
|
}
|
||||||
|
|
||||||
|
public getTrustedActiveKeys(): Array<any> {
|
||||||
|
return this.pubk['active'];
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public getEncryptKeys(): Array<any> {
|
||||||
|
return this.pubk['encrypt'];
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public getPrivateKey(): any {
|
||||||
|
return this.pk;
|
||||||
|
}
|
||||||
|
|
||||||
|
public getFingerprint(): string {
|
||||||
|
return this.fingerprint;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class PGPSigner implements Signer {
|
||||||
|
|
||||||
|
engine = 'pgp'
|
||||||
|
algo = 'sha256'
|
||||||
|
dgst: string
|
||||||
|
signature: Signature
|
||||||
|
keyStore: KeyStore
|
||||||
|
onsign: (Signature) => void
|
||||||
|
onverify: (boolean) => void
|
||||||
|
|
||||||
|
constructor(keyStore:KeyStore) {
|
||||||
|
this.keyStore = keyStore
|
||||||
|
this.onsign = (string) => {};
|
||||||
|
this.onverify = (boolean) => {};
|
||||||
|
}
|
||||||
|
|
||||||
|
public fingerprint(): string {
|
||||||
|
return this.keyStore.getFingerprint();
|
||||||
|
}
|
||||||
|
|
||||||
|
public prepare(material:Signable):boolean {
|
||||||
|
this.dgst = material.digest();
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public verify(digest:string, signature:Signature) {
|
||||||
|
pgp.signature.readArmored(signature.data).then((s) => {
|
||||||
|
const opts = {
|
||||||
|
message: pgp.cleartext.fromText(digest),
|
||||||
|
publicKeys: this.keyStore.getTrustedKeys(),
|
||||||
|
signature: s,
|
||||||
|
};
|
||||||
|
pgp.verify(opts).then((v) => {
|
||||||
|
let i = 0;
|
||||||
|
for (i = 0; i < v.signatures.length; i++) {
|
||||||
|
const s = v.signatures[i];
|
||||||
|
if (s.valid) {
|
||||||
|
this.onverify(s);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
console.error('checked ' + i + ' signature(s) but none valid');
|
||||||
|
this.onverify(false);
|
||||||
|
});
|
||||||
|
}).catch((e) => {
|
||||||
|
console.error(e);
|
||||||
|
this.onverify(false);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
public sign(digest:string) {
|
||||||
|
const m = pgp.cleartext.fromText(digest);
|
||||||
|
const pk = this.keyStore.getPrivateKey();
|
||||||
|
const opts = {
|
||||||
|
message: m,
|
||||||
|
privateKeys: [pk],
|
||||||
|
detached: true,
|
||||||
|
}
|
||||||
|
pgp.sign(opts).then((s) => {
|
||||||
|
this.signature = {
|
||||||
|
engine: this.engine,
|
||||||
|
algo: this.algo,
|
||||||
|
data: s.signature,
|
||||||
|
// TODO: fix for browser later
|
||||||
|
digest: digest,
|
||||||
|
};
|
||||||
|
this.onsign(this.signature);
|
||||||
|
}).catch((e) => {
|
||||||
|
console.error(e);
|
||||||
|
this.onsign(undefined);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
Signature,
|
||||||
|
Authoritative,
|
||||||
|
Signer,
|
||||||
|
KeyGetter,
|
||||||
|
Signable,
|
||||||
|
KeyStore,
|
||||||
|
PGPSigner,
|
||||||
|
PGPKeyStore,
|
||||||
|
};
|
||||||
71
apps/cic-meta/src/config.ts
Normal file
71
apps/cic-meta/src/config.ts
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
import * as fs from 'fs';
|
||||||
|
import * as ini from 'ini';
|
||||||
|
import * as path from 'path';
|
||||||
|
|
||||||
|
class Config {
|
||||||
|
|
||||||
|
filepath: string
|
||||||
|
store: Object
|
||||||
|
censor: Array<string>
|
||||||
|
require: Array<string>
|
||||||
|
env_prefix: string
|
||||||
|
|
||||||
|
constructor(filepath:string, env_prefix?:string) {
|
||||||
|
this.filepath = filepath;
|
||||||
|
this.store = {};
|
||||||
|
this.censor = [];
|
||||||
|
this.require = [];
|
||||||
|
this.env_prefix = '';
|
||||||
|
if (env_prefix !== undefined) {
|
||||||
|
this.env_prefix = env_prefix + "_";
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public process() {
|
||||||
|
const d = fs.readdirSync(this.filepath);
|
||||||
|
|
||||||
|
const r = /.*\.ini$/;
|
||||||
|
for (let i = 0; i < d.length; i++) {
|
||||||
|
const f = d[i];
|
||||||
|
if (!f.match(r)) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
const fp = path.join(this.filepath, f);
|
||||||
|
const v = fs.readFileSync(fp, 'utf-8');
|
||||||
|
const inid = ini.decode(v);
|
||||||
|
const inik = Object.keys(inid);
|
||||||
|
for (let j = 0; j < inik.length; j++) {
|
||||||
|
const k_section = inik[j]
|
||||||
|
const k = k_section.toUpperCase();
|
||||||
|
Object.keys(inid[k_section]).forEach((k_directive) => {
|
||||||
|
const kk = k_directive.toUpperCase();
|
||||||
|
const kkk = k + '_' + kk;
|
||||||
|
|
||||||
|
let r = inid[k_section][k_directive];
|
||||||
|
const k_env = this.env_prefix + kkk
|
||||||
|
const env = process.env[k_env];
|
||||||
|
if (env !== undefined) {
|
||||||
|
console.debug('Environment variable ' + k_env + ' overrides ' + kkk);
|
||||||
|
r = env;
|
||||||
|
}
|
||||||
|
this.store[kkk] = r;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public get(s:string) {
|
||||||
|
return this.store[s];
|
||||||
|
}
|
||||||
|
|
||||||
|
public toString() {
|
||||||
|
let s = '';
|
||||||
|
Object.keys(this.store).forEach((k) => {
|
||||||
|
s += k + '=' + this.store[k] + '\n';
|
||||||
|
});
|
||||||
|
return s;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export { Config };
|
||||||
38
apps/cic-meta/src/constants.ts
Normal file
38
apps/cic-meta/src/constants.ts
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
import { JSONSerializable } from './format';
|
||||||
|
|
||||||
|
const ENGINE_NAME = 'automerge';
|
||||||
|
const ENGINE_VERSION = '0.14.1';
|
||||||
|
|
||||||
|
const NETWORK_NAME = 'cic';
|
||||||
|
const NETWORK_VERSION = '1';
|
||||||
|
|
||||||
|
const CRYPTO_NAME = 'pgp';
|
||||||
|
const CRYPTO_VERSION = '2';
|
||||||
|
|
||||||
|
type VersionedSpec = {
|
||||||
|
name: string
|
||||||
|
version: string
|
||||||
|
ext?: Object
|
||||||
|
}
|
||||||
|
|
||||||
|
const engineSpec:VersionedSpec = {
|
||||||
|
name: ENGINE_NAME,
|
||||||
|
version: ENGINE_VERSION,
|
||||||
|
}
|
||||||
|
|
||||||
|
const cryptoSpec:VersionedSpec = {
|
||||||
|
name: CRYPTO_NAME,
|
||||||
|
version: CRYPTO_VERSION,
|
||||||
|
}
|
||||||
|
|
||||||
|
const networkSpec:VersionedSpec = {
|
||||||
|
name: NETWORK_NAME,
|
||||||
|
version: NETWORK_VERSION,
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
engineSpec,
|
||||||
|
cryptoSpec,
|
||||||
|
networkSpec,
|
||||||
|
VersionedSpec,
|
||||||
|
};
|
||||||
27
apps/cic-meta/src/crypto.ts
Normal file
27
apps/cic-meta/src/crypto.ts
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
import * as crypto from 'crypto';
|
||||||
|
|
||||||
|
const _algs = {
|
||||||
|
'SHA-256': 'sha256',
|
||||||
|
}
|
||||||
|
|
||||||
|
function cryptoWrapper() {
|
||||||
|
}
|
||||||
|
|
||||||
|
cryptoWrapper.prototype.digest = async function(s, d) {
|
||||||
|
const h = crypto.createHash(_algs[s]);
|
||||||
|
h.update(d);
|
||||||
|
return h.digest();
|
||||||
|
}
|
||||||
|
|
||||||
|
let subtle = undefined;
|
||||||
|
if (typeof window !== 'undefined') {
|
||||||
|
subtle = window.crypto.subtle;
|
||||||
|
} else {
|
||||||
|
subtle = new cryptoWrapper();
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
export {
|
||||||
|
subtle,
|
||||||
|
}
|
||||||
|
|
||||||
90
apps/cic-meta/src/db.ts
Normal file
90
apps/cic-meta/src/db.ts
Normal file
@@ -0,0 +1,90 @@
|
|||||||
|
import * as pg from 'pg';
|
||||||
|
import * as sqlite from 'sqlite3';
|
||||||
|
|
||||||
|
type DbConfig = {
|
||||||
|
name: string
|
||||||
|
host: string
|
||||||
|
port: number
|
||||||
|
user: string
|
||||||
|
password: string
|
||||||
|
}
|
||||||
|
|
||||||
|
interface DbAdapter {
|
||||||
|
query: (s:string, callback:(e:any, rs:any) => void) => void
|
||||||
|
close: () => void
|
||||||
|
}
|
||||||
|
|
||||||
|
const re_creatematch = /^(CREATE)/i
|
||||||
|
const re_getmatch = /^(SELECT)/i;
|
||||||
|
const re_setmatch = /^(INSERT|UPDATE)/i;
|
||||||
|
|
||||||
|
class SqliteAdapter implements DbAdapter {
|
||||||
|
|
||||||
|
db: any
|
||||||
|
|
||||||
|
constructor(dbConfig:DbConfig, callback?:(any) => void) {
|
||||||
|
this.db = new sqlite.Database(dbConfig.name); //, callback);
|
||||||
|
}
|
||||||
|
|
||||||
|
public query(s:string, callback:(e:any, rs?:any) => void): void {
|
||||||
|
const local_callback = (e, rs) => {
|
||||||
|
let r = undefined;
|
||||||
|
if (rs !== undefined) {
|
||||||
|
r = {
|
||||||
|
rowCount: rs.length,
|
||||||
|
rows: rs,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
callback(e, r);
|
||||||
|
};
|
||||||
|
if (s.match(re_getmatch)) {
|
||||||
|
this.db.all(s, local_callback);
|
||||||
|
} else if (s.match(re_setmatch)) {
|
||||||
|
this.db.run(s, local_callback);
|
||||||
|
} else if (s.match(re_creatematch)) {
|
||||||
|
this.db.run(s, callback);
|
||||||
|
} else {
|
||||||
|
throw 'unhandled query';
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public close() {
|
||||||
|
this.db.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class PostgresAdapter implements DbAdapter {
|
||||||
|
|
||||||
|
db: any
|
||||||
|
|
||||||
|
constructor(dbConfig:DbConfig) {
|
||||||
|
let o = dbConfig;
|
||||||
|
o['database'] = o.name;
|
||||||
|
this.db = new pg.Pool(o);
|
||||||
|
return this.db;
|
||||||
|
}
|
||||||
|
|
||||||
|
public query(s:string, callback:(e:any, rs:any) => void): void {
|
||||||
|
this.db.query(s, (e, rs) => {
|
||||||
|
let r = {
|
||||||
|
length: rs.rowCount,
|
||||||
|
}
|
||||||
|
rs.length = rs.rowCount;
|
||||||
|
if (e === undefined) {
|
||||||
|
e = null;
|
||||||
|
}
|
||||||
|
console.debug(e, rs);
|
||||||
|
callback(e, rs);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
public close() {
|
||||||
|
this.db.end();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
DbConfig,
|
||||||
|
SqliteAdapter,
|
||||||
|
PostgresAdapter,
|
||||||
|
}
|
||||||
68
apps/cic-meta/src/digest.ts
Normal file
68
apps/cic-meta/src/digest.ts
Normal file
@@ -0,0 +1,68 @@
|
|||||||
|
import * as crypto from './crypto';
|
||||||
|
|
||||||
|
interface Addressable {
|
||||||
|
key(): string
|
||||||
|
digest(): string
|
||||||
|
}
|
||||||
|
|
||||||
|
function stringToBytes(s:string) {
|
||||||
|
const a = new Uint8Array(20);
|
||||||
|
let j = 2;
|
||||||
|
for (let i = 0; i < a.byteLength; i++) {
|
||||||
|
const n = parseInt(s.substring(j, j+2), 16);
|
||||||
|
a[i] = n;
|
||||||
|
j += 2;
|
||||||
|
}
|
||||||
|
return a;
|
||||||
|
}
|
||||||
|
|
||||||
|
function bytesToHex(a:Uint8Array) {
|
||||||
|
let s = '';
|
||||||
|
for (let i = 0; i < a.byteLength; i++) {
|
||||||
|
const h = '00' + a[i].toString(16);
|
||||||
|
s += h.slice(-2);
|
||||||
|
}
|
||||||
|
return s;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function mergeKey(a:Uint8Array, s:Uint8Array) {
|
||||||
|
const y = new Uint8Array(a.byteLength + s.byteLength);
|
||||||
|
for (let i = 0; i < a.byteLength; i++) {
|
||||||
|
y[i] = a[i];
|
||||||
|
}
|
||||||
|
for (let i = 0; i < s.byteLength; i++) {
|
||||||
|
y[a.byteLength + i] = s[i];
|
||||||
|
}
|
||||||
|
const z = await crypto.subtle.digest('SHA-256', y);
|
||||||
|
return bytesToHex(new Uint8Array(z));
|
||||||
|
}
|
||||||
|
|
||||||
|
async function toKey(v:string, salt:string) {
|
||||||
|
const a = stringToBytes(v);
|
||||||
|
const s = new TextEncoder().encode(salt);
|
||||||
|
return await mergeKey(a, s);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
async function toAddressKey(zeroExHex:string, salt:string) {
|
||||||
|
const a = addressToBytes(zeroExHex);
|
||||||
|
const s = new TextEncoder().encode(salt);
|
||||||
|
return await mergeKey(a, s);
|
||||||
|
}
|
||||||
|
|
||||||
|
const re_addrHex = /^0[xX][a-fA-F0-9]{40}$/;
|
||||||
|
function addressToBytes(s:string) {
|
||||||
|
if (!s.match(re_addrHex)) {
|
||||||
|
throw 'invalid address hex';
|
||||||
|
}
|
||||||
|
return stringToBytes(s);
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
toKey,
|
||||||
|
toAddressKey,
|
||||||
|
mergeKey,
|
||||||
|
bytesToHex,
|
||||||
|
addressToBytes,
|
||||||
|
Addressable,
|
||||||
|
}
|
||||||
58
apps/cic-meta/src/dispatch.ts
Normal file
58
apps/cic-meta/src/dispatch.ts
Normal file
@@ -0,0 +1,58 @@
|
|||||||
|
import { v4 as uuidv4 } from 'uuid';
|
||||||
|
import { Syncable } from './sync';
|
||||||
|
import { Store } from './store';
|
||||||
|
import { PubSub } from './transport';
|
||||||
|
|
||||||
|
function toIndexKey(id:string):string {
|
||||||
|
const d = Date.now();
|
||||||
|
return d + '_' + id + '_' + uuidv4();
|
||||||
|
}
|
||||||
|
|
||||||
|
const _re_indexKey = /^\d+_(.+)_[-\d\w]+$/;
|
||||||
|
function fromIndexKey(s:string):string {
|
||||||
|
const m = s.match(_re_indexKey);
|
||||||
|
if (m === null) {
|
||||||
|
throw 'Invalid index key';
|
||||||
|
}
|
||||||
|
return m[1];
|
||||||
|
}
|
||||||
|
|
||||||
|
class Dispatcher {
|
||||||
|
|
||||||
|
idx: Array<string>
|
||||||
|
syncer: PubSub
|
||||||
|
store: Store
|
||||||
|
|
||||||
|
constructor(store:Store, syncer:PubSub) {
|
||||||
|
this.idx = new Array<string>()
|
||||||
|
this.syncer = syncer;
|
||||||
|
this.store = store;
|
||||||
|
}
|
||||||
|
|
||||||
|
public isDirty(): boolean {
|
||||||
|
return this.idx.length > 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
public add(id:string, item:Syncable): string {
|
||||||
|
const v = item.toJSON();
|
||||||
|
const k = toIndexKey(id);
|
||||||
|
this.store.put(k, v, true);
|
||||||
|
localStorage.setItem(k, v);
|
||||||
|
this.idx.push(k);
|
||||||
|
return k;
|
||||||
|
}
|
||||||
|
|
||||||
|
public sync(offset:number): number {
|
||||||
|
let i = 0;
|
||||||
|
this.idx.forEach((k) => {
|
||||||
|
const v = localStorage.getItem(k);
|
||||||
|
const k_id = fromIndexKey(k);
|
||||||
|
this.syncer.pub(v); // this must block until guaranteed delivery
|
||||||
|
localStorage.removeItem(k);
|
||||||
|
i++;
|
||||||
|
});
|
||||||
|
return i;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export { Dispatcher, toIndexKey, fromIndexKey }
|
||||||
5
apps/cic-meta/src/format.ts
Normal file
5
apps/cic-meta/src/format.ts
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
interface JSONSerializable {
|
||||||
|
toJSON(): string
|
||||||
|
}
|
||||||
|
|
||||||
|
export { JSONSerializable };
|
||||||
@@ -1,2 +1,5 @@
|
|||||||
export { User } from './user';
|
export { PGPSigner, PGPKeyStore, Signer, KeyStore } from './auth';
|
||||||
export { Phone } from './phone';
|
export { ArgPair, Envelope, Syncable } from './sync';
|
||||||
|
export { User } from './assets/user';
|
||||||
|
export { Phone } from './assets/phone';
|
||||||
|
export { Config } from './config';
|
||||||
|
|||||||
9
apps/cic-meta/src/store.ts
Normal file
9
apps/cic-meta/src/store.ts
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
import { Syncable } from './sync';
|
||||||
|
|
||||||
|
interface Store {
|
||||||
|
put(string, Syncable, boolean?)
|
||||||
|
get(string):Syncable
|
||||||
|
delete(string)
|
||||||
|
}
|
||||||
|
|
||||||
|
export { Store };
|
||||||
266
apps/cic-meta/src/sync.ts
Normal file
266
apps/cic-meta/src/sync.ts
Normal file
@@ -0,0 +1,266 @@
|
|||||||
|
import * as Automerge from 'automerge';
|
||||||
|
|
||||||
|
import { JSONSerializable } from './format';
|
||||||
|
|
||||||
|
import { Authoritative, Signer, PGPSigner, Signable, Signature } from './auth';
|
||||||
|
|
||||||
|
import { engineSpec, cryptoSpec, networkSpec, VersionedSpec } from './constants';
|
||||||
|
|
||||||
|
const fullSpec:VersionedSpec = {
|
||||||
|
name: 'cic',
|
||||||
|
version: '1',
|
||||||
|
ext: {
|
||||||
|
network: cryptoSpec,
|
||||||
|
engine: engineSpec,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
class Envelope {
|
||||||
|
|
||||||
|
o = fullSpec
|
||||||
|
|
||||||
|
constructor(payload:Object) {
|
||||||
|
this.set(payload);
|
||||||
|
}
|
||||||
|
|
||||||
|
public set(payload:Object) {
|
||||||
|
this.o['payload'] = payload
|
||||||
|
}
|
||||||
|
|
||||||
|
public get():string {
|
||||||
|
return this.o['payload'];
|
||||||
|
}
|
||||||
|
|
||||||
|
public toJSON() {
|
||||||
|
return JSON.stringify(this.o);
|
||||||
|
}
|
||||||
|
|
||||||
|
public static fromJSON(s:string): Envelope {
|
||||||
|
const e = new Envelope(undefined);
|
||||||
|
e.o = JSON.parse(s);
|
||||||
|
return e;
|
||||||
|
}
|
||||||
|
|
||||||
|
public unwrap(): Syncable {
|
||||||
|
return Syncable.fromJSON(this.o['payload']);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class ArgPair {
|
||||||
|
|
||||||
|
k:string
|
||||||
|
v:any
|
||||||
|
|
||||||
|
constructor(k:string, v:any) {
|
||||||
|
this.k = k;
|
||||||
|
this.v = v;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class SignablePart implements Signable {
|
||||||
|
|
||||||
|
s: string
|
||||||
|
|
||||||
|
constructor(s:string) {
|
||||||
|
this.s = s;
|
||||||
|
}
|
||||||
|
|
||||||
|
public digest():string {
|
||||||
|
return this.s;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function orderDict(src) {
|
||||||
|
let dst;
|
||||||
|
if (Array.isArray(src)) {
|
||||||
|
dst = [];
|
||||||
|
src.forEach((v) => {
|
||||||
|
if (typeof(v) == 'object') {
|
||||||
|
v = orderDict(v);
|
||||||
|
}
|
||||||
|
dst.push(v);
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
dst = {}
|
||||||
|
Object.keys(src).sort().forEach((k) => {
|
||||||
|
let v = src[k];
|
||||||
|
if (typeof(v) == 'object') {
|
||||||
|
v = orderDict(v);
|
||||||
|
}
|
||||||
|
dst[k] = v;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
return dst;
|
||||||
|
}
|
||||||
|
|
||||||
|
class Syncable implements JSONSerializable, Authoritative, Signable {
|
||||||
|
|
||||||
|
id: string
|
||||||
|
timestamp: number
|
||||||
|
m: any // automerge object
|
||||||
|
e: Envelope
|
||||||
|
signer: Signer
|
||||||
|
onwrap: (string) => void
|
||||||
|
onauthenticate: (boolean) => void
|
||||||
|
|
||||||
|
// TODO: Move data to sub-object so timestamp, id, signature don't collide
|
||||||
|
constructor(id:string, v:Object) {
|
||||||
|
this.id = id;
|
||||||
|
const o = {
|
||||||
|
'id': id,
|
||||||
|
'timestamp': Math.floor(Date.now() / 1000),
|
||||||
|
'data': v,
|
||||||
|
}
|
||||||
|
//this.m = Automerge.from(v)
|
||||||
|
this.m = Automerge.from(o)
|
||||||
|
}
|
||||||
|
|
||||||
|
public setSigner(signer:Signer) {
|
||||||
|
this.signer = signer;
|
||||||
|
this.signer.onsign = (s) => {
|
||||||
|
this.wrap(s);
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: To keep integrity, the non-link key/value pairs for each step also need to be hashed
|
||||||
|
public digest(): string {
|
||||||
|
const links = [];
|
||||||
|
Automerge.getAllChanges(this.m).forEach((ch:Object) => {
|
||||||
|
const op:Array<any> = ch['ops'];
|
||||||
|
ch['ops'].forEach((op:Array<Object>) => {
|
||||||
|
if (op['action'] == 'link') {
|
||||||
|
//console.log('op link', op);
|
||||||
|
links.push([op['obj'], op['value']]);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
//return JSON.stringify(links);
|
||||||
|
const j = JSON.stringify(links);
|
||||||
|
return Buffer.from(j).toString('base64');
|
||||||
|
}
|
||||||
|
|
||||||
|
private wrap(s:any) {
|
||||||
|
this.m = Automerge.change(this.m, 'sign', (doc) => {
|
||||||
|
doc['signature'] = s;
|
||||||
|
});
|
||||||
|
this.e = new Envelope(this.toJSON());
|
||||||
|
console.log('wrappin s', s, typeof(s));
|
||||||
|
this.e.o['digest'] = s.digest;
|
||||||
|
if (this.onwrap !== undefined) {
|
||||||
|
this.onwrap(this.e);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// private _verifyLoop(i:number, history:Array<any>, signable:Signable, result:boolean) {
|
||||||
|
// if (!result) {
|
||||||
|
// this.onauthenticate(false);
|
||||||
|
// return;
|
||||||
|
// } else if (history.length == 0) {
|
||||||
|
// this.onauthenticate(true);
|
||||||
|
// return;
|
||||||
|
// }
|
||||||
|
// const h = history.shift()
|
||||||
|
// if (i % 2 == 0) {
|
||||||
|
// i++;
|
||||||
|
// signable = {
|
||||||
|
// digest: () => {
|
||||||
|
// return Automerge.save(h.snapshot)
|
||||||
|
// },
|
||||||
|
// };
|
||||||
|
// this._verifyLoop(i, history, signable, true);
|
||||||
|
// } else {
|
||||||
|
// i++;
|
||||||
|
// const signature = h.snapshot['signature'];
|
||||||
|
// console.debug('signature', signature, signable.digest());
|
||||||
|
// this.signer.onverify = (v) => {
|
||||||
|
// this._verifyLoop(i, history, signable, v)
|
||||||
|
// }
|
||||||
|
// this.signer.verify(signable, signature);
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // TODO: This should replay the graph and check signatures on each step
|
||||||
|
// public _authenticate(full:boolean=false) {
|
||||||
|
// let h = Automerge.getHistory(this.m);
|
||||||
|
// h.forEach((m) => {
|
||||||
|
// //console.debug(m.snapshot);
|
||||||
|
// });
|
||||||
|
// const signable = {
|
||||||
|
// digest: () => { return '' },
|
||||||
|
// }
|
||||||
|
// if (!full) {
|
||||||
|
// h = h.slice(h.length-2);
|
||||||
|
// }
|
||||||
|
// this._verifyLoop(0, h, signable, true);
|
||||||
|
// }
|
||||||
|
|
||||||
|
public authenticate(full:boolean=false) {
|
||||||
|
if (full) {
|
||||||
|
console.warn('only doing shallow authentication for now, sorry');
|
||||||
|
}
|
||||||
|
//console.log('authenticating', signable.digest());
|
||||||
|
//console.log('signature', this.m.signature);
|
||||||
|
this.signer.onverify = (v) => {
|
||||||
|
//this._verifyLoop(i, history, signable, v)
|
||||||
|
this.onauthenticate(v);
|
||||||
|
}
|
||||||
|
this.signer.verify(this.m.signature.digest, this.m.signature);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
public sign() {
|
||||||
|
//this.signer.prepare(this);
|
||||||
|
this.signer.sign(this.digest());
|
||||||
|
}
|
||||||
|
|
||||||
|
public update(changes:Array<ArgPair>, changesDescription:string) {
|
||||||
|
this.m = Automerge.change(this.m, changesDescription, (m) => {
|
||||||
|
changes.forEach((c) => {
|
||||||
|
let path = c.k.split('.');
|
||||||
|
let target = m['data'];
|
||||||
|
while (path.length > 1) {
|
||||||
|
const part = path.shift();
|
||||||
|
target = target[part];
|
||||||
|
}
|
||||||
|
target[path[0]] = c.v;
|
||||||
|
});
|
||||||
|
m['timestamp'] = Math.floor(Date.now() / 1000);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
public replace(o:Object, changesDescription:string) {
|
||||||
|
this.m = Automerge.change(this.m, changesDescription, (m) => {
|
||||||
|
Object.keys(o).forEach((k) => {
|
||||||
|
m['data'][k] = o[k];
|
||||||
|
});
|
||||||
|
Object.keys(m).forEach((k) => {
|
||||||
|
if (o[k] == undefined) {
|
||||||
|
delete m['data'][k];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
m['timestamp'] = Math.floor(Date.now() / 1000);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
public merge(s:Syncable) {
|
||||||
|
this.m = Automerge.merge(s.m, this.m);
|
||||||
|
}
|
||||||
|
|
||||||
|
public toJSON(): string {
|
||||||
|
const s = Automerge.save(this.m);
|
||||||
|
const o = JSON.parse(s);
|
||||||
|
const oo = orderDict(o)
|
||||||
|
return JSON.stringify(oo);
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public static fromJSON(s:string): Syncable {
|
||||||
|
const doc = Automerge.load(s);
|
||||||
|
let y = new Syncable(doc['id'], {});
|
||||||
|
y.m = doc
|
||||||
|
return y
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export { JSONSerializable, Syncable, ArgPair, Envelope };
|
||||||
11
apps/cic-meta/src/transport.ts
Normal file
11
apps/cic-meta/src/transport.ts
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
interface SubConsumer {
|
||||||
|
post(string)
|
||||||
|
}
|
||||||
|
|
||||||
|
interface PubSub {
|
||||||
|
pub(v:string):boolean
|
||||||
|
close()
|
||||||
|
}
|
||||||
|
|
||||||
|
export { PubSub, SubConsumer };
|
||||||
|
|
||||||
50
apps/cic-meta/tests/1_basic.ts
Normal file
50
apps/cic-meta/tests/1_basic.ts
Normal file
@@ -0,0 +1,50 @@
|
|||||||
|
import * as Automerge from 'automerge';
|
||||||
|
import assert = require('assert');
|
||||||
|
|
||||||
|
import { Dispatcher, toIndexKey, fromIndexKey } from '../src/dispatch';
|
||||||
|
import { User } from '../src/assets/user';
|
||||||
|
import { Syncable, ArgPair } from '../src/sync';
|
||||||
|
|
||||||
|
import { MockSigner, MockStore } from './mock';
|
||||||
|
|
||||||
|
describe('basic', () => {
|
||||||
|
|
||||||
|
it('store', () => {
|
||||||
|
const store = new MockStore('s');
|
||||||
|
assert.equal(store.name, 's');
|
||||||
|
|
||||||
|
const mockSigner = new MockSigner();
|
||||||
|
const v = new Syncable('foo', {baz: 42});
|
||||||
|
v.setSigner(mockSigner);
|
||||||
|
store.put('foo', v);
|
||||||
|
const one = store.get('foo').toJSON();
|
||||||
|
const vv = new Syncable('bar', {baz: 666});
|
||||||
|
vv.setSigner(mockSigner);
|
||||||
|
assert.throws(() => {
|
||||||
|
store.put('foo', vv)
|
||||||
|
});
|
||||||
|
store.put('foo', vv, true);
|
||||||
|
const other = store.get('foo').toJSON();
|
||||||
|
assert.notEqual(one, other);
|
||||||
|
store.delete('foo');
|
||||||
|
assert.equal(store.get('foo'), undefined);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('add_doc_to_dispatcher', () => {
|
||||||
|
const store = new MockStore('s');
|
||||||
|
//const syncer = new MockSyncer();
|
||||||
|
const dispatcher = new Dispatcher(store, undefined);
|
||||||
|
const user = new User('foo');
|
||||||
|
dispatcher.add(user.id, user);
|
||||||
|
assert(dispatcher.isDirty());
|
||||||
|
});
|
||||||
|
|
||||||
|
it('dispatch_keyindex', () => {
|
||||||
|
const s = 'foo';
|
||||||
|
const k = toIndexKey(s);
|
||||||
|
const v = fromIndexKey(k);
|
||||||
|
assert.equal(s, v);
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
|
});
|
||||||
212
apps/cic-meta/tests/2_sync.ts
Normal file
212
apps/cic-meta/tests/2_sync.ts
Normal file
@@ -0,0 +1,212 @@
|
|||||||
|
import * as Automerge from 'automerge';
|
||||||
|
import assert = require('assert');
|
||||||
|
|
||||||
|
import * as pgp from 'openpgp';
|
||||||
|
import * as fs from 'fs';
|
||||||
|
|
||||||
|
import { PGPSigner } from '../src/auth';
|
||||||
|
|
||||||
|
import { Syncable, ArgPair } from '../src/sync';
|
||||||
|
|
||||||
|
import { MockKeyStore, MockSigner } from './mock';
|
||||||
|
|
||||||
|
|
||||||
|
describe('sync', async () => {
|
||||||
|
it('sync_merge', () => {
|
||||||
|
const mockSigner = new MockSigner();
|
||||||
|
const s = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
s.setSigner(mockSigner);
|
||||||
|
const changePair = new ArgPair('xyzzy', 42);
|
||||||
|
s.update([changePair], 'ch-ch-cha-changes');
|
||||||
|
assert.equal(s.m.data['xyzzy'], 42)
|
||||||
|
assert.equal(s.m.data['bar'], 'baz')
|
||||||
|
assert.equal(s.m['id'], 'foo')
|
||||||
|
assert.equal(Automerge.getHistory(s.m).length, 2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('sync_serialize', () => {
|
||||||
|
const mockSigner = new MockSigner();
|
||||||
|
const s = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
s.setSigner(mockSigner);
|
||||||
|
const j = s.toJSON();
|
||||||
|
const ss = Syncable.fromJSON(j);
|
||||||
|
assert.equal(ss.m['id'], 'foo');
|
||||||
|
assert.equal(ss.m['data']['bar'], 'baz');
|
||||||
|
assert.equal(Automerge.getHistory(ss.m).length, 1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('sync_sign_and_wrap', () => {
|
||||||
|
const mockSigner = new MockSigner();
|
||||||
|
const s = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
s.setSigner(mockSigner);
|
||||||
|
s.onwrap = (e) => {
|
||||||
|
const j = e.toJSON();
|
||||||
|
const v = JSON.parse(j);
|
||||||
|
assert.deepEqual(v.payload, e.o.payload);
|
||||||
|
|
||||||
|
}
|
||||||
|
s.sign();
|
||||||
|
});
|
||||||
|
it('sync_verify_success', async () => {
|
||||||
|
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||||
|
const pks = await pgp.key.readArmored(pksa);
|
||||||
|
await pks.keys[0].decrypt('merman');
|
||||||
|
await pks.keys[1].decrypt('beastman');
|
||||||
|
|
||||||
|
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||||
|
const pubks = await pgp.key.readArmored(pubksa);
|
||||||
|
|
||||||
|
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||||
|
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||||
|
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||||
|
|
||||||
|
const oneSigner = new PGPSigner(oneStore);
|
||||||
|
const twoSigner = new PGPSigner(twoStore);
|
||||||
|
const threeSigner = new PGPSigner(threeStore);
|
||||||
|
|
||||||
|
const x = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
x.setSigner(oneSigner);
|
||||||
|
|
||||||
|
// TODO: make this look better
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
let updateData = new ArgPair('bar', 'xyzzy');
|
||||||
|
x.update([updateData], 'change one');
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
x.setSigner(twoSigner);
|
||||||
|
updateData = new ArgPair('bar', 42);
|
||||||
|
x.update([updateData], 'change two');
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
const p = e.unwrap();
|
||||||
|
p.setSigner(twoSigner);
|
||||||
|
p.onauthenticate = (v) => {
|
||||||
|
assert(v);
|
||||||
|
}
|
||||||
|
p.authenticate();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
};
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
|
||||||
|
});
|
||||||
|
|
||||||
|
it('sync_verify_fail', async () => {
|
||||||
|
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||||
|
const pks = await pgp.key.readArmored(pksa);
|
||||||
|
await pks.keys[0].decrypt('merman');
|
||||||
|
await pks.keys[1].decrypt('beastman');
|
||||||
|
|
||||||
|
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||||
|
const pubks = await pgp.key.readArmored(pubksa);
|
||||||
|
|
||||||
|
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||||
|
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||||
|
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||||
|
|
||||||
|
const oneSigner = new PGPSigner(oneStore);
|
||||||
|
const twoSigner = new PGPSigner(twoStore);
|
||||||
|
const threeSigner = new PGPSigner(threeStore);
|
||||||
|
|
||||||
|
const x = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
x.setSigner(oneSigner);
|
||||||
|
|
||||||
|
// TODO: make this look better
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
let updateData = new ArgPair('bar', 'xyzzy');
|
||||||
|
x.update([updateData], 'change one');
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
x.setSigner(twoSigner);
|
||||||
|
updateData = new ArgPair('bar', 42);
|
||||||
|
x.update([updateData], 'change two');
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
const p = e.unwrap();
|
||||||
|
p.setSigner(threeSigner);
|
||||||
|
p.onauthenticate = (v) => {
|
||||||
|
assert(!v);
|
||||||
|
}
|
||||||
|
p.authenticate();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
};
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
|
||||||
|
});
|
||||||
|
|
||||||
|
xit('sync_verify_shallow_tricked', async () => {
|
||||||
|
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||||
|
const pks = await pgp.key.readArmored(pksa);
|
||||||
|
await pks.keys[0].decrypt('merman');
|
||||||
|
await pks.keys[1].decrypt('beastman');
|
||||||
|
|
||||||
|
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||||
|
const pubks = await pgp.key.readArmored(pubksa);
|
||||||
|
|
||||||
|
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||||
|
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||||
|
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||||
|
|
||||||
|
const oneSigner = new PGPSigner(oneStore);
|
||||||
|
const twoSigner = new PGPSigner(twoStore);
|
||||||
|
const threeSigner = new PGPSigner(threeStore);
|
||||||
|
|
||||||
|
const x = new Syncable('foo', {
|
||||||
|
bar: 'baz',
|
||||||
|
});
|
||||||
|
x.setSigner(twoSigner);
|
||||||
|
|
||||||
|
// TODO: make this look better
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
let updateData = new ArgPair('bar', 'xyzzy');
|
||||||
|
x.update([updateData], 'change one');
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
updateData = new ArgPair('bar', 42);
|
||||||
|
x.update([updateData], 'change two');
|
||||||
|
x.setSigner(oneSigner);
|
||||||
|
|
||||||
|
x.onwrap = (e) => {
|
||||||
|
const p = e.unwrap();
|
||||||
|
p.setSigner(threeSigner);
|
||||||
|
p.onauthenticate = (v) => {
|
||||||
|
assert(v);
|
||||||
|
p.onauthenticate = (v) => {
|
||||||
|
assert(!v);
|
||||||
|
}
|
||||||
|
p.authenticate(true);
|
||||||
|
}
|
||||||
|
p.authenticate();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
};
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
}
|
||||||
|
|
||||||
|
x.sign();
|
||||||
|
|
||||||
|
});
|
||||||
|
});
|
||||||
14
apps/cic-meta/tests/3_transport.ts
Normal file
14
apps/cic-meta/tests/3_transport.ts
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
import * as assert from 'assert';
|
||||||
|
|
||||||
|
import { MockPubSub, MockConsumer } from './mock';
|
||||||
|
|
||||||
|
describe('transport', () => {
|
||||||
|
it('pub_sub', () => {
|
||||||
|
const c = new MockConsumer();
|
||||||
|
const ps = new MockPubSub('foo', c);
|
||||||
|
ps.pub('foo');
|
||||||
|
ps.pub('bar');
|
||||||
|
ps.flush();
|
||||||
|
assert.deepEqual(c.omnoms, ['foo', 'bar']);
|
||||||
|
});
|
||||||
|
});
|
||||||
46
apps/cic-meta/tests/4_auth.ts
Normal file
46
apps/cic-meta/tests/4_auth.ts
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
import assert = require('assert');
|
||||||
|
import pgp = require('openpgp');
|
||||||
|
import crypto = require('crypto');
|
||||||
|
|
||||||
|
import { Syncable, ArgPair } from '../src/sync';
|
||||||
|
|
||||||
|
import { MockKeyStore, MockSignable } from './mock';
|
||||||
|
|
||||||
|
import { PGPSigner } from '../src/auth';
|
||||||
|
|
||||||
|
|
||||||
|
describe('auth', async () => {
|
||||||
|
await it('digest', async () => {
|
||||||
|
const opts = {
|
||||||
|
userIds: [
|
||||||
|
{
|
||||||
|
name: 'John Marston',
|
||||||
|
email: 'red@dead.com',
|
||||||
|
},
|
||||||
|
],
|
||||||
|
numBits: 2048,
|
||||||
|
passphrase: 'foo',
|
||||||
|
};
|
||||||
|
const pkgen = await pgp.generateKey(opts);
|
||||||
|
const pka = pkgen.privateKeyArmored;
|
||||||
|
const pks = await pgp.key.readArmored(pka);
|
||||||
|
await pks.keys[0].decrypt('foo');
|
||||||
|
const pubka = pkgen.publicKeyArmored;
|
||||||
|
const pubks = await pgp.key.readArmored(pubka);
|
||||||
|
const keyStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||||
|
const s = new PGPSigner(keyStore);
|
||||||
|
|
||||||
|
const message = await pgp.cleartext.fromText('foo');
|
||||||
|
s.onverify = (ok) => {
|
||||||
|
assert(ok);
|
||||||
|
}
|
||||||
|
s.onsign = (signature) => {
|
||||||
|
s.onverify((v) => {
|
||||||
|
console.log('bar', v);
|
||||||
|
});
|
||||||
|
s.verify('foo', signature);
|
||||||
|
}
|
||||||
|
|
||||||
|
await s.sign('foo');
|
||||||
|
});
|
||||||
|
});
|
||||||
47
apps/cic-meta/tests/999_functional.ts
Normal file
47
apps/cic-meta/tests/999_functional.ts
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
import * as assert from 'assert';
|
||||||
|
import * as pgp from 'openpgp';
|
||||||
|
|
||||||
|
import { Dispatcher } from '../src/dispatch';
|
||||||
|
import { User } from '../src/assets/user';
|
||||||
|
import { PGPSigner, KeyStore } from '../src/auth';
|
||||||
|
import { SubConsumer } from '../src/transport';
|
||||||
|
|
||||||
|
import { MockStore, MockPubSub, MockConsumer, MockKeyStore } from './mock';
|
||||||
|
|
||||||
|
async function createKeyStore() {
|
||||||
|
const opts = {
|
||||||
|
userIds: [
|
||||||
|
{
|
||||||
|
name: 'John Marston',
|
||||||
|
email: 'red@dead.com',
|
||||||
|
},
|
||||||
|
],
|
||||||
|
numBits: 2048,
|
||||||
|
passphrase: 'foo',
|
||||||
|
};
|
||||||
|
const pkgen = await pgp.generateKey(opts);
|
||||||
|
const pka = pkgen.privateKeyArmored;
|
||||||
|
const pks = await pgp.key.readArmored(pka);
|
||||||
|
await pks.keys[0].decrypt('foo');
|
||||||
|
return new MockKeyStore(pks.keys[0], []);
|
||||||
|
}
|
||||||
|
|
||||||
|
describe('fullchain', async () => {
|
||||||
|
it('dispatch_and_publish_user', async () => {
|
||||||
|
const g = await createKeyStore();
|
||||||
|
const n = new PGPSigner(g);
|
||||||
|
const u = new User('u1', {});
|
||||||
|
u.setSigner(n);
|
||||||
|
u.setName('Nico', 'Bellic');
|
||||||
|
const s = new MockStore('fooStore');
|
||||||
|
const c = new MockConsumer();
|
||||||
|
const p = new MockPubSub('fooPubSub', c);
|
||||||
|
const d = new Dispatcher(s, p);
|
||||||
|
u.onwrap = (e) => {
|
||||||
|
d.add(u.id, e);
|
||||||
|
d.sync(0);
|
||||||
|
assert.equal(p.pubs.length, 1);
|
||||||
|
};
|
||||||
|
u.sign();
|
||||||
|
});
|
||||||
|
});
|
||||||
150
apps/cic-meta/tests/mock.ts
Normal file
150
apps/cic-meta/tests/mock.ts
Normal file
@@ -0,0 +1,150 @@
|
|||||||
|
import * as crypto from 'crypto';
|
||||||
|
|
||||||
|
import { Signable, Signature, KeyStore } from '../src/auth';
|
||||||
|
import { Store } from '../src/store';
|
||||||
|
import { PubSub, SubConsumer } from '../src/transport';
|
||||||
|
import { Syncable } from '../src/sync';
|
||||||
|
|
||||||
|
class MockStore implements Store {
|
||||||
|
|
||||||
|
contents: Object
|
||||||
|
name: string
|
||||||
|
|
||||||
|
constructor(name:string) {
|
||||||
|
this.name = name;
|
||||||
|
this.contents = {};
|
||||||
|
}
|
||||||
|
|
||||||
|
public put(k:string, v:Syncable, existsOk = false) {
|
||||||
|
if (!existsOk && this.contents[k] !== undefined) {
|
||||||
|
throw '"' + k + '" already exists in store ' + this.name;
|
||||||
|
}
|
||||||
|
this.contents[k] = v;
|
||||||
|
}
|
||||||
|
|
||||||
|
public get(k:string): Syncable {
|
||||||
|
return this.contents[k];
|
||||||
|
}
|
||||||
|
|
||||||
|
public delete(k:string) {
|
||||||
|
delete this.contents[k];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MockSigner {
|
||||||
|
onsign: (string) => void
|
||||||
|
onverify: (boolean) => void
|
||||||
|
public verify(src:string, signature:Signature) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public sign(s:string):boolean {
|
||||||
|
this.onsign('there would be a signature here');
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public prepare(m:Signable):boolean {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public fingerprint():string {
|
||||||
|
return '';
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MockConsumer implements SubConsumer {
|
||||||
|
|
||||||
|
omnoms: Array<string>
|
||||||
|
|
||||||
|
constructor() {
|
||||||
|
this.omnoms = Array<string>();
|
||||||
|
}
|
||||||
|
|
||||||
|
public post(v:string) {
|
||||||
|
this.omnoms.push(v);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MockPubSub implements PubSub {
|
||||||
|
|
||||||
|
pubs: Array<string>
|
||||||
|
consumer: SubConsumer
|
||||||
|
|
||||||
|
constructor(name:string, consumer:SubConsumer) {
|
||||||
|
this.pubs = Array<string>();
|
||||||
|
this.consumer = consumer;
|
||||||
|
}
|
||||||
|
|
||||||
|
public pub(v:string): boolean {
|
||||||
|
this.pubs.push(v);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public flush() {
|
||||||
|
while (this.pubs.length > 0) {
|
||||||
|
const s = this.pubs.shift();
|
||||||
|
this.consumer.post(s);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public close() {
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MockSignable implements Signable {
|
||||||
|
|
||||||
|
src: string
|
||||||
|
dst: string
|
||||||
|
|
||||||
|
constructor(src:string) {
|
||||||
|
this.src = src;
|
||||||
|
}
|
||||||
|
|
||||||
|
public digest():string {
|
||||||
|
const h = crypto.createHash('sha256');
|
||||||
|
h.update(this.src);
|
||||||
|
this.dst= h.digest('hex');
|
||||||
|
return this.dst;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
class MockKeyStore implements KeyStore {
|
||||||
|
|
||||||
|
pk: any
|
||||||
|
pubks: Array<any>
|
||||||
|
|
||||||
|
constructor(pk:any, pubks:Array<any>) {
|
||||||
|
this.pk = pk;
|
||||||
|
this.pubks = pubks;
|
||||||
|
}
|
||||||
|
|
||||||
|
public getPrivateKey(): any {
|
||||||
|
return this.pk;
|
||||||
|
}
|
||||||
|
|
||||||
|
public getTrustedKeys(): Array<any> {
|
||||||
|
return this.pubks;
|
||||||
|
}
|
||||||
|
|
||||||
|
public getTrustedActiveKeys(): Array<any> {
|
||||||
|
return [];
|
||||||
|
}
|
||||||
|
|
||||||
|
public getEncryptKeys(): Array<any> {
|
||||||
|
return [];
|
||||||
|
}
|
||||||
|
|
||||||
|
public getFingerprint(): string {
|
||||||
|
return '';
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
MockStore,
|
||||||
|
MockPubSub,
|
||||||
|
MockConsumer,
|
||||||
|
MockSignable,
|
||||||
|
MockKeyStore,
|
||||||
|
MockSigner,
|
||||||
|
};
|
||||||
@@ -1,10 +1,13 @@
|
|||||||
|
import Automerge = require('automerge');
|
||||||
import assert = require('assert');
|
import assert = require('assert');
|
||||||
import fs = require('fs');
|
import fs = require('fs');
|
||||||
import pgp = require('openpgp');
|
import pgp = require('openpgp');
|
||||||
import sqlite = require('sqlite3');
|
import sqlite = require('sqlite3');
|
||||||
|
|
||||||
import * as handlers from '../scripts/server/handlers';
|
import * as handlers from '../scripts/server/handlers';
|
||||||
import { Envelope, Syncable, ArgPair, PGPKeyStore, PGPSigner, KeyStore, Signer, SqliteAdapter } from 'crdt-meta';
|
import { Envelope, Syncable, ArgPair } from '../src/sync';
|
||||||
|
import { PGPKeyStore, PGPSigner, KeyStore, Signer } from '../src/auth';
|
||||||
|
import { SqliteAdapter } from '../src/db';
|
||||||
|
|
||||||
function createKeystore() {
|
function createKeystore() {
|
||||||
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc', 'utf-8');
|
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc', 'utf-8');
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ def define_account_tx_metadata(user: Account):
|
|||||||
)
|
)
|
||||||
key = generate_metadata_pointer(
|
key = generate_metadata_pointer(
|
||||||
identifier=identifier,
|
identifier=identifier,
|
||||||
cic_type=':cic.person'
|
cic_type='cic.person'
|
||||||
)
|
)
|
||||||
account_metadata = get_cached_data(key=key)
|
account_metadata = get_cached_data(key=key)
|
||||||
|
|
||||||
|
|||||||
@@ -80,7 +80,7 @@ def get_cached_operational_balance(blockchain_address: str):
|
|||||||
"""
|
"""
|
||||||
key = create_cached_data_key(
|
key = create_cached_data_key(
|
||||||
identifier=bytes.fromhex(blockchain_address[2:]),
|
identifier=bytes.fromhex(blockchain_address[2:]),
|
||||||
salt=':cic.balances_data'
|
salt='cic.balances_data'
|
||||||
)
|
)
|
||||||
cached_balance = get_cached_data(key=key)
|
cached_balance = get_cached_data(key=key)
|
||||||
if cached_balance:
|
if cached_balance:
|
||||||
|
|||||||
@@ -38,13 +38,3 @@ class MetadataStoreError(Exception):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class SeppukuError(Exception):
|
|
||||||
"""Exception base class for all errors that should cause system shutdown"""
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class InitializationError(Exception):
|
|
||||||
"""Exception raised when initialization state is insufficient to run component"""
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user