Compare commits
2 Commits
spencer/re
...
lash/impro
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
c3ca86fcf4
|
||
|
|
e1d4ddae39
|
4
.gitignore
vendored
4
.gitignore
vendored
@@ -1,6 +1,2 @@
|
||||
service-configs/*
|
||||
!service-configs/.gitkeep
|
||||
**/node_modules/
|
||||
__pycache__
|
||||
*.pyc
|
||||
*.o
|
||||
|
||||
@@ -67,7 +67,6 @@ class ERC20TransferFilter(SyncFilter):
|
||||
tx.status == Status.SUCCESS,
|
||||
block.timestamp,
|
||||
)
|
||||
#db_session.flush()
|
||||
db_session.commit()
|
||||
db_session.flush()
|
||||
|
||||
return True
|
||||
|
||||
@@ -26,7 +26,7 @@ from chainlib.eth.block import (
|
||||
from hexathon import (
|
||||
strip_0x,
|
||||
)
|
||||
from chainsyncer.backend.sql import SQLBackend
|
||||
from chainsyncer.backend import SyncerBackend
|
||||
from chainsyncer.driver import (
|
||||
HeadSyncer,
|
||||
HistorySyncer,
|
||||
@@ -71,13 +71,13 @@ def main():
|
||||
|
||||
syncers = []
|
||||
|
||||
#if SQLBackend.first(chain_spec):
|
||||
# backend = SQLBackend.initial(chain_spec, block_offset)
|
||||
syncer_backends = SQLBackend.resume(chain_spec, block_offset)
|
||||
#if SyncerBackend.first(chain_spec):
|
||||
# backend = SyncerBackend.initial(chain_spec, block_offset)
|
||||
syncer_backends = SyncerBackend.resume(chain_spec, block_offset)
|
||||
|
||||
if len(syncer_backends) == 0:
|
||||
logg.info('found no backends to resume')
|
||||
syncer_backends.append(SQLBackend.initial(chain_spec, block_offset))
|
||||
syncers.append(SyncerBackend.initial(chain_spec, block_offset))
|
||||
else:
|
||||
for syncer_backend in syncer_backends:
|
||||
logg.info('resuming sync session {}'.format(syncer_backend))
|
||||
@@ -85,7 +85,7 @@ def main():
|
||||
for syncer_backend in syncer_backends:
|
||||
syncers.append(HistorySyncer(syncer_backend))
|
||||
|
||||
syncer_backend = SQLBackend.live(chain_spec, block_offset+1)
|
||||
syncer_backend = SyncerBackend.live(chain_spec, block_offset+1)
|
||||
syncers.append(HeadSyncer(syncer_backend))
|
||||
|
||||
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
||||
|
||||
@@ -17,7 +17,7 @@ RUN apt-get update && \
|
||||
|
||||
# Copy shared requirements from top of mono-repo
|
||||
RUN echo "copying root req file ${root_requirement_file}"
|
||||
RUN pip install $pip_extra_index_url_flag cic-base[full_graph]==0.1.2a76
|
||||
RUN pip install $pip_extra_index_url_flag cic-base[full_graph]==0.1.2a58
|
||||
|
||||
COPY cic-cache/requirements.txt ./
|
||||
COPY cic-cache/setup.cfg \
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
cic-base~=0.1.2a77
|
||||
cic-base~=0.1.2a76
|
||||
alembic==1.4.2
|
||||
confini~=0.3.6rc3
|
||||
uwsgi==2.0.19.1
|
||||
@@ -9,4 +9,4 @@ semver==2.13.0
|
||||
psycopg2==2.8.6
|
||||
celery==4.4.7
|
||||
redis==3.5.3
|
||||
chainsyncer[sql]~=0.0.2a2
|
||||
chainsyncer[sql]~=0.0.2a1
|
||||
|
||||
@@ -1,10 +1,6 @@
|
||||
# external imports
|
||||
# extended imports
|
||||
from chainlib.eth.constant import ZERO_ADDRESS
|
||||
from chainlib.status import Status as TxStatus
|
||||
from cic_eth_registry.erc20 import ERC20Token
|
||||
|
||||
# local imports
|
||||
from cic_eth.ext.address import translate_address
|
||||
|
||||
|
||||
class ExtendedTx:
|
||||
@@ -31,12 +27,12 @@ class ExtendedTx:
|
||||
self.status_code = TxStatus.PENDING.value
|
||||
|
||||
|
||||
def set_actors(self, sender, recipient, trusted_declarator_addresses=None, caller_address=ZERO_ADDRESS):
|
||||
def set_actors(self, sender, recipient, trusted_declarator_addresses=None):
|
||||
self.sender = sender
|
||||
self.recipient = recipient
|
||||
if trusted_declarator_addresses != None:
|
||||
self.sender_label = translate_address(sender, trusted_declarator_addresses, self.chain_spec, sender_address=caller_address)
|
||||
self.recipient_label = translate_address(recipient, trusted_declarator_addresses, self.chain_spec, sender_address=caller_address)
|
||||
self.sender_label = translate_address(sender, trusted_declarator_addresses, self.chain_spec)
|
||||
self.recipient_label = translate_address(recipient, trusted_declarator_addresses, self.chain_spec)
|
||||
|
||||
|
||||
def set_tokens(self, source, source_value, destination=None, destination_value=None):
|
||||
@@ -44,8 +40,8 @@ class ExtendedTx:
|
||||
destination = source
|
||||
if destination_value == None:
|
||||
destination_value = source_value
|
||||
st = ERC20Token(self.chain_spec, self.rpc, source)
|
||||
dt = ERC20Token(self.chain_spec, self.rpc, destination)
|
||||
st = ERC20Token(self.rpc, source)
|
||||
dt = ERC20Token(self.rpc, destination)
|
||||
self.source_token = source
|
||||
self.source_token_symbol = st.symbol
|
||||
self.source_token_name = st.name
|
||||
@@ -66,10 +62,10 @@ class ExtendedTx:
|
||||
self.status_code = n
|
||||
|
||||
|
||||
def asdict(self):
|
||||
def to_dict(self):
|
||||
o = {}
|
||||
for attr in dir(self):
|
||||
if attr[0] == '_' or attr in ['set_actors', 'set_tokens', 'set_status', 'asdict', 'rpc']:
|
||||
if attr[0] == '_' or attr in ['set_actors', 'set_tokens', 'set_status', 'to_dict']:
|
||||
continue
|
||||
o[attr] = getattr(self, attr)
|
||||
return o
|
||||
|
||||
@@ -15,6 +15,7 @@ from cic_eth_registry import CICRegistry
|
||||
from chainlib.chain import ChainSpec
|
||||
from chainlib.eth.tx import unpack
|
||||
from chainlib.connection import RPCConnection
|
||||
from chainsyncer.error import SyncDone
|
||||
from hexathon import strip_0x
|
||||
from chainqueue.db.enum import (
|
||||
StatusEnum,
|
||||
@@ -152,7 +153,10 @@ class DispatchSyncer:
|
||||
def main():
|
||||
syncer = DispatchSyncer(chain_spec)
|
||||
conn = RPCConnection.connect(chain_spec, 'default')
|
||||
syncer.loop(conn, float(config.get('DISPATCHER_LOOP_INTERVAL')))
|
||||
try:
|
||||
syncer.loop(conn, float(config.get('DISPATCHER_LOOP_INTERVAL')))
|
||||
except SyncDone as e:
|
||||
sys.stderr.write("dispatcher done at block {}\n".format(e))
|
||||
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
# standard imports
|
||||
import logging
|
||||
|
||||
# external imports
|
||||
# third-party imports
|
||||
import celery
|
||||
from cic_eth_registry.error import UnknownContractError
|
||||
from chainlib.status import Status as TxStatus
|
||||
@@ -9,13 +9,7 @@ from chainlib.eth.address import to_checksum_address
|
||||
from chainlib.eth.error import RequestMismatchException
|
||||
from chainlib.eth.constant import ZERO_ADDRESS
|
||||
from chainlib.eth.erc20 import ERC20
|
||||
from hexathon import (
|
||||
strip_0x,
|
||||
add_0x,
|
||||
)
|
||||
# TODO: use sarafu_Faucet for both when inheritance has been implemented
|
||||
from erc20_single_shot_faucet import SingleShotFaucet
|
||||
from sarafu_faucet import MinterFaucet as Faucet
|
||||
from hexathon import strip_0x
|
||||
|
||||
# local imports
|
||||
from .base import SyncFilter
|
||||
@@ -24,73 +18,65 @@ from cic_eth.eth.meta import ExtendedTx
|
||||
logg = logging.getLogger().getChild(__name__)
|
||||
|
||||
|
||||
def parse_transfer(tx):
|
||||
r = ERC20.parse_transfer_request(tx.payload)
|
||||
transfer_data = {}
|
||||
transfer_data['to'] = r[0]
|
||||
transfer_data['value'] = r[1]
|
||||
transfer_data['from'] = tx['from']
|
||||
transfer_data['token_address'] = tx['to']
|
||||
return ('transfer', transfer_data)
|
||||
|
||||
|
||||
def parse_transferfrom(tx):
|
||||
r = ERC20.parse_transfer_request(tx.payload)
|
||||
transfer_data = unpack_transferfrom(tx.payload)
|
||||
transfer_data['from'] = r[0]
|
||||
transfer_data['to'] = r[1]
|
||||
transfer_data['value'] = r[2]
|
||||
transfer_data['token_address'] = tx['to']
|
||||
return ('transferfrom', transfer_data)
|
||||
|
||||
|
||||
def parse_giftto(tx):
|
||||
# TODO: broken
|
||||
logg.error('broken')
|
||||
return
|
||||
transfer_data = unpack_gift(tx.payload)
|
||||
transfer_data['from'] = tx.inputs[0]
|
||||
transfer_data['value'] = 0
|
||||
transfer_data['token_address'] = ZERO_ADDRESS
|
||||
# TODO: would be better to query the gift amount from the block state
|
||||
for l in tx.logs:
|
||||
topics = l['topics']
|
||||
logg.debug('topixx {}'.format(topics))
|
||||
if strip_0x(topics[0]) == '45c201a59ac545000ead84f30b2db67da23353aa1d58ac522c48505412143ffa':
|
||||
#transfer_data['value'] = web3.Web3.toInt(hexstr=strip_0x(l['data']))
|
||||
transfer_data['value'] = int.from_bytes(bytes.fromhex(strip_0x(l_data)))
|
||||
#token_address_bytes = topics[2][32-20:]
|
||||
token_address = strip_0x(topics[2])[64-40:]
|
||||
transfer_data['token_address'] = to_checksum_address(token_address)
|
||||
return ('tokengift', transfer_data)
|
||||
|
||||
|
||||
class CallbackFilter(SyncFilter):
|
||||
|
||||
trusted_addresses = []
|
||||
|
||||
def __init__(self, chain_spec, method, queue, caller_address=ZERO_ADDRESS):
|
||||
def __init__(self, chain_spec, method, queue):
|
||||
self.queue = queue
|
||||
self.method = method
|
||||
self.chain_spec = chain_spec
|
||||
self.caller_address = caller_address
|
||||
|
||||
|
||||
def parse_transfer(self, tx, conn):
|
||||
if not tx.payload:
|
||||
return (None, None)
|
||||
r = ERC20.parse_transfer_request(tx.payload)
|
||||
transfer_data = {}
|
||||
transfer_data['to'] = r[0]
|
||||
transfer_data['value'] = r[1]
|
||||
transfer_data['from'] = tx.outputs[0]
|
||||
transfer_data['token_address'] = tx.inputs[0]
|
||||
return ('transfer', transfer_data)
|
||||
|
||||
|
||||
def parse_transferfrom(self, tx, conn):
|
||||
if not tx.payload:
|
||||
return (None, None)
|
||||
r = ERC20.parse_transfer_from_request(tx.payload)
|
||||
transfer_data = {}
|
||||
transfer_data['from'] = r[0]
|
||||
transfer_data['to'] = r[1]
|
||||
transfer_data['value'] = r[2]
|
||||
transfer_data['token_address'] = tx.inputs[0]
|
||||
return ('transferfrom', transfer_data)
|
||||
|
||||
|
||||
def parse_giftto(self, tx, conn):
|
||||
if not tx.payload:
|
||||
return (None, None)
|
||||
r = Faucet.parse_give_to_request(tx.payload)
|
||||
transfer_data = {}
|
||||
transfer_data['to'] = r[0]
|
||||
transfer_data['value'] = tx.value
|
||||
transfer_data['from'] = tx.outputs[0]
|
||||
#transfer_data['token_address'] = tx.inputs[0]
|
||||
faucet_contract = tx.inputs[0]
|
||||
|
||||
c = SingleShotFaucet(self.chain_spec)
|
||||
o = c.token(faucet_contract, sender_address=self.caller_address)
|
||||
r = conn.do(o)
|
||||
transfer_data['token_address'] = add_0x(c.parse_token(r))
|
||||
|
||||
o = c.amount(faucet_contract, sender_address=self.caller_address)
|
||||
r = conn.do(o)
|
||||
transfer_data['value'] = c.parse_amount(r)
|
||||
|
||||
return ('tokengift', transfer_data)
|
||||
|
||||
|
||||
def call_back(self, transfer_type, result):
|
||||
result['chain_spec'] = result['chain_spec'].asdict()
|
||||
logg.debug('result {}'.format(result))
|
||||
s = celery.signature(
|
||||
self.method,
|
||||
[
|
||||
result,
|
||||
transfer_type,
|
||||
int(result['status_code'] != 0),
|
||||
int(result['status_code'] == 0),
|
||||
],
|
||||
queue=self.queue,
|
||||
)
|
||||
@@ -106,29 +92,26 @@ class CallbackFilter(SyncFilter):
|
||||
# s_translate.link(s)
|
||||
# s_translate.apply_async()
|
||||
t = s.apply_async()
|
||||
return t
|
||||
return s
|
||||
|
||||
|
||||
def parse_data(self, tx, conn):
|
||||
def parse_data(self, tx):
|
||||
transfer_type = None
|
||||
transfer_data = None
|
||||
# TODO: what's with the mix of attributes and dict keys
|
||||
logg.debug('have payload {}'.format(tx.payload))
|
||||
method_signature = tx.payload[:8]
|
||||
|
||||
logg.debug('tx status {}'.format(tx.status))
|
||||
|
||||
for parser in [
|
||||
self.parse_transfer,
|
||||
self.parse_transferfrom,
|
||||
self.parse_giftto,
|
||||
parse_transfer,
|
||||
parse_transferfrom,
|
||||
parse_giftto,
|
||||
]:
|
||||
try:
|
||||
if tx:
|
||||
(transfer_type, transfer_data) = parser(tx, conn)
|
||||
if transfer_type == None:
|
||||
continue
|
||||
else:
|
||||
pass
|
||||
(transfer_type, transfer_data) = parser(tx)
|
||||
break
|
||||
except RequestMismatchException:
|
||||
continue
|
||||
|
||||
@@ -145,7 +128,7 @@ class CallbackFilter(SyncFilter):
|
||||
transfer_data = None
|
||||
transfer_type = None
|
||||
try:
|
||||
(transfer_type, transfer_data) = self.parse_data(tx, conn)
|
||||
(transfer_type, transfer_data) = self.parse_data(tx)
|
||||
except TypeError:
|
||||
logg.debug('invalid method data length for tx {}'.format(tx.hash))
|
||||
return
|
||||
@@ -161,17 +144,16 @@ class CallbackFilter(SyncFilter):
|
||||
result = None
|
||||
try:
|
||||
tokentx = ExtendedTx(conn, tx.hash, self.chain_spec)
|
||||
tokentx.set_actors(transfer_data['from'], transfer_data['to'], self.trusted_addresses, caller_address=self.caller_address)
|
||||
tokentx.set_actors(transfer_data['from'], transfer_data['to'], self.trusted_addresses)
|
||||
tokentx.set_tokens(transfer_data['token_address'], transfer_data['value'])
|
||||
if transfer_data['status'] == 0:
|
||||
tokentx.set_status(1)
|
||||
else:
|
||||
tokentx.set_status(0)
|
||||
result = tokentx.asdict()
|
||||
t = self.call_back(transfer_type, result)
|
||||
logg.info('callback success task id {} tx {} queue {}'.format(t, tx.hash, t.queue))
|
||||
t = self.call_back(transfer_type, tokentx.to_dict())
|
||||
logg.info('callback success task id {} tx {}'.format(t, tx.hash))
|
||||
except UnknownContractError:
|
||||
logg.debug('callback filter {}:{} skipping "transfer" method on unknown contract {} tx {}'.format(tx.queue, tx.method, transfer_data['to'], tx.hash))
|
||||
logg.debug('callback filter {}:{} skipping "transfer" method on unknown contract {} tx {}'.format(tc.queue, tc.method, transfer_data['to'], tx.hash))
|
||||
|
||||
|
||||
def __str__(self):
|
||||
|
||||
@@ -39,7 +39,6 @@ from cic_eth.queue import (
|
||||
from cic_eth.callbacks import (
|
||||
Callback,
|
||||
http,
|
||||
noop,
|
||||
#tcp,
|
||||
redis,
|
||||
)
|
||||
|
||||
@@ -15,6 +15,7 @@ import cic_base.config
|
||||
import cic_base.log
|
||||
import cic_base.argparse
|
||||
import cic_base.rpc
|
||||
from cic_eth_registry import CICRegistry
|
||||
from cic_eth_registry.error import UnknownContractError
|
||||
from chainlib.chain import ChainSpec
|
||||
from chainlib.eth.constant import ZERO_ADDRESS
|
||||
@@ -25,7 +26,7 @@ from chainlib.eth.block import (
|
||||
from hexathon import (
|
||||
strip_0x,
|
||||
)
|
||||
from chainsyncer.backend.sql import SQLBackend
|
||||
from chainsyncer.backend import SyncerBackend
|
||||
from chainsyncer.driver import (
|
||||
HeadSyncer,
|
||||
HistorySyncer,
|
||||
@@ -42,12 +43,6 @@ from cic_eth.runnable.daemons.filters import (
|
||||
TransferAuthFilter,
|
||||
)
|
||||
from cic_eth.stat import init_chain_stat
|
||||
from cic_eth.registry import (
|
||||
connect as connect_registry,
|
||||
connect_declarator,
|
||||
connect_token_registry,
|
||||
)
|
||||
|
||||
|
||||
script_dir = os.path.realpath(os.path.dirname(__file__))
|
||||
|
||||
@@ -93,18 +88,18 @@ def main():
|
||||
|
||||
syncers = []
|
||||
|
||||
#if SQLBackend.first(chain_spec):
|
||||
# backend = SQLBackend.initial(chain_spec, block_offset)
|
||||
syncer_backends = SQLBackend.resume(chain_spec, block_offset)
|
||||
#if SyncerBackend.first(chain_spec):
|
||||
# backend = SyncerBackend.initial(chain_spec, block_offset)
|
||||
syncer_backends = SyncerBackend.resume(chain_spec, block_offset)
|
||||
|
||||
if len(syncer_backends) == 0:
|
||||
logg.info('found no backends to resume')
|
||||
syncer_backends.append(SQLBackend.initial(chain_spec, block_offset))
|
||||
syncer_backends.append(SyncerBackend.initial(chain_spec, block_offset))
|
||||
else:
|
||||
for syncer_backend in syncer_backends:
|
||||
logg.info('resuming sync session {}'.format(syncer_backend))
|
||||
|
||||
syncer_backends.append(SQLBackend.live(chain_spec, block_offset+1))
|
||||
syncer_backends.append(SyncerBackend.live(chain_spec, block_offset+1))
|
||||
|
||||
for syncer_backend in syncer_backends:
|
||||
try:
|
||||
@@ -114,8 +109,6 @@ def main():
|
||||
logg.info('Initializing HEAD syncer on backend {}'.format(syncer_backend))
|
||||
syncers.append(HeadSyncer(syncer_backend))
|
||||
|
||||
connect_registry(rpc, chain_spec, config.get('CIC_REGISTRY_ADDRESS'))
|
||||
|
||||
trusted_addresses_src = config.get('CIC_TRUST_ADDRESS')
|
||||
if trusted_addresses_src == None:
|
||||
logg.critical('At least one trusted address must be declared in CIC_TRUST_ADDRESS')
|
||||
@@ -123,8 +116,6 @@ def main():
|
||||
trusted_addresses = trusted_addresses_src.split(',')
|
||||
for address in trusted_addresses:
|
||||
logg.info('using trusted address {}'.format(address))
|
||||
connect_declarator(rpc, chain_spec, trusted_addresses)
|
||||
connect_token_registry(rpc, chain_spec)
|
||||
CallbackFilter.trusted_addresses = trusted_addresses
|
||||
|
||||
callback_filters = []
|
||||
|
||||
@@ -4,7 +4,7 @@ import datetime
|
||||
|
||||
# external imports
|
||||
from chainsyncer.driver import HeadSyncer
|
||||
from chainsyncer.backend.memory import MemBackend
|
||||
from chainsyncer.backend import MemBackend
|
||||
from chainsyncer.error import NoBlockForYou
|
||||
from chainlib.eth.block import (
|
||||
block_by_number,
|
||||
|
||||
@@ -10,7 +10,7 @@ version = (
|
||||
0,
|
||||
11,
|
||||
0,
|
||||
'beta.6',
|
||||
'beta.3',
|
||||
)
|
||||
|
||||
version_object = semver.VersionInfo(
|
||||
|
||||
@@ -29,7 +29,7 @@ RUN /usr/local/bin/python -m pip install --upgrade pip
|
||||
# python merge_requirements.py | tee merged_requirements.txt
|
||||
#RUN cd cic-base && \
|
||||
# pip install $pip_extra_index_url_flag -r ./merged_requirements.txt
|
||||
RUN pip install $pip_extra_index_url_flag cic-base[full_graph]==0.1.2a77
|
||||
RUN pip install $pip_extra_index_url_flag cic-base[full_graph]==0.1.2a62
|
||||
|
||||
COPY cic-eth/scripts/ scripts/
|
||||
COPY cic-eth/setup.cfg cic-eth/setup.py ./
|
||||
|
||||
@@ -1,25 +1,25 @@
|
||||
cic-base~=0.1.2a76
|
||||
cic-base~=0.1.2a67
|
||||
celery==4.4.7
|
||||
crypto-dev-signer~=0.4.14b2
|
||||
crypto-dev-signer~=0.4.14a17
|
||||
confini~=0.3.6rc3
|
||||
cic-eth-registry~=0.5.4a16
|
||||
cic-eth-registry~=0.5.4a13
|
||||
#cic-bancor~=0.0.6
|
||||
redis==3.5.3
|
||||
alembic==1.4.2
|
||||
websockets==8.1
|
||||
requests~=2.24.0
|
||||
eth_accounts_index~=0.0.11a9
|
||||
erc20-transfer-authorization~=0.3.1a5
|
||||
eth_accounts_index~=0.0.11a7
|
||||
erc20-transfer-authorization~=0.3.1a3
|
||||
#simple-rlp==0.1.2
|
||||
uWSGI==2.0.19.1
|
||||
semver==2.13.0
|
||||
websocket-client==0.57.0
|
||||
moolb~=0.1.1b2
|
||||
eth-address-index~=0.1.1a9
|
||||
chainlib~=0.0.2a13
|
||||
eth-address-index~=0.1.1a7
|
||||
chainlib~=0.0.2a5
|
||||
hexathon~=0.0.1a7
|
||||
chainsyncer[sql]~=0.0.2a2
|
||||
chainsyncer[sql]~=0.0.2a1
|
||||
chainqueue~=0.0.1a7
|
||||
pysha3==1.0.2
|
||||
coincurve==15.0.0
|
||||
sarafu-faucet==0.0.2a28
|
||||
potaahto~=0.0.1a1
|
||||
sarafu-faucet~=0.0.2a19
|
||||
|
||||
@@ -4,4 +4,4 @@ pytest-mock==3.3.1
|
||||
pytest-cov==2.10.1
|
||||
eth-tester==0.5.0b3
|
||||
py-evm==0.3.0a20
|
||||
giftable-erc20-token==0.0.8a9
|
||||
giftable-erc20-token==0.0.8a4
|
||||
|
||||
@@ -1,223 +0,0 @@
|
||||
# standard import
|
||||
import logging
|
||||
import datetime
|
||||
import os
|
||||
|
||||
# external imports
|
||||
import pytest
|
||||
from chainlib.connection import RPCConnection
|
||||
from chainlib.eth.nonce import RPCNonceOracle
|
||||
from chainlib.eth.gas import OverrideGasOracle
|
||||
from chainlib.eth.tx import (
|
||||
receipt,
|
||||
transaction,
|
||||
Tx,
|
||||
)
|
||||
from chainlib.eth.block import Block
|
||||
from chainlib.eth.erc20 import ERC20
|
||||
from sarafu_faucet import MinterFaucet
|
||||
from eth_accounts_index import AccountRegistry
|
||||
from potaahto.symbols import snake_and_camel
|
||||
from hexathon import add_0x
|
||||
|
||||
# local imports
|
||||
from cic_eth.runnable.daemons.filters.callback import CallbackFilter
|
||||
|
||||
logg = logging.getLogger()
|
||||
|
||||
|
||||
@pytest.mark.skip()
|
||||
def test_transfer_tx(
|
||||
default_chain_spec,
|
||||
init_database,
|
||||
eth_rpc,
|
||||
eth_signer,
|
||||
foo_token,
|
||||
agent_roles,
|
||||
token_roles,
|
||||
contract_roles,
|
||||
celery_session_worker,
|
||||
):
|
||||
|
||||
rpc = RPCConnection.connect(default_chain_spec, 'default')
|
||||
nonce_oracle = RPCNonceOracle(token_roles['FOO_TOKEN_OWNER'], rpc)
|
||||
gas_oracle = OverrideGasOracle(conn=rpc, limit=200000)
|
||||
|
||||
txf = ERC20(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
(tx_hash_hex, o) = txf.transfer(foo_token, token_roles['FOO_TOKEN_OWNER'], agent_roles['ALICE'], 1024)
|
||||
r = rpc.do(o)
|
||||
|
||||
o = transaction(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
logg.debug(r)
|
||||
tx_src = snake_and_camel(r)
|
||||
tx = Tx(tx_src)
|
||||
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
rcpt = snake_and_camel(r)
|
||||
tx.apply_receipt(rcpt)
|
||||
|
||||
fltr = CallbackFilter(default_chain_spec, None, None, caller_address=contract_roles['CONTRACT_DEPLOYER'])
|
||||
(transfer_type, transfer_data) = fltr.parse_transfer(tx, eth_rpc)
|
||||
|
||||
assert transfer_type == 'transfer'
|
||||
|
||||
|
||||
@pytest.mark.skip()
|
||||
def test_transfer_from_tx(
|
||||
default_chain_spec,
|
||||
init_database,
|
||||
eth_rpc,
|
||||
eth_signer,
|
||||
foo_token,
|
||||
agent_roles,
|
||||
token_roles,
|
||||
contract_roles,
|
||||
celery_session_worker,
|
||||
):
|
||||
|
||||
rpc = RPCConnection.connect(default_chain_spec, 'default')
|
||||
nonce_oracle = RPCNonceOracle(token_roles['FOO_TOKEN_OWNER'], rpc)
|
||||
gas_oracle = OverrideGasOracle(conn=rpc, limit=200000)
|
||||
|
||||
txf = ERC20(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
|
||||
(tx_hash_hex, o) = txf.approve(foo_token, token_roles['FOO_TOKEN_OWNER'], agent_roles['ALICE'], 1024)
|
||||
r = rpc.do(o)
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
nonce_oracle = RPCNonceOracle(agent_roles['ALICE'], rpc)
|
||||
txf = ERC20(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
(tx_hash_hex, o) = txf.transfer_from(foo_token, agent_roles['ALICE'], token_roles['FOO_TOKEN_OWNER'], agent_roles['BOB'], 1024)
|
||||
r = rpc.do(o)
|
||||
|
||||
o = transaction(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
tx_src = snake_and_camel(r)
|
||||
tx = Tx(tx_src)
|
||||
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
rcpt = snake_and_camel(r)
|
||||
tx.apply_receipt(rcpt)
|
||||
|
||||
fltr = CallbackFilter(default_chain_spec, None, None, caller_address=contract_roles['CONTRACT_DEPLOYER'])
|
||||
(transfer_type, transfer_data) = fltr.parse_transferfrom(tx, eth_rpc)
|
||||
|
||||
assert transfer_type == 'transferfrom'
|
||||
|
||||
|
||||
def test_faucet_gift_to_tx(
|
||||
default_chain_spec,
|
||||
init_database,
|
||||
eth_rpc,
|
||||
eth_signer,
|
||||
foo_token,
|
||||
agent_roles,
|
||||
contract_roles,
|
||||
faucet,
|
||||
account_registry,
|
||||
celery_session_worker,
|
||||
):
|
||||
|
||||
rpc = RPCConnection.connect(default_chain_spec, 'default')
|
||||
gas_oracle = OverrideGasOracle(conn=rpc, limit=800000)
|
||||
|
||||
nonce_oracle = RPCNonceOracle(contract_roles['ACCOUNT_REGISTRY_WRITER'], rpc)
|
||||
txf = AccountRegistry(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
(tx_hash_hex, o) = txf.add(account_registry, contract_roles['ACCOUNT_REGISTRY_WRITER'], agent_roles['ALICE'])
|
||||
r = rpc.do(o)
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
nonce_oracle = RPCNonceOracle(agent_roles['ALICE'], rpc)
|
||||
txf = MinterFaucet(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
(tx_hash_hex, o) = txf.give_to(faucet, agent_roles['ALICE'], agent_roles['ALICE'])
|
||||
r = rpc.do(o)
|
||||
|
||||
o = transaction(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
tx_src = snake_and_camel(r)
|
||||
tx = Tx(tx_src)
|
||||
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
rcpt = snake_and_camel(r)
|
||||
tx.apply_receipt(rcpt)
|
||||
|
||||
fltr = CallbackFilter(default_chain_spec, None, None, caller_address=contract_roles['CONTRACT_DEPLOYER'])
|
||||
(transfer_type, transfer_data) = fltr.parse_giftto(tx, eth_rpc)
|
||||
|
||||
assert transfer_type == 'tokengift'
|
||||
assert transfer_data['token_address'] == foo_token
|
||||
|
||||
|
||||
def test_callback_filter(
|
||||
default_chain_spec,
|
||||
init_database,
|
||||
eth_rpc,
|
||||
eth_signer,
|
||||
foo_token,
|
||||
token_roles,
|
||||
agent_roles,
|
||||
contract_roles,
|
||||
register_lookups,
|
||||
):
|
||||
|
||||
rpc = RPCConnection.connect(default_chain_spec, 'default')
|
||||
nonce_oracle = RPCNonceOracle(token_roles['FOO_TOKEN_OWNER'], rpc)
|
||||
gas_oracle = OverrideGasOracle(conn=rpc, limit=200000)
|
||||
|
||||
txf = ERC20(default_chain_spec, signer=eth_signer, nonce_oracle=nonce_oracle, gas_oracle=gas_oracle)
|
||||
(tx_hash_hex, o) = txf.transfer(foo_token, token_roles['FOO_TOKEN_OWNER'], agent_roles['ALICE'], 1024)
|
||||
r = rpc.do(o)
|
||||
|
||||
o = transaction(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
logg.debug(r)
|
||||
|
||||
mockblock_src = {
|
||||
'hash': add_0x(os.urandom(32).hex()),
|
||||
'number': '0x2a',
|
||||
'transactions': [tx_hash_hex],
|
||||
'timestamp': datetime.datetime.utcnow().timestamp(),
|
||||
}
|
||||
mockblock = Block(mockblock_src)
|
||||
|
||||
tx_src = snake_and_camel(r)
|
||||
tx = Tx(tx_src, block=mockblock)
|
||||
|
||||
o = receipt(tx_hash_hex)
|
||||
r = rpc.do(o)
|
||||
assert r['status'] == 1
|
||||
|
||||
rcpt = snake_and_camel(r)
|
||||
tx.apply_receipt(rcpt)
|
||||
|
||||
fltr = CallbackFilter(default_chain_spec, None, None, caller_address=contract_roles['CONTRACT_DEPLOYER'])
|
||||
|
||||
class CallbackMock:
|
||||
|
||||
def __init__(self):
|
||||
self.results = {}
|
||||
|
||||
def call_back(self, transfer_type, result):
|
||||
self.results[transfer_type] = result
|
||||
|
||||
mock = CallbackMock()
|
||||
fltr.call_back = mock.call_back
|
||||
|
||||
fltr.filter(eth_rpc, mockblock, tx, init_database)
|
||||
|
||||
assert mock.results.get('transfer') != None
|
||||
assert mock.results['transfer']['destination_token'] == foo_token
|
||||
1
apps/cic-meta/.gitignore
vendored
1
apps/cic-meta/.gitignore
vendored
@@ -3,3 +3,4 @@ dist
|
||||
dist-web
|
||||
dist-server
|
||||
scratch
|
||||
tests
|
||||
|
||||
101
apps/cic-meta/package-lock.json
generated
101
apps/cic-meta/package-lock.json
generated
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "cic-client-meta",
|
||||
"version": "0.0.7-alpha.7",
|
||||
"version": "0.0.7-alpha.2",
|
||||
"lockfileVersion": 1,
|
||||
"requires": true,
|
||||
"dependencies": {
|
||||
@@ -852,75 +852,6 @@
|
||||
"printj": "~1.1.0"
|
||||
}
|
||||
},
|
||||
"crdt-meta": {
|
||||
"version": "0.0.8",
|
||||
"resolved": "https://registry.npmjs.org/crdt-meta/-/crdt-meta-0.0.8.tgz",
|
||||
"integrity": "sha512-CS0sS0L2QWthz7vmu6vzl3p4kcpJ+IKILBJ4tbgN4A3iNG8wnBeuDIv/z3KFFQjcfuP4QAh6E9LywKUTxtDc3g==",
|
||||
"requires": {
|
||||
"automerge": "^0.14.2",
|
||||
"ini": "^1.3.8",
|
||||
"openpgp": "^4.10.8",
|
||||
"pg": "^8.5.1",
|
||||
"sqlite3": "^5.0.2"
|
||||
},
|
||||
"dependencies": {
|
||||
"automerge": {
|
||||
"version": "0.14.2",
|
||||
"resolved": "https://registry.npmjs.org/automerge/-/automerge-0.14.2.tgz",
|
||||
"integrity": "sha512-shiwuJHCbNRI23WZyIECLV4Ovf3WiAFJ7P9BH4l5gON1In/UUbjcSJKRygtIirObw2UQumeYxp3F2XBdSvQHnA==",
|
||||
"requires": {
|
||||
"immutable": "^3.8.2",
|
||||
"transit-immutable-js": "^0.7.0",
|
||||
"transit-js": "^0.8.861",
|
||||
"uuid": "^3.4.0"
|
||||
}
|
||||
},
|
||||
"node-addon-api": {
|
||||
"version": "3.1.0",
|
||||
"resolved": "https://registry.npmjs.org/node-addon-api/-/node-addon-api-3.1.0.tgz",
|
||||
"integrity": "sha512-flmrDNB06LIl5lywUz7YlNGZH/5p0M7W28k8hzd9Lshtdh1wshD2Y+U4h9LD6KObOy1f+fEVdgprPrEymjM5uw=="
|
||||
},
|
||||
"pg": {
|
||||
"version": "8.6.0",
|
||||
"resolved": "https://registry.npmjs.org/pg/-/pg-8.6.0.tgz",
|
||||
"integrity": "sha512-qNS9u61lqljTDFvmk/N66EeGq3n6Ujzj0FFyNMGQr6XuEv4tgNTXvJQTfJdcvGit5p5/DWPu+wj920hAJFI+QQ==",
|
||||
"requires": {
|
||||
"buffer-writer": "2.0.0",
|
||||
"packet-reader": "1.0.0",
|
||||
"pg-connection-string": "^2.5.0",
|
||||
"pg-pool": "^3.3.0",
|
||||
"pg-protocol": "^1.5.0",
|
||||
"pg-types": "^2.1.0",
|
||||
"pgpass": "1.x"
|
||||
}
|
||||
},
|
||||
"pg-connection-string": {
|
||||
"version": "2.5.0",
|
||||
"resolved": "https://registry.npmjs.org/pg-connection-string/-/pg-connection-string-2.5.0.tgz",
|
||||
"integrity": "sha512-r5o/V/ORTA6TmUnyWZR9nCj1klXCO2CEKNRlVuJptZe85QuhFayC7WeMic7ndayT5IRIR0S0xFxFi2ousartlQ=="
|
||||
},
|
||||
"pg-pool": {
|
||||
"version": "3.3.0",
|
||||
"resolved": "https://registry.npmjs.org/pg-pool/-/pg-pool-3.3.0.tgz",
|
||||
"integrity": "sha512-0O5huCql8/D6PIRFAlmccjphLYWC+JIzvUhSzXSpGaf+tjTZc4nn+Lr7mLXBbFJfvwbP0ywDv73EiaBsxn7zdg=="
|
||||
},
|
||||
"pg-protocol": {
|
||||
"version": "1.5.0",
|
||||
"resolved": "https://registry.npmjs.org/pg-protocol/-/pg-protocol-1.5.0.tgz",
|
||||
"integrity": "sha512-muRttij7H8TqRNu/DxrAJQITO4Ac7RmX3Klyr/9mJEOBeIpgnF8f9jAfRz5d3XwQZl5qBjF9gLsUtMPJE0vezQ=="
|
||||
},
|
||||
"sqlite3": {
|
||||
"version": "5.0.2",
|
||||
"resolved": "https://registry.npmjs.org/sqlite3/-/sqlite3-5.0.2.tgz",
|
||||
"integrity": "sha512-1SdTNo+BVU211Xj1csWa8lV6KM0CtucDwRyA0VHl91wEH1Mgh7RxUpI4rVvG7OhHrzCSGaVyW5g8vKvlrk9DJA==",
|
||||
"requires": {
|
||||
"node-addon-api": "^3.0.0",
|
||||
"node-gyp": "3.x",
|
||||
"node-pre-gyp": "^0.11.0"
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"create-hash": {
|
||||
"version": "1.2.0",
|
||||
"resolved": "https://registry.npmjs.org/create-hash/-/create-hash-1.2.0.tgz",
|
||||
@@ -1035,17 +966,17 @@
|
||||
"dev": true
|
||||
},
|
||||
"elliptic": {
|
||||
"version": "6.5.4",
|
||||
"resolved": "https://registry.npmjs.org/elliptic/-/elliptic-6.5.4.tgz",
|
||||
"integrity": "sha512-iLhC6ULemrljPZb+QutR5TQGB+pdW6KGD5RSegS+8sorOZT+rdQFbsQFJgvN3eRqNALqJer4oQ16YvJHlU8hzQ==",
|
||||
"version": "6.5.3",
|
||||
"resolved": "https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz",
|
||||
"integrity": "sha512-IMqzv5wNQf+E6aHeIqATs0tOLeOTwj1QKbRcS3jBbYkl5oLAserA8yJTT7/VyHUYG91PRmPyeQDObKLPpeS4dw==",
|
||||
"requires": {
|
||||
"bn.js": "^4.11.9",
|
||||
"brorand": "^1.1.0",
|
||||
"bn.js": "^4.4.0",
|
||||
"brorand": "^1.0.1",
|
||||
"hash.js": "^1.0.0",
|
||||
"hmac-drbg": "^1.0.1",
|
||||
"inherits": "^2.0.4",
|
||||
"minimalistic-assert": "^1.0.1",
|
||||
"minimalistic-crypto-utils": "^1.0.1"
|
||||
"hmac-drbg": "^1.0.0",
|
||||
"inherits": "^2.0.1",
|
||||
"minimalistic-assert": "^1.0.0",
|
||||
"minimalistic-crypto-utils": "^1.0.0"
|
||||
}
|
||||
},
|
||||
"emoji-regex": {
|
||||
@@ -1558,9 +1489,9 @@
|
||||
"integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ=="
|
||||
},
|
||||
"ini": {
|
||||
"version": "1.3.8",
|
||||
"resolved": "https://registry.npmjs.org/ini/-/ini-1.3.8.tgz",
|
||||
"integrity": "sha512-JV/yugV2uzW5iMRSiZAyDtQd+nxtUnjeLt0acNdw98kKLrvuRVyB80tsREOE7yvGVgalhZ6RNXCmEHkUKBKxew=="
|
||||
"version": "1.3.5",
|
||||
"resolved": "https://registry.npmjs.org/ini/-/ini-1.3.5.tgz",
|
||||
"integrity": "sha512-RZY5huIKCMRWDUqZlEi72f/lmXKMvuszcMBduliQ3nnWbx9X/ZBQO7DijMEYS9EhHBb2qacRUMtC7svLwe0lcw=="
|
||||
},
|
||||
"interpret": {
|
||||
"version": "2.2.0",
|
||||
@@ -2026,9 +1957,9 @@
|
||||
}
|
||||
},
|
||||
"y18n": {
|
||||
"version": "4.0.3",
|
||||
"resolved": "https://registry.npmjs.org/y18n/-/y18n-4.0.3.tgz",
|
||||
"integrity": "sha512-JKhqTOwSrqNA1NY5lSztJ1GrBiUodLMmIZuLiDaMRJ+itFd+ABVE8XBjOvIWL+rSqNDC74LCSFmlb/U4UZ4hJQ==",
|
||||
"version": "4.0.0",
|
||||
"resolved": "https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz",
|
||||
"integrity": "sha512-r9S/ZyXu/Xu9q1tYlpsLIsa3EeLXXk0VwlxqTcFRfg9EhMW+17kbt9G0NrgCmhGb5vT2hyhJZLfDGx+7+5Uj/w==",
|
||||
"dev": true
|
||||
},
|
||||
"yargs": {
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "cic-client-meta",
|
||||
"version": "0.0.7-alpha.8",
|
||||
"version": "0.0.7-alpha.6",
|
||||
"description": "Signed CRDT metadata graphs for the CIC network",
|
||||
"main": "dist/index.js",
|
||||
"types": "dist/index.d.ts",
|
||||
@@ -15,9 +15,8 @@
|
||||
"dependencies": {
|
||||
"@ethereumjs/tx": "^3.0.0-beta.1",
|
||||
"automerge": "^0.14.1",
|
||||
"crdt-meta": "0.0.8",
|
||||
"ethereumjs-wallet": "^1.0.1",
|
||||
"ini": "^1.3.8",
|
||||
"ini": "^1.3.5",
|
||||
"openpgp": "^4.10.8",
|
||||
"pg": "^8.4.2",
|
||||
"sqlite3": "^5.0.0",
|
||||
@@ -41,6 +40,6 @@
|
||||
],
|
||||
"license": "GPL-3.0-or-later",
|
||||
"engines": {
|
||||
"node": ">=14.16.1"
|
||||
"node": "~14.16.1"
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { Config } from 'crdt-meta';
|
||||
const config = require('./src/config');
|
||||
const fs = require('fs');
|
||||
|
||||
if (process.argv[2] === undefined) {
|
||||
@@ -15,6 +15,6 @@ try {
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
const c = new Config(process.argv[2], process.env['CONFINI_ENV_PREFIX']);
|
||||
const c = new config.Config(process.argv[2], process.env['CONFINI_ENV_PREFIX']);
|
||||
c.process();
|
||||
process.stdout.write(c.toString());
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
import * as Automerge from 'automerge';
|
||||
import * as pgp from 'openpgp';
|
||||
import * as pg from 'pg';
|
||||
|
||||
import { Envelope, Syncable } from 'crdt-meta';
|
||||
import { Envelope, Syncable } from '../../src/sync';
|
||||
|
||||
|
||||
function handleNoMergeGet(db, digest, keystore) {
|
||||
|
||||
@@ -1,11 +1,15 @@
|
||||
import * as http from 'http';
|
||||
import * as fs from 'fs';
|
||||
import * as path from 'path';
|
||||
import * as pgp from 'openpgp';
|
||||
|
||||
import * as handlers from './handlers';
|
||||
import { PGPKeyStore, PGPSigner, Config, SqliteAdapter, PostgresAdapter } from 'crdt-meta';
|
||||
import { Envelope, Syncable } from '../../src/sync';
|
||||
import { PGPKeyStore, PGPSigner } from '../../src/auth';
|
||||
|
||||
import { standardArgs } from './args';
|
||||
import { Config } from '../../src/config';
|
||||
import { SqliteAdapter, PostgresAdapter } from '../../src/db';
|
||||
|
||||
let configPath = '/usr/local/etc/cic-meta';
|
||||
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
import { Syncable, Addressable, mergeKey } from 'crdt-meta';
|
||||
import { ArgPair, Syncable } from '../sync';
|
||||
import { Addressable, mergeKey } from '../digest';
|
||||
|
||||
class Phone extends Syncable implements Addressable {
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
import { Syncable, Addressable, toAddressKey } from 'crdt-meta';
|
||||
import { ArgPair, Syncable } from '../sync';
|
||||
import { Addressable, addressToBytes, bytesToHex, toAddressKey } from '../digest';
|
||||
|
||||
const keySalt = new TextEncoder().encode(':cic.person');
|
||||
class User extends Syncable implements Addressable {
|
||||
191
apps/cic-meta/src/auth.ts
Normal file
191
apps/cic-meta/src/auth.ts
Normal file
@@ -0,0 +1,191 @@
|
||||
import * as pgp from 'openpgp';
|
||||
import * as crypto from 'crypto';
|
||||
|
||||
interface Signable {
|
||||
digest():string;
|
||||
}
|
||||
|
||||
type KeyGetter = () => any;
|
||||
|
||||
type Signature = {
|
||||
engine:string
|
||||
algo:string
|
||||
data:string
|
||||
digest:string
|
||||
}
|
||||
|
||||
interface Signer {
|
||||
prepare(Signable):boolean;
|
||||
onsign(Signature):void;
|
||||
onverify(boolean):void;
|
||||
sign(digest:string):void
|
||||
verify(digest:string, signature:Signature):void
|
||||
fingerprint():string
|
||||
}
|
||||
|
||||
interface Authoritative {
|
||||
}
|
||||
|
||||
interface KeyStore {
|
||||
getPrivateKey: KeyGetter
|
||||
getFingerprint: () => string
|
||||
getTrustedKeys: () => Array<any>
|
||||
getTrustedActiveKeys: () => Array<any>
|
||||
getEncryptKeys: () => Array<any>
|
||||
}
|
||||
|
||||
class PGPKeyStore implements KeyStore {
|
||||
|
||||
fingerprint: string
|
||||
pk: any
|
||||
|
||||
pubk = {
|
||||
active: [],
|
||||
trusted: [],
|
||||
encrypt: [],
|
||||
}
|
||||
loads = 0x00;
|
||||
loadsTarget = 0x0f;
|
||||
onload: (k:KeyStore) => void;
|
||||
|
||||
constructor(passphrase:string, pkArmor:string, pubkActiveArmor:string, pubkTrustedArmor:string, pubkEncryptArmor:string, onload = (ks:KeyStore) => {}) {
|
||||
this._readKey(pkArmor, undefined, 1, passphrase);
|
||||
this._readKey(pubkActiveArmor, 'active', 2);
|
||||
this._readKey(pubkTrustedArmor, 'trusted', 4);
|
||||
this._readKey(pubkEncryptArmor, 'encrypt', 8);
|
||||
this.onload = onload;
|
||||
}
|
||||
|
||||
private _readKey(a:string, x:any, n:number, pass?:string) {
|
||||
pgp.key.readArmored(a).then((k) => {
|
||||
if (pass !== undefined) {
|
||||
this.pk = k.keys[0];
|
||||
this.pk.decrypt(pass).then(() => {
|
||||
this.fingerprint = this.pk.getFingerprint();
|
||||
console.log('private key (sign)', this.fingerprint);
|
||||
this._registerLoad(n);
|
||||
});
|
||||
} else {
|
||||
this.pubk[x] = k.keys;
|
||||
k.keys.forEach((pubk) => {
|
||||
console.log('public key (' + x + ')', pubk.getFingerprint());
|
||||
});
|
||||
this._registerLoad(n);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
private _registerLoad(b:number) {
|
||||
this.loads |= b;
|
||||
if (this.loads == this.loadsTarget) {
|
||||
this.onload(this);
|
||||
}
|
||||
}
|
||||
|
||||
public getTrustedKeys(): Array<any> {
|
||||
return this.pubk['trusted'];
|
||||
}
|
||||
|
||||
public getTrustedActiveKeys(): Array<any> {
|
||||
return this.pubk['active'];
|
||||
|
||||
}
|
||||
|
||||
public getEncryptKeys(): Array<any> {
|
||||
return this.pubk['encrypt'];
|
||||
|
||||
}
|
||||
|
||||
public getPrivateKey(): any {
|
||||
return this.pk;
|
||||
}
|
||||
|
||||
public getFingerprint(): string {
|
||||
return this.fingerprint;
|
||||
}
|
||||
}
|
||||
|
||||
class PGPSigner implements Signer {
|
||||
|
||||
engine = 'pgp'
|
||||
algo = 'sha256'
|
||||
dgst: string
|
||||
signature: Signature
|
||||
keyStore: KeyStore
|
||||
onsign: (Signature) => void
|
||||
onverify: (boolean) => void
|
||||
|
||||
constructor(keyStore:KeyStore) {
|
||||
this.keyStore = keyStore
|
||||
this.onsign = (string) => {};
|
||||
this.onverify = (boolean) => {};
|
||||
}
|
||||
|
||||
public fingerprint(): string {
|
||||
return this.keyStore.getFingerprint();
|
||||
}
|
||||
|
||||
public prepare(material:Signable):boolean {
|
||||
this.dgst = material.digest();
|
||||
return true;
|
||||
}
|
||||
|
||||
public verify(digest:string, signature:Signature) {
|
||||
pgp.signature.readArmored(signature.data).then((s) => {
|
||||
const opts = {
|
||||
message: pgp.cleartext.fromText(digest),
|
||||
publicKeys: this.keyStore.getTrustedKeys(),
|
||||
signature: s,
|
||||
};
|
||||
pgp.verify(opts).then((v) => {
|
||||
let i = 0;
|
||||
for (i = 0; i < v.signatures.length; i++) {
|
||||
const s = v.signatures[i];
|
||||
if (s.valid) {
|
||||
this.onverify(s);
|
||||
return;
|
||||
}
|
||||
}
|
||||
console.error('checked ' + i + ' signature(s) but none valid');
|
||||
this.onverify(false);
|
||||
});
|
||||
}).catch((e) => {
|
||||
console.error(e);
|
||||
this.onverify(false);
|
||||
});
|
||||
}
|
||||
|
||||
public sign(digest:string) {
|
||||
const m = pgp.cleartext.fromText(digest);
|
||||
const pk = this.keyStore.getPrivateKey();
|
||||
const opts = {
|
||||
message: m,
|
||||
privateKeys: [pk],
|
||||
detached: true,
|
||||
}
|
||||
pgp.sign(opts).then((s) => {
|
||||
this.signature = {
|
||||
engine: this.engine,
|
||||
algo: this.algo,
|
||||
data: s.signature,
|
||||
// TODO: fix for browser later
|
||||
digest: digest,
|
||||
};
|
||||
this.onsign(this.signature);
|
||||
}).catch((e) => {
|
||||
console.error(e);
|
||||
this.onsign(undefined);
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
export {
|
||||
Signature,
|
||||
Authoritative,
|
||||
Signer,
|
||||
KeyGetter,
|
||||
Signable,
|
||||
KeyStore,
|
||||
PGPSigner,
|
||||
PGPKeyStore,
|
||||
};
|
||||
71
apps/cic-meta/src/config.ts
Normal file
71
apps/cic-meta/src/config.ts
Normal file
@@ -0,0 +1,71 @@
|
||||
import * as fs from 'fs';
|
||||
import * as ini from 'ini';
|
||||
import * as path from 'path';
|
||||
|
||||
class Config {
|
||||
|
||||
filepath: string
|
||||
store: Object
|
||||
censor: Array<string>
|
||||
require: Array<string>
|
||||
env_prefix: string
|
||||
|
||||
constructor(filepath:string, env_prefix?:string) {
|
||||
this.filepath = filepath;
|
||||
this.store = {};
|
||||
this.censor = [];
|
||||
this.require = [];
|
||||
this.env_prefix = '';
|
||||
if (env_prefix !== undefined) {
|
||||
this.env_prefix = env_prefix + "_";
|
||||
}
|
||||
}
|
||||
|
||||
public process() {
|
||||
const d = fs.readdirSync(this.filepath);
|
||||
|
||||
const r = /.*\.ini$/;
|
||||
for (let i = 0; i < d.length; i++) {
|
||||
const f = d[i];
|
||||
if (!f.match(r)) {
|
||||
return;
|
||||
}
|
||||
|
||||
const fp = path.join(this.filepath, f);
|
||||
const v = fs.readFileSync(fp, 'utf-8');
|
||||
const inid = ini.decode(v);
|
||||
const inik = Object.keys(inid);
|
||||
for (let j = 0; j < inik.length; j++) {
|
||||
const k_section = inik[j]
|
||||
const k = k_section.toUpperCase();
|
||||
Object.keys(inid[k_section]).forEach((k_directive) => {
|
||||
const kk = k_directive.toUpperCase();
|
||||
const kkk = k + '_' + kk;
|
||||
|
||||
let r = inid[k_section][k_directive];
|
||||
const k_env = this.env_prefix + kkk
|
||||
const env = process.env[k_env];
|
||||
if (env !== undefined) {
|
||||
console.debug('Environment variable ' + k_env + ' overrides ' + kkk);
|
||||
r = env;
|
||||
}
|
||||
this.store[kkk] = r;
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public get(s:string) {
|
||||
return this.store[s];
|
||||
}
|
||||
|
||||
public toString() {
|
||||
let s = '';
|
||||
Object.keys(this.store).forEach((k) => {
|
||||
s += k + '=' + this.store[k] + '\n';
|
||||
});
|
||||
return s;
|
||||
}
|
||||
}
|
||||
|
||||
export { Config };
|
||||
38
apps/cic-meta/src/constants.ts
Normal file
38
apps/cic-meta/src/constants.ts
Normal file
@@ -0,0 +1,38 @@
|
||||
import { JSONSerializable } from './format';
|
||||
|
||||
const ENGINE_NAME = 'automerge';
|
||||
const ENGINE_VERSION = '0.14.1';
|
||||
|
||||
const NETWORK_NAME = 'cic';
|
||||
const NETWORK_VERSION = '1';
|
||||
|
||||
const CRYPTO_NAME = 'pgp';
|
||||
const CRYPTO_VERSION = '2';
|
||||
|
||||
type VersionedSpec = {
|
||||
name: string
|
||||
version: string
|
||||
ext?: Object
|
||||
}
|
||||
|
||||
const engineSpec:VersionedSpec = {
|
||||
name: ENGINE_NAME,
|
||||
version: ENGINE_VERSION,
|
||||
}
|
||||
|
||||
const cryptoSpec:VersionedSpec = {
|
||||
name: CRYPTO_NAME,
|
||||
version: CRYPTO_VERSION,
|
||||
}
|
||||
|
||||
const networkSpec:VersionedSpec = {
|
||||
name: NETWORK_NAME,
|
||||
version: NETWORK_VERSION,
|
||||
}
|
||||
|
||||
export {
|
||||
engineSpec,
|
||||
cryptoSpec,
|
||||
networkSpec,
|
||||
VersionedSpec,
|
||||
};
|
||||
27
apps/cic-meta/src/crypto.ts
Normal file
27
apps/cic-meta/src/crypto.ts
Normal file
@@ -0,0 +1,27 @@
|
||||
import * as crypto from 'crypto';
|
||||
|
||||
const _algs = {
|
||||
'SHA-256': 'sha256',
|
||||
}
|
||||
|
||||
function cryptoWrapper() {
|
||||
}
|
||||
|
||||
cryptoWrapper.prototype.digest = async function(s, d) {
|
||||
const h = crypto.createHash(_algs[s]);
|
||||
h.update(d);
|
||||
return h.digest();
|
||||
}
|
||||
|
||||
let subtle = undefined;
|
||||
if (typeof window !== 'undefined') {
|
||||
subtle = window.crypto.subtle;
|
||||
} else {
|
||||
subtle = new cryptoWrapper();
|
||||
}
|
||||
|
||||
|
||||
export {
|
||||
subtle,
|
||||
}
|
||||
|
||||
90
apps/cic-meta/src/db.ts
Normal file
90
apps/cic-meta/src/db.ts
Normal file
@@ -0,0 +1,90 @@
|
||||
import * as pg from 'pg';
|
||||
import * as sqlite from 'sqlite3';
|
||||
|
||||
type DbConfig = {
|
||||
name: string
|
||||
host: string
|
||||
port: number
|
||||
user: string
|
||||
password: string
|
||||
}
|
||||
|
||||
interface DbAdapter {
|
||||
query: (s:string, callback:(e:any, rs:any) => void) => void
|
||||
close: () => void
|
||||
}
|
||||
|
||||
const re_creatematch = /^(CREATE)/i
|
||||
const re_getmatch = /^(SELECT)/i;
|
||||
const re_setmatch = /^(INSERT|UPDATE)/i;
|
||||
|
||||
class SqliteAdapter implements DbAdapter {
|
||||
|
||||
db: any
|
||||
|
||||
constructor(dbConfig:DbConfig, callback?:(any) => void) {
|
||||
this.db = new sqlite.Database(dbConfig.name); //, callback);
|
||||
}
|
||||
|
||||
public query(s:string, callback:(e:any, rs?:any) => void): void {
|
||||
const local_callback = (e, rs) => {
|
||||
let r = undefined;
|
||||
if (rs !== undefined) {
|
||||
r = {
|
||||
rowCount: rs.length,
|
||||
rows: rs,
|
||||
}
|
||||
}
|
||||
callback(e, r);
|
||||
};
|
||||
if (s.match(re_getmatch)) {
|
||||
this.db.all(s, local_callback);
|
||||
} else if (s.match(re_setmatch)) {
|
||||
this.db.run(s, local_callback);
|
||||
} else if (s.match(re_creatematch)) {
|
||||
this.db.run(s, callback);
|
||||
} else {
|
||||
throw 'unhandled query';
|
||||
}
|
||||
}
|
||||
|
||||
public close() {
|
||||
this.db.close();
|
||||
}
|
||||
}
|
||||
|
||||
class PostgresAdapter implements DbAdapter {
|
||||
|
||||
db: any
|
||||
|
||||
constructor(dbConfig:DbConfig) {
|
||||
let o = dbConfig;
|
||||
o['database'] = o.name;
|
||||
this.db = new pg.Pool(o);
|
||||
return this.db;
|
||||
}
|
||||
|
||||
public query(s:string, callback:(e:any, rs:any) => void): void {
|
||||
this.db.query(s, (e, rs) => {
|
||||
let r = {
|
||||
length: rs.rowCount,
|
||||
}
|
||||
rs.length = rs.rowCount;
|
||||
if (e === undefined) {
|
||||
e = null;
|
||||
}
|
||||
console.debug(e, rs);
|
||||
callback(e, rs);
|
||||
});
|
||||
}
|
||||
|
||||
public close() {
|
||||
this.db.end();
|
||||
}
|
||||
}
|
||||
|
||||
export {
|
||||
DbConfig,
|
||||
SqliteAdapter,
|
||||
PostgresAdapter,
|
||||
}
|
||||
68
apps/cic-meta/src/digest.ts
Normal file
68
apps/cic-meta/src/digest.ts
Normal file
@@ -0,0 +1,68 @@
|
||||
import * as crypto from './crypto';
|
||||
|
||||
interface Addressable {
|
||||
key(): string
|
||||
digest(): string
|
||||
}
|
||||
|
||||
function stringToBytes(s:string) {
|
||||
const a = new Uint8Array(20);
|
||||
let j = 2;
|
||||
for (let i = 0; i < a.byteLength; i++) {
|
||||
const n = parseInt(s.substring(j, j+2), 16);
|
||||
a[i] = n;
|
||||
j += 2;
|
||||
}
|
||||
return a;
|
||||
}
|
||||
|
||||
function bytesToHex(a:Uint8Array) {
|
||||
let s = '';
|
||||
for (let i = 0; i < a.byteLength; i++) {
|
||||
const h = '00' + a[i].toString(16);
|
||||
s += h.slice(-2);
|
||||
}
|
||||
return s;
|
||||
}
|
||||
|
||||
async function mergeKey(a:Uint8Array, s:Uint8Array) {
|
||||
const y = new Uint8Array(a.byteLength + s.byteLength);
|
||||
for (let i = 0; i < a.byteLength; i++) {
|
||||
y[i] = a[i];
|
||||
}
|
||||
for (let i = 0; i < s.byteLength; i++) {
|
||||
y[a.byteLength + i] = s[i];
|
||||
}
|
||||
const z = await crypto.subtle.digest('SHA-256', y);
|
||||
return bytesToHex(new Uint8Array(z));
|
||||
}
|
||||
|
||||
async function toKey(v:string, salt:string) {
|
||||
const a = stringToBytes(v);
|
||||
const s = new TextEncoder().encode(salt);
|
||||
return await mergeKey(a, s);
|
||||
}
|
||||
|
||||
|
||||
async function toAddressKey(zeroExHex:string, salt:string) {
|
||||
const a = addressToBytes(zeroExHex);
|
||||
const s = new TextEncoder().encode(salt);
|
||||
return await mergeKey(a, s);
|
||||
}
|
||||
|
||||
const re_addrHex = /^0[xX][a-fA-F0-9]{40}$/;
|
||||
function addressToBytes(s:string) {
|
||||
if (!s.match(re_addrHex)) {
|
||||
throw 'invalid address hex';
|
||||
}
|
||||
return stringToBytes(s);
|
||||
}
|
||||
|
||||
export {
|
||||
toKey,
|
||||
toAddressKey,
|
||||
mergeKey,
|
||||
bytesToHex,
|
||||
addressToBytes,
|
||||
Addressable,
|
||||
}
|
||||
58
apps/cic-meta/src/dispatch.ts
Normal file
58
apps/cic-meta/src/dispatch.ts
Normal file
@@ -0,0 +1,58 @@
|
||||
import { v4 as uuidv4 } from 'uuid';
|
||||
import { Syncable } from './sync';
|
||||
import { Store } from './store';
|
||||
import { PubSub } from './transport';
|
||||
|
||||
function toIndexKey(id:string):string {
|
||||
const d = Date.now();
|
||||
return d + '_' + id + '_' + uuidv4();
|
||||
}
|
||||
|
||||
const _re_indexKey = /^\d+_(.+)_[-\d\w]+$/;
|
||||
function fromIndexKey(s:string):string {
|
||||
const m = s.match(_re_indexKey);
|
||||
if (m === null) {
|
||||
throw 'Invalid index key';
|
||||
}
|
||||
return m[1];
|
||||
}
|
||||
|
||||
class Dispatcher {
|
||||
|
||||
idx: Array<string>
|
||||
syncer: PubSub
|
||||
store: Store
|
||||
|
||||
constructor(store:Store, syncer:PubSub) {
|
||||
this.idx = new Array<string>()
|
||||
this.syncer = syncer;
|
||||
this.store = store;
|
||||
}
|
||||
|
||||
public isDirty(): boolean {
|
||||
return this.idx.length > 0;
|
||||
}
|
||||
|
||||
public add(id:string, item:Syncable): string {
|
||||
const v = item.toJSON();
|
||||
const k = toIndexKey(id);
|
||||
this.store.put(k, v, true);
|
||||
localStorage.setItem(k, v);
|
||||
this.idx.push(k);
|
||||
return k;
|
||||
}
|
||||
|
||||
public sync(offset:number): number {
|
||||
let i = 0;
|
||||
this.idx.forEach((k) => {
|
||||
const v = localStorage.getItem(k);
|
||||
const k_id = fromIndexKey(k);
|
||||
this.syncer.pub(v); // this must block until guaranteed delivery
|
||||
localStorage.removeItem(k);
|
||||
i++;
|
||||
});
|
||||
return i;
|
||||
}
|
||||
}
|
||||
|
||||
export { Dispatcher, toIndexKey, fromIndexKey }
|
||||
5
apps/cic-meta/src/format.ts
Normal file
5
apps/cic-meta/src/format.ts
Normal file
@@ -0,0 +1,5 @@
|
||||
interface JSONSerializable {
|
||||
toJSON(): string
|
||||
}
|
||||
|
||||
export { JSONSerializable };
|
||||
@@ -1,2 +1,5 @@
|
||||
export { User } from './user';
|
||||
export { Phone } from './phone';
|
||||
export { PGPSigner, PGPKeyStore, Signer, KeyStore } from './auth';
|
||||
export { ArgPair, Envelope, Syncable } from './sync';
|
||||
export { User } from './assets/user';
|
||||
export { Phone } from './assets/phone';
|
||||
export { Config } from './config';
|
||||
|
||||
9
apps/cic-meta/src/store.ts
Normal file
9
apps/cic-meta/src/store.ts
Normal file
@@ -0,0 +1,9 @@
|
||||
import { Syncable } from './sync';
|
||||
|
||||
interface Store {
|
||||
put(string, Syncable, boolean?)
|
||||
get(string):Syncable
|
||||
delete(string)
|
||||
}
|
||||
|
||||
export { Store };
|
||||
266
apps/cic-meta/src/sync.ts
Normal file
266
apps/cic-meta/src/sync.ts
Normal file
@@ -0,0 +1,266 @@
|
||||
import * as Automerge from 'automerge';
|
||||
|
||||
import { JSONSerializable } from './format';
|
||||
|
||||
import { Authoritative, Signer, PGPSigner, Signable, Signature } from './auth';
|
||||
|
||||
import { engineSpec, cryptoSpec, networkSpec, VersionedSpec } from './constants';
|
||||
|
||||
const fullSpec:VersionedSpec = {
|
||||
name: 'cic',
|
||||
version: '1',
|
||||
ext: {
|
||||
network: cryptoSpec,
|
||||
engine: engineSpec,
|
||||
},
|
||||
}
|
||||
|
||||
class Envelope {
|
||||
|
||||
o = fullSpec
|
||||
|
||||
constructor(payload:Object) {
|
||||
this.set(payload);
|
||||
}
|
||||
|
||||
public set(payload:Object) {
|
||||
this.o['payload'] = payload
|
||||
}
|
||||
|
||||
public get():string {
|
||||
return this.o['payload'];
|
||||
}
|
||||
|
||||
public toJSON() {
|
||||
return JSON.stringify(this.o);
|
||||
}
|
||||
|
||||
public static fromJSON(s:string): Envelope {
|
||||
const e = new Envelope(undefined);
|
||||
e.o = JSON.parse(s);
|
||||
return e;
|
||||
}
|
||||
|
||||
public unwrap(): Syncable {
|
||||
return Syncable.fromJSON(this.o['payload']);
|
||||
}
|
||||
}
|
||||
|
||||
class ArgPair {
|
||||
|
||||
k:string
|
||||
v:any
|
||||
|
||||
constructor(k:string, v:any) {
|
||||
this.k = k;
|
||||
this.v = v;
|
||||
}
|
||||
}
|
||||
|
||||
class SignablePart implements Signable {
|
||||
|
||||
s: string
|
||||
|
||||
constructor(s:string) {
|
||||
this.s = s;
|
||||
}
|
||||
|
||||
public digest():string {
|
||||
return this.s;
|
||||
}
|
||||
}
|
||||
|
||||
function orderDict(src) {
|
||||
let dst;
|
||||
if (Array.isArray(src)) {
|
||||
dst = [];
|
||||
src.forEach((v) => {
|
||||
if (typeof(v) == 'object') {
|
||||
v = orderDict(v);
|
||||
}
|
||||
dst.push(v);
|
||||
});
|
||||
} else {
|
||||
dst = {}
|
||||
Object.keys(src).sort().forEach((k) => {
|
||||
let v = src[k];
|
||||
if (typeof(v) == 'object') {
|
||||
v = orderDict(v);
|
||||
}
|
||||
dst[k] = v;
|
||||
});
|
||||
}
|
||||
return dst;
|
||||
}
|
||||
|
||||
class Syncable implements JSONSerializable, Authoritative, Signable {
|
||||
|
||||
id: string
|
||||
timestamp: number
|
||||
m: any // automerge object
|
||||
e: Envelope
|
||||
signer: Signer
|
||||
onwrap: (string) => void
|
||||
onauthenticate: (boolean) => void
|
||||
|
||||
// TODO: Move data to sub-object so timestamp, id, signature don't collide
|
||||
constructor(id:string, v:Object) {
|
||||
this.id = id;
|
||||
const o = {
|
||||
'id': id,
|
||||
'timestamp': Math.floor(Date.now() / 1000),
|
||||
'data': v,
|
||||
}
|
||||
//this.m = Automerge.from(v)
|
||||
this.m = Automerge.from(o)
|
||||
}
|
||||
|
||||
public setSigner(signer:Signer) {
|
||||
this.signer = signer;
|
||||
this.signer.onsign = (s) => {
|
||||
this.wrap(s);
|
||||
};
|
||||
}
|
||||
|
||||
// TODO: To keep integrity, the non-link key/value pairs for each step also need to be hashed
|
||||
public digest(): string {
|
||||
const links = [];
|
||||
Automerge.getAllChanges(this.m).forEach((ch:Object) => {
|
||||
const op:Array<any> = ch['ops'];
|
||||
ch['ops'].forEach((op:Array<Object>) => {
|
||||
if (op['action'] == 'link') {
|
||||
//console.log('op link', op);
|
||||
links.push([op['obj'], op['value']]);
|
||||
}
|
||||
});
|
||||
});
|
||||
//return JSON.stringify(links);
|
||||
const j = JSON.stringify(links);
|
||||
return Buffer.from(j).toString('base64');
|
||||
}
|
||||
|
||||
private wrap(s:any) {
|
||||
this.m = Automerge.change(this.m, 'sign', (doc) => {
|
||||
doc['signature'] = s;
|
||||
});
|
||||
this.e = new Envelope(this.toJSON());
|
||||
console.log('wrappin s', s, typeof(s));
|
||||
this.e.o['digest'] = s.digest;
|
||||
if (this.onwrap !== undefined) {
|
||||
this.onwrap(this.e);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
// private _verifyLoop(i:number, history:Array<any>, signable:Signable, result:boolean) {
|
||||
// if (!result) {
|
||||
// this.onauthenticate(false);
|
||||
// return;
|
||||
// } else if (history.length == 0) {
|
||||
// this.onauthenticate(true);
|
||||
// return;
|
||||
// }
|
||||
// const h = history.shift()
|
||||
// if (i % 2 == 0) {
|
||||
// i++;
|
||||
// signable = {
|
||||
// digest: () => {
|
||||
// return Automerge.save(h.snapshot)
|
||||
// },
|
||||
// };
|
||||
// this._verifyLoop(i, history, signable, true);
|
||||
// } else {
|
||||
// i++;
|
||||
// const signature = h.snapshot['signature'];
|
||||
// console.debug('signature', signature, signable.digest());
|
||||
// this.signer.onverify = (v) => {
|
||||
// this._verifyLoop(i, history, signable, v)
|
||||
// }
|
||||
// this.signer.verify(signable, signature);
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// // TODO: This should replay the graph and check signatures on each step
|
||||
// public _authenticate(full:boolean=false) {
|
||||
// let h = Automerge.getHistory(this.m);
|
||||
// h.forEach((m) => {
|
||||
// //console.debug(m.snapshot);
|
||||
// });
|
||||
// const signable = {
|
||||
// digest: () => { return '' },
|
||||
// }
|
||||
// if (!full) {
|
||||
// h = h.slice(h.length-2);
|
||||
// }
|
||||
// this._verifyLoop(0, h, signable, true);
|
||||
// }
|
||||
|
||||
public authenticate(full:boolean=false) {
|
||||
if (full) {
|
||||
console.warn('only doing shallow authentication for now, sorry');
|
||||
}
|
||||
//console.log('authenticating', signable.digest());
|
||||
//console.log('signature', this.m.signature);
|
||||
this.signer.onverify = (v) => {
|
||||
//this._verifyLoop(i, history, signable, v)
|
||||
this.onauthenticate(v);
|
||||
}
|
||||
this.signer.verify(this.m.signature.digest, this.m.signature);
|
||||
}
|
||||
|
||||
|
||||
public sign() {
|
||||
//this.signer.prepare(this);
|
||||
this.signer.sign(this.digest());
|
||||
}
|
||||
|
||||
public update(changes:Array<ArgPair>, changesDescription:string) {
|
||||
this.m = Automerge.change(this.m, changesDescription, (m) => {
|
||||
changes.forEach((c) => {
|
||||
let path = c.k.split('.');
|
||||
let target = m['data'];
|
||||
while (path.length > 1) {
|
||||
const part = path.shift();
|
||||
target = target[part];
|
||||
}
|
||||
target[path[0]] = c.v;
|
||||
});
|
||||
m['timestamp'] = Math.floor(Date.now() / 1000);
|
||||
});
|
||||
}
|
||||
|
||||
public replace(o:Object, changesDescription:string) {
|
||||
this.m = Automerge.change(this.m, changesDescription, (m) => {
|
||||
Object.keys(o).forEach((k) => {
|
||||
m['data'][k] = o[k];
|
||||
});
|
||||
Object.keys(m).forEach((k) => {
|
||||
if (o[k] == undefined) {
|
||||
delete m['data'][k];
|
||||
}
|
||||
});
|
||||
m['timestamp'] = Math.floor(Date.now() / 1000);
|
||||
});
|
||||
}
|
||||
|
||||
public merge(s:Syncable) {
|
||||
this.m = Automerge.merge(s.m, this.m);
|
||||
}
|
||||
|
||||
public toJSON(): string {
|
||||
const s = Automerge.save(this.m);
|
||||
const o = JSON.parse(s);
|
||||
const oo = orderDict(o)
|
||||
return JSON.stringify(oo);
|
||||
|
||||
}
|
||||
|
||||
public static fromJSON(s:string): Syncable {
|
||||
const doc = Automerge.load(s);
|
||||
let y = new Syncable(doc['id'], {});
|
||||
y.m = doc
|
||||
return y
|
||||
}
|
||||
}
|
||||
|
||||
export { JSONSerializable, Syncable, ArgPair, Envelope };
|
||||
11
apps/cic-meta/src/transport.ts
Normal file
11
apps/cic-meta/src/transport.ts
Normal file
@@ -0,0 +1,11 @@
|
||||
interface SubConsumer {
|
||||
post(string)
|
||||
}
|
||||
|
||||
interface PubSub {
|
||||
pub(v:string):boolean
|
||||
close()
|
||||
}
|
||||
|
||||
export { PubSub, SubConsumer };
|
||||
|
||||
50
apps/cic-meta/tests/1_basic.ts
Normal file
50
apps/cic-meta/tests/1_basic.ts
Normal file
@@ -0,0 +1,50 @@
|
||||
import * as Automerge from 'automerge';
|
||||
import assert = require('assert');
|
||||
|
||||
import { Dispatcher, toIndexKey, fromIndexKey } from '../src/dispatch';
|
||||
import { User } from '../src/assets/user';
|
||||
import { Syncable, ArgPair } from '../src/sync';
|
||||
|
||||
import { MockSigner, MockStore } from './mock';
|
||||
|
||||
describe('basic', () => {
|
||||
|
||||
it('store', () => {
|
||||
const store = new MockStore('s');
|
||||
assert.equal(store.name, 's');
|
||||
|
||||
const mockSigner = new MockSigner();
|
||||
const v = new Syncable('foo', {baz: 42});
|
||||
v.setSigner(mockSigner);
|
||||
store.put('foo', v);
|
||||
const one = store.get('foo').toJSON();
|
||||
const vv = new Syncable('bar', {baz: 666});
|
||||
vv.setSigner(mockSigner);
|
||||
assert.throws(() => {
|
||||
store.put('foo', vv)
|
||||
});
|
||||
store.put('foo', vv, true);
|
||||
const other = store.get('foo').toJSON();
|
||||
assert.notEqual(one, other);
|
||||
store.delete('foo');
|
||||
assert.equal(store.get('foo'), undefined);
|
||||
});
|
||||
|
||||
it('add_doc_to_dispatcher', () => {
|
||||
const store = new MockStore('s');
|
||||
//const syncer = new MockSyncer();
|
||||
const dispatcher = new Dispatcher(store, undefined);
|
||||
const user = new User('foo');
|
||||
dispatcher.add(user.id, user);
|
||||
assert(dispatcher.isDirty());
|
||||
});
|
||||
|
||||
it('dispatch_keyindex', () => {
|
||||
const s = 'foo';
|
||||
const k = toIndexKey(s);
|
||||
const v = fromIndexKey(k);
|
||||
assert.equal(s, v);
|
||||
});
|
||||
|
||||
|
||||
});
|
||||
212
apps/cic-meta/tests/2_sync.ts
Normal file
212
apps/cic-meta/tests/2_sync.ts
Normal file
@@ -0,0 +1,212 @@
|
||||
import * as Automerge from 'automerge';
|
||||
import assert = require('assert');
|
||||
|
||||
import * as pgp from 'openpgp';
|
||||
import * as fs from 'fs';
|
||||
|
||||
import { PGPSigner } from '../src/auth';
|
||||
|
||||
import { Syncable, ArgPair } from '../src/sync';
|
||||
|
||||
import { MockKeyStore, MockSigner } from './mock';
|
||||
|
||||
|
||||
describe('sync', async () => {
|
||||
it('sync_merge', () => {
|
||||
const mockSigner = new MockSigner();
|
||||
const s = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
s.setSigner(mockSigner);
|
||||
const changePair = new ArgPair('xyzzy', 42);
|
||||
s.update([changePair], 'ch-ch-cha-changes');
|
||||
assert.equal(s.m.data['xyzzy'], 42)
|
||||
assert.equal(s.m.data['bar'], 'baz')
|
||||
assert.equal(s.m['id'], 'foo')
|
||||
assert.equal(Automerge.getHistory(s.m).length, 2);
|
||||
});
|
||||
|
||||
it('sync_serialize', () => {
|
||||
const mockSigner = new MockSigner();
|
||||
const s = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
s.setSigner(mockSigner);
|
||||
const j = s.toJSON();
|
||||
const ss = Syncable.fromJSON(j);
|
||||
assert.equal(ss.m['id'], 'foo');
|
||||
assert.equal(ss.m['data']['bar'], 'baz');
|
||||
assert.equal(Automerge.getHistory(ss.m).length, 1);
|
||||
});
|
||||
|
||||
it('sync_sign_and_wrap', () => {
|
||||
const mockSigner = new MockSigner();
|
||||
const s = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
s.setSigner(mockSigner);
|
||||
s.onwrap = (e) => {
|
||||
const j = e.toJSON();
|
||||
const v = JSON.parse(j);
|
||||
assert.deepEqual(v.payload, e.o.payload);
|
||||
|
||||
}
|
||||
s.sign();
|
||||
});
|
||||
it('sync_verify_success', async () => {
|
||||
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||
const pks = await pgp.key.readArmored(pksa);
|
||||
await pks.keys[0].decrypt('merman');
|
||||
await pks.keys[1].decrypt('beastman');
|
||||
|
||||
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||
const pubks = await pgp.key.readArmored(pubksa);
|
||||
|
||||
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||
|
||||
const oneSigner = new PGPSigner(oneStore);
|
||||
const twoSigner = new PGPSigner(twoStore);
|
||||
const threeSigner = new PGPSigner(threeStore);
|
||||
|
||||
const x = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
x.setSigner(oneSigner);
|
||||
|
||||
// TODO: make this look better
|
||||
x.onwrap = (e) => {
|
||||
let updateData = new ArgPair('bar', 'xyzzy');
|
||||
x.update([updateData], 'change one');
|
||||
|
||||
x.onwrap = (e) => {
|
||||
x.setSigner(twoSigner);
|
||||
updateData = new ArgPair('bar', 42);
|
||||
x.update([updateData], 'change two');
|
||||
|
||||
x.onwrap = (e) => {
|
||||
const p = e.unwrap();
|
||||
p.setSigner(twoSigner);
|
||||
p.onauthenticate = (v) => {
|
||||
assert(v);
|
||||
}
|
||||
p.authenticate();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
};
|
||||
|
||||
x.sign();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
|
||||
});
|
||||
|
||||
it('sync_verify_fail', async () => {
|
||||
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||
const pks = await pgp.key.readArmored(pksa);
|
||||
await pks.keys[0].decrypt('merman');
|
||||
await pks.keys[1].decrypt('beastman');
|
||||
|
||||
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||
const pubks = await pgp.key.readArmored(pubksa);
|
||||
|
||||
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||
|
||||
const oneSigner = new PGPSigner(oneStore);
|
||||
const twoSigner = new PGPSigner(twoStore);
|
||||
const threeSigner = new PGPSigner(threeStore);
|
||||
|
||||
const x = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
x.setSigner(oneSigner);
|
||||
|
||||
// TODO: make this look better
|
||||
x.onwrap = (e) => {
|
||||
let updateData = new ArgPair('bar', 'xyzzy');
|
||||
x.update([updateData], 'change one');
|
||||
|
||||
x.onwrap = (e) => {
|
||||
x.setSigner(twoSigner);
|
||||
updateData = new ArgPair('bar', 42);
|
||||
x.update([updateData], 'change two');
|
||||
|
||||
x.onwrap = (e) => {
|
||||
const p = e.unwrap();
|
||||
p.setSigner(threeSigner);
|
||||
p.onauthenticate = (v) => {
|
||||
assert(!v);
|
||||
}
|
||||
p.authenticate();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
};
|
||||
|
||||
x.sign();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
|
||||
});
|
||||
|
||||
xit('sync_verify_shallow_tricked', async () => {
|
||||
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc');
|
||||
const pks = await pgp.key.readArmored(pksa);
|
||||
await pks.keys[0].decrypt('merman');
|
||||
await pks.keys[1].decrypt('beastman');
|
||||
|
||||
const pubksa = fs.readFileSync(__dirname + '/publickeys.asc');
|
||||
const pubks = await pgp.key.readArmored(pubksa);
|
||||
|
||||
const oneStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||
const twoStore = new MockKeyStore(pks.keys[1], pubks.keys);
|
||||
const threeStore = new MockKeyStore(pks.keys[2], [pubks.keys[0], pubks.keys[2]]);
|
||||
|
||||
const oneSigner = new PGPSigner(oneStore);
|
||||
const twoSigner = new PGPSigner(twoStore);
|
||||
const threeSigner = new PGPSigner(threeStore);
|
||||
|
||||
const x = new Syncable('foo', {
|
||||
bar: 'baz',
|
||||
});
|
||||
x.setSigner(twoSigner);
|
||||
|
||||
// TODO: make this look better
|
||||
x.onwrap = (e) => {
|
||||
let updateData = new ArgPair('bar', 'xyzzy');
|
||||
x.update([updateData], 'change one');
|
||||
|
||||
x.onwrap = (e) => {
|
||||
updateData = new ArgPair('bar', 42);
|
||||
x.update([updateData], 'change two');
|
||||
x.setSigner(oneSigner);
|
||||
|
||||
x.onwrap = (e) => {
|
||||
const p = e.unwrap();
|
||||
p.setSigner(threeSigner);
|
||||
p.onauthenticate = (v) => {
|
||||
assert(v);
|
||||
p.onauthenticate = (v) => {
|
||||
assert(!v);
|
||||
}
|
||||
p.authenticate(true);
|
||||
}
|
||||
p.authenticate();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
};
|
||||
|
||||
x.sign();
|
||||
}
|
||||
|
||||
x.sign();
|
||||
|
||||
});
|
||||
});
|
||||
14
apps/cic-meta/tests/3_transport.ts
Normal file
14
apps/cic-meta/tests/3_transport.ts
Normal file
@@ -0,0 +1,14 @@
|
||||
import * as assert from 'assert';
|
||||
|
||||
import { MockPubSub, MockConsumer } from './mock';
|
||||
|
||||
describe('transport', () => {
|
||||
it('pub_sub', () => {
|
||||
const c = new MockConsumer();
|
||||
const ps = new MockPubSub('foo', c);
|
||||
ps.pub('foo');
|
||||
ps.pub('bar');
|
||||
ps.flush();
|
||||
assert.deepEqual(c.omnoms, ['foo', 'bar']);
|
||||
});
|
||||
});
|
||||
46
apps/cic-meta/tests/4_auth.ts
Normal file
46
apps/cic-meta/tests/4_auth.ts
Normal file
@@ -0,0 +1,46 @@
|
||||
import assert = require('assert');
|
||||
import pgp = require('openpgp');
|
||||
import crypto = require('crypto');
|
||||
|
||||
import { Syncable, ArgPair } from '../src/sync';
|
||||
|
||||
import { MockKeyStore, MockSignable } from './mock';
|
||||
|
||||
import { PGPSigner } from '../src/auth';
|
||||
|
||||
|
||||
describe('auth', async () => {
|
||||
await it('digest', async () => {
|
||||
const opts = {
|
||||
userIds: [
|
||||
{
|
||||
name: 'John Marston',
|
||||
email: 'red@dead.com',
|
||||
},
|
||||
],
|
||||
numBits: 2048,
|
||||
passphrase: 'foo',
|
||||
};
|
||||
const pkgen = await pgp.generateKey(opts);
|
||||
const pka = pkgen.privateKeyArmored;
|
||||
const pks = await pgp.key.readArmored(pka);
|
||||
await pks.keys[0].decrypt('foo');
|
||||
const pubka = pkgen.publicKeyArmored;
|
||||
const pubks = await pgp.key.readArmored(pubka);
|
||||
const keyStore = new MockKeyStore(pks.keys[0], pubks.keys);
|
||||
const s = new PGPSigner(keyStore);
|
||||
|
||||
const message = await pgp.cleartext.fromText('foo');
|
||||
s.onverify = (ok) => {
|
||||
assert(ok);
|
||||
}
|
||||
s.onsign = (signature) => {
|
||||
s.onverify((v) => {
|
||||
console.log('bar', v);
|
||||
});
|
||||
s.verify('foo', signature);
|
||||
}
|
||||
|
||||
await s.sign('foo');
|
||||
});
|
||||
});
|
||||
47
apps/cic-meta/tests/999_functional.ts
Normal file
47
apps/cic-meta/tests/999_functional.ts
Normal file
@@ -0,0 +1,47 @@
|
||||
import * as assert from 'assert';
|
||||
import * as pgp from 'openpgp';
|
||||
|
||||
import { Dispatcher } from '../src/dispatch';
|
||||
import { User } from '../src/assets/user';
|
||||
import { PGPSigner, KeyStore } from '../src/auth';
|
||||
import { SubConsumer } from '../src/transport';
|
||||
|
||||
import { MockStore, MockPubSub, MockConsumer, MockKeyStore } from './mock';
|
||||
|
||||
async function createKeyStore() {
|
||||
const opts = {
|
||||
userIds: [
|
||||
{
|
||||
name: 'John Marston',
|
||||
email: 'red@dead.com',
|
||||
},
|
||||
],
|
||||
numBits: 2048,
|
||||
passphrase: 'foo',
|
||||
};
|
||||
const pkgen = await pgp.generateKey(opts);
|
||||
const pka = pkgen.privateKeyArmored;
|
||||
const pks = await pgp.key.readArmored(pka);
|
||||
await pks.keys[0].decrypt('foo');
|
||||
return new MockKeyStore(pks.keys[0], []);
|
||||
}
|
||||
|
||||
describe('fullchain', async () => {
|
||||
it('dispatch_and_publish_user', async () => {
|
||||
const g = await createKeyStore();
|
||||
const n = new PGPSigner(g);
|
||||
const u = new User('u1', {});
|
||||
u.setSigner(n);
|
||||
u.setName('Nico', 'Bellic');
|
||||
const s = new MockStore('fooStore');
|
||||
const c = new MockConsumer();
|
||||
const p = new MockPubSub('fooPubSub', c);
|
||||
const d = new Dispatcher(s, p);
|
||||
u.onwrap = (e) => {
|
||||
d.add(u.id, e);
|
||||
d.sync(0);
|
||||
assert.equal(p.pubs.length, 1);
|
||||
};
|
||||
u.sign();
|
||||
});
|
||||
});
|
||||
150
apps/cic-meta/tests/mock.ts
Normal file
150
apps/cic-meta/tests/mock.ts
Normal file
@@ -0,0 +1,150 @@
|
||||
import * as crypto from 'crypto';
|
||||
|
||||
import { Signable, Signature, KeyStore } from '../src/auth';
|
||||
import { Store } from '../src/store';
|
||||
import { PubSub, SubConsumer } from '../src/transport';
|
||||
import { Syncable } from '../src/sync';
|
||||
|
||||
class MockStore implements Store {
|
||||
|
||||
contents: Object
|
||||
name: string
|
||||
|
||||
constructor(name:string) {
|
||||
this.name = name;
|
||||
this.contents = {};
|
||||
}
|
||||
|
||||
public put(k:string, v:Syncable, existsOk = false) {
|
||||
if (!existsOk && this.contents[k] !== undefined) {
|
||||
throw '"' + k + '" already exists in store ' + this.name;
|
||||
}
|
||||
this.contents[k] = v;
|
||||
}
|
||||
|
||||
public get(k:string): Syncable {
|
||||
return this.contents[k];
|
||||
}
|
||||
|
||||
public delete(k:string) {
|
||||
delete this.contents[k];
|
||||
}
|
||||
}
|
||||
|
||||
class MockSigner {
|
||||
onsign: (string) => void
|
||||
onverify: (boolean) => void
|
||||
public verify(src:string, signature:Signature) {
|
||||
return true;
|
||||
}
|
||||
|
||||
public sign(s:string):boolean {
|
||||
this.onsign('there would be a signature here');
|
||||
return true;
|
||||
}
|
||||
|
||||
public prepare(m:Signable):boolean {
|
||||
return true;
|
||||
}
|
||||
|
||||
public fingerprint():string {
|
||||
return '';
|
||||
}
|
||||
}
|
||||
|
||||
class MockConsumer implements SubConsumer {
|
||||
|
||||
omnoms: Array<string>
|
||||
|
||||
constructor() {
|
||||
this.omnoms = Array<string>();
|
||||
}
|
||||
|
||||
public post(v:string) {
|
||||
this.omnoms.push(v);
|
||||
}
|
||||
}
|
||||
|
||||
class MockPubSub implements PubSub {
|
||||
|
||||
pubs: Array<string>
|
||||
consumer: SubConsumer
|
||||
|
||||
constructor(name:string, consumer:SubConsumer) {
|
||||
this.pubs = Array<string>();
|
||||
this.consumer = consumer;
|
||||
}
|
||||
|
||||
public pub(v:string): boolean {
|
||||
this.pubs.push(v);
|
||||
return true;
|
||||
}
|
||||
|
||||
public flush() {
|
||||
while (this.pubs.length > 0) {
|
||||
const s = this.pubs.shift();
|
||||
this.consumer.post(s);
|
||||
}
|
||||
}
|
||||
|
||||
public close() {
|
||||
}
|
||||
}
|
||||
|
||||
class MockSignable implements Signable {
|
||||
|
||||
src: string
|
||||
dst: string
|
||||
|
||||
constructor(src:string) {
|
||||
this.src = src;
|
||||
}
|
||||
|
||||
public digest():string {
|
||||
const h = crypto.createHash('sha256');
|
||||
h.update(this.src);
|
||||
this.dst= h.digest('hex');
|
||||
return this.dst;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
class MockKeyStore implements KeyStore {
|
||||
|
||||
pk: any
|
||||
pubks: Array<any>
|
||||
|
||||
constructor(pk:any, pubks:Array<any>) {
|
||||
this.pk = pk;
|
||||
this.pubks = pubks;
|
||||
}
|
||||
|
||||
public getPrivateKey(): any {
|
||||
return this.pk;
|
||||
}
|
||||
|
||||
public getTrustedKeys(): Array<any> {
|
||||
return this.pubks;
|
||||
}
|
||||
|
||||
public getTrustedActiveKeys(): Array<any> {
|
||||
return [];
|
||||
}
|
||||
|
||||
public getEncryptKeys(): Array<any> {
|
||||
return [];
|
||||
}
|
||||
|
||||
public getFingerprint(): string {
|
||||
return '';
|
||||
}
|
||||
}
|
||||
|
||||
export {
|
||||
MockStore,
|
||||
MockPubSub,
|
||||
MockConsumer,
|
||||
MockSignable,
|
||||
MockKeyStore,
|
||||
MockSigner,
|
||||
};
|
||||
@@ -1,10 +1,13 @@
|
||||
import Automerge = require('automerge');
|
||||
import assert = require('assert');
|
||||
import fs = require('fs');
|
||||
import pgp = require('openpgp');
|
||||
import sqlite = require('sqlite3');
|
||||
|
||||
import * as handlers from '../scripts/server/handlers';
|
||||
import { Envelope, Syncable, ArgPair, PGPKeyStore, PGPSigner, KeyStore, Signer, SqliteAdapter } from 'crdt-meta';
|
||||
import { Envelope, Syncable, ArgPair } from '../src/sync';
|
||||
import { PGPKeyStore, PGPSigner, KeyStore, Signer } from '../src/auth';
|
||||
import { SqliteAdapter } from '../src/db';
|
||||
|
||||
function createKeystore() {
|
||||
const pksa = fs.readFileSync(__dirname + '/privatekeys.asc', 'utf-8');
|
||||
|
||||
@@ -33,9 +33,7 @@ elif args.v:
|
||||
|
||||
config = confini.Config(args.c, args.env_prefix)
|
||||
config.process()
|
||||
config.add(args.q, '_CELERY_QUEUE', True)
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
logg.debug('config loaded from {}:\n{}'.format(args.c, config))
|
||||
|
||||
# connect to database
|
||||
dsn = dsn_from_config(config)
|
||||
|
||||
@@ -6,10 +6,11 @@ import time
|
||||
import semver
|
||||
|
||||
# local imports
|
||||
from cic_notify.error import PleaseCommitFirstError
|
||||
|
||||
logg = logging.getLogger()
|
||||
|
||||
version = (0, 4, 0, 'alpha.4')
|
||||
version = (0, 4, 0, 'alpha.3')
|
||||
|
||||
version_object = semver.VersionInfo(
|
||||
major=version[0],
|
||||
@@ -17,4 +18,27 @@ version_object = semver.VersionInfo(
|
||||
patch=version[2],
|
||||
prerelease=version[3],
|
||||
)
|
||||
|
||||
version_string = str(version_object)
|
||||
|
||||
|
||||
def git_hash():
|
||||
import subprocess
|
||||
|
||||
git_hash = subprocess.run(['git', 'rev-parse', 'HEAD'], capture_output=True)
|
||||
git_hash_brief = git_hash.stdout.decode('utf-8')[:8]
|
||||
return git_hash_brief
|
||||
|
||||
|
||||
try:
|
||||
version_git = git_hash()
|
||||
version_string += '+build.{}'.format(version_git)
|
||||
except FileNotFoundError:
|
||||
time_string_pair = str(time.time()).split('.')
|
||||
version_string += '+build.{}{:<09d}'.format(
|
||||
time_string_pair[0],
|
||||
int(time_string_pair[1]),
|
||||
)
|
||||
logg.info(f'Final version string will be {version_string}')
|
||||
|
||||
__version_string__ = version_string
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
[metadata]
|
||||
name = cic-notify
|
||||
version= 0.4.0a3
|
||||
description = CIC notifications service
|
||||
author = Louis Holbrook
|
||||
author_email = dev@holbrook.no
|
||||
|
||||
@@ -1,31 +1,9 @@
|
||||
# standard imports
|
||||
import logging
|
||||
import subprocess
|
||||
import time
|
||||
from setuptools import setup
|
||||
|
||||
# third-party imports
|
||||
|
||||
# local imports
|
||||
from cic_notify.version import version_string
|
||||
|
||||
logg = logging.getLogger()
|
||||
|
||||
|
||||
def git_hash():
|
||||
git_hash = subprocess.run(['git', 'rev-parse', 'HEAD'], capture_output=True)
|
||||
git_hash_brief = git_hash.stdout.decode('utf-8')[:8]
|
||||
return git_hash_brief
|
||||
|
||||
|
||||
try:
|
||||
version_git = git_hash()
|
||||
version_string += '+build.{}'.format(version_git)
|
||||
except FileNotFoundError:
|
||||
time_string_pair = str(time.time()).split('.')
|
||||
version_string += '+build.{}{:<09d}'.format(
|
||||
time_string_pair[0],
|
||||
int(time_string_pair[1]),
|
||||
)
|
||||
logg.info(f'Final version string will be {version_string}')
|
||||
|
||||
|
||||
requirements = []
|
||||
@@ -47,6 +25,6 @@ while True:
|
||||
test_requirements_file.close()
|
||||
|
||||
setup(
|
||||
version=version_string,
|
||||
install_requires=requirements,
|
||||
tests_require=test_requirements)
|
||||
tests_require=test_requirements,
|
||||
)
|
||||
|
||||
@@ -20,5 +20,5 @@ TRANSITIONS=/usr/src/cic-ussd/transitions/
|
||||
|
||||
[client]
|
||||
host =
|
||||
port =
|
||||
port =
|
||||
ssl =
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
[celery]
|
||||
BROKER_URL=redis://
|
||||
RESULT_URL=redis://
|
||||
BROKER_URL=redis://redis:6379
|
||||
RESULT_URL=redis://redis:6379
|
||||
|
||||
[redis]
|
||||
HOSTNAME=redis
|
||||
|
||||
@@ -8,12 +8,12 @@ from cic_types.processor import generate_metadata_pointer
|
||||
|
||||
# local imports
|
||||
from cic_ussd.chain import Chain
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.metadata import blockchain_address_to_metadata_pointer
|
||||
from cic_ussd.redis import get_cached_data
|
||||
|
||||
|
||||
def define_account_tx_metadata(user: Account):
|
||||
def define_account_tx_metadata(user: User):
|
||||
# get sender metadata
|
||||
identifier = blockchain_address_to_metadata_pointer(
|
||||
blockchain_address=user.blockchain_address
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
"""Create account table
|
||||
"""Create user table
|
||||
|
||||
Revision ID: f289e8510444
|
||||
Revises:
|
||||
@@ -17,7 +17,7 @@ depends_on = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
op.create_table('account',
|
||||
op.create_table('user',
|
||||
sa.Column('id', sa.Integer(), nullable=False),
|
||||
sa.Column('blockchain_address', sa.String(), nullable=False),
|
||||
sa.Column('phone_number', sa.String(), nullable=False),
|
||||
@@ -29,11 +29,11 @@ def upgrade():
|
||||
sa.Column('updated', sa.DateTime(), nullable=False),
|
||||
sa.PrimaryKeyConstraint('id')
|
||||
)
|
||||
op.create_index(op.f('ix_account_phone_number'), 'account', ['phone_number'], unique=True)
|
||||
op.create_index(op.f('ix_account_blockchain_address'), 'account', ['blockchain_address'], unique=True)
|
||||
op.create_index(op.f('ix_user_phone_number'), 'user', ['phone_number'], unique=True)
|
||||
op.create_index(op.f('ix_user_blockchain_address'), 'user', ['blockchain_address'], unique=True)
|
||||
|
||||
|
||||
def downgrade():
|
||||
op.drop_index(op.f('ix_account_blockchain_address'), table_name='account')
|
||||
op.drop_index(op.f('ix_account_phone_number'), table_name='account')
|
||||
op.drop_table('account')
|
||||
op.drop_index(op.f('ix_user_blockchain_address'), table_name='user')
|
||||
op.drop_index(op.f('ix_user_phone_number'), table_name='user')
|
||||
op.drop_table('user')
|
||||
|
||||
@@ -16,12 +16,12 @@ class AccountStatus(IntEnum):
|
||||
RESET = 4
|
||||
|
||||
|
||||
class Account(SessionBase):
|
||||
class User(SessionBase):
|
||||
"""
|
||||
This class defines a user record along with functions responsible for hashing the user's corresponding password and
|
||||
subsequently verifying a password's validity given an input to compare against the persisted hash.
|
||||
"""
|
||||
__tablename__ = 'account'
|
||||
__tablename__ = 'user'
|
||||
|
||||
blockchain_address = Column(String)
|
||||
phone_number = Column(String)
|
||||
@@ -38,7 +38,7 @@ class Account(SessionBase):
|
||||
self.account_status = AccountStatus.PENDING.value
|
||||
|
||||
def __repr__(self):
|
||||
return f'<Account: {self.blockchain_address}>'
|
||||
return f'<User: {self.blockchain_address}>'
|
||||
|
||||
def create_password(self, password):
|
||||
"""This method takes a password value and hashes the value before assigning it to the corresponding
|
||||
@@ -10,7 +10,7 @@ from tinydb.table import Document
|
||||
from typing import Optional
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.db.models.ussd_session import UssdSession
|
||||
from cic_ussd.db.models.task_tracker import TaskTracker
|
||||
from cic_ussd.menu.ussd_menu import UssdMenu
|
||||
@@ -143,10 +143,10 @@ def get_account_status(phone_number) -> str:
|
||||
:return: The user account status.
|
||||
:rtype: str
|
||||
"""
|
||||
user = Account.session.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = User.session.query(User).filter_by(phone_number=phone_number).first()
|
||||
status = user.get_account_status()
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
return status
|
||||
|
||||
@@ -269,12 +269,12 @@ def cache_account_creation_task_id(phone_number: str, task_id: str):
|
||||
redis_cache.persist(name=task_id)
|
||||
|
||||
|
||||
def process_current_menu(ussd_session: Optional[dict], user: Account, user_input: str) -> Document:
|
||||
def process_current_menu(ussd_session: Optional[dict], user: User, user_input: str) -> Document:
|
||||
"""This function checks user input and returns a corresponding ussd menu
|
||||
:param ussd_session: An in db ussd session object.
|
||||
:type ussd_session: UssdSession
|
||||
:param user: A user object.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param user_input: The user's input.
|
||||
:type user_input: str
|
||||
:return: An in memory ussd menu object.
|
||||
@@ -324,7 +324,7 @@ def process_menu_interaction_requests(chain_str: str,
|
||||
|
||||
else:
|
||||
# get user
|
||||
user = Account.session.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = User.session.query(User).filter_by(phone_number=phone_number).first()
|
||||
|
||||
# find any existing ussd session
|
||||
existing_ussd_session = UssdSession.session.query(UssdSession).filter_by(
|
||||
@@ -390,10 +390,10 @@ def reset_pin(phone_number: str) -> str:
|
||||
:return: The status of the pin reset.
|
||||
:rtype: str
|
||||
"""
|
||||
user = Account.session.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = User.session.query(User).filter_by(phone_number=phone_number).first()
|
||||
user.reset_account_pin()
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
response = f'Pin reset for user {phone_number} is successful!'
|
||||
return response
|
||||
|
||||
@@ -5,7 +5,7 @@ from typing import Optional
|
||||
import phonenumbers
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
|
||||
def process_phone_number(phone_number: str, region: str):
|
||||
@@ -30,14 +30,14 @@ def process_phone_number(phone_number: str, region: str):
|
||||
return parsed_phone_number
|
||||
|
||||
|
||||
def get_user_by_phone_number(phone_number: str) -> Optional[Account]:
|
||||
def get_user_by_phone_number(phone_number: str) -> Optional[User]:
|
||||
"""This function queries the database for a user based on the provided phone number.
|
||||
:param phone_number: A valid phone number.
|
||||
:type phone_number: str
|
||||
:return: A user object matching a given phone number
|
||||
:rtype: Account|None
|
||||
:rtype: User|None
|
||||
"""
|
||||
# consider adding region to user's metadata
|
||||
phone_number = process_phone_number(phone_number=phone_number, region='KE')
|
||||
user = Account.session.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = User.session.query(User).filter_by(phone_number=phone_number).first()
|
||||
return user
|
||||
|
||||
@@ -13,7 +13,7 @@ from tinydb.table import Document
|
||||
from cic_ussd.account import define_account_tx_metadata, retrieve_account_statement
|
||||
from cic_ussd.balance import BalanceManager, compute_operational_balance, get_cached_operational_balance
|
||||
from cic_ussd.chain import Chain
|
||||
from cic_ussd.db.models.account import AccountStatus, Account
|
||||
from cic_ussd.db.models.user import AccountStatus, User
|
||||
from cic_ussd.db.models.ussd_session import UssdSession
|
||||
from cic_ussd.error import MetadataNotFoundError
|
||||
from cic_ussd.menu.ussd_menu import UssdMenu
|
||||
@@ -28,13 +28,13 @@ from cic_types.models.person import generate_metadata_pointer, get_contact_data_
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def process_pin_authorization(display_key: str, user: Account, **kwargs) -> str:
|
||||
def process_pin_authorization(display_key: str, user: User, **kwargs) -> str:
|
||||
"""
|
||||
This method provides translation for all ussd menu entries that follow the pin authorization pattern.
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
:type display_key: str
|
||||
:param user: The user in a running USSD session.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param kwargs: Any additional information required by the text values in the internationalization files.
|
||||
:type kwargs
|
||||
:return: A string value corresponding the ussd menu's text value.
|
||||
@@ -55,13 +55,13 @@ def process_pin_authorization(display_key: str, user: Account, **kwargs) -> str:
|
||||
)
|
||||
|
||||
|
||||
def process_exit_insufficient_balance(display_key: str, user: Account, ussd_session: dict):
|
||||
def process_exit_insufficient_balance(display_key: str, user: User, ussd_session: dict):
|
||||
"""This function processes the exit menu letting users their account balance is insufficient to perform a specific
|
||||
transaction.
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
:type display_key: str
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param ussd_session: A JSON serialized in-memory ussd session object
|
||||
:type ussd_session: dict
|
||||
:return: Corresponding translation text response
|
||||
@@ -90,12 +90,12 @@ def process_exit_insufficient_balance(display_key: str, user: Account, ussd_sess
|
||||
)
|
||||
|
||||
|
||||
def process_exit_successful_transaction(display_key: str, user: Account, ussd_session: dict):
|
||||
def process_exit_successful_transaction(display_key: str, user: User, ussd_session: dict):
|
||||
"""This function processes the exit menu after a successful initiation for a transfer of tokens.
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
:type display_key: str
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param ussd_session: A JSON serialized in-memory ussd session object
|
||||
:type ussd_session: dict
|
||||
:return: Corresponding translation text response
|
||||
@@ -118,11 +118,11 @@ def process_exit_successful_transaction(display_key: str, user: Account, ussd_se
|
||||
)
|
||||
|
||||
|
||||
def process_transaction_pin_authorization(user: Account, display_key: str, ussd_session: dict):
|
||||
def process_transaction_pin_authorization(user: User, display_key: str, ussd_session: dict):
|
||||
"""This function processes pin authorization where making a transaction is concerned. It constructs a
|
||||
pre-transaction response menu that shows the details of the transaction.
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
:type display_key: str
|
||||
:param ussd_session: The USSD session determining what user data needs to be extracted and added to the menu's
|
||||
@@ -151,7 +151,7 @@ def process_transaction_pin_authorization(user: Account, display_key: str, ussd_
|
||||
)
|
||||
|
||||
|
||||
def process_account_balances(user: Account, display_key: str, ussd_session: dict):
|
||||
def process_account_balances(user: User, display_key: str, ussd_session: dict):
|
||||
"""
|
||||
:param user:
|
||||
:type user:
|
||||
@@ -205,7 +205,7 @@ def format_transactions(transactions: list, preferred_language: str):
|
||||
return formatted_transactions
|
||||
|
||||
|
||||
def process_display_user_metadata(user: Account, display_key: str):
|
||||
def process_display_user_metadata(user: User, display_key: str):
|
||||
"""
|
||||
:param user:
|
||||
:type user:
|
||||
@@ -238,7 +238,7 @@ def process_display_user_metadata(user: Account, display_key: str):
|
||||
raise MetadataNotFoundError(f'Expected person metadata but found none in cache for key: {key}')
|
||||
|
||||
|
||||
def process_account_statement(user: Account, display_key: str, ussd_session: dict):
|
||||
def process_account_statement(user: User, display_key: str, ussd_session: dict):
|
||||
"""
|
||||
:param user:
|
||||
:type user:
|
||||
@@ -301,12 +301,12 @@ def process_account_statement(user: Account, display_key: str, ussd_session: dic
|
||||
)
|
||||
|
||||
|
||||
def process_start_menu(display_key: str, user: Account):
|
||||
def process_start_menu(display_key: str, user: User):
|
||||
"""This function gets data on an account's balance and token in order to append it to the start of the start menu's
|
||||
title. It passes said arguments to the translation function and returns the appropriate corresponding text from the
|
||||
translation files.
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
:type display_key: str
|
||||
:return: Corresponding translation text response
|
||||
@@ -361,13 +361,13 @@ def retrieve_most_recent_ussd_session(phone_number: str) -> UssdSession:
|
||||
return last_ussd_session
|
||||
|
||||
|
||||
def process_request(user_input: str, user: Account, ussd_session: Optional[dict] = None) -> Document:
|
||||
def process_request(user_input: str, user: User, ussd_session: Optional[dict] = None) -> Document:
|
||||
"""This function assesses a request based on the user from the request comes, the session_id and the user's
|
||||
input. It determines whether the request translates to a return to an existing session by checking whether the
|
||||
provided session id exists in the database or whether the creation of a new ussd session object is warranted.
|
||||
It then returns the appropriate ussd menu text values.
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param user_input: The value a user enters in the ussd menu.
|
||||
:type user_input: str
|
||||
:param ussd_session: A JSON serialized in-memory ussd session object
|
||||
@@ -415,14 +415,14 @@ def process_request(user_input: str, user: Account, ussd_session: Optional[dict]
|
||||
return UssdMenu.find_by_name(name='initial_pin_entry')
|
||||
|
||||
|
||||
def next_state(ussd_session: dict, user: Account, user_input: str) -> str:
|
||||
def next_state(ussd_session: dict, user: User, user_input: str) -> str:
|
||||
"""This function navigates the state machine based on the ussd session object and user inputs it receives.
|
||||
It checks the user input and provides the successive state in the state machine. It then updates the session's
|
||||
state attribute with the new state.
|
||||
:param ussd_session: A JSON serialized in-memory ussd session object
|
||||
:type ussd_session: dict
|
||||
:param user: The user requesting access to the ussd menu.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param user_input: The value a user enters in the ussd menu.
|
||||
:type user_input: str
|
||||
:return: A string value corresponding the successive give a specific state in the state machine.
|
||||
@@ -438,7 +438,7 @@ def custom_display_text(
|
||||
display_key: str,
|
||||
menu_name: str,
|
||||
ussd_session: dict,
|
||||
user: Account) -> str:
|
||||
user: User) -> str:
|
||||
"""This function extracts the appropriate session data based on the current menu name. It then inserts them as
|
||||
keywords in the i18n function.
|
||||
:param display_key: The path in the translation files defining an appropriate ussd response
|
||||
@@ -446,7 +446,7 @@ def custom_display_text(
|
||||
:param menu_name: The name by which a specific menu can be identified.
|
||||
:type menu_name: str
|
||||
:param user: The user in a running USSD session.
|
||||
:type user: Account
|
||||
:type user: User
|
||||
:param ussd_session: A JSON serialized in-memory ussd session object
|
||||
:type ussd_session: dict
|
||||
:return: A string value corresponding the ussd menu's text value.
|
||||
|
||||
@@ -10,7 +10,7 @@ from urllib.parse import urlparse, parse_qs
|
||||
from sqlalchemy import desc
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import AccountStatus, Account
|
||||
from cic_ussd.db.models.user import AccountStatus, User
|
||||
from cic_ussd.operations import get_account_status, reset_pin
|
||||
from cic_ussd.validator import check_known_user
|
||||
|
||||
@@ -123,9 +123,9 @@ def process_locked_accounts_requests(env: dict) -> tuple:
|
||||
else:
|
||||
limit = r[1]
|
||||
|
||||
locked_accounts = Account.session.query(Account.blockchain_address).filter(
|
||||
Account.account_status == AccountStatus.LOCKED.value,
|
||||
Account.failed_pin_attempts >= 3).order_by(desc(Account.updated)).offset(offset).limit(limit).all()
|
||||
locked_accounts = User.session.query(User.blockchain_address).filter(
|
||||
User.account_status == AccountStatus.LOCKED.value,
|
||||
User.failed_pin_attempts >= 3).order_by(desc(User.updated)).offset(offset).limit(limit).all()
|
||||
|
||||
# convert lists to scalar blockchain addresses
|
||||
locked_accounts = [blockchain_address for (blockchain_address, ) in locked_accounts]
|
||||
|
||||
@@ -57,17 +57,19 @@ arg_parser.add_argument('--env-prefix',
|
||||
help='environment prefix for variables to overwrite configuration')
|
||||
args = arg_parser.parse_args()
|
||||
|
||||
# parse config
|
||||
config = Config(config_dir=args.c, env_prefix=args.env_prefix)
|
||||
config.process()
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
|
||||
# define log levels
|
||||
if args.vv:
|
||||
logging.getLogger().setLevel(logging.DEBUG)
|
||||
elif args.v:
|
||||
logging.getLogger().setLevel(logging.INFO)
|
||||
|
||||
# parse config
|
||||
config = Config(config_dir=args.c, env_prefix=args.env_prefix)
|
||||
config.process()
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
logg.debug('config loaded from {}:\n{}'.format(args.c, config))
|
||||
# log config vars
|
||||
logg.debug(config)
|
||||
|
||||
# initialize elements
|
||||
# set up translations
|
||||
|
||||
@@ -6,7 +6,6 @@ import tempfile
|
||||
|
||||
# third party imports
|
||||
import celery
|
||||
import i18n
|
||||
import redis
|
||||
from confini import Config
|
||||
|
||||
@@ -34,18 +33,18 @@ arg_parser.add_argument('-vv', action='store_true', help='be more verbose')
|
||||
arg_parser.add_argument('--env-prefix', default=os.environ.get('CONFINI_ENV_PREFIX'), dest='env_prefix', type=str, help='environment prefix for variables to overwrite configuration')
|
||||
args = arg_parser.parse_args()
|
||||
|
||||
# parse config
|
||||
config = Config(config_dir=args.c, env_prefix=args.env_prefix)
|
||||
config.process()
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
|
||||
# define log levels
|
||||
if args.vv:
|
||||
logging.getLogger().setLevel(logging.DEBUG)
|
||||
elif args.v:
|
||||
logging.getLogger().setLevel(logging.INFO)
|
||||
|
||||
# parse config
|
||||
config = Config(args.c, args.env_prefix)
|
||||
config.process()
|
||||
config.add(args.q, '_CELERY_QUEUE', True)
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
logg.debug('config loaded from {}:\n{}'.format(args.c, config))
|
||||
logg.debug(config)
|
||||
|
||||
# connect to database
|
||||
data_source_name = dsn_from_config(config)
|
||||
@@ -78,10 +77,6 @@ if key_file_path:
|
||||
validate_presence(path=key_file_path)
|
||||
Signer.key_file_path = key_file_path
|
||||
|
||||
# set up translations
|
||||
i18n.load_path.append(config.get('APP_LOCALE_PATH'))
|
||||
i18n.set('fallback', config.get('APP_LOCALE_FALLBACK'))
|
||||
|
||||
# set up celery
|
||||
current_app = celery.Celery(__name__)
|
||||
|
||||
|
||||
@@ -5,12 +5,12 @@ from typing import Tuple
|
||||
# third-party imports
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
logg = logging.getLogger(__file__)
|
||||
|
||||
|
||||
def process_mini_statement_request(state_machine_data: Tuple[str, dict, Account]):
|
||||
def process_mini_statement_request(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function compiles a brief statement of a user's last three inbound and outbound transactions and send the
|
||||
same as a message on their selected avenue for notification.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
|
||||
@@ -6,10 +6,10 @@ ussd menu facilitating the return of appropriate menu responses based on said us
|
||||
from typing import Tuple
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
|
||||
def menu_one_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_one_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that user input matches a string with value '1'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
@@ -20,7 +20,7 @@ def menu_one_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user_input == '1'
|
||||
|
||||
|
||||
def menu_two_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_two_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that user input matches a string with value '2'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -31,7 +31,7 @@ def menu_two_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user_input == '2'
|
||||
|
||||
|
||||
def menu_three_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_three_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that user input matches a string with value '3'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -42,7 +42,7 @@ def menu_three_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user_input == '3'
|
||||
|
||||
|
||||
def menu_four_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_four_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""
|
||||
This function checks that user input matches a string with value '4'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -54,7 +54,7 @@ def menu_four_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user_input == '4'
|
||||
|
||||
|
||||
def menu_five_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_five_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""
|
||||
This function checks that user input matches a string with value '5'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -66,7 +66,7 @@ def menu_five_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user_input == '5'
|
||||
|
||||
|
||||
def menu_zero_zero_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_zero_zero_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""
|
||||
This function checks that user input matches a string with value '00'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -78,7 +78,7 @@ def menu_zero_zero_selected(state_machine_data: Tuple[str, dict, Account]) -> bo
|
||||
return user_input == '00'
|
||||
|
||||
|
||||
def menu_ninety_nine_selected(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def menu_ninety_nine_selected(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""
|
||||
This function checks that user input matches a string with value '99'
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
|
||||
@@ -12,7 +12,7 @@ from typing import Tuple
|
||||
import bcrypt
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import AccountStatus, Account
|
||||
from cic_ussd.db.models.user import AccountStatus, User
|
||||
from cic_ussd.encoder import PasswordEncoder, create_password_hash
|
||||
from cic_ussd.operations import persist_session_to_db_task, create_or_update_session
|
||||
from cic_ussd.redis import InMemoryStore
|
||||
@@ -21,7 +21,7 @@ from cic_ussd.redis import InMemoryStore
|
||||
logg = logging.getLogger(__file__)
|
||||
|
||||
|
||||
def is_valid_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_valid_pin(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks a pin's validity by ensuring it has a length of for characters and the characters are
|
||||
numeric.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -37,7 +37,7 @@ def is_valid_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return pin_is_valid
|
||||
|
||||
|
||||
def is_authorized_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_authorized_pin(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks whether the user input confirming a specific pin matches the initial pin entered.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -48,7 +48,7 @@ def is_authorized_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user.verify_password(password=user_input)
|
||||
|
||||
|
||||
def is_locked_account(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_locked_account(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks whether a user's account is locked due to too many failed attempts.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -59,7 +59,7 @@ def is_locked_account(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user.get_account_status() == AccountStatus.LOCKED.name
|
||||
|
||||
|
||||
def save_initial_pin_to_session_data(state_machine_data: Tuple[str, dict, Account]):
|
||||
def save_initial_pin_to_session_data(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function hashes a pin and stores it in session data.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -94,7 +94,7 @@ def save_initial_pin_to_session_data(state_machine_data: Tuple[str, dict, Accoun
|
||||
persist_session_to_db_task(external_session_id=external_session_id, queue='cic-ussd')
|
||||
|
||||
|
||||
def pins_match(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def pins_match(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks whether the user input confirming a specific pin matches the initial pin entered.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -108,7 +108,7 @@ def pins_match(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return bcrypt.checkpw(user_input.encode(), initial_pin)
|
||||
|
||||
|
||||
def complete_pin_change(state_machine_data: Tuple[str, dict, Account]):
|
||||
def complete_pin_change(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function persists the user's pin to the database
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -116,11 +116,11 @@ def complete_pin_change(state_machine_data: Tuple[str, dict, Account]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
password_hash = ussd_session.get('session_data').get('initial_pin')
|
||||
user.password_hash = password_hash
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
|
||||
def is_blocked_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_blocked_pin(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks whether the user input confirming a specific pin matches the initial pin entered.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -131,7 +131,7 @@ def is_blocked_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return user.get_account_status() == AccountStatus.LOCKED.name
|
||||
|
||||
|
||||
def is_valid_new_pin(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_valid_new_pin(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks whether the user's new pin is a valid pin and that it isn't the same as the old one.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
|
||||
@@ -3,21 +3,21 @@ import logging
|
||||
from typing import Tuple
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
logg = logging.getLogger()
|
||||
|
||||
|
||||
def send_terms_to_user_if_required(state_machine_data: Tuple[str, dict, Account]):
|
||||
def send_terms_to_user_if_required(state_machine_data: Tuple[str, dict, User]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
logg.debug('Requires integration to cic-notify.')
|
||||
|
||||
|
||||
def process_mini_statement_request(state_machine_data: Tuple[str, dict, Account]):
|
||||
def process_mini_statement_request(state_machine_data: Tuple[str, dict, User]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
logg.debug('Requires integration to cic-notify.')
|
||||
|
||||
|
||||
def upsell_unregistered_recipient(state_machine_data: Tuple[str, dict, Account]):
|
||||
def upsell_unregistered_recipient(state_machine_data: Tuple[str, dict, User]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
logg.debug('Requires integration to cic-notify.')
|
||||
@@ -9,7 +9,7 @@ import celery
|
||||
# local imports
|
||||
from cic_ussd.balance import BalanceManager, compute_operational_balance
|
||||
from cic_ussd.chain import Chain
|
||||
from cic_ussd.db.models.account import AccountStatus, Account
|
||||
from cic_ussd.db.models.user import AccountStatus, User
|
||||
from cic_ussd.operations import save_to_in_memory_ussd_session_data
|
||||
from cic_ussd.phone_number import get_user_by_phone_number
|
||||
from cic_ussd.redis import create_cached_data_key, get_cached_data
|
||||
@@ -19,7 +19,7 @@ from cic_ussd.transactions import OutgoingTransactionProcessor
|
||||
logg = logging.getLogger(__file__)
|
||||
|
||||
|
||||
def is_valid_recipient(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_valid_recipient(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that a user exists, is not the initiator of the transaction, has an active account status
|
||||
and is authorized to perform standard transactions.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -34,7 +34,7 @@ def is_valid_recipient(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
return is_not_initiator and has_active_account_status and recipient is not None
|
||||
|
||||
|
||||
def is_valid_transaction_amount(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def is_valid_transaction_amount(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that the transaction amount provided is valid as per the criteria for the transaction
|
||||
being attempted.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -49,7 +49,7 @@ def is_valid_transaction_amount(state_machine_data: Tuple[str, dict, Account]) -
|
||||
return False
|
||||
|
||||
|
||||
def has_sufficient_balance(state_machine_data: Tuple[str, dict, Account]) -> bool:
|
||||
def has_sufficient_balance(state_machine_data: Tuple[str, dict, User]) -> bool:
|
||||
"""This function checks that the transaction amount provided is valid as per the criteria for the transaction
|
||||
being attempted.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
@@ -72,7 +72,7 @@ def has_sufficient_balance(state_machine_data: Tuple[str, dict, Account]) -> boo
|
||||
return int(user_input) <= operational_balance
|
||||
|
||||
|
||||
def save_recipient_phone_to_session_data(state_machine_data: Tuple[str, dict, Account]):
|
||||
def save_recipient_phone_to_session_data(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function saves the phone number corresponding the intended recipients blockchain account.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
@@ -85,7 +85,7 @@ def save_recipient_phone_to_session_data(state_machine_data: Tuple[str, dict, Ac
|
||||
save_to_in_memory_ussd_session_data(queue='cic-ussd', session_data=session_data, ussd_session=ussd_session)
|
||||
|
||||
|
||||
def retrieve_recipient_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
def retrieve_recipient_metadata(state_machine_data: Tuple[str, dict, User]):
|
||||
"""
|
||||
:param state_machine_data:
|
||||
:type state_machine_data:
|
||||
@@ -104,7 +104,7 @@ def retrieve_recipient_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
s_query_person_metadata.apply_async(queue='cic-ussd')
|
||||
|
||||
|
||||
def save_transaction_amount_to_session_data(state_machine_data: Tuple[str, dict, Account]):
|
||||
def save_transaction_amount_to_session_data(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function saves the phone number corresponding the intended recipients blockchain account.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
@@ -117,7 +117,7 @@ def save_transaction_amount_to_session_data(state_machine_data: Tuple[str, dict,
|
||||
save_to_in_memory_ussd_session_data(queue='cic-ussd', session_data=session_data, ussd_session=ussd_session)
|
||||
|
||||
|
||||
def process_transaction_request(state_machine_data: Tuple[str, dict, Account]):
|
||||
def process_transaction_request(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function saves the phone number corresponding the intended recipients blockchain account.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
|
||||
@@ -10,7 +10,7 @@ from cic_types.models.person import generate_vcard_from_contact_data, manage_ide
|
||||
|
||||
# local imports
|
||||
from cic_ussd.chain import Chain
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.error import MetadataNotFoundError
|
||||
from cic_ussd.metadata import blockchain_address_to_metadata_pointer
|
||||
from cic_ussd.operations import save_to_in_memory_ussd_session_data
|
||||
@@ -19,40 +19,40 @@ from cic_ussd.redis import get_cached_data
|
||||
logg = logging.getLogger(__file__)
|
||||
|
||||
|
||||
def change_preferred_language_to_en(state_machine_data: Tuple[str, dict, Account]):
|
||||
def change_preferred_language_to_en(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function changes the user's preferred language to english.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
"""
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
user.preferred_language = 'en'
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
|
||||
def change_preferred_language_to_sw(state_machine_data: Tuple[str, dict, Account]):
|
||||
def change_preferred_language_to_sw(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function changes the user's preferred language to swahili.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
"""
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
user.preferred_language = 'sw'
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
|
||||
def update_account_status_to_active(state_machine_data: Tuple[str, dict, Account]):
|
||||
def update_account_status_to_active(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function sets user's account to active.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
"""
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
user.activate_account()
|
||||
Account.session.add(user)
|
||||
Account.session.commit()
|
||||
User.session.add(user)
|
||||
User.session.commit()
|
||||
|
||||
|
||||
def process_gender_user_input(user: Account, user_input: str):
|
||||
def process_gender_user_input(user: User, user_input: str):
|
||||
"""
|
||||
:param user:
|
||||
:type user:
|
||||
@@ -74,7 +74,7 @@ def process_gender_user_input(user: Account, user_input: str):
|
||||
return gender
|
||||
|
||||
|
||||
def save_metadata_attribute_to_session_data(state_machine_data: Tuple[str, dict, Account]):
|
||||
def save_metadata_attribute_to_session_data(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function saves first name data to the ussd session in the redis cache.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -109,7 +109,7 @@ def save_metadata_attribute_to_session_data(state_machine_data: Tuple[str, dict,
|
||||
save_to_in_memory_ussd_session_data(queue='cic-ussd', session_data=session_data, ussd_session=ussd_session)
|
||||
|
||||
|
||||
def format_user_metadata(metadata: dict, user: Account):
|
||||
def format_user_metadata(metadata: dict, user: User):
|
||||
"""
|
||||
:param metadata:
|
||||
:type metadata:
|
||||
@@ -150,7 +150,7 @@ def format_user_metadata(metadata: dict, user: Account):
|
||||
}
|
||||
|
||||
|
||||
def save_complete_user_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
def save_complete_user_metadata(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function persists elements of the user metadata stored in session data
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: tuple
|
||||
@@ -171,7 +171,7 @@ def save_complete_user_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
s_create_person_metadata.apply_async(queue='cic-ussd')
|
||||
|
||||
|
||||
def edit_user_metadata_attribute(state_machine_data: Tuple[str, dict, Account]):
|
||||
def edit_user_metadata_attribute(state_machine_data: Tuple[str, dict, User]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
blockchain_address = user.blockchain_address
|
||||
key = generate_metadata_pointer(
|
||||
@@ -218,7 +218,7 @@ def edit_user_metadata_attribute(state_machine_data: Tuple[str, dict, Account]):
|
||||
s_edit_person_metadata.apply_async(queue='cic-ussd')
|
||||
|
||||
|
||||
def get_user_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
def get_user_metadata(state_machine_data: Tuple[str, dict, User]):
|
||||
user_input, ussd_session, user = state_machine_data
|
||||
blockchain_address = user.blockchain_address
|
||||
s_get_user_metadata = celery.signature(
|
||||
|
||||
@@ -7,14 +7,14 @@ from typing import Tuple
|
||||
from cic_types.models.person import generate_metadata_pointer
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.metadata import blockchain_address_to_metadata_pointer
|
||||
from cic_ussd.redis import get_cached_data
|
||||
|
||||
logg = logging.getLogger()
|
||||
|
||||
|
||||
def has_cached_user_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
def has_cached_user_metadata(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function checks whether the attributes of the user's metadata constituting a profile are filled out.
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
@@ -29,7 +29,7 @@ def has_cached_user_metadata(state_machine_data: Tuple[str, dict, Account]):
|
||||
return user_metadata is not None
|
||||
|
||||
|
||||
def is_valid_name(state_machine_data: Tuple[str, dict, Account]):
|
||||
def is_valid_name(state_machine_data: Tuple[str, dict, User]):
|
||||
"""This function checks that a user provided name is valid
|
||||
:param state_machine_data: A tuple containing user input, a ussd session and user object.
|
||||
:type state_machine_data: str
|
||||
@@ -43,7 +43,7 @@ def is_valid_name(state_machine_data: Tuple[str, dict, Account]):
|
||||
return False
|
||||
|
||||
|
||||
def is_valid_gender_selection(state_machine_data: Tuple[str, dict, Account]):
|
||||
def is_valid_gender_selection(state_machine_data: Tuple[str, dict, User]):
|
||||
"""
|
||||
:param state_machine_data:
|
||||
:type state_machine_data:
|
||||
|
||||
@@ -13,7 +13,7 @@ class UssdStateMachine(Machine):
|
||||
"""This class describes a finite state machine responsible for maintaining all the states that describe the ussd
|
||||
menu as well as providing a means for navigating through these states based on different user inputs.
|
||||
It defines different helper functions that co-ordinate with the stakeholder components of the ussd menu: i.e the
|
||||
Account, UssdSession, UssdMenu to facilitate user interaction with ussd menu.
|
||||
User, UssdSession, UssdMenu to facilitate user interaction with ussd menu.
|
||||
:cvar states: A list of pre-defined states.
|
||||
:type states: list
|
||||
:cvar transitions: A list of pre-defined transitions.
|
||||
|
||||
@@ -9,7 +9,7 @@ import celery
|
||||
# local imports
|
||||
from cic_ussd.conversions import from_wei
|
||||
from cic_ussd.db.models.base import SessionBase
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.account import define_account_tx_metadata
|
||||
from cic_ussd.error import ActionDataNotFoundError
|
||||
from cic_ussd.redis import InMemoryStore, cache_data, create_cached_data_key
|
||||
@@ -49,7 +49,7 @@ def process_account_creation_callback(self, result: str, url: str, status_code:
|
||||
phone_number = account_creation_data.get('phone_number')
|
||||
|
||||
# create user
|
||||
user = Account(blockchain_address=result, phone_number=phone_number)
|
||||
user = User(blockchain_address=result, phone_number=phone_number)
|
||||
session.add(user)
|
||||
session.commit()
|
||||
session.close()
|
||||
@@ -83,12 +83,12 @@ def process_incoming_transfer_callback(result: dict, param: str, status_code: in
|
||||
# collect result data
|
||||
recipient_blockchain_address = result.get('recipient')
|
||||
sender_blockchain_address = result.get('sender')
|
||||
token_symbol = result.get('destination_token_symbol')
|
||||
value = result.get('destination_token_value')
|
||||
token_symbol = result.get('token_symbol')
|
||||
value = result.get('destination_value')
|
||||
|
||||
# try to find users in system
|
||||
recipient_user = session.query(Account).filter_by(blockchain_address=recipient_blockchain_address).first()
|
||||
sender_user = session.query(Account).filter_by(blockchain_address=sender_blockchain_address).first()
|
||||
recipient_user = session.query(User).filter_by(blockchain_address=recipient_blockchain_address).first()
|
||||
sender_user = session.query(User).filter_by(blockchain_address=sender_blockchain_address).first()
|
||||
|
||||
# check whether recipient is in the system
|
||||
if not recipient_user:
|
||||
@@ -188,8 +188,8 @@ def process_statement_callback(result, param: str, status_code: int):
|
||||
processed_transaction = {}
|
||||
|
||||
# check if sender is in the system
|
||||
sender: Account = session.query(Account).filter_by(blockchain_address=sender_blockchain_address).first()
|
||||
owner: Account = session.query(Account).filter_by(blockchain_address=param).first()
|
||||
sender: User = session.query(User).filter_by(blockchain_address=sender_blockchain_address).first()
|
||||
owner: User = session.query(User).filter_by(blockchain_address=param).first()
|
||||
if sender:
|
||||
processed_transaction['sender_phone_number'] = sender.phone_number
|
||||
|
||||
@@ -205,7 +205,7 @@ def process_statement_callback(result, param: str, status_code: int):
|
||||
processed_transaction['sender_phone_number'] = 'GRASSROOTS ECONOMICS'
|
||||
|
||||
# check if recipient is in the system
|
||||
recipient: Account = session.query(Account).filter_by(blockchain_address=recipient_address).first()
|
||||
recipient: User = session.query(User).filter_by(blockchain_address=recipient_address).first()
|
||||
if recipient:
|
||||
processed_transaction['recipient_phone_number'] = recipient.phone_number
|
||||
|
||||
|
||||
@@ -8,7 +8,7 @@ from typing import Optional
|
||||
def translation_for(key: str, preferred_language: Optional[str] = None, **kwargs) -> str:
|
||||
"""
|
||||
Translates text mapped to a specific YAML key into the user's set preferred language.
|
||||
:param preferred_language: Account's preferred language in which to view the ussd menu.
|
||||
:param preferred_language: User's preferred language in which to view the ussd menu.
|
||||
:type preferred_language str
|
||||
:param key: Key to a specific YAML test entry
|
||||
:type key: str
|
||||
|
||||
@@ -8,7 +8,7 @@ import ipaddress
|
||||
from confini import Config
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
logg = logging.getLogger(__file__)
|
||||
|
||||
@@ -68,7 +68,7 @@ def check_known_user(phone: str):
|
||||
:return: Is known phone number
|
||||
:rtype: boolean
|
||||
"""
|
||||
user = Account.session.query(Account).filter_by(phone_number=phone).first()
|
||||
user = User.session.query(User).filter_by(phone_number=phone).first()
|
||||
return user is not None
|
||||
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
# standard imports
|
||||
import semver
|
||||
|
||||
version = (0, 3, 0, 'alpha.9')
|
||||
version = (0, 3, 0, 'alpha.8')
|
||||
|
||||
version_object = semver.VersionInfo(
|
||||
major=version[0],
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
cic_base[full_graph]~=0.1.2a68
|
||||
cic-eth~=0.11.0b3
|
||||
cic-notify~=0.4.0a4
|
||||
cic-notify~=0.4.0a3
|
||||
cic-types~=0.1.0a10
|
||||
|
||||
@@ -4,19 +4,19 @@
|
||||
import pytest
|
||||
|
||||
# platform imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
|
||||
|
||||
def test_user(init_database, set_fernet_key):
|
||||
user = Account(blockchain_address='0x417f5962fc52dc33ff0689659b25848680dec6dcedc6785b03d1df60fc6d5c51',
|
||||
phone_number='+254700000000')
|
||||
user = User(blockchain_address='0x417f5962fc52dc33ff0689659b25848680dec6dcedc6785b03d1df60fc6d5c51',
|
||||
phone_number='+254700000000')
|
||||
user.create_password('0000')
|
||||
|
||||
session = Account.session
|
||||
session = User.session
|
||||
session.add(user)
|
||||
session.commit()
|
||||
|
||||
queried_user = session.query(Account).get(1)
|
||||
queried_user = session.query(User).get(1)
|
||||
assert queried_user.blockchain_address == '0x417f5962fc52dc33ff0689659b25848680dec6dcedc6785b03d1df60fc6d5c51'
|
||||
assert queried_user.phone_number == '+254700000000'
|
||||
assert queried_user.failed_pin_attempts == 0
|
||||
@@ -25,7 +25,7 @@ def test_user(init_database, set_fernet_key):
|
||||
|
||||
def test_user_state_transition(create_pending_user):
|
||||
user = create_pending_user
|
||||
session = Account.session
|
||||
session = User.session
|
||||
|
||||
assert user.get_account_status() == 'PENDING'
|
||||
user.activate_account()
|
||||
|
||||
@@ -8,7 +8,7 @@ import celery
|
||||
import pytest
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.error import ActionDataNotFoundError
|
||||
from cic_ussd.conversions import from_wei
|
||||
|
||||
@@ -29,7 +29,7 @@ def test_successful_process_account_creation_callback_task(account_creation_acti
|
||||
# WARNING: [THE SETTING OF THE ROOT ID IS A HACK AND SHOULD BE REVIEWED OR IMPROVED]
|
||||
mocked_task_request.root_id = task_id
|
||||
|
||||
user = init_database.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = init_database.query(User).filter_by(phone_number=phone_number).first()
|
||||
assert user is None
|
||||
|
||||
redis_cache = init_redis_cache
|
||||
@@ -48,7 +48,7 @@ def test_successful_process_account_creation_callback_task(account_creation_acti
|
||||
)
|
||||
s_process_callback_request.apply_async().get()
|
||||
|
||||
user = init_database.query(Account).filter_by(phone_number=phone_number).first()
|
||||
user = init_database.query(User).filter_by(phone_number=phone_number).first()
|
||||
assert user.blockchain_address == result
|
||||
|
||||
action_data = redis_cache.get(task_id)
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
import json
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import Account
|
||||
from cic_ussd.db.models.user import User
|
||||
from cic_ussd.requests import (get_query_parameters,
|
||||
get_request_endpoint,
|
||||
get_request_method,
|
||||
@@ -58,8 +58,8 @@ def test_process_locked_accounts_requests(create_locked_accounts, valid_locked_a
|
||||
assert len(locked_account_addresses) == 10
|
||||
|
||||
# check that blockchain addresses are ordered by most recently accessed
|
||||
user_1 = Account.session.query(Account).filter_by(blockchain_address=locked_account_addresses[2]).first()
|
||||
user_2 = Account.session.query(Account).filter_by(blockchain_address=locked_account_addresses[7]).first()
|
||||
user_1 = User.session.query(User).filter_by(blockchain_address=locked_account_addresses[2]).first()
|
||||
user_2 = User.session.query(User).filter_by(blockchain_address=locked_account_addresses[7]).first()
|
||||
|
||||
assert user_1.updated > user_2.updated
|
||||
|
||||
|
||||
14
apps/cic-ussd/tests/fixtures/user.py
vendored
14
apps/cic-ussd/tests/fixtures/user.py
vendored
@@ -9,7 +9,7 @@ from cic_types.models.person import generate_metadata_pointer
|
||||
from faker import Faker
|
||||
|
||||
# local imports
|
||||
from cic_ussd.db.models.account import AccountStatus, Account
|
||||
from cic_ussd.db.models.user import AccountStatus, User
|
||||
from cic_ussd.redis import cache_data
|
||||
from cic_ussd.metadata import blockchain_address_to_metadata_pointer
|
||||
|
||||
@@ -19,7 +19,7 @@ fake = Faker()
|
||||
|
||||
@pytest.fixture(scope='function')
|
||||
def create_activated_user(init_database, set_fernet_key):
|
||||
user = Account(
|
||||
user = User(
|
||||
blockchain_address='0xFD9c5aD15C72C6F60f1a119A608931226674243f',
|
||||
phone_number='+25498765432'
|
||||
)
|
||||
@@ -33,7 +33,7 @@ def create_activated_user(init_database, set_fernet_key):
|
||||
|
||||
@pytest.fixture(scope='function')
|
||||
def create_valid_tx_recipient(init_database, set_fernet_key):
|
||||
user = Account(
|
||||
user = User(
|
||||
blockchain_address='0xd6204101012270Bf2558EDcFEd595938d1847bf0',
|
||||
phone_number='+25498765432'
|
||||
)
|
||||
@@ -47,7 +47,7 @@ def create_valid_tx_recipient(init_database, set_fernet_key):
|
||||
|
||||
@pytest.fixture(scope='function')
|
||||
def create_valid_tx_sender(init_database, set_fernet_key):
|
||||
user = Account(
|
||||
user = User(
|
||||
blockchain_address='0xd6204101012270Bf2558EDcFEd595938d1847bf1',
|
||||
phone_number='+25498765433'
|
||||
)
|
||||
@@ -61,7 +61,7 @@ def create_valid_tx_sender(init_database, set_fernet_key):
|
||||
|
||||
@pytest.fixture(scope='function')
|
||||
def create_pending_user(init_database, set_fernet_key):
|
||||
user = Account(
|
||||
user = User(
|
||||
blockchain_address='0x0ebdea8612c1b05d952c036859266c7f2cfcd6a29842d9c6cce3b9f1ba427588',
|
||||
phone_number='+25498765432'
|
||||
)
|
||||
@@ -72,7 +72,7 @@ def create_pending_user(init_database, set_fernet_key):
|
||||
|
||||
@pytest.fixture(scope='function')
|
||||
def create_pin_blocked_user(init_database, set_fernet_key):
|
||||
user = Account(
|
||||
user = User(
|
||||
blockchain_address='0x0ebdea8612c1b05d952c036859266c7f2cfcd6a29842d9c6cce3b9f1ba427588',
|
||||
phone_number='+25498765432'
|
||||
)
|
||||
@@ -90,7 +90,7 @@ def create_locked_accounts(init_database, set_fernet_key):
|
||||
blockchain_address = str(uuid.uuid4())
|
||||
phone_number = fake.phone_number()
|
||||
pin = f'{randint(1000, 9999)}'
|
||||
user = Account(phone_number=phone_number, blockchain_address=blockchain_address)
|
||||
user = User(phone_number=phone_number, blockchain_address=blockchain_address)
|
||||
user.create_password(password=pin)
|
||||
user.failed_pin_attempts = 3
|
||||
user.account_status = AccountStatus.LOCKED.value
|
||||
|
||||
@@ -57,22 +57,19 @@ WORKDIR /home/grassroots
|
||||
USER grassroots
|
||||
|
||||
ARG pip_extra_index_url=https://pip.grassrootseconomics.net:8433
|
||||
ARG cic_base_version=0.1.2a77
|
||||
ARG cic_eth_version=0.11.0b6
|
||||
ARG sarafu_faucet_version=0.0.2a28
|
||||
ARG sarafu_token_version==0.0.1a6
|
||||
ARG cic_base_version=0.1.2a67
|
||||
ARG cic_eth_version=0.11.0b1
|
||||
ARG sarafu_faucet_version=0.0.2a19
|
||||
ARG cic_contracts_version=0.0.2a2
|
||||
RUN pip install --user --extra-index-url $pip_extra_index_url cic-base[full_graph]==$cic_base_version \
|
||||
cic-eth==$cic_eth_version \
|
||||
cic-contracts==$cic_contracts_version \
|
||||
sarafu-faucet==$sarafu_faucet_version \
|
||||
sarafu-token==$sarafu_token_version
|
||||
sarafu-faucet==$sarafu_faucet_version
|
||||
|
||||
FROM python:3.8.6-slim-buster as runtime-image
|
||||
|
||||
RUN apt-get update
|
||||
RUN apt-get install -y --no-install-recommends gnupg libpq-dev
|
||||
RUN apt-get install -y --no-install-recommends jq
|
||||
RUN apt-get install -y --no-install-recommends gnupg libpq-dev
|
||||
|
||||
COPY --from=compile-image /usr/local/bin/ /usr/local/bin/
|
||||
COPY --from=compile-image /usr/local/etc/cic/ /usr/local/etc/cic/
|
||||
|
||||
@@ -2,112 +2,82 @@
|
||||
|
||||
set -a
|
||||
|
||||
CIC_CHAIN_SPEC=${CIC_CHAIN_SPEC:-evm:bloxberg:8995}
|
||||
DEV_TOKEN_TYPE=${DEV_TOKEN_TYPE:-giftable}
|
||||
DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER=0xEb3907eCad74a0013c259D5874AE7f22DcBcC95C
|
||||
DEV_ETH_ACCOUNT_RESERVE_MINTER=${DEV_ETH_ACCOUNT_RESERVE_MINTER:-$DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER}
|
||||
DEV_ETH_ACCOUNT_ACCOUNTS_INDEX_WRITER=${DEV_ETH_ACCOUNT_RESERVE_MINTER:-$DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER}
|
||||
DEV_RESERVE_AMOUNT=${DEV_ETH_RESERVE_AMOUNT:-""10000000000000000000000000000000000}
|
||||
DEV_FAUCET_AMOUNT=${DEV_FAUCET_AMOUNT:-0}
|
||||
DEV_ETH_KEYSTORE_FILE=${DEV_ETH_KEYSTORE_FILE:-`realpath ./keystore/UTC--2021-01-08T17-18-44.521011372Z--eb3907ecad74a0013c259d5874ae7f22dcbcc95c`}
|
||||
|
||||
set -e
|
||||
|
||||
DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER=`eth-checksum $(cat $DEV_ETH_KEYSTORE_FILE | jq -r .address)`
|
||||
|
||||
if [ ! -z $DEV_ETH_GAS_PRICE ]; then
|
||||
gas_price_arg="--gas-price $DEV_ETH_GAS_PRICE"
|
||||
>&2 echo using static gas price $DEV_ETH_GAS_PRICE
|
||||
fi
|
||||
|
||||
if [[ $DEV_TOKEN_TYPE != 'giftable' && $DEV_TOKEN_TYPE != 'sarafu' ]]; then
|
||||
echo $DEV_TOKEN_TYPE
|
||||
>&2 echo DEV_TOKEN_TYPE must be one of [giftable,sarafu]
|
||||
exit 1
|
||||
fi
|
||||
faucet_amount=${DEV_FAUCET_AMOUNT:-0}
|
||||
keystore_file=$(realpath ./keystore/UTC--2021-01-08T17-18-44.521011372Z--eb3907ecad74a0013c259d5874ae7f22dcbcc95c)
|
||||
|
||||
echo "environment:"
|
||||
printenv
|
||||
echo \n
|
||||
|
||||
echo "using wallet address '$DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER' from keystore file $DEV_ETH_KEYSTORE_FILE"
|
||||
|
||||
# This is a grassroots team convention for building the Bancor contracts using the bancor protocol repository truffle setup
|
||||
# Running this in docker-internal dev container (built from Docker folder in this repo) will write a
|
||||
# source-able env file to CIC_DATA_DIR. Services dependent on these contracts can mount this file OR
|
||||
# define these parameters at runtime
|
||||
# pushd /usr/src
|
||||
|
||||
if [ -z $CIC_DATA_DIR ]; then
|
||||
CIC_DATA_DIR=`mktemp -d`
|
||||
fi
|
||||
>&2 echo using data dir $CIC_DATA_DIR
|
||||
|
||||
init_level_file=${CIC_DATA_DIR}/.init
|
||||
if [ ! -f ${CIC_DATA_DIR}/.init ]; then
|
||||
echo "Creating .init file..."
|
||||
mkdir -p $CIC_DATA_DIR
|
||||
touch $CIC_DATA_DIR/.init
|
||||
touch /tmp/cic/config/.init
|
||||
# touch $init_level_file
|
||||
fi
|
||||
echo -n 1 > $init_level_file
|
||||
|
||||
# Abort on any error (including if wait-for-it fails).
|
||||
set -e
|
||||
|
||||
# Wait for the backend to be up, if we know where it is.
|
||||
if [[ -n "${ETH_PROVIDER}" ]]; then
|
||||
echo "waiting for ${ETH_PROVIDER}..."
|
||||
./wait-for-it.sh "${ETH_PROVIDER_HOST}:${ETH_PROVIDER_PORT}"
|
||||
|
||||
if [ ! -z "$DEV_USE_DOCKER_WAIT_SCRIPT" ]; then
|
||||
echo "waiting for ${ETH_PROVIDER}..."
|
||||
./wait-for-it.sh "${ETH_PROVIDER_HOST}:${ETH_PROVIDER_PORT}"
|
||||
fi
|
||||
DEV_RESERVE_ADDRESS=`giftable-token-deploy -p $ETH_PROVIDER -y $keystore_file -i $CIC_CHAIN_SPEC -v -w --name "Sarafu" --symbol "SRF" --decimals 6`
|
||||
giftable-token-gift -p $ETH_PROVIDER -y $keystore_file -i $CIC_CHAIN_SPEC -v -w -a $DEV_RESERVE_ADDRESS $DEV_RESERVE_AMOUNT
|
||||
|
||||
if [ $DEV_TOKEN_TYPE == 'giftable' ]; then
|
||||
>&2 echo "deploying 'giftable token'"
|
||||
DEV_RESERVE_ADDRESS=`giftable-token-deploy $gas_price_arg -p $ETH_PROVIDER -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -vv -w --name "Giftable Token" --symbol "GFT" --decimals 6 -vv`
|
||||
else
|
||||
>&2 echo "deploying 'sarafu' token'"
|
||||
DEV_RESERVE_ADDRESS=`sarafu-token-deploy $gas_price_arg -p $ETH_PROVIDER -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -vv -w --name "Sarafu" --decimals 6 -vv SRF $DEV_SARAFU_DEMURRAGE_LEVEL`
|
||||
fi
|
||||
giftable-token-gift $gas_price_arg -p $ETH_PROVIDER -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -vv -w -a $DEV_RESERVE_ADDRESS $DEV_RESERVE_AMOUNT
|
||||
|
||||
#BANCOR_REGISTRY_ADDRESS=`cic-bancor-deploy $gas_price_arg --bancor-dir /usr/local/share/cic/bancor -z $DEV_ETH_RESERVE_ADDRESS -p $ETH_PROVIDER -o $DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER`
|
||||
#BANCOR_REGISTRY_ADDRESS=`cic-bancor-deploy --bancor-dir /usr/local/share/cic/bancor -z $DEV_ETH_RESERVE_ADDRESS -p $ETH_PROVIDER -o $DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER`
|
||||
|
||||
>&2 echo "deploy account index contract"
|
||||
DEV_ACCOUNT_INDEX_ADDRESS=`eth-accounts-index-deploy $gas_price_arg -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -y $DEV_ETH_KEYSTORE_FILE -vv -w`
|
||||
DEV_ACCOUNT_INDEX_ADDRESS=`eth-accounts-index-deploy -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -y $keystore_file -vv -w`
|
||||
>&2 echo "add deployer address as account index writer"
|
||||
eth-accounts-index-writer $gas_price_arg -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -a $DEV_ACCOUNT_INDEX_ADDRESS -ww -vv $debug $DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER
|
||||
eth-accounts-index-writer -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -a $DEV_ACCOUNT_INDEX_ADDRESS -ww $debug $DEV_ETH_ACCOUNT_CONTRACT_DEPLOYER
|
||||
|
||||
CIC_REGISTRY_ADDRESS=`eth-contract-registry-deploy $gas_price_arg -i $CIC_CHAIN_SPEC -y $DEV_ETH_KEYSTORE_FILE --identifier BancorRegistry --identifier AccountRegistry --identifier TokenRegistry --identifier AddressDeclarator --identifier Faucet --identifier TransferAuthorization -p $ETH_PROVIDER -vv -w`
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv ContractRegistry $CIC_REGISTRY_ADDRESS
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv AccountRegistry $DEV_ACCOUNT_INDEX_ADDRESS
|
||||
CIC_REGISTRY_ADDRESS=`eth-contract-registry-deploy -i $CIC_CHAIN_SPEC -y $keystore_file --identifier BancorRegistry --identifier AccountRegistry --identifier TokenRegistry --identifier AddressDeclarator --identifier Faucet --identifier TransferAuthorization -p $ETH_PROVIDER -vv -w`
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv ContractRegistry $CIC_REGISTRY_ADDRESS
|
||||
#cic-registry-set -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -k BancorRegistry -p $ETH_PROVIDER $BANCOR_REGISTRY_ADDRESS -vv
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv AccountRegistry $DEV_ACCOUNT_INDEX_ADDRESS
|
||||
|
||||
# Deploy address declarator registry
|
||||
>&2 echo "deploy address declarator contract"
|
||||
declarator_description=0x546869732069732074686520434943206e6574776f726b000000000000000000
|
||||
DEV_DECLARATOR_ADDRESS=`eth-address-declarator-deploy -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -vv $declarator_description`
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv AddressDeclarator $DEV_DECLARATOR_ADDRESS
|
||||
DEV_DECLARATOR_ADDRESS=`eth-address-declarator-deploy -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -v $declarator_description`
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv AddressDeclarator $DEV_DECLARATOR_ADDRESS
|
||||
|
||||
# Deploy transfer authorization contact
|
||||
>&2 echo "deploy address declarator contract"
|
||||
DEV_TRANSFER_AUTHORIZATION_ADDRESS=`erc20-transfer-auth-deploy $gas_price_arg -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -vv`
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv TransferAuthorization $DEV_TRANSFER_AUTHORIZATION_ADDRESS
|
||||
DEV_TRANSFER_AUTHORIZATION_ADDRESS=`erc20-transfer-auth-deploy -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -v`
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv TransferAuthorization $DEV_TRANSFER_AUTHORIZATION_ADDRESS
|
||||
|
||||
# Deploy token index contract
|
||||
>&2 echo "deploy token index contract"
|
||||
DEV_TOKEN_INDEX_ADDRESS=`eth-token-index-deploy $gas_price_arg -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -vv`
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv TokenRegistry $DEV_TOKEN_INDEX_ADDRESS
|
||||
DEV_TOKEN_INDEX_ADDRESS=`eth-token-index-deploy -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -v`
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv TokenRegistry $DEV_TOKEN_INDEX_ADDRESS
|
||||
>&2 echo "add reserve token to token index"
|
||||
eth-token-index-add $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv -a $DEV_TOKEN_INDEX_ADDRESS $DEV_RESERVE_ADDRESS
|
||||
eth-token-index-add -w -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv -a $DEV_TOKEN_INDEX_ADDRESS $DEV_RESERVE_ADDRESS
|
||||
|
||||
# Sarafu faucet contract
|
||||
>&2 echo "deploy token faucet contract"
|
||||
DEV_FAUCET_ADDRESS=`sarafu-faucet-deploy $gas_price_arg -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -vv --account-index-address $DEV_ACCOUNT_INDEX_ADDRESS $DEV_RESERVE_ADDRESS`
|
||||
eth-contract-registry-set $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv Faucet $DEV_FAUCET_ADDRESS
|
||||
DEV_FAUCET_ADDRESS=`sarafu-faucet-deploy -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -w -v --account-index-address $DEV_ACCOUNT_INDEX_ADDRESS $DEV_RESERVE_ADDRESS`
|
||||
eth-contract-registry-set -w -y $keystore_file -r $CIC_REGISTRY_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv Faucet $DEV_FAUCET_ADDRESS
|
||||
>&2 echo "set faucet as token minter"
|
||||
giftable-token-minter $gas_price_arg -w -y $DEV_ETH_KEYSTORE_FILE -a $DEV_RESERVE_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv $DEV_FAUCET_ADDRESS
|
||||
giftable-token-minter -w -y $keystore_file -a $DEV_RESERVE_ADDRESS -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -vv $DEV_FAUCET_ADDRESS
|
||||
|
||||
>&2 echo "set token faucet amount"
|
||||
sarafu-faucet-set $gas_price_arg -y $DEV_ETH_KEYSTORE_FILE -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -a $DEV_FAUCET_ADDRESS -vv $DEV_FAUCET_AMOUNT
|
||||
sarafu-faucet-set -y $keystore_file -i $CIC_CHAIN_SPEC -p $ETH_PROVIDER -a $DEV_FAUCET_ADDRESS $faucet_amount
|
||||
|
||||
|
||||
else
|
||||
|
||||
@@ -17,7 +17,7 @@ from hexathon import (
|
||||
strip_0x,
|
||||
add_0x,
|
||||
)
|
||||
from chainsyncer.backend.memory import MemBackend
|
||||
from chainsyncer.backend import MemBackend
|
||||
from chainsyncer.driver import HeadSyncer
|
||||
from chainlib.eth.connection import EthHTTPConnection
|
||||
from chainlib.eth.block import (
|
||||
@@ -25,13 +25,13 @@ from chainlib.eth.block import (
|
||||
block_by_number,
|
||||
Block,
|
||||
)
|
||||
from chainlib.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.address import to_checksum_address
|
||||
from chainlib.eth.erc20 import ERC20
|
||||
from chainlib.eth.gas import OverrideGasOracle
|
||||
from chainlib.eth.nonce import RPCNonceOracle
|
||||
from chainlib.eth.tx import TxFactory
|
||||
from chainlib.jsonrpc import jsonrpc_template
|
||||
from chainlib.eth.rpc import jsonrpc_template
|
||||
from chainlib.eth.error import EthException
|
||||
from chainlib.chain import ChainSpec
|
||||
from crypto_dev_signer.eth.signer import ReferenceSigner as EIP155Signer
|
||||
|
||||
@@ -3,11 +3,10 @@ const path = require('path');
|
||||
const http = require('http');
|
||||
|
||||
const cic = require('cic-client-meta');
|
||||
const crdt = require('crdt-meta');
|
||||
|
||||
//const conf = JSON.parse(fs.readFileSync('./cic.conf'));
|
||||
|
||||
const config = new crdt.Config('./config');
|
||||
const config = new cic.Config('./config');
|
||||
config.process();
|
||||
console.log(config);
|
||||
|
||||
@@ -42,7 +41,7 @@ function sendit(uid, envelope) {
|
||||
}
|
||||
|
||||
function doOne(keystore, filePath) {
|
||||
const signer = new crdt.PGPSigner(keystore);
|
||||
const signer = new cic.PGPSigner(keystore);
|
||||
const parts = path.basename(filePath).split('.');
|
||||
const ethereum_address = path.basename(parts[0]);
|
||||
|
||||
@@ -52,7 +51,7 @@ function doOne(keystore, filePath) {
|
||||
//console.log(o);
|
||||
fs.unlinkSync(filePath);
|
||||
|
||||
const s = new crdt.Syncable(uid, o);
|
||||
const s = new cic.Syncable(uid, o);
|
||||
s.setSigner(signer);
|
||||
s.onwrap = (env) => {
|
||||
sendit(uid, env);
|
||||
@@ -66,7 +65,7 @@ const publicKeyPath = path.join(config.get('PGP_EXPORTS_DIR'), config.get('PGP_P
|
||||
pk = fs.readFileSync(privateKeyPath);
|
||||
pubk = fs.readFileSync(publicKeyPath);
|
||||
|
||||
new crdt.PGPKeyStore(
|
||||
new cic.PGPKeyStore(
|
||||
config.get('PGP_PASSPHRASE'),
|
||||
pk,
|
||||
pubk,
|
||||
|
||||
@@ -4,11 +4,10 @@ const http = require('http');
|
||||
|
||||
const cic = require('cic-client-meta');
|
||||
const vcfp = require('vcard-parser');
|
||||
const crdt = require('crdt-meta');
|
||||
|
||||
//const conf = JSON.parse(fs.readFileSync('./cic.conf'));
|
||||
|
||||
const config = new crdt.Config('./config');
|
||||
const config = new cic.Config('./config');
|
||||
config.process();
|
||||
console.log(config);
|
||||
|
||||
@@ -43,7 +42,7 @@ function sendit(uid, envelope) {
|
||||
}
|
||||
|
||||
function doOne(keystore, filePath, address) {
|
||||
const signer = new crdt.PGPSigner(keystore);
|
||||
const signer = new cic.PGPSigner(keystore);
|
||||
|
||||
const j = JSON.parse(fs.readFileSync(filePath).toString());
|
||||
const b = Buffer.from(j['vcard'], 'base64');
|
||||
@@ -52,8 +51,9 @@ function doOne(keystore, filePath, address) {
|
||||
const phone = o.tel[0].value;
|
||||
|
||||
cic.Phone.toKey(phone).then((uid) => {
|
||||
const o = fs.readFileSync(filePath, 'utf-8');
|
||||
|
||||
const s = new crdt.Syncable(uid, address);
|
||||
const s = new cic.Syncable(uid, o);
|
||||
s.setSigner(signer);
|
||||
s.onwrap = (env) => {
|
||||
sendit(uid, env);
|
||||
@@ -67,7 +67,7 @@ const publicKeyPath = path.join(config.get('PGP_EXPORTS_DIR'), config.get('PGP_P
|
||||
pk = fs.readFileSync(privateKeyPath);
|
||||
pubk = fs.readFileSync(publicKeyPath);
|
||||
|
||||
new crdt.PGPKeyStore(
|
||||
new cic.PGPKeyStore(
|
||||
config.get('PGP_PASSPHRASE'),
|
||||
pk,
|
||||
pubk,
|
||||
@@ -123,7 +123,7 @@ function importMetaPhone(keystore) {
|
||||
if (batchCount == batchSize) {
|
||||
console.debug('reached batch size, breathing');
|
||||
batchCount=0;
|
||||
setTimeout(importMetaPhone, batchDelay, keystore);
|
||||
setTimeout(importMeta, batchDelay, keystore);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -17,7 +17,7 @@ from hexathon import (
|
||||
strip_0x,
|
||||
add_0x,
|
||||
)
|
||||
from chainsyncer.backend.memory import MemBackend
|
||||
from chainsyncer.backend import MemBackend
|
||||
from chainsyncer.driver import HeadSyncer
|
||||
from chainlib.eth.connection import EthHTTPConnection
|
||||
from chainlib.eth.block import (
|
||||
@@ -25,13 +25,13 @@ from chainlib.eth.block import (
|
||||
block_by_number,
|
||||
Block,
|
||||
)
|
||||
from chainlib.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.address import to_checksum_address
|
||||
from chainlib.eth.erc20 import ERC20
|
||||
from chainlib.eth.gas import OverrideGasOracle
|
||||
from chainlib.eth.nonce import RPCNonceOracle
|
||||
from chainlib.eth.tx import TxFactory
|
||||
from chainlib.jsonrpc import jsonrpc_template
|
||||
from chainlib.eth.rpc import jsonrpc_template
|
||||
from chainlib.eth.error import EthException
|
||||
from chainlib.chain import ChainSpec
|
||||
from crypto_dev_signer.eth.signer import ReferenceSigner as EIP155Signer
|
||||
|
||||
@@ -24,7 +24,7 @@ from chainlib.eth.gas import RPCGasOracle
|
||||
from chainlib.eth.nonce import RPCNonceOracle
|
||||
from cic_types.processor import generate_metadata_pointer
|
||||
from eth_accounts_index import AccountRegistry
|
||||
from eth_contract_registry import Registry
|
||||
from contract_registry import Registry
|
||||
from crypto_dev_signer.keystore.dict import DictKeystore
|
||||
from crypto_dev_signer.eth.signer.defaultsigner import ReferenceSigner as EIP155Signer
|
||||
from crypto_dev_signer.keystore.keyfile import to_dict as to_keyfile_dict
|
||||
|
||||
2276
apps/contract-migration/scripts/package-lock.json
generated
2276
apps/contract-migration/scripts/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@@ -1,7 +1,6 @@
|
||||
{
|
||||
"dependencies": {
|
||||
"cic-client-meta": "^0.0.7-alpha.8",
|
||||
"crdt-meta": "0.0.8",
|
||||
"cic-client-meta": "^0.0.7-alpha.6",
|
||||
"vcard-parser": "^1.0.0"
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
cic-base[full_graph]==0.1.2a77
|
||||
sarafu-faucet==0.0.2a28
|
||||
cic-eth==0.11.0b6
|
||||
cic-base[full_graph]==0.1.2a67
|
||||
sarafu-faucet==0.0.2a20
|
||||
cic-eth==0.11.0b3
|
||||
cic-types==0.1.0a10
|
||||
crypto-dev-signer==0.4.14b2
|
||||
crypto-dev-signer==0.4.14b1
|
||||
|
||||
@@ -22,7 +22,7 @@ from hexathon import (
|
||||
strip_0x,
|
||||
add_0x,
|
||||
)
|
||||
from chainsyncer.backend.memory import MemBackend
|
||||
from chainsyncer.backend import MemBackend
|
||||
from chainsyncer.driver import HeadSyncer
|
||||
from chainlib.chain import ChainSpec
|
||||
from chainlib.eth.connection import EthHTTPConnection
|
||||
@@ -32,7 +32,7 @@ from chainlib.eth.block import (
|
||||
block_by_number,
|
||||
Block,
|
||||
)
|
||||
from chainlib.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.hash import keccak256_string_to_hex
|
||||
from chainlib.eth.address import to_checksum_address
|
||||
from chainlib.eth.erc20 import ERC20
|
||||
from chainlib.eth.gas import (
|
||||
@@ -40,7 +40,7 @@ from chainlib.eth.gas import (
|
||||
balance,
|
||||
)
|
||||
from chainlib.eth.tx import TxFactory
|
||||
from chainlib.jsonrpc import jsonrpc_template
|
||||
from chainlib.eth.rpc import jsonrpc_template
|
||||
from chainlib.eth.error import EthException
|
||||
from cic_types.models.person import (
|
||||
Person,
|
||||
@@ -57,7 +57,6 @@ custodial_tests = [
|
||||
'local_key',
|
||||
'gas',
|
||||
'faucet',
|
||||
'ussd'
|
||||
]
|
||||
|
||||
metadata_tests = [
|
||||
@@ -345,11 +344,12 @@ class Verifier:
|
||||
address_recovered = address_recovered.replace('"', '')
|
||||
|
||||
try:
|
||||
upper_address_recovered = strip_0x(address_recovered).upper()
|
||||
address = strip_0x(address)
|
||||
address_recovered = strip_0x(address_recovered)
|
||||
except ValueError:
|
||||
raise VerifierError(address_recovered, 'metadata (phone) address {} address recovered {}'.format(address, address_recovered))
|
||||
|
||||
if upper_address != upper_address_recovered:
|
||||
if address != address_recovered:
|
||||
raise VerifierError(address_recovered, 'metadata (phone)')
|
||||
|
||||
|
||||
|
||||
@@ -24,7 +24,7 @@ variables:
|
||||
IMAGE_TAG_BASE: $CI_REGISTRY_IMAGE/$APP_NAME:$CI_COMMIT_BRANCH-$CI_COMMIT_SHORT_SHA
|
||||
LATEST_TAG: $CI_REGISTRY_IMAGE/$APP_NAME:latest
|
||||
script:
|
||||
- export IMAGE_TAG="$IMAGE_TAG_BASE-$(date +%s)"
|
||||
- export IMAGE_TAG="$IMAGE_TAG_BASE-$(date +%F.%H%M%S)"
|
||||
- mkdir -p /kaniko/.docker
|
||||
- echo "{\"auths\":{\"$CI_REGISTRY\":{\"username\":\"$CI_REGISTRY_USER\",\"password\":\"$CI_REGISTRY_PASSWORD\"}}}" > "/kaniko/.docker/config.json"
|
||||
# - /kaniko/executor --context $CONTEXT --dockerfile $DOCKERFILE_PATH $KANIKO_CACHE_ARGS --destination $IMAGE_TAG
|
||||
|
||||
@@ -85,7 +85,6 @@ services:
|
||||
# ETH_PROVIDER should be broken out into host/port but cic-eth expects this
|
||||
ETH_PROVIDER: http://eth:8545
|
||||
# And these two are for wait-for-it (could parse this)
|
||||
DEV_USE_DOCKER_WAIT_SCRIPT: 1
|
||||
ETH_PROVIDER_HOST: eth
|
||||
ETH_PROVIDER_PORT: 8545
|
||||
CIC_CHAIN_SPEC: ${CIC_CHAIN_SPEC:-evm:bloxberg:8996}
|
||||
@@ -104,9 +103,6 @@ services:
|
||||
DEV_PIP_EXTRA_INDEX_URL: ${DEV_PIP_EXTRA_INDEX_URL:-https://pip.grassrootseconomics.net:8433}
|
||||
RUN_MASK: ${RUN_MASK:-0} # bit flags; 1: contract migrations 2: seed data
|
||||
DEV_FAUCET_AMOUNT: ${DEV_FAUCET_AMOUNT:-0}
|
||||
DEV_TOKEN_TYPE: ${DEV_TOKEN_TYPE:-giftable}
|
||||
DEV_SARAFU_DEMURRAGE_LEVEL: ${DEV_SARAFU_DEMURRAGE_LEVEL:-196454828847045000000000000000000}
|
||||
DEV_ETH_GAS_PRICE: ${DEV_ETH_GAS_PRICE:-1}
|
||||
command: ["./run_job.sh"]
|
||||
#command: ["./reset.sh"]
|
||||
depends_on:
|
||||
@@ -450,7 +446,7 @@ services:
|
||||
deploy:
|
||||
restart_policy:
|
||||
condition: on-failure
|
||||
command: "/root/start_tasker.sh -q cic-notify -vv"
|
||||
command: "/root/start_tasker.sh -q cic-notify"
|
||||
|
||||
|
||||
cic-meta-server:
|
||||
@@ -498,8 +494,6 @@ services:
|
||||
DATABASE_NAME: cic_ussd
|
||||
DATABASE_ENGINE: postgresql
|
||||
DATABASE_DRIVER: psycopg2
|
||||
CELERY_BROKER_URL: ${CELERY_BROKER_URL:-redis://redis}
|
||||
CELERY_RESULT_URL: ${CELERY_BROKER_URL:-redis://redis}
|
||||
PGP_PASSPHRASE: merman
|
||||
SERVER_PORT: 9000
|
||||
CIC_META_URL: ${CIC_META_URL:-http://meta:8000}
|
||||
|
||||
@@ -5,7 +5,6 @@ CREATE DATABASE "cic_notify";
|
||||
CREATE DATABASE "cic_meta";
|
||||
CREATE DATABASE "cic_signer";
|
||||
CREATE DATABASE "cic_ussd";
|
||||
CREATE DATABASE "chain_sync";
|
||||
GRANT ALL PRIVILEGES
|
||||
ON DATABASE "cic_cache", "cic_eth", "cic_notify", "cic_meta", "cic_signer", "cic_ussd", "chain_sync"
|
||||
ON DATABASE "cic_cache", "cic_eth", "cic_notify", "cic_meta", "cic_signer", "cic_ussd"
|
||||
TO grassroots;
|
||||
|
||||
Reference in New Issue
Block a user