Compare commits
27 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
465d692956
|
||
|
|
81c1207828
|
||
|
|
f33ba13d74
|
||
|
|
5459d4c3f8
|
||
|
|
3e05717395
|
||
|
|
54d10ee40b
|
||
|
|
9cffdc5867
|
||
|
|
4f96be2024
|
||
|
|
32e1bc6aa5
|
||
|
|
387014f77b
|
||
|
|
c3a592c0f6
|
||
|
|
5d61506133
|
||
|
|
5102b4ac6e
|
||
|
|
9b98703f24
|
||
|
|
4be5325df2
|
||
|
|
96bdca20cc
|
||
|
|
14fce6f63f
|
||
|
|
9ed3bad0c4
|
||
|
|
e87ec0cd4c
|
||
|
|
b52d69c3f9
|
||
|
|
1d2656aaed
|
||
|
|
b198e3f6b1
|
||
|
|
a55591f243
|
||
|
|
800b70680b
|
||
|
|
e54f03a8f6
|
||
|
|
18d10dc8b7
|
||
|
|
7e78fd0da2
|
23
CHANGELOG
23
CHANGELOG
@@ -1,3 +1,26 @@
|
||||
- 0.2.10
|
||||
* Upgrade shep to guarantee state lock atomicity
|
||||
- 0.2.9
|
||||
* Minimize instantiations of adapters in filter execution
|
||||
- 0.2.8
|
||||
* Upgrade chainsyncer
|
||||
- 0.2.7
|
||||
* Upgrade chainlib
|
||||
- 0.2.6
|
||||
* Deps upgrade
|
||||
- 0.2.5
|
||||
* Deps upgrade
|
||||
- 0.2.4
|
||||
* Allow omission of state store sync in queue store backend
|
||||
- 0.2.2
|
||||
* Fix missing symbol crashes related to race conditions
|
||||
- 0.2.1
|
||||
* Receive removed race checks from chainqueue
|
||||
- 0.2.0
|
||||
* primitive race condition handling between fs access of sync and queue
|
||||
* re-enable throttling based on in-flight transaction count
|
||||
- 0.1.2
|
||||
* add settings object
|
||||
- 0.1.0
|
||||
* consume non chain-specific code
|
||||
- 0.0.1
|
||||
|
||||
@@ -1,10 +1,27 @@
|
||||
# standard imports
|
||||
import logging
|
||||
import time
|
||||
|
||||
# external imports
|
||||
from chainqueue import Store as QueueStore
|
||||
|
||||
# local imports
|
||||
from chaind.lock import StoreLock
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ChaindAdapter:
|
||||
|
||||
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0):
|
||||
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, store_sync=True):
|
||||
self.cache_adapter = cache_adapter
|
||||
self.dispatcher = dispatcher
|
||||
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache)
|
||||
store_lock = StoreLock()
|
||||
while True:
|
||||
try:
|
||||
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache, sync=store_sync)
|
||||
break
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('queuestore instantiation failed, possible race condition (will try again): {}'.format(e))
|
||||
store_lock.again()
|
||||
continue
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
# standard imports
|
||||
import logging
|
||||
import os
|
||||
import time
|
||||
|
||||
# external imports
|
||||
from chainlib.error import RPCException
|
||||
@@ -10,36 +12,67 @@ from chainqueue.store.fs import (
|
||||
CounterStore,
|
||||
)
|
||||
from shep.store.file import SimpleFileStoreFactory
|
||||
from shep.error import (
|
||||
StateInvalid,
|
||||
StateLockedKey,
|
||||
)
|
||||
|
||||
# local imports
|
||||
from .base import ChaindAdapter
|
||||
from chaind.lock import StoreLock
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ChaindFsAdapter(ChaindAdapter):
|
||||
|
||||
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32):
|
||||
factory = SimpleFileStoreFactory(path).add
|
||||
state_store = Status(factory)
|
||||
index_store = IndexStore(path, digest_bytes=digest_bytes)
|
||||
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32, event_callback=None, store_sync=True):
|
||||
factory = SimpleFileStoreFactory(path, use_lock=True).add
|
||||
state_store = Status(factory, allow_invalid=True, event_callback=event_callback)
|
||||
index_path = os.path.join(path, 'tx')
|
||||
index_store = IndexStore(index_path, digest_bytes=digest_bytes)
|
||||
counter_store = CounterStore(path)
|
||||
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold)
|
||||
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold, store_sync=store_sync)
|
||||
|
||||
|
||||
def put(self, signed_tx):
|
||||
#cache_tx = self.deserialize(signed_tx)
|
||||
(s, tx_hash,) = self.store.put(signed_tx, cache_adapter=self.cache_adapter)
|
||||
return tx_hash
|
||||
|
||||
|
||||
def get(self, tx_hash):
|
||||
v = self.store.get(tx_hash)
|
||||
v = None
|
||||
store_lock = StoreLock()
|
||||
while True:
|
||||
try:
|
||||
v = self.store.get(tx_hash)
|
||||
break
|
||||
except StateInvalid as e:
|
||||
logg.error('I am just a simple syncer and do not know how to handle the state which the tx {} is in: {}'.format(tx_hash, e))
|
||||
return None
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('queuestore get (file missing) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
||||
store_lock.again()
|
||||
continue
|
||||
except StateLockedKey as e:
|
||||
logg.debug('queuestore get (statelock) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
||||
store_lock.again()
|
||||
continue
|
||||
|
||||
return v[1]
|
||||
|
||||
|
||||
def upcoming(self):
|
||||
return self.store.upcoming()
|
||||
def upcoming(self, limit=0):
|
||||
real_limit = 0
|
||||
in_flight = []
|
||||
if limit > 0:
|
||||
in_flight = self.store.by_state(state=self.store.IN_NETWORK, not_state=self.store.FINAL)
|
||||
real_limit = limit - len(in_flight)
|
||||
if real_limit <= 0:
|
||||
return []
|
||||
r = self.store.upcoming(limit=real_limit)
|
||||
logg.info('upcoming returning {} upcoming from limit {} less {} active in-flight txs'.format(len(r), limit, len(in_flight)))
|
||||
return r
|
||||
|
||||
|
||||
def pending(self):
|
||||
@@ -50,12 +83,30 @@ class ChaindFsAdapter(ChaindAdapter):
|
||||
return self.store.deferred()
|
||||
|
||||
|
||||
def failed(self):
|
||||
return self.store.failed()
|
||||
|
||||
|
||||
def succeed(self, block, tx):
|
||||
return self.store.final(tx.hash, block, tx, error=False)
|
||||
if self.store.is_reserved(tx.hash):
|
||||
raise QueueLockError(tx.hash)
|
||||
r = self.store.final(tx.hash, block, tx, error=False)
|
||||
(k, v) = self.store.get(tx.hash)
|
||||
self.store.purge(k)
|
||||
return r
|
||||
|
||||
|
||||
def fail(self, block, tx):
|
||||
return self.store.final(tx.hash, block, tx, error=True)
|
||||
if self.store.is_reserved(tx.hash):
|
||||
raise QueueLockError(tx.hash)
|
||||
r = self.store.final(tx.hash, block, tx, error=True)
|
||||
(k, v) = self.store.get(tx.hash)
|
||||
self.store.purge(k)
|
||||
return r
|
||||
|
||||
|
||||
def sendfail(self):
|
||||
return self.store.fail(tx.hash)
|
||||
|
||||
|
||||
def enqueue(self, tx_hash):
|
||||
@@ -63,15 +114,44 @@ class ChaindFsAdapter(ChaindAdapter):
|
||||
|
||||
|
||||
def dispatch(self, tx_hash):
|
||||
entry = self.store.send_start(tx_hash)
|
||||
entry = None
|
||||
|
||||
store_lock = StoreLock()
|
||||
while True:
|
||||
try:
|
||||
entry = self.store.send_start(tx_hash)
|
||||
break
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||
store_lock.again()
|
||||
continue
|
||||
except StateLockedKey as e:
|
||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||
store_lock.again()
|
||||
continue
|
||||
|
||||
tx_wire = entry.serialize()
|
||||
|
||||
r = None
|
||||
try:
|
||||
r = self.dispatcher.send(tx_wire)
|
||||
except RPCException:
|
||||
except RPCException as e:
|
||||
logg.error('dispatch send failed for {}: {}'.format(tx_hash, e))
|
||||
self.store.fail(tx_hash)
|
||||
return False
|
||||
|
||||
self.store.send_end(tx_hash)
|
||||
store_lock = StoreLock()
|
||||
while True:
|
||||
try:
|
||||
self.store.send_end(tx_hash)
|
||||
break
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||
store_lock.again(e)
|
||||
continue
|
||||
except StateLockedKey as e:
|
||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||
store_lock.again(e)
|
||||
continue
|
||||
|
||||
return True
|
||||
|
||||
12
chaind/cli/__init__.py
Normal file
12
chaind/cli/__init__.py
Normal file
@@ -0,0 +1,12 @@
|
||||
# standard imports
|
||||
import os
|
||||
|
||||
# local imports
|
||||
from .base import *
|
||||
from .arg import process_flags
|
||||
from .config import process_config
|
||||
|
||||
|
||||
__script_dir = os.path.dirname(os.path.realpath(__file__))
|
||||
data_dir = os.path.join(os.path.dirname(__script_dir), 'data')
|
||||
config_dir = os.path.join(data_dir, 'config')
|
||||
17
chaind/cli/arg.py
Normal file
17
chaind/cli/arg.py
Normal file
@@ -0,0 +1,17 @@
|
||||
# local imports
|
||||
from .base import ChaindFlag
|
||||
|
||||
def process_flags(argparser, flags):
|
||||
if flags & ChaindFlag.SESSION > 0:
|
||||
argparser.add_argument('--session-id', dest='session_id', type=str, help='Session to store state and data under')
|
||||
argparser.add_argument('--runtime-dir', dest='runtime_dir', type=str, help='Directory to store volatile data')
|
||||
argparser.add_argument('--data-dir', dest='data_dir', type=str, help='Directory to store persistent data')
|
||||
|
||||
if flags & ChaindFlag.SOCKET > 0:
|
||||
argparser.add_argument('--socket-path', dest='socket', type=str, help='UNIX socket path')
|
||||
|
||||
if flags & ChaindFlag.SOCKET_CLIENT > 0:
|
||||
argparser.add_argument('--send-socket', dest='socket_send', action='store_true', help='Send to UNIX socket')
|
||||
|
||||
if flags & ChaindFlag.TOKEN > 0:
|
||||
argparser.add_argument('--token-module', dest='token_module', type=str, help='Python module path to resolve tokens from identifiers')
|
||||
13
chaind/cli/base.py
Normal file
13
chaind/cli/base.py
Normal file
@@ -0,0 +1,13 @@
|
||||
# standard imports
|
||||
import enum
|
||||
|
||||
|
||||
class ChaindFlag(enum.IntEnum):
|
||||
SESSION = 1
|
||||
DISPATCH = 2
|
||||
SOCKET = 16
|
||||
SOCKET_CLIENT = 32
|
||||
TOKEN = 256
|
||||
|
||||
argflag_local_base = ChaindFlag.SESSION
|
||||
argflag_local_socket_client = ChaindFlag.SESSION | ChaindFlag.SOCKET | ChaindFlag.SOCKET_CLIENT
|
||||
23
chaind/cli/config.py
Normal file
23
chaind/cli/config.py
Normal file
@@ -0,0 +1,23 @@
|
||||
# external imports
|
||||
from chaind.cli import ChaindFlag
|
||||
|
||||
|
||||
def process_config(config, args, flags):
|
||||
args_override = {}
|
||||
if flags & ChaindFlag.SESSION:
|
||||
args_override['SESSION_ID'] = getattr(args, 'session_id')
|
||||
args_override['SESSION_RUNTIME_DIR'] = getattr(args, 'runtime_dir')
|
||||
args_override['SESSION_DATA_DIR'] = getattr(args, 'data_dir')
|
||||
|
||||
if flags & ChaindFlag.SOCKET:
|
||||
args_override['SESSION_SOCKET_PATH'] = getattr(args, 'socket')
|
||||
|
||||
if flags & ChaindFlag.TOKEN:
|
||||
args_override['TOKEN_MODULE'] = getattr(args, 'token_module')
|
||||
|
||||
config.dict_override(args_override, 'local cli args')
|
||||
|
||||
if flags & ChaindFlag.SOCKET_CLIENT:
|
||||
config.add(getattr(args, 'socket_send'), '_SOCKET_SEND', False)
|
||||
|
||||
return config
|
||||
@@ -3,13 +3,3 @@ socket_path =
|
||||
runtime_dir =
|
||||
id =
|
||||
data_dir =
|
||||
|
||||
[database]
|
||||
engine =
|
||||
name = chaind
|
||||
driver =
|
||||
user =
|
||||
password =
|
||||
host =
|
||||
port =
|
||||
debug = 0
|
||||
|
||||
2
chaind/data/config/token.ini
Normal file
2
chaind/data/config/token.ini
Normal file
@@ -0,0 +1,2 @@
|
||||
[token]
|
||||
module =
|
||||
33
chaind/dispatch.py
Normal file
33
chaind/dispatch.py
Normal file
@@ -0,0 +1,33 @@
|
||||
# standard imports
|
||||
import logging
|
||||
|
||||
# local ipmorts
|
||||
from chaind.adapters.fs import ChaindFsAdapter
|
||||
from chaind.eth.cache import EthCacheTx
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class DispatchProcessor:
|
||||
|
||||
def __init__(self, chain_spec, queue_dir, dispatcher):
|
||||
self.dispatcher = dispatcher
|
||||
self.chain_spec = chain_spec,
|
||||
self.queue_dir = queue_dir
|
||||
|
||||
|
||||
def process(self, rpc, limit=50):
|
||||
adapter = ChaindFsAdapter(
|
||||
self.chain_spec,
|
||||
self.queue_dir,
|
||||
EthCacheTx,
|
||||
self.dispatcher,
|
||||
)
|
||||
|
||||
upcoming = adapter.upcoming(limit=limit)
|
||||
logg.info('processor has {} candidates for {}, processing with limit {}'.format(len(upcoming), self.chain_spec, limit))
|
||||
i = 0
|
||||
for tx_hash in upcoming:
|
||||
if adapter.dispatch(tx_hash):
|
||||
i += 1
|
||||
return i
|
||||
@@ -16,3 +16,11 @@ class ClientBlockError(BlockingIOError):
|
||||
|
||||
class ClientInputError(ValueError):
|
||||
pass
|
||||
|
||||
|
||||
class QueueLockError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class BackendError(Exception):
|
||||
pass
|
||||
|
||||
115
chaind/filter.py
115
chaind/filter.py
@@ -1,30 +1,123 @@
|
||||
# standard imports
|
||||
import logging
|
||||
import time
|
||||
|
||||
# external imports
|
||||
from chainlib.status import Status as TxStatus
|
||||
from chainsyncer.filter import SyncFilter
|
||||
from chainqueue.error import NotLocalTxError
|
||||
from chaind.adapters.fs import ChaindFsAdapter
|
||||
from shep.error import StateLockedKey
|
||||
|
||||
# local imports
|
||||
from .error import (
|
||||
QueueLockError,
|
||||
BackendError,
|
||||
)
|
||||
from chaind.lock import StoreLock
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class StateFilter(SyncFilter):
|
||||
|
||||
def __init__(self, adapter, throttler=None):
|
||||
self.adapter = adapter
|
||||
def __init__(self, chain_spec, adapter_path, tx_adapter, throttler=None):
|
||||
self.chain_spec = chain_spec
|
||||
self.adapter_path = adapter_path
|
||||
self.tx_adapter = tx_adapter
|
||||
self.throttler = throttler
|
||||
self.last_block_height = 0
|
||||
self.adapter = None
|
||||
self.store_lock = None
|
||||
|
||||
|
||||
def __get_adapter(self, block, force_reload=False):
|
||||
if self.store_lock == None:
|
||||
self.store_lock = StoreLock()
|
||||
|
||||
reload = False
|
||||
if block.number != self.last_block_height:
|
||||
reload = True
|
||||
elif self.adapter == None:
|
||||
reload = True
|
||||
elif force_reload:
|
||||
reload = True
|
||||
|
||||
self.last_block_height = block.number
|
||||
|
||||
if reload:
|
||||
while True:
|
||||
logg.info('reloading adapter')
|
||||
try:
|
||||
self.adapter = ChaindFsAdapter(
|
||||
self.chain_spec,
|
||||
self.adapter_path,
|
||||
self.tx_adapter,
|
||||
None,
|
||||
)
|
||||
break
|
||||
except BackendError as e:
|
||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
||||
self.store_lock.again()
|
||||
continue
|
||||
|
||||
return self.adapter
|
||||
|
||||
|
||||
def filter(self, conn, block, tx, session=None):
|
||||
try:
|
||||
cache_tx = self.adapter.get(tx.hash)
|
||||
except NotLocalTxError:
|
||||
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
||||
return False
|
||||
if tx.status == TxStatus.SUCCESS:
|
||||
self.adapter.succeed(block, tx)
|
||||
else:
|
||||
self.adapter.fail(block, tx)
|
||||
cache_tx = None
|
||||
queue_adapter = self.__get_adapter(block)
|
||||
|
||||
self.store_lock.reset()
|
||||
|
||||
while True:
|
||||
try:
|
||||
cache_tx = queue_adapter.get(tx.hash)
|
||||
break
|
||||
except NotLocalTxError:
|
||||
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
||||
self.__stop_adapter()
|
||||
return False
|
||||
except BackendError as e:
|
||||
logg.error('adapter get failed: {}, one more try'.format(e))
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
|
||||
if cache_tx == None:
|
||||
raise NotLocalTxError(tx.hash)
|
||||
|
||||
self.store_lock.reset()
|
||||
|
||||
queue_lock = StoreLock(error=QueueLockError)
|
||||
while True:
|
||||
try:
|
||||
if tx.status == TxStatus.SUCCESS:
|
||||
queue_adapter.succeed(block, tx)
|
||||
else:
|
||||
queue_adapter.fail(block, tx)
|
||||
break
|
||||
except QueueLockError as e:
|
||||
logg.debug('queue item {} is blocked, will retry: {}'.format(tx.hash, e))
|
||||
queue_lock.again()
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
except NotLocalTxError as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
except StateLockedKey as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
|
||||
logg.info('filter registered {} for {} in {}'.format(tx.status.name, tx.hash, block))
|
||||
|
||||
if self.throttler != None:
|
||||
self.throttler.dec(tx.hash)
|
||||
|
||||
|
||||
34
chaind/lock.py
Normal file
34
chaind/lock.py
Normal file
@@ -0,0 +1,34 @@
|
||||
# standard imports
|
||||
import time
|
||||
|
||||
# local imports
|
||||
from .error import BackendError
|
||||
|
||||
BASE_DELAY = 0.01
|
||||
BASE_DELAY_LIMIT = 10.0
|
||||
|
||||
|
||||
class StoreLock:
|
||||
|
||||
def __init__(self, delay=BASE_DELAY, delay_limit=BASE_DELAY_LIMIT, error=BackendError, description=None):
|
||||
self.base_delay = delay
|
||||
self.delay = delay
|
||||
self.delay_limit = delay_limit
|
||||
self.error = error
|
||||
self.description = description
|
||||
|
||||
|
||||
def again(self, e=None):
|
||||
if self.delay > self.delay_limit:
|
||||
err = None
|
||||
if e != None:
|
||||
err = str(e)
|
||||
else:
|
||||
err = self.description
|
||||
raise self.error(err)
|
||||
time.sleep(self.delay)
|
||||
self.delay *= 2
|
||||
|
||||
|
||||
def reset(self):
|
||||
self.delay = self.base_delay
|
||||
@@ -8,19 +8,21 @@ import stat
|
||||
from hexathon import strip_0x
|
||||
|
||||
# local imports
|
||||
from chaind.error import (
|
||||
from .error import (
|
||||
NothingToDoError,
|
||||
ClientGoneError,
|
||||
ClientBlockError,
|
||||
ClientInputError,
|
||||
)
|
||||
from .lock import StoreLock
|
||||
from .error import BackendError
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class SessionController:
|
||||
|
||||
def __init__(self, config, adapter, processor):
|
||||
def __init__(self, config, processor):
|
||||
self.dead = False
|
||||
os.makedirs(os.path.dirname(config.get('SESSION_SOCKET_PATH')), exist_ok=True)
|
||||
try:
|
||||
@@ -35,7 +37,6 @@ class SessionController:
|
||||
self.srv.settimeout(float(config.get('SESSION_DISPATCH_DELAY')))
|
||||
self.processor = processor
|
||||
self.chain_spec = config.get('CHAIN_SPEC')
|
||||
self.adapter = adapter
|
||||
|
||||
|
||||
def shutdown(self, signo, frame):
|
||||
@@ -59,7 +60,16 @@ class SessionController:
|
||||
|
||||
|
||||
def process(self, conn):
|
||||
r = self.processor(self.chain_spec, self.adapter, conn)
|
||||
state_lock = StoreLock()
|
||||
r = None
|
||||
while True:
|
||||
try:
|
||||
r = self.processor(conn)
|
||||
break
|
||||
except BackendError as e:
|
||||
state_lock.again(e)
|
||||
continue
|
||||
|
||||
if r > 0:
|
||||
self.srv.settimeout(0.1)
|
||||
else:
|
||||
@@ -104,7 +114,6 @@ class SessionController:
|
||||
logg.error('invalid input "{}"'.format(data_in_str))
|
||||
raise ClientInputError()
|
||||
|
||||
logg.info('recv {} bytes'.format(len(data)))
|
||||
return (srvs, data,)
|
||||
|
||||
|
||||
@@ -117,3 +126,4 @@ class SessionController:
|
||||
logg.debug('{} bytes sent'.format(len(v)))
|
||||
except BrokenPipeError:
|
||||
logg.debug('they just hung up. how rude.')
|
||||
srvs.close()
|
||||
|
||||
125
chaind/settings.py
Normal file
125
chaind/settings.py
Normal file
@@ -0,0 +1,125 @@
|
||||
# standard imports
|
||||
import logging
|
||||
import os
|
||||
import uuid
|
||||
|
||||
# external imports
|
||||
from chainsyncer.settings import ChainsyncerSettings
|
||||
from chainqueue.settings import ChainqueueSettings
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ChaindSettings(ChainsyncerSettings, ChainqueueSettings):
|
||||
|
||||
def __init__(self, include_sync=False, include_queue=False):
|
||||
super(ChaindSettings, self).__init__()
|
||||
self.include_sync = include_sync
|
||||
self.include_queue = include_queue
|
||||
|
||||
|
||||
def process_session(self, config):
|
||||
session_id = config.get('SESSION_ID')
|
||||
|
||||
base_dir = os.getcwd()
|
||||
data_dir = config.get('SESSION_DATA_DIR')
|
||||
if data_dir == None:
|
||||
data_dir = os.path.join(base_dir, '.chaind', 'chaind', self.o.get('CHAIND_BACKEND'))
|
||||
data_engine_dir = os.path.join(data_dir, config.get('CHAIND_ENGINE'))
|
||||
os.makedirs(data_engine_dir, exist_ok=True)
|
||||
|
||||
# check if existing session
|
||||
if session_id == None:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
try:
|
||||
os.stat(fp)
|
||||
fp = os.path.realpath(fp)
|
||||
except FileNotFoundError:
|
||||
fp = None
|
||||
if fp != None:
|
||||
session_id = os.path.basename(fp)
|
||||
|
||||
make_default = False
|
||||
if session_id == None:
|
||||
session_id = str(uuid.uuid4())
|
||||
make_default = True
|
||||
|
||||
# create the session persistent dir
|
||||
session_dir = os.path.join(data_engine_dir, session_id)
|
||||
if make_default:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
os.symlink(session_dir, fp)
|
||||
|
||||
#data_dir = os.path.join(session_dir, config.get('CHAIND_COMPONENT'))
|
||||
data_dir = session_dir
|
||||
os.makedirs(data_dir, exist_ok=True)
|
||||
|
||||
# create volatile dir
|
||||
uid = os.getuid()
|
||||
runtime_dir = config.get('SESSION_RUNTIME_DIR')
|
||||
if runtime_dir == None:
|
||||
runtime_dir = os.path.join('/run', 'user', str(uid), 'chaind', self.o.get('CHAIND_BACKEND'))
|
||||
#runtime_dir = os.path.join(runtime_dir, config.get('CHAIND_ENGINE'), session_id, config.get('CHAIND_COMPONENT'))
|
||||
runtime_dir = os.path.join(runtime_dir, config.get('CHAIND_ENGINE'), session_id)
|
||||
os.makedirs(runtime_dir, exist_ok=True)
|
||||
|
||||
self.o['SESSION_RUNTIME_DIR'] = runtime_dir
|
||||
self.o['SESSION_DIR'] = session_dir
|
||||
self.o['SESSION_DATA_DIR'] = data_dir
|
||||
self.o['SESSION_ID'] = session_id
|
||||
|
||||
|
||||
def process_sync_interface(self, config):
|
||||
raise NotImplementedError('no sync interface implementation defined')
|
||||
|
||||
|
||||
def process_sync(self, config):
|
||||
self.process_sync_interface(config)
|
||||
self.process_sync_range(config)
|
||||
|
||||
|
||||
def process_socket(self, config):
|
||||
socket_path = config.get('SESSION_SOCKET_PATH')
|
||||
if socket_path == None:
|
||||
socket_path = os.path.join(self.o['SESSION_RUNTIME_DIR'], 'chaind.sock')
|
||||
self.o['SESSION_SOCKET_PATH'] = socket_path
|
||||
|
||||
|
||||
def process_dispatch(self, config):
|
||||
self.o['SESSION_DISPATCH_DELAY'] = 0.01
|
||||
|
||||
|
||||
def process_token(self, config):
|
||||
self.o['TOKEN_MODULE'] = config.get('TOKEN_MODULE')
|
||||
|
||||
|
||||
def process_backend(self, config):
|
||||
if self.include_sync and self.include_queue:
|
||||
if self.o['QUEUE_BACKEND'] != self.o['SYNCER_BACKEND']:
|
||||
raise ValueError('queue and syncer backends must match. queue "{}" != syncer "{}"'.format(self.o['QUEUE_BACKEND'], self.o['SYNCER_BACKEND']))
|
||||
self.o['CHAIND_BACKEND'] = self.o['SYNCER_BACKEND']
|
||||
elif self.include_sync:
|
||||
self.o['CHAIND_BACKEND'] = self.o['SYNCER_BACKEND']
|
||||
elif self.include_queue:
|
||||
self.o['CHAIND_BACKEND'] = self.o['QUEUE_BACKEND']
|
||||
else:
|
||||
raise ValueError('at least one backend must be set')
|
||||
|
||||
|
||||
def process(self, config):
|
||||
super(ChaindSettings, self).process(config)
|
||||
if self.include_sync:
|
||||
self.process_sync(config)
|
||||
self.process_sync_backend(config)
|
||||
if self.include_queue:
|
||||
self.process_queue_backend(config)
|
||||
self.process_dispatch(config)
|
||||
self.process_token(config)
|
||||
|
||||
self.process_backend(config)
|
||||
self.process_session(config)
|
||||
self.process_socket(config)
|
||||
|
||||
|
||||
def dir_for(self, k):
|
||||
return os.path.join(self.o['SESSION_DIR'], k)
|
||||
@@ -1,5 +1,4 @@
|
||||
# standard imports
|
||||
import unittest
|
||||
import hashlib
|
||||
import tempfile
|
||||
|
||||
@@ -9,9 +8,6 @@ from chainlib.status import Status as TxStatus
|
||||
from chainlib.chain import ChainSpec
|
||||
from chainlib.error import RPCException
|
||||
|
||||
# local imports
|
||||
from chaind.adapters.fs import ChaindFsAdapter
|
||||
|
||||
|
||||
class MockCacheAdapter(CacheTokenTx):
|
||||
|
||||
@@ -33,7 +29,7 @@ class MockDispatcher:
|
||||
|
||||
|
||||
def send(self, v):
|
||||
if v not in self.fails:
|
||||
if v in self.fails:
|
||||
raise RPCException('{} is in fails'.format(v))
|
||||
pass
|
||||
|
||||
@@ -45,10 +41,7 @@ class MockTx:
|
||||
self.status = status
|
||||
|
||||
|
||||
class TestChaindFsBase(unittest.TestCase):
|
||||
|
||||
def setUp(self):
|
||||
self.chain_spec = ChainSpec('foo', 'bar', 42, 'baz')
|
||||
self.path = tempfile.mkdtemp()
|
||||
self.adapter = ChaindFsAdapter(self.chain_spec, self.path, self.cache_adapter, self.dispatcher)
|
||||
class MockBlock:
|
||||
|
||||
def __init__(self, number):
|
||||
self.number = number
|
||||
|
||||
31
chaind/unittest/fs.py
Normal file
31
chaind/unittest/fs.py
Normal file
@@ -0,0 +1,31 @@
|
||||
# standard imports
|
||||
import unittest
|
||||
import tempfile
|
||||
import logging
|
||||
|
||||
# external imports
|
||||
from chainlib.chain import ChainSpec
|
||||
|
||||
# local imports
|
||||
from chaind.adapters.fs import ChaindFsAdapter
|
||||
|
||||
logging.STATETRACE = 5
|
||||
logg = logging.getLogger(__name__)
|
||||
logg.setLevel(logging.STATETRACE)
|
||||
|
||||
|
||||
class TestChaindFsBase(unittest.TestCase):
|
||||
|
||||
def setUp(self):
|
||||
self.chain_spec = ChainSpec('foo', 'bar', 42, 'baz')
|
||||
self.path = tempfile.mkdtemp()
|
||||
self.adapter = ChaindFsAdapter(self.chain_spec, self.path, self.cache_adapter, self.dispatcher, event_callback=self.log_state)
|
||||
|
||||
|
||||
def log_state(self, k, from_state, to_state):
|
||||
logg.log(logging.STATETRACE, 'state change {}: {} -> {}'.format(
|
||||
k,
|
||||
from_state,
|
||||
to_state,
|
||||
)
|
||||
)
|
||||
@@ -1,6 +1,6 @@
|
||||
chainlib~=0.1.0
|
||||
chainqueue~=0.1.1
|
||||
chainsyncer~=0.3.5
|
||||
chainlib~=0.1.2
|
||||
chainqueue~=0.1.13
|
||||
chainsyncer~=0.4.5
|
||||
confini~=0.6.0
|
||||
funga~=0.5.2
|
||||
pyxdg~=0.26
|
||||
|
||||
@@ -1,80 +0,0 @@
|
||||
#!/usr/bin/python
|
||||
import os
|
||||
import argparse
|
||||
import logging
|
||||
|
||||
# external imports
|
||||
import alembic
|
||||
from alembic.config import Config as AlembicConfig
|
||||
import confini
|
||||
import chainqueue.db
|
||||
import chainsyncer.db
|
||||
|
||||
# local imports
|
||||
from chaind import Environment
|
||||
|
||||
logging.basicConfig(level=logging.WARNING)
|
||||
logg = logging.getLogger()
|
||||
|
||||
rootdir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
||||
dbdir = os.path.join(rootdir, 'chaind', 'db')
|
||||
configdir = os.path.join(rootdir, 'chaind', 'data', 'config')
|
||||
default_migrations_dir = os.path.join(dbdir, 'migrations')
|
||||
|
||||
env = Environment(env=os.environ)
|
||||
|
||||
argparser = argparse.ArgumentParser()
|
||||
argparser.add_argument('-c', type=str, default=env.config_dir, help='config directory')
|
||||
argparser.add_argument('--env-prefix', default=os.environ.get('CONFINI_ENV_PREFIX'), dest='env_prefix', type=str, help='environment prefix for variables to overwrite configuration')
|
||||
argparser.add_argument('--data-dir', dest='data_dir', type=str, help='data directory')
|
||||
argparser.add_argument('--migrations-dir', dest='migrations_dir', default=default_migrations_dir, type=str, help='path to alembic migrations directory')
|
||||
argparser.add_argument('--reset', action='store_true', help='reset exsting database')
|
||||
argparser.add_argument('-v', action='store_true', help='be verbose')
|
||||
argparser.add_argument('-vv', action='store_true', help='be more verbose')
|
||||
args = argparser.parse_args()
|
||||
|
||||
if args.vv:
|
||||
logging.getLogger().setLevel(logging.DEBUG)
|
||||
elif args.v:
|
||||
logging.getLogger().setLevel(logging.INFO)
|
||||
|
||||
# process config
|
||||
logg.debug('loading config from {}'.format(args.c))
|
||||
config = confini.Config(configdir, args.env_prefix, override_dirs=[args.c])
|
||||
config.process()
|
||||
args_override = {
|
||||
'SESSION_DATA_DIR': getattr(args, 'data_dir'),
|
||||
}
|
||||
config.dict_override(args_override, 'cli args')
|
||||
|
||||
if config.get('DATABASE_ENGINE') == 'sqlite':
|
||||
config.add(os.path.join(config.get('SESSION_DATA_DIR'), config.get('DATABASE_NAME') + '.sqlite'), 'DATABASE_NAME', True)
|
||||
|
||||
config.censor('PASSWORD', 'DATABASE')
|
||||
|
||||
logg.debug('config loaded:\n{}'.format(config))
|
||||
config.add(os.path.join(args.migrations_dir, config.get('DATABASE_ENGINE')), '_MIGRATIONS_DIR', True)
|
||||
if not os.path.isdir(config.get('_MIGRATIONS_DIR')):
|
||||
logg.debug('migrations dir for engine {} not found, reverting to default'.format(config.get('DATABASE_ENGINE')))
|
||||
config.add(os.path.join(args.migrations_dir, 'default'), '_MIGRATIONS_DIR', True)
|
||||
|
||||
os.makedirs(config.get('SESSION_DATA_DIR'), exist_ok=True)
|
||||
|
||||
dsn = chainqueue.db.dsn_from_config(config)
|
||||
|
||||
def main():
|
||||
logg.info('using migrations dir {}'.format(config.get('_MIGRATIONS_DIR')))
|
||||
logg.info('using db {}'.format(dsn))
|
||||
ac = AlembicConfig(os.path.join(config.get('_MIGRATIONS_DIR'), 'alembic.ini'))
|
||||
ac.set_main_option('sqlalchemy.url', dsn)
|
||||
ac.set_main_option('script_location', config.get('_MIGRATIONS_DIR'))
|
||||
|
||||
if args.reset:
|
||||
logg.debug('reset is set, purging existing content')
|
||||
alembic.command.downgrade(ac, 'base')
|
||||
|
||||
alembic.command.upgrade(ac, 'head')
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
@@ -1,7 +1,7 @@
|
||||
[metadata]
|
||||
name = chaind
|
||||
version = 0.1.0
|
||||
description = Base package for chain queue servicek
|
||||
version = 0.2.10
|
||||
description = Base package for chain queue service
|
||||
author = Louis Holbrook
|
||||
author_email = dev@holbrook.no
|
||||
url = https://gitlab.com/chaintool/chaind
|
||||
@@ -23,14 +23,14 @@ licence_files =
|
||||
LICENSE
|
||||
|
||||
[options]
|
||||
python_requires = >= 3.6
|
||||
python_requires = >= 3.7
|
||||
include_package_data = True
|
||||
packages =
|
||||
chaind
|
||||
# chaind.sql
|
||||
# chaind.runnable
|
||||
chaind.adapters
|
||||
chaind.unittest
|
||||
chaind.cli
|
||||
|
||||
#[options.entry_points]
|
||||
#console_scripts =
|
||||
|
||||
@@ -14,16 +14,16 @@ from chaind.filter import StateFilter
|
||||
# test imports
|
||||
from chaind.unittest.common import (
|
||||
MockTx,
|
||||
MockBlock,
|
||||
MockCacheAdapter,
|
||||
TestChaindFsBase,
|
||||
MockDispatcher,
|
||||
)
|
||||
|
||||
from chaind.unittest.fs import TestChaindFsBase
|
||||
|
||||
logging.basicConfig(level=logging.DEBUG)
|
||||
logg = logging.getLogger()
|
||||
|
||||
|
||||
|
||||
class TestChaindFs(TestChaindFsBase):
|
||||
|
||||
def setUp(self):
|
||||
@@ -43,12 +43,15 @@ class TestChaindFs(TestChaindFsBase):
|
||||
self.assertEqual(data, v)
|
||||
|
||||
|
||||
def test_fs_defer(self):
|
||||
def test_fs_fail(self):
|
||||
data = os.urandom(128).hex()
|
||||
hsh = self.adapter.put(data)
|
||||
self.dispatcher.add_fail(hsh)
|
||||
self.adapter.dispatch(hsh)
|
||||
txs = self.adapter.deferred()
|
||||
self.dispatcher.add_fail(data)
|
||||
|
||||
r = self.adapter.dispatch(hsh)
|
||||
self.assertFalse(r)
|
||||
|
||||
txs = self.adapter.failed()
|
||||
self.assertEqual(len(txs), 1)
|
||||
|
||||
|
||||
@@ -72,9 +75,10 @@ class TestChaindFs(TestChaindFsBase):
|
||||
data = os.urandom(128).hex()
|
||||
hsh = self.adapter.put(data)
|
||||
|
||||
fltr = StateFilter(self.adapter)
|
||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||
tx = MockTx(hsh)
|
||||
fltr.filter(None, None, tx)
|
||||
block = MockBlock(42)
|
||||
fltr.filter(None, block, tx)
|
||||
|
||||
|
||||
def test_fs_filter_fail(self):
|
||||
@@ -83,9 +87,30 @@ class TestChaindFs(TestChaindFsBase):
|
||||
data = os.urandom(128).hex()
|
||||
hsh = self.adapter.put(data)
|
||||
|
||||
fltr = StateFilter(self.adapter)
|
||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||
tx = MockTx(hsh, TxStatus.ERROR)
|
||||
fltr.filter(None, None, tx)
|
||||
block = MockBlock(42)
|
||||
fltr.filter(None, block, tx)
|
||||
|
||||
|
||||
def test_upcoming(self):
|
||||
drv = QueueDriver(self.adapter)
|
||||
|
||||
txs = []
|
||||
for i in range(10):
|
||||
data = os.urandom(128).hex()
|
||||
hsh = self.adapter.put(data)
|
||||
txs.append(hsh)
|
||||
self.adapter.enqueue(hsh)
|
||||
|
||||
r = self.adapter.upcoming(limit=5)
|
||||
self.assertEqual(len(r), 5)
|
||||
|
||||
r = self.adapter.dispatch(txs[0])
|
||||
self.assertTrue(r)
|
||||
|
||||
r = self.adapter.upcoming(limit=5)
|
||||
self.assertEqual(len(r), 4)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
|
||||
Reference in New Issue
Block a user