Compare commits
17 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
443e64f3e9
|
||
|
|
bb5e0b359f
|
||
|
|
8bc49e7408
|
||
|
|
c83aade144
|
||
|
|
63a1d07a28
|
||
|
|
6da2a1ced9
|
||
|
|
9f2a791b1f
|
||
|
|
ba894044d8
|
||
|
|
e5a18d1abf
|
||
|
|
139e2772af
|
||
|
|
0a5818ebf1
|
||
|
|
5d2d73fa64
|
||
|
|
dd1879bb91
|
||
|
|
e264ed5c37
|
||
|
|
d472bd4f7c
|
||
|
|
465d692956
|
||
|
|
81c1207828
|
11
CHANGELOG
11
CHANGELOG
@@ -1,3 +1,14 @@
|
||||
- 0.3.0
|
||||
* Implement on chainlib 0.3.0
|
||||
- 0.2.12
|
||||
* Breaking upgrade of chainlib.
|
||||
* Implement generic block and tx.
|
||||
- 0.2.11
|
||||
* Upgrade shep to handle exception in filestore list
|
||||
- 0.2.10
|
||||
* Upgrade shep to guarantee state lock atomicity
|
||||
- 0.2.9
|
||||
* Minimize instantiations of adapters in filter execution
|
||||
- 0.2.8
|
||||
* Upgrade chainsyncer
|
||||
- 0.2.7
|
||||
|
||||
@@ -1,12 +0,0 @@
|
||||
# standard imports
|
||||
import os
|
||||
|
||||
# local imports
|
||||
from .base import *
|
||||
from .arg import process_flags
|
||||
from .config import process_config
|
||||
|
||||
|
||||
__script_dir = os.path.dirname(os.path.realpath(__file__))
|
||||
data_dir = os.path.join(os.path.dirname(__script_dir), 'data')
|
||||
config_dir = os.path.join(data_dir, 'config')
|
||||
@@ -1,17 +1,19 @@
|
||||
# local imports
|
||||
from .base import ChaindFlag
|
||||
def apply_flag(flag):
|
||||
flag.add('session')
|
||||
flag.add('dispatch')
|
||||
flag.add('socket')
|
||||
flag.add('socket_client')
|
||||
flag.add('token')
|
||||
|
||||
def process_flags(argparser, flags):
|
||||
if flags & ChaindFlag.SESSION > 0:
|
||||
argparser.add_argument('--session-id', dest='session_id', type=str, help='Session to store state and data under')
|
||||
argparser.add_argument('--runtime-dir', dest='runtime_dir', type=str, help='Directory to store volatile data')
|
||||
argparser.add_argument('--data-dir', dest='data_dir', type=str, help='Directory to store persistent data')
|
||||
flag.alias('chaind_base', 'session')
|
||||
flag.alias('chaind_socket_client', 'session', 'socket', 'socket_client')
|
||||
|
||||
if flags & ChaindFlag.SOCKET > 0:
|
||||
argparser.add_argument('--socket-path', dest='socket', type=str, help='UNIX socket path')
|
||||
return flag
|
||||
|
||||
if flags & ChaindFlag.SOCKET_CLIENT > 0:
|
||||
argparser.add_argument('--send-socket', dest='socket_send', action='store_true', help='Send to UNIX socket')
|
||||
|
||||
if flags & ChaindFlag.TOKEN > 0:
|
||||
argparser.add_argument('--token-module', dest='token_module', type=str, help='Python module path to resolve tokens from identifiers')
|
||||
def apply_arg(arg):
|
||||
arg.add_long('session-id', 'session', help='Session to store state and data under')
|
||||
arg.add_long('socket-path', 'socket', help='UNIX socket path')
|
||||
arg.add_long('send-socket', 'socket_client', typ=bool, help='Send to UNIX socket')
|
||||
arg.add_long('token-module', 'token', help='Python module path to resolve tokens from identifiers')
|
||||
return arg
|
||||
|
||||
@@ -1,23 +1,19 @@
|
||||
# external imports
|
||||
from chaind.cli import ChaindFlag
|
||||
|
||||
|
||||
def process_config(config, args, flags):
|
||||
def process_config(config, arg, args, flags):
|
||||
args_override = {}
|
||||
if flags & ChaindFlag.SESSION:
|
||||
if arg.match('session', flags):
|
||||
args_override['SESSION_ID'] = getattr(args, 'session_id')
|
||||
args_override['SESSION_RUNTIME_DIR'] = getattr(args, 'runtime_dir')
|
||||
args_override['SESSION_DATA_DIR'] = getattr(args, 'data_dir')
|
||||
args_override['SESSION_RUNTIME_DIR'] = getattr(args, 'runtime_path')
|
||||
args_override['SESSION_DATA_DIR'] = getattr(args, 'state_path')
|
||||
|
||||
if flags & ChaindFlag.SOCKET:
|
||||
if arg.match('socket', flags):
|
||||
args_override['SESSION_SOCKET_PATH'] = getattr(args, 'socket')
|
||||
|
||||
if flags & ChaindFlag.TOKEN:
|
||||
if arg.match('token', flags):
|
||||
args_override['TOKEN_MODULE'] = getattr(args, 'token_module')
|
||||
|
||||
config.dict_override(args_override, 'local cli args')
|
||||
|
||||
if flags & ChaindFlag.SOCKET_CLIENT:
|
||||
config.add(getattr(args, 'socket_send'), '_SOCKET_SEND', False)
|
||||
if arg.match('socket_client', flags):
|
||||
config.add(getattr(args, 'send_socket'), '_SOCKET_SEND', False)
|
||||
|
||||
return config
|
||||
|
||||
6
chaind/data/__init__.py
Normal file
6
chaind/data/__init__.py
Normal file
@@ -0,0 +1,6 @@
|
||||
# standard imports
|
||||
import os
|
||||
|
||||
|
||||
data_dir = os.path.realpath(os.path.dirname(__file__))
|
||||
config_dir = os.path.join(data_dir, 'config')
|
||||
@@ -1,5 +1,5 @@
|
||||
[session]
|
||||
socket_path =
|
||||
runtime_dir =
|
||||
runtime_path =
|
||||
id =
|
||||
data_dir =
|
||||
data_path =
|
||||
|
||||
@@ -26,27 +26,51 @@ class StateFilter(SyncFilter):
|
||||
self.adapter_path = adapter_path
|
||||
self.tx_adapter = tx_adapter
|
||||
self.throttler = throttler
|
||||
self.last_block_height = 0
|
||||
self.adapter = None
|
||||
self.store_lock = None
|
||||
|
||||
|
||||
def __get_adapter(self, block, force_reload=False):
|
||||
if self.store_lock == None:
|
||||
self.store_lock = StoreLock()
|
||||
|
||||
reload = False
|
||||
if block.number != self.last_block_height:
|
||||
reload = True
|
||||
elif self.adapter == None:
|
||||
reload = True
|
||||
elif force_reload:
|
||||
reload = True
|
||||
|
||||
self.last_block_height = block.number
|
||||
|
||||
if reload:
|
||||
while True:
|
||||
logg.info('reloading adapter')
|
||||
try:
|
||||
self.adapter = ChaindFsAdapter(
|
||||
self.chain_spec,
|
||||
self.adapter_path,
|
||||
self.tx_adapter,
|
||||
None,
|
||||
)
|
||||
break
|
||||
except BackendError as e:
|
||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
||||
self.store_lock.again()
|
||||
continue
|
||||
|
||||
return self.adapter
|
||||
|
||||
|
||||
def filter(self, conn, block, tx, session=None):
|
||||
cache_tx = None
|
||||
store_lock = StoreLock()
|
||||
queue_adapter = None
|
||||
queue_adapter = self.__get_adapter(block)
|
||||
|
||||
self.store_lock.reset()
|
||||
|
||||
while True:
|
||||
try:
|
||||
queue_adapter = ChaindFsAdapter(
|
||||
self.chain_spec,
|
||||
self.adapter_path,
|
||||
self.tx_adapter,
|
||||
None,
|
||||
)
|
||||
except BackendError as e:
|
||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
||||
store_lock.again()
|
||||
continue
|
||||
|
||||
store_lock.reset()
|
||||
|
||||
try:
|
||||
cache_tx = queue_adapter.get(tx.hash)
|
||||
break
|
||||
@@ -55,14 +79,15 @@ class StateFilter(SyncFilter):
|
||||
return False
|
||||
except BackendError as e:
|
||||
logg.error('adapter get failed: {}, one more try'.format(e))
|
||||
queue_adapter = None
|
||||
store_lock.again()
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
|
||||
if cache_tx == None:
|
||||
raise NotLocalTxError(tx.hash)
|
||||
|
||||
store_lock = StoreLock()
|
||||
self.store_lock.reset()
|
||||
|
||||
queue_lock = StoreLock(error=QueueLockError)
|
||||
while True:
|
||||
try:
|
||||
@@ -76,18 +101,21 @@ class StateFilter(SyncFilter):
|
||||
queue_lock.again()
|
||||
except FileNotFoundError as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
store_lock.again()
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
except NotLocalTxError as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
store_lock.again()
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
except StateLockedKey as e:
|
||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||
store_lock.again()
|
||||
self.store_lock.again()
|
||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
||||
continue
|
||||
|
||||
logg.info('filter registered {} for {} in {}'.format(tx.status.name, tx.hash, block))
|
||||
logg.info('filter registered {} for {} in {}'.format(tx.status_name, tx.hash, block))
|
||||
|
||||
if self.throttler != None:
|
||||
self.throttler.dec(tx.hash)
|
||||
|
||||
@@ -4,122 +4,123 @@ import os
|
||||
import uuid
|
||||
|
||||
# external imports
|
||||
from chainsyncer.settings import ChainsyncerSettings
|
||||
from chainqueue.settings import ChainqueueSettings
|
||||
from chainlib.settings import ChainSettings
|
||||
from chainqueue.settings import *
|
||||
|
||||
logg = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ChaindSettings(ChainsyncerSettings, ChainqueueSettings):
|
||||
class ChaindSettings(ChainSettings):
|
||||
|
||||
def __init__(self, include_sync=False, include_queue=False):
|
||||
super(ChaindSettings, self).__init__()
|
||||
self.include_sync = include_sync
|
||||
self.include_queue = include_queue
|
||||
|
||||
|
||||
def process_session(self, config):
|
||||
session_id = config.get('SESSION_ID')
|
||||
|
||||
base_dir = os.getcwd()
|
||||
data_dir = config.get('SESSION_DATA_DIR')
|
||||
if data_dir == None:
|
||||
data_dir = os.path.join(base_dir, '.chaind', 'chaind', self.o.get('CHAIND_BACKEND'))
|
||||
data_engine_dir = os.path.join(data_dir, config.get('CHAIND_ENGINE'))
|
||||
os.makedirs(data_engine_dir, exist_ok=True)
|
||||
|
||||
# check if existing session
|
||||
if session_id == None:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
try:
|
||||
os.stat(fp)
|
||||
fp = os.path.realpath(fp)
|
||||
except FileNotFoundError:
|
||||
fp = None
|
||||
if fp != None:
|
||||
session_id = os.path.basename(fp)
|
||||
|
||||
make_default = False
|
||||
if session_id == None:
|
||||
session_id = str(uuid.uuid4())
|
||||
make_default = True
|
||||
|
||||
# create the session persistent dir
|
||||
session_dir = os.path.join(data_engine_dir, session_id)
|
||||
if make_default:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
os.symlink(session_dir, fp)
|
||||
|
||||
#data_dir = os.path.join(session_dir, config.get('CHAIND_COMPONENT'))
|
||||
data_dir = session_dir
|
||||
os.makedirs(data_dir, exist_ok=True)
|
||||
|
||||
# create volatile dir
|
||||
uid = os.getuid()
|
||||
runtime_dir = config.get('SESSION_RUNTIME_DIR')
|
||||
if runtime_dir == None:
|
||||
runtime_dir = os.path.join('/run', 'user', str(uid), 'chaind', self.o.get('CHAIND_BACKEND'))
|
||||
#runtime_dir = os.path.join(runtime_dir, config.get('CHAIND_ENGINE'), session_id, config.get('CHAIND_COMPONENT'))
|
||||
runtime_dir = os.path.join(runtime_dir, config.get('CHAIND_ENGINE'), session_id)
|
||||
os.makedirs(runtime_dir, exist_ok=True)
|
||||
|
||||
self.o['SESSION_RUNTIME_DIR'] = runtime_dir
|
||||
self.o['SESSION_DIR'] = session_dir
|
||||
self.o['SESSION_DATA_DIR'] = data_dir
|
||||
self.o['SESSION_ID'] = session_id
|
||||
|
||||
|
||||
def process_sync_interface(self, config):
|
||||
raise NotImplementedError('no sync interface implementation defined')
|
||||
|
||||
|
||||
def process_sync(self, config):
|
||||
self.process_sync_interface(config)
|
||||
self.process_sync_range(config)
|
||||
|
||||
|
||||
def process_socket(self, config):
|
||||
socket_path = config.get('SESSION_SOCKET_PATH')
|
||||
if socket_path == None:
|
||||
socket_path = os.path.join(self.o['SESSION_RUNTIME_DIR'], 'chaind.sock')
|
||||
self.o['SESSION_SOCKET_PATH'] = socket_path
|
||||
|
||||
|
||||
def process_dispatch(self, config):
|
||||
self.o['SESSION_DISPATCH_DELAY'] = 0.01
|
||||
|
||||
|
||||
def process_token(self, config):
|
||||
self.o['TOKEN_MODULE'] = config.get('TOKEN_MODULE')
|
||||
|
||||
|
||||
def process_backend(self, config):
|
||||
if self.include_sync and self.include_queue:
|
||||
if self.o['QUEUE_BACKEND'] != self.o['SYNCER_BACKEND']:
|
||||
raise ValueError('queue and syncer backends must match. queue "{}" != syncer "{}"'.format(self.o['QUEUE_BACKEND'], self.o['SYNCER_BACKEND']))
|
||||
self.o['CHAIND_BACKEND'] = self.o['SYNCER_BACKEND']
|
||||
elif self.include_sync:
|
||||
self.o['CHAIND_BACKEND'] = self.o['SYNCER_BACKEND']
|
||||
elif self.include_queue:
|
||||
self.o['CHAIND_BACKEND'] = self.o['QUEUE_BACKEND']
|
||||
else:
|
||||
raise ValueError('at least one backend must be set')
|
||||
|
||||
|
||||
def process(self, config):
|
||||
super(ChaindSettings, self).process(config)
|
||||
if self.include_sync:
|
||||
self.process_sync(config)
|
||||
self.process_sync_backend(config)
|
||||
if self.include_queue:
|
||||
self.process_queue_backend(config)
|
||||
self.process_dispatch(config)
|
||||
self.process_token(config)
|
||||
|
||||
self.process_backend(config)
|
||||
self.process_session(config)
|
||||
self.process_socket(config)
|
||||
def __init__(settings, include_sync=False, include_queue=False):
|
||||
super(ChaindSettings, settings).__init__()
|
||||
settings.include_sync = include_sync
|
||||
settings.include_queue = include_queue
|
||||
|
||||
|
||||
def dir_for(self, k):
|
||||
return os.path.join(self.o['SESSION_DIR'], k)
|
||||
return os.path.join(self.o['SESSION_PATH'], k)
|
||||
|
||||
|
||||
def process_session(settings, config):
|
||||
session_id = config.get('SESSION_ID')
|
||||
|
||||
base_dir = os.getcwd()
|
||||
data_dir = config.get('SESSION_DATA_PATH')
|
||||
if data_dir == None:
|
||||
data_dir = os.path.join(base_dir, '.chaind', 'chaind', settings.get('CHAIND_BACKEND'))
|
||||
data_engine_dir = os.path.join(data_dir, config.get('CHAIND_ENGINE'))
|
||||
os.makedirs(data_engine_dir, exist_ok=True)
|
||||
|
||||
# check if existing session
|
||||
if session_id == None:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
try:
|
||||
os.stat(fp)
|
||||
fp = os.path.realpath(fp)
|
||||
except FileNotFoundError:
|
||||
fp = None
|
||||
if fp != None:
|
||||
session_id = os.path.basename(fp)
|
||||
|
||||
make_default = False
|
||||
if session_id == None:
|
||||
session_id = str(uuid.uuid4())
|
||||
make_default = True
|
||||
|
||||
chain_spec = settings.get('CHAIN_SPEC')
|
||||
network_id_str = str(chain_spec.network_id())
|
||||
# create the session persistent dir
|
||||
session_path = os.path.join(
|
||||
data_engine_dir,
|
||||
chain_spec.arch(),
|
||||
chain_spec.fork(),
|
||||
network_id_str,
|
||||
session_id,
|
||||
)
|
||||
if make_default:
|
||||
fp = os.path.join(data_engine_dir, 'default')
|
||||
os.symlink(session_path, fp)
|
||||
|
||||
data_path = session_path
|
||||
os.makedirs(data_path, exist_ok=True)
|
||||
|
||||
# create volatile dir
|
||||
uid = os.getuid()
|
||||
runtime_path = config.get('SESSION_RUNTIME_PATH')
|
||||
if runtime_path == None:
|
||||
runtime_path = os.path.join('/run', 'user', str(uid), 'chaind', settings.get('CHAIND_BACKEND'))
|
||||
runtime_path = os.path.join(
|
||||
runtime_path,
|
||||
config.get('CHAIND_ENGINE'),
|
||||
chain_spec.arch(),
|
||||
chain_spec.fork(),
|
||||
str(chain_spec.network_id()),
|
||||
session_id,
|
||||
)
|
||||
os.makedirs(runtime_path, exist_ok=True)
|
||||
|
||||
settings.set('SESSION_RUNTIME_PATH', runtime_path)
|
||||
settings.set('SESSION_PATH', session_path)
|
||||
settings.set('SESSION_DATA_PATH', data_path)
|
||||
settings.set('SESSION_ID', session_id)
|
||||
|
||||
return settings
|
||||
|
||||
|
||||
def process_socket(settings, config):
|
||||
socket_path = config.get('SESSION_SOCKET_PATH')
|
||||
if socket_path == None:
|
||||
socket_path = os.path.join(settings.get('SESSION_RUNTIME_PATH'), 'chaind.sock')
|
||||
settings.set('SESSION_SOCKET_PATH', socket_path)
|
||||
return settings
|
||||
|
||||
|
||||
def process_dispatch(settings, config):
|
||||
settings.set('SESSION_DISPATCH_DELAY', 0.01)
|
||||
return settings
|
||||
|
||||
|
||||
def process_token(settings, config):
|
||||
settings.set('TOKEN_MODULE', config.get('TOKEN_MODULE'))
|
||||
return settings
|
||||
|
||||
|
||||
def process_backend(settings, config):
|
||||
settings.set('CHAIND_BACKEND', config.get('STATE_BACKEND')) #backend)
|
||||
return settings
|
||||
|
||||
|
||||
def process_queue(settings, config):
|
||||
if config.get('STATE_PATH') == None:
|
||||
queue_state_dir = settings.dir_for('queue')
|
||||
config.add(queue_state_dir, 'STATE_PATH', False)
|
||||
logg.debug('setting queue state path {}'.format(queue_state_dir))
|
||||
|
||||
settings = process_queue_tx(settings, config)
|
||||
settings = process_queue_paths(settings, config)
|
||||
if config.get('STATE_BACKEND') == 'fs':
|
||||
settings = process_queue_backend_fs(settings, config)
|
||||
settings = process_queue_store(settings, config)
|
||||
|
||||
return settings
|
||||
|
||||
@@ -7,6 +7,11 @@ from chainqueue.cache import CacheTokenTx
|
||||
from chainlib.status import Status as TxStatus
|
||||
from chainlib.chain import ChainSpec
|
||||
from chainlib.error import RPCException
|
||||
from chainlib.tx import (
|
||||
Tx,
|
||||
TxResult,
|
||||
)
|
||||
from chainlib.block import Block
|
||||
|
||||
|
||||
class MockCacheAdapter(CacheTokenTx):
|
||||
@@ -34,8 +39,17 @@ class MockDispatcher:
|
||||
pass
|
||||
|
||||
|
||||
class MockTx:
|
||||
class MockTx(Tx):
|
||||
|
||||
def __init__(self, tx_hash, status=TxStatus.SUCCESS):
|
||||
self.hash = tx_hash
|
||||
self.status = status
|
||||
result = TxResult()
|
||||
result.status = status
|
||||
super(MockTx, self).__init__(result=result)
|
||||
self.set_hash(tx_hash)
|
||||
|
||||
|
||||
class MockBlock(Block):
|
||||
|
||||
def __init__(self, number):
|
||||
super(MockBlock, self).__init__()
|
||||
self.number = number
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
chainlib~=0.1.2
|
||||
chainqueue~=0.1.12
|
||||
chainsyncer~=0.4.4
|
||||
confini~=0.6.0
|
||||
chainlib~=0.3.0
|
||||
chainqueue~=0.2.0
|
||||
chainsyncer~=0.5.0
|
||||
confini~=0.6.1
|
||||
funga~=0.5.2
|
||||
pyxdg~=0.26
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
[metadata]
|
||||
name = chaind
|
||||
version = 0.2.8
|
||||
version = 0.3.0
|
||||
description = Base package for chain queue service
|
||||
author = Louis Holbrook
|
||||
author_email = dev@holbrook.no
|
||||
@@ -30,6 +30,7 @@ packages =
|
||||
# chaind.runnable
|
||||
chaind.adapters
|
||||
chaind.unittest
|
||||
chaind.data
|
||||
chaind.cli
|
||||
|
||||
#[options.entry_points]
|
||||
|
||||
@@ -14,6 +14,7 @@ from chaind.filter import StateFilter
|
||||
# test imports
|
||||
from chaind.unittest.common import (
|
||||
MockTx,
|
||||
MockBlock,
|
||||
MockCacheAdapter,
|
||||
MockDispatcher,
|
||||
)
|
||||
@@ -76,7 +77,8 @@ class TestChaindFs(TestChaindFsBase):
|
||||
|
||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||
tx = MockTx(hsh)
|
||||
fltr.filter(None, None, tx)
|
||||
block = MockBlock(42)
|
||||
fltr.filter(None, block, tx)
|
||||
|
||||
|
||||
def test_fs_filter_fail(self):
|
||||
@@ -87,7 +89,8 @@ class TestChaindFs(TestChaindFsBase):
|
||||
|
||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||
tx = MockTx(hsh, TxStatus.ERROR)
|
||||
fltr.filter(None, None, tx)
|
||||
block = MockBlock(42)
|
||||
fltr.filter(None, block, tx)
|
||||
|
||||
|
||||
def test_upcoming(self):
|
||||
|
||||
Reference in New Issue
Block a user