Compare commits
12 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
f33ba13d74
|
||
|
|
5459d4c3f8
|
||
|
|
3e05717395
|
||
|
|
54d10ee40b
|
||
|
|
9cffdc5867
|
||
|
|
4f96be2024
|
||
|
|
32e1bc6aa5
|
||
|
|
387014f77b
|
||
|
|
c3a592c0f6
|
||
|
|
5d61506133
|
||
|
|
5102b4ac6e
|
||
|
|
9b98703f24
|
17
CHANGELOG
17
CHANGELOG
@@ -1,3 +1,20 @@
|
|||||||
|
- 0.2.8
|
||||||
|
* Upgrade chainsyncer
|
||||||
|
- 0.2.7
|
||||||
|
* Upgrade chainlib
|
||||||
|
- 0.2.6
|
||||||
|
* Deps upgrade
|
||||||
|
- 0.2.5
|
||||||
|
* Deps upgrade
|
||||||
|
- 0.2.4
|
||||||
|
* Allow omission of state store sync in queue store backend
|
||||||
|
- 0.2.2
|
||||||
|
* Fix missing symbol crashes related to race conditions
|
||||||
|
- 0.2.1
|
||||||
|
* Receive removed race checks from chainqueue
|
||||||
|
- 0.2.0
|
||||||
|
* primitive race condition handling between fs access of sync and queue
|
||||||
|
* re-enable throttling based on in-flight transaction count
|
||||||
- 0.1.2
|
- 0.1.2
|
||||||
* add settings object
|
* add settings object
|
||||||
- 0.1.0
|
- 0.1.0
|
||||||
|
|||||||
@@ -1,10 +1,27 @@
|
|||||||
|
# standard imports
|
||||||
|
import logging
|
||||||
|
import time
|
||||||
|
|
||||||
# external imports
|
# external imports
|
||||||
from chainqueue import Store as QueueStore
|
from chainqueue import Store as QueueStore
|
||||||
|
|
||||||
|
# local imports
|
||||||
|
from chaind.lock import StoreLock
|
||||||
|
|
||||||
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class ChaindAdapter:
|
class ChaindAdapter:
|
||||||
|
|
||||||
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0):
|
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, store_sync=True):
|
||||||
self.cache_adapter = cache_adapter
|
self.cache_adapter = cache_adapter
|
||||||
self.dispatcher = dispatcher
|
self.dispatcher = dispatcher
|
||||||
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache)
|
store_lock = StoreLock()
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache, sync=store_sync)
|
||||||
|
break
|
||||||
|
except FileNotFoundError as e:
|
||||||
|
logg.debug('queuestore instantiation failed, possible race condition (will try again): {}'.format(e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
# standard imports
|
# standard imports
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
|
import time
|
||||||
|
|
||||||
# external imports
|
# external imports
|
||||||
from chainlib.error import RPCException
|
from chainlib.error import RPCException
|
||||||
@@ -10,25 +11,28 @@ from chainqueue.store.fs import (
|
|||||||
IndexStore,
|
IndexStore,
|
||||||
CounterStore,
|
CounterStore,
|
||||||
)
|
)
|
||||||
from chainqueue.error import BackendIntegrityError
|
|
||||||
from shep.store.file import SimpleFileStoreFactory
|
from shep.store.file import SimpleFileStoreFactory
|
||||||
from shep.error import StateInvalid
|
from shep.error import (
|
||||||
|
StateInvalid,
|
||||||
|
StateLockedKey,
|
||||||
|
)
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .base import ChaindAdapter
|
from .base import ChaindAdapter
|
||||||
|
from chaind.lock import StoreLock
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class ChaindFsAdapter(ChaindAdapter):
|
class ChaindFsAdapter(ChaindAdapter):
|
||||||
|
|
||||||
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32, event_callback=None):
|
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32, event_callback=None, store_sync=True):
|
||||||
factory = SimpleFileStoreFactory(path).add
|
factory = SimpleFileStoreFactory(path, use_lock=True).add
|
||||||
state_store = Status(factory, allow_invalid=True, event_callback=event_callback)
|
state_store = Status(factory, allow_invalid=True, event_callback=event_callback)
|
||||||
index_path = os.path.join(path, 'tx')
|
index_path = os.path.join(path, 'tx')
|
||||||
index_store = IndexStore(index_path, digest_bytes=digest_bytes)
|
index_store = IndexStore(index_path, digest_bytes=digest_bytes)
|
||||||
counter_store = CounterStore(path)
|
counter_store = CounterStore(path)
|
||||||
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold)
|
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold, store_sync=store_sync)
|
||||||
|
|
||||||
|
|
||||||
def put(self, signed_tx):
|
def put(self, signed_tx):
|
||||||
@@ -38,21 +42,29 @@ class ChaindFsAdapter(ChaindAdapter):
|
|||||||
|
|
||||||
def get(self, tx_hash):
|
def get(self, tx_hash):
|
||||||
v = None
|
v = None
|
||||||
try:
|
store_lock = StoreLock()
|
||||||
v = self.store.get(tx_hash)
|
while True:
|
||||||
except StateInvalid as e:
|
try:
|
||||||
logg.error('I am just a simple syncer and do not know how to handle the state which the tx {} is in: {}'.format(tx_hash, e))
|
v = self.store.get(tx_hash)
|
||||||
return None
|
break
|
||||||
except FileNotFoundError:
|
except StateInvalid as e:
|
||||||
pass
|
logg.error('I am just a simple syncer and do not know how to handle the state which the tx {} is in: {}'.format(tx_hash, e))
|
||||||
if v ==None:
|
return None
|
||||||
raise BackendIntegrityError(tx_hash)
|
except FileNotFoundError as e:
|
||||||
|
logg.debug('queuestore get (file missing) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
except StateLockedKey as e:
|
||||||
|
logg.debug('queuestore get (statelock) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
|
||||||
return v[1]
|
return v[1]
|
||||||
|
|
||||||
|
|
||||||
def upcoming(self, limit=0):
|
def upcoming(self, limit=0):
|
||||||
real_limit = 0
|
real_limit = 0
|
||||||
in_flight = 0
|
in_flight = []
|
||||||
if limit > 0:
|
if limit > 0:
|
||||||
in_flight = self.store.by_state(state=self.store.IN_NETWORK, not_state=self.store.FINAL)
|
in_flight = self.store.by_state(state=self.store.IN_NETWORK, not_state=self.store.FINAL)
|
||||||
real_limit = limit - len(in_flight)
|
real_limit = limit - len(in_flight)
|
||||||
@@ -78,12 +90,19 @@ class ChaindFsAdapter(ChaindAdapter):
|
|||||||
def succeed(self, block, tx):
|
def succeed(self, block, tx):
|
||||||
if self.store.is_reserved(tx.hash):
|
if self.store.is_reserved(tx.hash):
|
||||||
raise QueueLockError(tx.hash)
|
raise QueueLockError(tx.hash)
|
||||||
|
r = self.store.final(tx.hash, block, tx, error=False)
|
||||||
return self.store.final(tx.hash, block, tx, error=False)
|
(k, v) = self.store.get(tx.hash)
|
||||||
|
self.store.purge(k)
|
||||||
|
return r
|
||||||
|
|
||||||
|
|
||||||
def fail(self, block, tx):
|
def fail(self, block, tx):
|
||||||
return self.store.final(tx.hash, block, tx, error=True)
|
if self.store.is_reserved(tx.hash):
|
||||||
|
raise QueueLockError(tx.hash)
|
||||||
|
r = self.store.final(tx.hash, block, tx, error=True)
|
||||||
|
(k, v) = self.store.get(tx.hash)
|
||||||
|
self.store.purge(k)
|
||||||
|
return r
|
||||||
|
|
||||||
|
|
||||||
def sendfail(self):
|
def sendfail(self):
|
||||||
@@ -95,15 +114,44 @@ class ChaindFsAdapter(ChaindAdapter):
|
|||||||
|
|
||||||
|
|
||||||
def dispatch(self, tx_hash):
|
def dispatch(self, tx_hash):
|
||||||
entry = self.store.send_start(tx_hash)
|
entry = None
|
||||||
|
|
||||||
|
store_lock = StoreLock()
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
entry = self.store.send_start(tx_hash)
|
||||||
|
break
|
||||||
|
except FileNotFoundError as e:
|
||||||
|
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
except StateLockedKey as e:
|
||||||
|
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
|
||||||
tx_wire = entry.serialize()
|
tx_wire = entry.serialize()
|
||||||
|
|
||||||
r = None
|
r = None
|
||||||
try:
|
try:
|
||||||
r = self.dispatcher.send(tx_wire)
|
r = self.dispatcher.send(tx_wire)
|
||||||
except RPCException:
|
except RPCException as e:
|
||||||
|
logg.error('dispatch send failed for {}: {}'.format(tx_hash, e))
|
||||||
self.store.fail(tx_hash)
|
self.store.fail(tx_hash)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
self.store.send_end(tx_hash)
|
store_lock = StoreLock()
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
self.store.send_end(tx_hash)
|
||||||
|
break
|
||||||
|
except FileNotFoundError as e:
|
||||||
|
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||||
|
store_lock.again(e)
|
||||||
|
continue
|
||||||
|
except StateLockedKey as e:
|
||||||
|
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
||||||
|
store_lock.again(e)
|
||||||
|
continue
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|||||||
33
chaind/dispatch.py
Normal file
33
chaind/dispatch.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# standard imports
|
||||||
|
import logging
|
||||||
|
|
||||||
|
# local ipmorts
|
||||||
|
from chaind.adapters.fs import ChaindFsAdapter
|
||||||
|
from chaind.eth.cache import EthCacheTx
|
||||||
|
|
||||||
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class DispatchProcessor:
|
||||||
|
|
||||||
|
def __init__(self, chain_spec, queue_dir, dispatcher):
|
||||||
|
self.dispatcher = dispatcher
|
||||||
|
self.chain_spec = chain_spec,
|
||||||
|
self.queue_dir = queue_dir
|
||||||
|
|
||||||
|
|
||||||
|
def process(self, rpc, limit=50):
|
||||||
|
adapter = ChaindFsAdapter(
|
||||||
|
self.chain_spec,
|
||||||
|
self.queue_dir,
|
||||||
|
EthCacheTx,
|
||||||
|
self.dispatcher,
|
||||||
|
)
|
||||||
|
|
||||||
|
upcoming = adapter.upcoming(limit=limit)
|
||||||
|
logg.info('processor has {} candidates for {}, processing with limit {}'.format(len(upcoming), self.chain_spec, limit))
|
||||||
|
i = 0
|
||||||
|
for tx_hash in upcoming:
|
||||||
|
if adapter.dispatch(tx_hash):
|
||||||
|
i += 1
|
||||||
|
return i
|
||||||
@@ -20,3 +20,7 @@ class ClientInputError(ValueError):
|
|||||||
|
|
||||||
class QueueLockError(Exception):
|
class QueueLockError(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class BackendError(Exception):
|
||||||
|
pass
|
||||||
|
|||||||
@@ -5,22 +5,22 @@ import time
|
|||||||
# external imports
|
# external imports
|
||||||
from chainlib.status import Status as TxStatus
|
from chainlib.status import Status as TxStatus
|
||||||
from chainsyncer.filter import SyncFilter
|
from chainsyncer.filter import SyncFilter
|
||||||
from chainqueue.error import (
|
from chainqueue.error import NotLocalTxError
|
||||||
NotLocalTxError,
|
|
||||||
BackendIntegrityError,
|
|
||||||
)
|
|
||||||
from chaind.adapters.fs import ChaindFsAdapter
|
from chaind.adapters.fs import ChaindFsAdapter
|
||||||
|
from shep.error import StateLockedKey
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .error import QueueLockError
|
from .error import (
|
||||||
|
QueueLockError,
|
||||||
|
BackendError,
|
||||||
|
)
|
||||||
|
from chaind.lock import StoreLock
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class StateFilter(SyncFilter):
|
class StateFilter(SyncFilter):
|
||||||
|
|
||||||
delay_limit = 3.0
|
|
||||||
|
|
||||||
def __init__(self, chain_spec, adapter_path, tx_adapter, throttler=None):
|
def __init__(self, chain_spec, adapter_path, tx_adapter, throttler=None):
|
||||||
self.chain_spec = chain_spec
|
self.chain_spec = chain_spec
|
||||||
self.adapter_path = adapter_path
|
self.adapter_path = adapter_path
|
||||||
@@ -30,8 +30,9 @@ class StateFilter(SyncFilter):
|
|||||||
|
|
||||||
def filter(self, conn, block, tx, session=None):
|
def filter(self, conn, block, tx, session=None):
|
||||||
cache_tx = None
|
cache_tx = None
|
||||||
for i in range(3):
|
store_lock = StoreLock()
|
||||||
queue_adapter = None
|
queue_adapter = None
|
||||||
|
while True:
|
||||||
try:
|
try:
|
||||||
queue_adapter = ChaindFsAdapter(
|
queue_adapter = ChaindFsAdapter(
|
||||||
self.chain_spec,
|
self.chain_spec,
|
||||||
@@ -39,28 +40,31 @@ class StateFilter(SyncFilter):
|
|||||||
self.tx_adapter,
|
self.tx_adapter,
|
||||||
None,
|
None,
|
||||||
)
|
)
|
||||||
except BackendIntegrityError as e:
|
except BackendError as e:
|
||||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
||||||
|
store_lock.again()
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
store_lock.reset()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cache_tx = queue_adapter.get(tx.hash)
|
cache_tx = queue_adapter.get(tx.hash)
|
||||||
|
break
|
||||||
except NotLocalTxError:
|
except NotLocalTxError:
|
||||||
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
||||||
return False
|
return False
|
||||||
except BackendIntegrityError as e:
|
except BackendError as e:
|
||||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
logg.error('adapter get failed: {}, one more try'.format(e))
|
||||||
|
queue_adapter = None
|
||||||
|
store_lock.again()
|
||||||
continue
|
continue
|
||||||
|
|
||||||
break
|
|
||||||
|
|
||||||
if cache_tx == None:
|
if cache_tx == None:
|
||||||
raise NotLocalTxError(tx.hash)
|
raise NotLocalTxError(tx.hash)
|
||||||
|
|
||||||
delay = 0.01
|
store_lock = StoreLock()
|
||||||
|
queue_lock = StoreLock(error=QueueLockError)
|
||||||
while True:
|
while True:
|
||||||
if delay > self.delay_limit:
|
|
||||||
raise QueueLockError('The queue lock for tx {} seems to be stuck. Human meddling needed.'.format(tx.hash))
|
|
||||||
try:
|
try:
|
||||||
if tx.status == TxStatus.SUCCESS:
|
if tx.status == TxStatus.SUCCESS:
|
||||||
queue_adapter.succeed(block, tx)
|
queue_adapter.succeed(block, tx)
|
||||||
@@ -69,8 +73,21 @@ class StateFilter(SyncFilter):
|
|||||||
break
|
break
|
||||||
except QueueLockError as e:
|
except QueueLockError as e:
|
||||||
logg.debug('queue item {} is blocked, will retry: {}'.format(tx.hash, e))
|
logg.debug('queue item {} is blocked, will retry: {}'.format(tx.hash, e))
|
||||||
time.sleep(delay)
|
queue_lock.again()
|
||||||
delay *= 2
|
except FileNotFoundError as e:
|
||||||
|
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
except NotLocalTxError as e:
|
||||||
|
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
except StateLockedKey as e:
|
||||||
|
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
||||||
|
store_lock.again()
|
||||||
|
continue
|
||||||
|
|
||||||
|
logg.info('filter registered {} for {} in {}'.format(tx.status.name, tx.hash, block))
|
||||||
|
|
||||||
if self.throttler != None:
|
if self.throttler != None:
|
||||||
self.throttler.dec(tx.hash)
|
self.throttler.dec(tx.hash)
|
||||||
|
|||||||
34
chaind/lock.py
Normal file
34
chaind/lock.py
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
# standard imports
|
||||||
|
import time
|
||||||
|
|
||||||
|
# local imports
|
||||||
|
from .error import BackendError
|
||||||
|
|
||||||
|
BASE_DELAY = 0.01
|
||||||
|
BASE_DELAY_LIMIT = 10.0
|
||||||
|
|
||||||
|
|
||||||
|
class StoreLock:
|
||||||
|
|
||||||
|
def __init__(self, delay=BASE_DELAY, delay_limit=BASE_DELAY_LIMIT, error=BackendError, description=None):
|
||||||
|
self.base_delay = delay
|
||||||
|
self.delay = delay
|
||||||
|
self.delay_limit = delay_limit
|
||||||
|
self.error = error
|
||||||
|
self.description = description
|
||||||
|
|
||||||
|
|
||||||
|
def again(self, e=None):
|
||||||
|
if self.delay > self.delay_limit:
|
||||||
|
err = None
|
||||||
|
if e != None:
|
||||||
|
err = str(e)
|
||||||
|
else:
|
||||||
|
err = self.description
|
||||||
|
raise self.error(err)
|
||||||
|
time.sleep(self.delay)
|
||||||
|
self.delay *= 2
|
||||||
|
|
||||||
|
|
||||||
|
def reset(self):
|
||||||
|
self.delay = self.base_delay
|
||||||
@@ -8,19 +8,21 @@ import stat
|
|||||||
from hexathon import strip_0x
|
from hexathon import strip_0x
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from chaind.error import (
|
from .error import (
|
||||||
NothingToDoError,
|
NothingToDoError,
|
||||||
ClientGoneError,
|
ClientGoneError,
|
||||||
ClientBlockError,
|
ClientBlockError,
|
||||||
ClientInputError,
|
ClientInputError,
|
||||||
)
|
)
|
||||||
|
from .lock import StoreLock
|
||||||
|
from .error import BackendError
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class SessionController:
|
class SessionController:
|
||||||
|
|
||||||
def __init__(self, config, adapter, processor):
|
def __init__(self, config, processor):
|
||||||
self.dead = False
|
self.dead = False
|
||||||
os.makedirs(os.path.dirname(config.get('SESSION_SOCKET_PATH')), exist_ok=True)
|
os.makedirs(os.path.dirname(config.get('SESSION_SOCKET_PATH')), exist_ok=True)
|
||||||
try:
|
try:
|
||||||
@@ -35,7 +37,6 @@ class SessionController:
|
|||||||
self.srv.settimeout(float(config.get('SESSION_DISPATCH_DELAY')))
|
self.srv.settimeout(float(config.get('SESSION_DISPATCH_DELAY')))
|
||||||
self.processor = processor
|
self.processor = processor
|
||||||
self.chain_spec = config.get('CHAIN_SPEC')
|
self.chain_spec = config.get('CHAIN_SPEC')
|
||||||
self.adapter = adapter
|
|
||||||
|
|
||||||
|
|
||||||
def shutdown(self, signo, frame):
|
def shutdown(self, signo, frame):
|
||||||
@@ -59,7 +60,16 @@ class SessionController:
|
|||||||
|
|
||||||
|
|
||||||
def process(self, conn):
|
def process(self, conn):
|
||||||
r = self.processor(self.chain_spec, self.adapter, conn)
|
state_lock = StoreLock()
|
||||||
|
r = None
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
r = self.processor(conn)
|
||||||
|
break
|
||||||
|
except BackendError as e:
|
||||||
|
state_lock.again(e)
|
||||||
|
continue
|
||||||
|
|
||||||
if r > 0:
|
if r > 0:
|
||||||
self.srv.settimeout(0.1)
|
self.srv.settimeout(0.1)
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -29,8 +29,6 @@ class MockDispatcher:
|
|||||||
|
|
||||||
|
|
||||||
def send(self, v):
|
def send(self, v):
|
||||||
import sys
|
|
||||||
sys.stderr.write('susu v {} {}\n'.format(v, self.fails))
|
|
||||||
if v in self.fails:
|
if v in self.fails:
|
||||||
raise RPCException('{} is in fails'.format(v))
|
raise RPCException('{} is in fails'.format(v))
|
||||||
pass
|
pass
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
chainlib~=0.1.1
|
chainlib~=0.1.2
|
||||||
chainqueue~=0.1.6
|
chainqueue~=0.1.12
|
||||||
chainsyncer~=0.4.3
|
chainsyncer~=0.4.4
|
||||||
confini~=0.6.0
|
confini~=0.6.0
|
||||||
funga~=0.5.2
|
funga~=0.5.2
|
||||||
pyxdg~=0.26
|
pyxdg~=0.26
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
[metadata]
|
[metadata]
|
||||||
name = chaind
|
name = chaind
|
||||||
version = 0.2.0
|
version = 0.2.8
|
||||||
description = Base package for chain queue service
|
description = Base package for chain queue service
|
||||||
author = Louis Holbrook
|
author = Louis Holbrook
|
||||||
author_email = dev@holbrook.no
|
author_email = dev@holbrook.no
|
||||||
|
|||||||
@@ -74,7 +74,7 @@ class TestChaindFs(TestChaindFsBase):
|
|||||||
data = os.urandom(128).hex()
|
data = os.urandom(128).hex()
|
||||||
hsh = self.adapter.put(data)
|
hsh = self.adapter.put(data)
|
||||||
|
|
||||||
fltr = StateFilter(self.adapter)
|
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||||
tx = MockTx(hsh)
|
tx = MockTx(hsh)
|
||||||
fltr.filter(None, None, tx)
|
fltr.filter(None, None, tx)
|
||||||
|
|
||||||
@@ -85,7 +85,7 @@ class TestChaindFs(TestChaindFsBase):
|
|||||||
data = os.urandom(128).hex()
|
data = os.urandom(128).hex()
|
||||||
hsh = self.adapter.put(data)
|
hsh = self.adapter.put(data)
|
||||||
|
|
||||||
fltr = StateFilter(self.adapter)
|
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
||||||
tx = MockTx(hsh, TxStatus.ERROR)
|
tx = MockTx(hsh, TxStatus.ERROR)
|
||||||
fltr.filter(None, None, tx)
|
fltr.filter(None, None, tx)
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user