Compare commits
No commits in common. "master" and "lash/chainqueue-shep" have entirely different histories.
master
...
lash/chain
32
CHANGELOG
32
CHANGELOG
@ -1,35 +1,3 @@
|
|||||||
- 0.3.1
|
|
||||||
* Change license to AGPL3 and copyright waived to public domain
|
|
||||||
- 0.3.0
|
|
||||||
* Implement on chainlib 0.3.0
|
|
||||||
- 0.2.12
|
|
||||||
* Breaking upgrade of chainlib.
|
|
||||||
* Implement generic block and tx.
|
|
||||||
- 0.2.11
|
|
||||||
* Upgrade shep to handle exception in filestore list
|
|
||||||
- 0.2.10
|
|
||||||
* Upgrade shep to guarantee state lock atomicity
|
|
||||||
- 0.2.9
|
|
||||||
* Minimize instantiations of adapters in filter execution
|
|
||||||
- 0.2.8
|
|
||||||
* Upgrade chainsyncer
|
|
||||||
- 0.2.7
|
|
||||||
* Upgrade chainlib
|
|
||||||
- 0.2.6
|
|
||||||
* Deps upgrade
|
|
||||||
- 0.2.5
|
|
||||||
* Deps upgrade
|
|
||||||
- 0.2.4
|
|
||||||
* Allow omission of state store sync in queue store backend
|
|
||||||
- 0.2.2
|
|
||||||
* Fix missing symbol crashes related to race conditions
|
|
||||||
- 0.2.1
|
|
||||||
* Receive removed race checks from chainqueue
|
|
||||||
- 0.2.0
|
|
||||||
* primitive race condition handling between fs access of sync and queue
|
|
||||||
* re-enable throttling based on in-flight transaction count
|
|
||||||
- 0.1.2
|
|
||||||
* add settings object
|
|
||||||
- 0.1.0
|
- 0.1.0
|
||||||
* consume non chain-specific code
|
* consume non chain-specific code
|
||||||
- 0.0.1
|
- 0.0.1
|
||||||
|
141
LICENSE
141
LICENSE
@ -1,5 +1,5 @@
|
|||||||
GNU AFFERO GENERAL PUBLIC LICENSE
|
GNU GENERAL PUBLIC LICENSE
|
||||||
Version 3, 19 November 2007
|
Version 3, 29 June 2007
|
||||||
|
|
||||||
Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/>
|
Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/>
|
||||||
Everyone is permitted to copy and distribute verbatim copies
|
Everyone is permitted to copy and distribute verbatim copies
|
||||||
@ -7,15 +7,17 @@
|
|||||||
|
|
||||||
Preamble
|
Preamble
|
||||||
|
|
||||||
The GNU Affero General Public License is a free, copyleft license for
|
The GNU General Public License is a free, copyleft license for
|
||||||
software and other kinds of works, specifically designed to ensure
|
software and other kinds of works.
|
||||||
cooperation with the community in the case of network server software.
|
|
||||||
|
|
||||||
The licenses for most software and other practical works are designed
|
The licenses for most software and other practical works are designed
|
||||||
to take away your freedom to share and change the works. By contrast,
|
to take away your freedom to share and change the works. By contrast,
|
||||||
our General Public Licenses are intended to guarantee your freedom to
|
the GNU General Public License is intended to guarantee your freedom to
|
||||||
share and change all versions of a program--to make sure it remains free
|
share and change all versions of a program--to make sure it remains free
|
||||||
software for all its users.
|
software for all its users. We, the Free Software Foundation, use the
|
||||||
|
GNU General Public License for most of our software; it applies also to
|
||||||
|
any other work released this way by its authors. You can apply it to
|
||||||
|
your programs, too.
|
||||||
|
|
||||||
When we speak of free software, we are referring to freedom, not
|
When we speak of free software, we are referring to freedom, not
|
||||||
price. Our General Public Licenses are designed to make sure that you
|
price. Our General Public Licenses are designed to make sure that you
|
||||||
@ -24,34 +26,44 @@ them if you wish), that you receive source code or can get it if you
|
|||||||
want it, that you can change the software or use pieces of it in new
|
want it, that you can change the software or use pieces of it in new
|
||||||
free programs, and that you know you can do these things.
|
free programs, and that you know you can do these things.
|
||||||
|
|
||||||
Developers that use our General Public Licenses protect your rights
|
To protect your rights, we need to prevent others from denying you
|
||||||
with two steps: (1) assert copyright on the software, and (2) offer
|
these rights or asking you to surrender the rights. Therefore, you have
|
||||||
you this License which gives you legal permission to copy, distribute
|
certain responsibilities if you distribute copies of the software, or if
|
||||||
and/or modify the software.
|
you modify it: responsibilities to respect the freedom of others.
|
||||||
|
|
||||||
A secondary benefit of defending all users' freedom is that
|
For example, if you distribute copies of such a program, whether
|
||||||
improvements made in alternate versions of the program, if they
|
gratis or for a fee, you must pass on to the recipients the same
|
||||||
receive widespread use, become available for other developers to
|
freedoms that you received. You must make sure that they, too, receive
|
||||||
incorporate. Many developers of free software are heartened and
|
or can get the source code. And you must show them these terms so they
|
||||||
encouraged by the resulting cooperation. However, in the case of
|
know their rights.
|
||||||
software used on network servers, this result may fail to come about.
|
|
||||||
The GNU General Public License permits making a modified version and
|
|
||||||
letting the public access it on a server without ever releasing its
|
|
||||||
source code to the public.
|
|
||||||
|
|
||||||
The GNU Affero General Public License is designed specifically to
|
Developers that use the GNU GPL protect your rights with two steps:
|
||||||
ensure that, in such cases, the modified source code becomes available
|
(1) assert copyright on the software, and (2) offer you this License
|
||||||
to the community. It requires the operator of a network server to
|
giving you legal permission to copy, distribute and/or modify it.
|
||||||
provide the source code of the modified version running there to the
|
|
||||||
users of that server. Therefore, public use of a modified version, on
|
|
||||||
a publicly accessible server, gives the public access to the source
|
|
||||||
code of the modified version.
|
|
||||||
|
|
||||||
An older license, called the Affero General Public License and
|
For the developers' and authors' protection, the GPL clearly explains
|
||||||
published by Affero, was designed to accomplish similar goals. This is
|
that there is no warranty for this free software. For both users' and
|
||||||
a different license, not a version of the Affero GPL, but Affero has
|
authors' sake, the GPL requires that modified versions be marked as
|
||||||
released a new version of the Affero GPL which permits relicensing under
|
changed, so that their problems will not be attributed erroneously to
|
||||||
this license.
|
authors of previous versions.
|
||||||
|
|
||||||
|
Some devices are designed to deny users access to install or run
|
||||||
|
modified versions of the software inside them, although the manufacturer
|
||||||
|
can do so. This is fundamentally incompatible with the aim of
|
||||||
|
protecting users' freedom to change the software. The systematic
|
||||||
|
pattern of such abuse occurs in the area of products for individuals to
|
||||||
|
use, which is precisely where it is most unacceptable. Therefore, we
|
||||||
|
have designed this version of the GPL to prohibit the practice for those
|
||||||
|
products. If such problems arise substantially in other domains, we
|
||||||
|
stand ready to extend this provision to those domains in future versions
|
||||||
|
of the GPL, as needed to protect the freedom of users.
|
||||||
|
|
||||||
|
Finally, every program is threatened constantly by software patents.
|
||||||
|
States should not allow patents to restrict development and use of
|
||||||
|
software on general-purpose computers, but in those that do, we wish to
|
||||||
|
avoid the special danger that patents applied to a free program could
|
||||||
|
make it effectively proprietary. To prevent this, the GPL assures that
|
||||||
|
patents cannot be used to render the program non-free.
|
||||||
|
|
||||||
The precise terms and conditions for copying, distribution and
|
The precise terms and conditions for copying, distribution and
|
||||||
modification follow.
|
modification follow.
|
||||||
@ -60,7 +72,7 @@ modification follow.
|
|||||||
|
|
||||||
0. Definitions.
|
0. Definitions.
|
||||||
|
|
||||||
"This License" refers to version 3 of the GNU Affero General Public License.
|
"This License" refers to version 3 of the GNU General Public License.
|
||||||
|
|
||||||
"Copyright" also means copyright-like laws that apply to other kinds of
|
"Copyright" also means copyright-like laws that apply to other kinds of
|
||||||
works, such as semiconductor masks.
|
works, such as semiconductor masks.
|
||||||
@ -537,45 +549,35 @@ to collect a royalty for further conveying from those to whom you convey
|
|||||||
the Program, the only way you could satisfy both those terms and this
|
the Program, the only way you could satisfy both those terms and this
|
||||||
License would be to refrain entirely from conveying the Program.
|
License would be to refrain entirely from conveying the Program.
|
||||||
|
|
||||||
13. Remote Network Interaction; Use with the GNU General Public License.
|
13. Use with the GNU Affero General Public License.
|
||||||
|
|
||||||
Notwithstanding any other provision of this License, if you modify the
|
|
||||||
Program, your modified version must prominently offer all users
|
|
||||||
interacting with it remotely through a computer network (if your version
|
|
||||||
supports such interaction) an opportunity to receive the Corresponding
|
|
||||||
Source of your version by providing access to the Corresponding Source
|
|
||||||
from a network server at no charge, through some standard or customary
|
|
||||||
means of facilitating copying of software. This Corresponding Source
|
|
||||||
shall include the Corresponding Source for any work covered by version 3
|
|
||||||
of the GNU General Public License that is incorporated pursuant to the
|
|
||||||
following paragraph.
|
|
||||||
|
|
||||||
Notwithstanding any other provision of this License, you have
|
Notwithstanding any other provision of this License, you have
|
||||||
permission to link or combine any covered work with a work licensed
|
permission to link or combine any covered work with a work licensed
|
||||||
under version 3 of the GNU General Public License into a single
|
under version 3 of the GNU Affero General Public License into a single
|
||||||
combined work, and to convey the resulting work. The terms of this
|
combined work, and to convey the resulting work. The terms of this
|
||||||
License will continue to apply to the part which is the covered work,
|
License will continue to apply to the part which is the covered work,
|
||||||
but the work with which it is combined will remain governed by version
|
but the special requirements of the GNU Affero General Public License,
|
||||||
3 of the GNU General Public License.
|
section 13, concerning interaction through a network will apply to the
|
||||||
|
combination as such.
|
||||||
|
|
||||||
14. Revised Versions of this License.
|
14. Revised Versions of this License.
|
||||||
|
|
||||||
The Free Software Foundation may publish revised and/or new versions of
|
The Free Software Foundation may publish revised and/or new versions of
|
||||||
the GNU Affero General Public License from time to time. Such new versions
|
the GNU General Public License from time to time. Such new versions will
|
||||||
will be similar in spirit to the present version, but may differ in detail to
|
be similar in spirit to the present version, but may differ in detail to
|
||||||
address new problems or concerns.
|
address new problems or concerns.
|
||||||
|
|
||||||
Each version is given a distinguishing version number. If the
|
Each version is given a distinguishing version number. If the
|
||||||
Program specifies that a certain numbered version of the GNU Affero General
|
Program specifies that a certain numbered version of the GNU General
|
||||||
Public License "or any later version" applies to it, you have the
|
Public License "or any later version" applies to it, you have the
|
||||||
option of following the terms and conditions either of that numbered
|
option of following the terms and conditions either of that numbered
|
||||||
version or of any later version published by the Free Software
|
version or of any later version published by the Free Software
|
||||||
Foundation. If the Program does not specify a version number of the
|
Foundation. If the Program does not specify a version number of the
|
||||||
GNU Affero General Public License, you may choose any version ever published
|
GNU General Public License, you may choose any version ever published
|
||||||
by the Free Software Foundation.
|
by the Free Software Foundation.
|
||||||
|
|
||||||
If the Program specifies that a proxy can decide which future
|
If the Program specifies that a proxy can decide which future
|
||||||
versions of the GNU Affero General Public License can be used, that proxy's
|
versions of the GNU General Public License can be used, that proxy's
|
||||||
public statement of acceptance of a version permanently authorizes you
|
public statement of acceptance of a version permanently authorizes you
|
||||||
to choose that version for the Program.
|
to choose that version for the Program.
|
||||||
|
|
||||||
@ -633,29 +635,40 @@ the "copyright" line and a pointer to where the full notice is found.
|
|||||||
Copyright (C) <year> <name of author>
|
Copyright (C) <year> <name of author>
|
||||||
|
|
||||||
This program is free software: you can redistribute it and/or modify
|
This program is free software: you can redistribute it and/or modify
|
||||||
it under the terms of the GNU Affero General Public License as published by
|
it under the terms of the GNU General Public License as published by
|
||||||
the Free Software Foundation, either version 3 of the License, or
|
the Free Software Foundation, either version 3 of the License, or
|
||||||
(at your option) any later version.
|
(at your option) any later version.
|
||||||
|
|
||||||
This program is distributed in the hope that it will be useful,
|
This program is distributed in the hope that it will be useful,
|
||||||
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
GNU Affero General Public License for more details.
|
GNU General Public License for more details.
|
||||||
|
|
||||||
You should have received a copy of the GNU Affero General Public License
|
You should have received a copy of the GNU General Public License
|
||||||
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
Also add information on how to contact you by electronic and paper mail.
|
Also add information on how to contact you by electronic and paper mail.
|
||||||
|
|
||||||
If your software can interact with users remotely through a computer
|
If the program does terminal interaction, make it output a short
|
||||||
network, you should also make sure that it provides a way for users to
|
notice like this when it starts in an interactive mode:
|
||||||
get its source. For example, if your program is a web application, its
|
|
||||||
interface could display a "Source" link that leads users to an archive
|
<program> Copyright (C) <year> <name of author>
|
||||||
of the code. There are many ways you could offer source, and different
|
This program comes with ABSOLUTELY NO WARRANTY; for details type `show w'.
|
||||||
solutions will be better for different programs; see section 13 for the
|
This is free software, and you are welcome to redistribute it
|
||||||
specific requirements.
|
under certain conditions; type `show c' for details.
|
||||||
|
|
||||||
|
The hypothetical commands `show w' and `show c' should show the appropriate
|
||||||
|
parts of the General Public License. Of course, your program's commands
|
||||||
|
might be different; for a GUI interface, you would use an "about box".
|
||||||
|
|
||||||
You should also get your employer (if you work as a programmer) or school,
|
You should also get your employer (if you work as a programmer) or school,
|
||||||
if any, to sign a "copyright disclaimer" for the program, if necessary.
|
if any, to sign a "copyright disclaimer" for the program, if necessary.
|
||||||
For more information on this, and how to apply and follow the GNU AGPL, see
|
For more information on this, and how to apply and follow the GNU GPL, see
|
||||||
<https://www.gnu.org/licenses/>.
|
<https://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
The GNU General Public License does not permit incorporating your program
|
||||||
|
into proprietary programs. If your program is a subroutine library, you
|
||||||
|
may consider it more useful to permit linking proprietary applications with
|
||||||
|
the library. If this is what you want to do, use the GNU Lesser General
|
||||||
|
Public License instead of this License. But first, please read
|
||||||
|
<https://www.gnu.org/licenses/why-not-lgpl.html>.
|
||||||
|
@ -1 +1 @@
|
|||||||
include *requirements.txt LICENSE WAIVER WAIVER.asc CHANGELOG chaind/data/config/* chaind/data/config/syncer/*
|
include *requirements.txt chaind/data/config/* chaind/data/config/syncer/*
|
||||||
|
17
WAIVER
17
WAIVER
@ -1,17 +0,0 @@
|
|||||||
# Copyright waiver for the python package "chaind"
|
|
||||||
|
|
||||||
I dedicate any and all copyright interest in this software to the
|
|
||||||
public domain. I make this dedication for the benefit of the public at
|
|
||||||
large and to the detriment of my heirs and successors. I intend this
|
|
||||||
dedication to be an overt act of relinquishment in perpetuity of all
|
|
||||||
present and future rights to this software under copyright law.
|
|
||||||
|
|
||||||
To the best of my knowledge and belief, my contributions are either
|
|
||||||
originally authored by me or are derived from prior works which I have
|
|
||||||
verified are also in the public domain and are not subject to claims
|
|
||||||
of copyright by other parties.
|
|
||||||
|
|
||||||
To the best of my knowledge and belief, no individual, business,
|
|
||||||
organization, government, or other entity has any copyright interest
|
|
||||||
in my contributions, and I affirm that I will not make contributions
|
|
||||||
that are otherwise encumbered.
|
|
28
WAIVER.asc
28
WAIVER.asc
@ -1,28 +0,0 @@
|
|||||||
-----BEGIN PGP MESSAGE-----
|
|
||||||
|
|
||||||
owGVU2tQVFUcX0BkvY6jTQFjBJ4FA2tWgl6YyZLD8BIFiYeQOXT37lnuifviPnbd
|
|
||||||
AkIweRnhCAbpEDBWyHNsoJkm0QYICpTZScIv4IAPHlEqZTU0gp27C2r1qQ87s/ee
|
|
||||||
///3PLdynZtG66L7Yfrrkp7Po13a3Djj6n0749KiXqdybl8KAJG8YBNRFi0DK4ks
|
|
||||||
UARmXgQyDYFgk2meAwJJZZNZEPhTNIk4kz9BxAETNCGKlCEgORv+mQDJMIB6AIQ4
|
|
||||||
GYpQUv9gJCQBiTfLVlKEQOZVaEJQjAyigIlnMWQwiAMsmQ2do8vQCDOvCDFCDpqR
|
|
||||||
DHizU5dzmZQJhhSzoIPfiYuXZRGxkHPMsjZAQyRKjgFJoSgoSbwoqXSqQHULExKP
|
|
||||||
EGIUo4oHeJyDDEjKgSNCBnE5CpJoBzL2JEBRgLKCZJt6jr0TArarHqpUZkVWsFVH
|
|
||||||
FJJT2aMRKJwJh/wwLIa0BhNECr9sVVrRns3xVgaalg0asQho1qsHFM9hl0ZFlYzN
|
|
||||||
YUiI8K5I8BgQcVgO7kTB3YkQ79kACwEOUp3DxLhhrFDkWSCICL+28mK2BKw0omic
|
|
||||||
C01aIIG9IzPCY+oKyUi8s8YHwTtbc9aOJzhexuka34Y4LWyWYkjESgT28NAiFsGr
|
|
||||||
CvFdEmUEpf/hl1PZTciCTArJ6IFRkRCHe9Rjs1kkh95xFKcHWWpjnNqPXjXrZMNP
|
|
||||||
akU0KTnu6X/vJ4Ft/DtQvYM+DpBmMxJZLJKU8ZMV4QuuOnVc1H8sEI4RNQkHqxVJ
|
|
||||||
uBCOUlgj5jAFE6WuL6zSuGg1vl66VdrAXxqErX7F/deKk1a+SHdX9TvUEGs2rLz5
|
|
||||||
8sT6++kJ1+OPLe74jGlMjj5M+Obp7L+3921dXaWZ+Nk7sN2jubh1Ls/65kD5YCIb
|
|
||||||
USRMCEc2kImvFBYacgOT6wa1jQHD9ry5xYqhvu+fGR5dd4jtqk+3HCBTH6ub3p17
|
|
||||||
cMvuK4XdMaGV8wl0WpR7fG3DpwtTZ5VWu2z5KPPdn5gztye3HzxRWlrt1x4WuQf2
|
|
||||||
pUy2E6H+sSdv2OHxXV0ex9tCaWlpku5cmz9U00nQixUdcWlzpfaRJ+WWWo/W87MX
|
|
||||||
N22M6j2WEX0+vSCiXvx4gyE/dn/NtpHMsjtrqmPGerq2B53y/OTH2VsGw+jpirVF
|
|
||||||
C6476QmXHZqae16VS2dto/xTTQfSzK9e0Dz3wRSXOfINsfelsj/aky2DxTfOlX8X
|
|
||||||
YprouzCli6mN+O3Zy8arh2D9E/d3hRf4Td1L6Viafc0eftJt/nSnT0vWW3HMuSt3
|
|
||||||
iyuo8Nw/xzx9P2wzzle5N5fHdGi/8Aj71hIemxLi4v3r0eTJJK/h1CbLzaKBW2Tr
|
|
||||||
V4szM1pD0+XenIik2fUNOqH5TC/dFTXSX7kxhKpyObLp5Yp9wYk+qfnTGd3XDFzH
|
|
||||||
4Za/xsf33DW835bSe9Q7drNWDHtxf+1A/RsLngWFmrSJgchTnmU943XRNwu6rzam
|
|
||||||
PG5IzEhYuFMyu6Vy296gGh82ngvo181cD4p+enNDbmGJUG26NDT8fMbF98b+Bg==
|
|
||||||
=vZcN
|
|
||||||
-----END PGP MESSAGE-----
|
|
@ -1,27 +1,10 @@
|
|||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
import time
|
|
||||||
|
|
||||||
# external imports
|
# external imports
|
||||||
from chainqueue import Store as QueueStore
|
from chainqueue import Store as QueueStore
|
||||||
|
|
||||||
# local imports
|
|
||||||
from chaind.lock import StoreLock
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class ChaindAdapter:
|
class ChaindAdapter:
|
||||||
|
|
||||||
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, store_sync=True):
|
def __init__(self, chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0):
|
||||||
self.cache_adapter = cache_adapter
|
self.cache_adapter = cache_adapter
|
||||||
self.dispatcher = dispatcher
|
self.dispatcher = dispatcher
|
||||||
store_lock = StoreLock()
|
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache)
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
self.store = QueueStore(chain_spec, state_store, index_store, counter_store, cache=cache, sync=store_sync)
|
|
||||||
break
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
logg.debug('queuestore instantiation failed, possible race condition (will try again): {}'.format(e))
|
|
||||||
store_lock.again()
|
|
||||||
continue
|
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
# standard imports
|
# standard imports
|
||||||
import logging
|
import logging
|
||||||
import os
|
|
||||||
import time
|
|
||||||
|
|
||||||
# external imports
|
# external imports
|
||||||
from chainlib.error import RPCException
|
from chainlib.error import RPCException
|
||||||
@ -12,67 +10,36 @@ from chainqueue.store.fs import (
|
|||||||
CounterStore,
|
CounterStore,
|
||||||
)
|
)
|
||||||
from shep.store.file import SimpleFileStoreFactory
|
from shep.store.file import SimpleFileStoreFactory
|
||||||
from shep.error import (
|
|
||||||
StateInvalid,
|
|
||||||
StateLockedKey,
|
|
||||||
)
|
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .base import ChaindAdapter
|
from .base import ChaindAdapter
|
||||||
from chaind.lock import StoreLock
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class ChaindFsAdapter(ChaindAdapter):
|
class ChaindFsAdapter(ChaindAdapter):
|
||||||
|
|
||||||
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32, event_callback=None, store_sync=True):
|
def __init__(self, chain_spec, path, cache_adapter, dispatcher, cache=None, pending_retry_threshold=0, error_retry_threshold=0, digest_bytes=32):
|
||||||
factory = SimpleFileStoreFactory(path, use_lock=True).add
|
factory = SimpleFileStoreFactory(path).add
|
||||||
state_store = Status(factory, allow_invalid=True, event_callback=event_callback)
|
state_store = Status(factory)
|
||||||
index_path = os.path.join(path, 'tx')
|
index_store = IndexStore(path, digest_bytes=digest_bytes)
|
||||||
index_store = IndexStore(index_path, digest_bytes=digest_bytes)
|
|
||||||
counter_store = CounterStore(path)
|
counter_store = CounterStore(path)
|
||||||
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold, store_sync=store_sync)
|
super(ChaindFsAdapter, self).__init__(chain_spec, state_store, index_store, counter_store, cache_adapter, dispatcher, cache=cache, pending_retry_threshold=pending_retry_threshold, error_retry_threshold=error_retry_threshold)
|
||||||
|
|
||||||
|
|
||||||
def put(self, signed_tx):
|
def put(self, signed_tx):
|
||||||
|
#cache_tx = self.deserialize(signed_tx)
|
||||||
(s, tx_hash,) = self.store.put(signed_tx, cache_adapter=self.cache_adapter)
|
(s, tx_hash,) = self.store.put(signed_tx, cache_adapter=self.cache_adapter)
|
||||||
return tx_hash
|
return tx_hash
|
||||||
|
|
||||||
|
|
||||||
def get(self, tx_hash):
|
def get(self, tx_hash):
|
||||||
v = None
|
|
||||||
store_lock = StoreLock()
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
v = self.store.get(tx_hash)
|
v = self.store.get(tx_hash)
|
||||||
break
|
|
||||||
except StateInvalid as e:
|
|
||||||
logg.error('I am just a simple syncer and do not know how to handle the state which the tx {} is in: {}'.format(tx_hash, e))
|
|
||||||
return None
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
logg.debug('queuestore get (file missing) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
|
||||||
store_lock.again()
|
|
||||||
continue
|
|
||||||
except StateLockedKey as e:
|
|
||||||
logg.debug('queuestore get (statelock) {} failed, possible race condition (will try again): {}'.format(tx_hash, e))
|
|
||||||
store_lock.again()
|
|
||||||
continue
|
|
||||||
|
|
||||||
return v[1]
|
return v[1]
|
||||||
|
|
||||||
|
|
||||||
def upcoming(self, limit=0):
|
def upcoming(self):
|
||||||
real_limit = 0
|
return self.store.upcoming()
|
||||||
in_flight = []
|
|
||||||
if limit > 0:
|
|
||||||
in_flight = self.store.by_state(state=self.store.IN_NETWORK, not_state=self.store.FINAL)
|
|
||||||
real_limit = limit - len(in_flight)
|
|
||||||
if real_limit <= 0:
|
|
||||||
return []
|
|
||||||
r = self.store.upcoming(limit=real_limit)
|
|
||||||
logg.info('upcoming returning {} upcoming from limit {} less {} active in-flight txs'.format(len(r), limit, len(in_flight)))
|
|
||||||
return r
|
|
||||||
|
|
||||||
|
|
||||||
def pending(self):
|
def pending(self):
|
||||||
@ -83,30 +50,12 @@ class ChaindFsAdapter(ChaindAdapter):
|
|||||||
return self.store.deferred()
|
return self.store.deferred()
|
||||||
|
|
||||||
|
|
||||||
def failed(self):
|
|
||||||
return self.store.failed()
|
|
||||||
|
|
||||||
|
|
||||||
def succeed(self, block, tx):
|
def succeed(self, block, tx):
|
||||||
if self.store.is_reserved(tx.hash):
|
return self.store.final(tx.hash, block, tx, error=False)
|
||||||
raise QueueLockError(tx.hash)
|
|
||||||
r = self.store.final(tx.hash, block, tx, error=False)
|
|
||||||
(k, v) = self.store.get(tx.hash)
|
|
||||||
self.store.purge(k)
|
|
||||||
return r
|
|
||||||
|
|
||||||
|
|
||||||
def fail(self, block, tx):
|
def fail(self, block, tx):
|
||||||
if self.store.is_reserved(tx.hash):
|
return self.store.final(tx.hash, block, tx, error=True)
|
||||||
raise QueueLockError(tx.hash)
|
|
||||||
r = self.store.final(tx.hash, block, tx, error=True)
|
|
||||||
(k, v) = self.store.get(tx.hash)
|
|
||||||
self.store.purge(k)
|
|
||||||
return r
|
|
||||||
|
|
||||||
|
|
||||||
def sendfail(self):
|
|
||||||
return self.store.fail(tx.hash)
|
|
||||||
|
|
||||||
|
|
||||||
def enqueue(self, tx_hash):
|
def enqueue(self, tx_hash):
|
||||||
@ -114,44 +63,15 @@ class ChaindFsAdapter(ChaindAdapter):
|
|||||||
|
|
||||||
|
|
||||||
def dispatch(self, tx_hash):
|
def dispatch(self, tx_hash):
|
||||||
entry = None
|
|
||||||
|
|
||||||
store_lock = StoreLock()
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
entry = self.store.send_start(tx_hash)
|
entry = self.store.send_start(tx_hash)
|
||||||
break
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
|
||||||
store_lock.again()
|
|
||||||
continue
|
|
||||||
except StateLockedKey as e:
|
|
||||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
|
||||||
store_lock.again()
|
|
||||||
continue
|
|
||||||
|
|
||||||
tx_wire = entry.serialize()
|
tx_wire = entry.serialize()
|
||||||
|
|
||||||
r = None
|
r = None
|
||||||
try:
|
try:
|
||||||
r = self.dispatcher.send(tx_wire)
|
r = self.dispatcher.send(tx_wire)
|
||||||
except RPCException as e:
|
except RPCException:
|
||||||
logg.error('dispatch send failed for {}: {}'.format(tx_hash, e))
|
|
||||||
self.store.fail(tx_hash)
|
self.store.fail(tx_hash)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
store_lock = StoreLock()
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
self.store.send_end(tx_hash)
|
self.store.send_end(tx_hash)
|
||||||
break
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
|
||||||
store_lock.again(e)
|
|
||||||
continue
|
|
||||||
except StateLockedKey as e:
|
|
||||||
logg.debug('dispatch failed to find {} in backend, will try again: {}'.format(tx_hash, e))
|
|
||||||
store_lock.again(e)
|
|
||||||
continue
|
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
@ -1,19 +0,0 @@
|
|||||||
def apply_flag(flag):
|
|
||||||
flag.add('session')
|
|
||||||
flag.add('dispatch')
|
|
||||||
flag.add('socket')
|
|
||||||
flag.add('socket_client')
|
|
||||||
flag.add('token')
|
|
||||||
|
|
||||||
flag.alias('chaind_base', 'session')
|
|
||||||
flag.alias('chaind_socket_client', 'session', 'socket', 'socket_client')
|
|
||||||
|
|
||||||
return flag
|
|
||||||
|
|
||||||
|
|
||||||
def apply_arg(arg):
|
|
||||||
arg.add_long('session-id', 'session', help='Session to store state and data under')
|
|
||||||
arg.add_long('socket-path', 'socket', help='UNIX socket path')
|
|
||||||
arg.add_long('send-socket', 'socket_client', typ=bool, help='Send to UNIX socket')
|
|
||||||
arg.add_long('token-module', 'token', help='Python module path to resolve tokens from identifiers')
|
|
||||||
return arg
|
|
@ -1,13 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import enum
|
|
||||||
|
|
||||||
|
|
||||||
class ChaindFlag(enum.IntEnum):
|
|
||||||
SESSION = 1
|
|
||||||
DISPATCH = 2
|
|
||||||
SOCKET = 16
|
|
||||||
SOCKET_CLIENT = 32
|
|
||||||
TOKEN = 256
|
|
||||||
|
|
||||||
argflag_local_base = ChaindFlag.SESSION
|
|
||||||
argflag_local_socket_client = ChaindFlag.SESSION | ChaindFlag.SOCKET | ChaindFlag.SOCKET_CLIENT
|
|
@ -1,19 +0,0 @@
|
|||||||
def process_config(config, arg, args, flags):
|
|
||||||
args_override = {}
|
|
||||||
if arg.match('session', flags):
|
|
||||||
args_override['SESSION_ID'] = getattr(args, 'session_id')
|
|
||||||
args_override['SESSION_RUNTIME_DIR'] = getattr(args, 'runtime_path')
|
|
||||||
args_override['SESSION_DATA_DIR'] = getattr(args, 'state_path')
|
|
||||||
|
|
||||||
if arg.match('socket', flags):
|
|
||||||
args_override['SESSION_SOCKET_PATH'] = getattr(args, 'socket')
|
|
||||||
|
|
||||||
if arg.match('token', flags):
|
|
||||||
args_override['TOKEN_MODULE'] = getattr(args, 'token_module')
|
|
||||||
|
|
||||||
config.dict_override(args_override, 'local cli args')
|
|
||||||
|
|
||||||
if arg.match('socket_client', flags):
|
|
||||||
config.add(getattr(args, 'send_socket'), '_SOCKET_SEND', False)
|
|
||||||
|
|
||||||
return config
|
|
@ -1,6 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import os
|
|
||||||
|
|
||||||
|
|
||||||
data_dir = os.path.realpath(os.path.dirname(__file__))
|
|
||||||
config_dir = os.path.join(data_dir, 'config')
|
|
@ -1,5 +1,15 @@
|
|||||||
[session]
|
[session]
|
||||||
socket_path =
|
socket_path =
|
||||||
runtime_path =
|
runtime_dir =
|
||||||
id =
|
id =
|
||||||
data_path =
|
data_dir =
|
||||||
|
|
||||||
|
[database]
|
||||||
|
engine =
|
||||||
|
name = chaind
|
||||||
|
driver =
|
||||||
|
user =
|
||||||
|
password =
|
||||||
|
host =
|
||||||
|
port =
|
||||||
|
debug = 0
|
||||||
|
@ -1,2 +0,0 @@
|
|||||||
[token]
|
|
||||||
module =
|
|
@ -1,33 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
|
|
||||||
# local ipmorts
|
|
||||||
from chaind.adapters.fs import ChaindFsAdapter
|
|
||||||
from chaind.eth.cache import EthCacheTx
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class DispatchProcessor:
|
|
||||||
|
|
||||||
def __init__(self, chain_spec, queue_dir, dispatcher):
|
|
||||||
self.dispatcher = dispatcher
|
|
||||||
self.chain_spec = chain_spec,
|
|
||||||
self.queue_dir = queue_dir
|
|
||||||
|
|
||||||
|
|
||||||
def process(self, rpc, limit=50):
|
|
||||||
adapter = ChaindFsAdapter(
|
|
||||||
self.chain_spec,
|
|
||||||
self.queue_dir,
|
|
||||||
EthCacheTx,
|
|
||||||
self.dispatcher,
|
|
||||||
)
|
|
||||||
|
|
||||||
upcoming = adapter.upcoming(limit=limit)
|
|
||||||
logg.info('processor has {} candidates for {}, processing with limit {}'.format(len(upcoming), self.chain_spec, limit))
|
|
||||||
i = 0
|
|
||||||
for tx_hash in upcoming:
|
|
||||||
if adapter.dispatch(tx_hash):
|
|
||||||
i += 1
|
|
||||||
return i
|
|
@ -16,11 +16,3 @@ class ClientBlockError(BlockingIOError):
|
|||||||
|
|
||||||
class ClientInputError(ValueError):
|
class ClientInputError(ValueError):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class QueueLockError(Exception):
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class BackendError(Exception):
|
|
||||||
pass
|
|
||||||
|
102
chaind/filter.py
102
chaind/filter.py
@ -1,122 +1,30 @@
|
|||||||
# standard imports
|
# standard imports
|
||||||
import logging
|
import logging
|
||||||
import time
|
|
||||||
|
|
||||||
# external imports
|
# external imports
|
||||||
from chainlib.status import Status as TxStatus
|
from chainlib.status import Status as TxStatus
|
||||||
from chainsyncer.filter import SyncFilter
|
from chainsyncer.filter import SyncFilter
|
||||||
from chainqueue.error import NotLocalTxError
|
from chainqueue.error import NotLocalTxError
|
||||||
from chaind.adapters.fs import ChaindFsAdapter
|
|
||||||
from shep.error import StateLockedKey
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from .error import (
|
|
||||||
QueueLockError,
|
|
||||||
BackendError,
|
|
||||||
)
|
|
||||||
from chaind.lock import StoreLock
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class StateFilter(SyncFilter):
|
class StateFilter(SyncFilter):
|
||||||
|
|
||||||
def __init__(self, chain_spec, adapter_path, tx_adapter, throttler=None):
|
def __init__(self, adapter, throttler=None):
|
||||||
self.chain_spec = chain_spec
|
self.adapter = adapter
|
||||||
self.adapter_path = adapter_path
|
|
||||||
self.tx_adapter = tx_adapter
|
|
||||||
self.throttler = throttler
|
self.throttler = throttler
|
||||||
self.last_block_height = 0
|
|
||||||
self.adapter = None
|
|
||||||
self.store_lock = None
|
|
||||||
|
|
||||||
|
|
||||||
def __get_adapter(self, block, force_reload=False):
|
|
||||||
if self.store_lock == None:
|
|
||||||
self.store_lock = StoreLock()
|
|
||||||
|
|
||||||
reload = False
|
|
||||||
if block.number != self.last_block_height:
|
|
||||||
reload = True
|
|
||||||
elif self.adapter == None:
|
|
||||||
reload = True
|
|
||||||
elif force_reload:
|
|
||||||
reload = True
|
|
||||||
|
|
||||||
self.last_block_height = block.number
|
|
||||||
|
|
||||||
if reload:
|
|
||||||
while True:
|
|
||||||
logg.info('reloading adapter')
|
|
||||||
try:
|
|
||||||
self.adapter = ChaindFsAdapter(
|
|
||||||
self.chain_spec,
|
|
||||||
self.adapter_path,
|
|
||||||
self.tx_adapter,
|
|
||||||
None,
|
|
||||||
)
|
|
||||||
break
|
|
||||||
except BackendError as e:
|
|
||||||
logg.error('adapter instantiation failed: {}, one more try'.format(e))
|
|
||||||
self.store_lock.again()
|
|
||||||
continue
|
|
||||||
|
|
||||||
return self.adapter
|
|
||||||
|
|
||||||
|
|
||||||
def filter(self, conn, block, tx, session=None):
|
def filter(self, conn, block, tx, session=None):
|
||||||
cache_tx = None
|
|
||||||
queue_adapter = self.__get_adapter(block)
|
|
||||||
|
|
||||||
self.store_lock.reset()
|
|
||||||
|
|
||||||
while True:
|
|
||||||
try:
|
try:
|
||||||
cache_tx = queue_adapter.get(tx.hash)
|
cache_tx = self.adapter.get(tx.hash)
|
||||||
break
|
|
||||||
except NotLocalTxError:
|
except NotLocalTxError:
|
||||||
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
logg.debug('skipping not local transaction {}'.format(tx.hash))
|
||||||
return False
|
return False
|
||||||
except BackendError as e:
|
|
||||||
logg.error('adapter get failed: {}, one more try'.format(e))
|
|
||||||
self.store_lock.again()
|
|
||||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
|
||||||
continue
|
|
||||||
|
|
||||||
if cache_tx == None:
|
|
||||||
raise NotLocalTxError(tx.hash)
|
|
||||||
|
|
||||||
self.store_lock.reset()
|
|
||||||
|
|
||||||
queue_lock = StoreLock(error=QueueLockError)
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
if tx.status == TxStatus.SUCCESS:
|
if tx.status == TxStatus.SUCCESS:
|
||||||
queue_adapter.succeed(block, tx)
|
self.adapter.succeed(block, tx)
|
||||||
else:
|
else:
|
||||||
queue_adapter.fail(block, tx)
|
self.adapter.fail(block, tx)
|
||||||
break
|
|
||||||
except QueueLockError as e:
|
|
||||||
logg.debug('queue item {} is blocked, will retry: {}'.format(tx.hash, e))
|
|
||||||
queue_lock.again()
|
|
||||||
except FileNotFoundError as e:
|
|
||||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
|
||||||
self.store_lock.again()
|
|
||||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
|
||||||
continue
|
|
||||||
except NotLocalTxError as e:
|
|
||||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
|
||||||
self.store_lock.again()
|
|
||||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
|
||||||
continue
|
|
||||||
except StateLockedKey as e:
|
|
||||||
logg.debug('queue item {} not found, possible race condition, will retry: {}'.format(tx.hash, e))
|
|
||||||
self.store_lock.again()
|
|
||||||
queue_adapter = self.__get_adapter(block, force_reload=True)
|
|
||||||
continue
|
|
||||||
|
|
||||||
logg.info('filter registered {} for {} in {}'.format(tx.status_name, tx.hash, block))
|
|
||||||
|
|
||||||
if self.throttler != None:
|
if self.throttler != None:
|
||||||
self.throttler.dec(tx.hash)
|
self.throttler.dec(tx.hash)
|
||||||
|
|
||||||
|
@ -1,34 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import time
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from .error import BackendError
|
|
||||||
|
|
||||||
BASE_DELAY = 0.01
|
|
||||||
BASE_DELAY_LIMIT = 10.0
|
|
||||||
|
|
||||||
|
|
||||||
class StoreLock:
|
|
||||||
|
|
||||||
def __init__(self, delay=BASE_DELAY, delay_limit=BASE_DELAY_LIMIT, error=BackendError, description=None):
|
|
||||||
self.base_delay = delay
|
|
||||||
self.delay = delay
|
|
||||||
self.delay_limit = delay_limit
|
|
||||||
self.error = error
|
|
||||||
self.description = description
|
|
||||||
|
|
||||||
|
|
||||||
def again(self, e=None):
|
|
||||||
if self.delay > self.delay_limit:
|
|
||||||
err = None
|
|
||||||
if e != None:
|
|
||||||
err = str(e)
|
|
||||||
else:
|
|
||||||
err = self.description
|
|
||||||
raise self.error(err)
|
|
||||||
time.sleep(self.delay)
|
|
||||||
self.delay *= 2
|
|
||||||
|
|
||||||
|
|
||||||
def reset(self):
|
|
||||||
self.delay = self.base_delay
|
|
@ -8,21 +8,19 @@ import stat
|
|||||||
from hexathon import strip_0x
|
from hexathon import strip_0x
|
||||||
|
|
||||||
# local imports
|
# local imports
|
||||||
from .error import (
|
from chaind.error import (
|
||||||
NothingToDoError,
|
NothingToDoError,
|
||||||
ClientGoneError,
|
ClientGoneError,
|
||||||
ClientBlockError,
|
ClientBlockError,
|
||||||
ClientInputError,
|
ClientInputError,
|
||||||
)
|
)
|
||||||
from .lock import StoreLock
|
|
||||||
from .error import BackendError
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
logg = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class SessionController:
|
class SessionController:
|
||||||
|
|
||||||
def __init__(self, config, processor):
|
def __init__(self, config, adapter, processor):
|
||||||
self.dead = False
|
self.dead = False
|
||||||
os.makedirs(os.path.dirname(config.get('SESSION_SOCKET_PATH')), exist_ok=True)
|
os.makedirs(os.path.dirname(config.get('SESSION_SOCKET_PATH')), exist_ok=True)
|
||||||
try:
|
try:
|
||||||
@ -37,6 +35,7 @@ class SessionController:
|
|||||||
self.srv.settimeout(float(config.get('SESSION_DISPATCH_DELAY')))
|
self.srv.settimeout(float(config.get('SESSION_DISPATCH_DELAY')))
|
||||||
self.processor = processor
|
self.processor = processor
|
||||||
self.chain_spec = config.get('CHAIN_SPEC')
|
self.chain_spec = config.get('CHAIN_SPEC')
|
||||||
|
self.adapter = adapter
|
||||||
|
|
||||||
|
|
||||||
def shutdown(self, signo, frame):
|
def shutdown(self, signo, frame):
|
||||||
@ -60,16 +59,7 @@ class SessionController:
|
|||||||
|
|
||||||
|
|
||||||
def process(self, conn):
|
def process(self, conn):
|
||||||
state_lock = StoreLock()
|
r = self.processor(self.chain_spec, self.adapter, conn)
|
||||||
r = None
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
r = self.processor(conn)
|
|
||||||
break
|
|
||||||
except BackendError as e:
|
|
||||||
state_lock.again(e)
|
|
||||||
continue
|
|
||||||
|
|
||||||
if r > 0:
|
if r > 0:
|
||||||
self.srv.settimeout(0.1)
|
self.srv.settimeout(0.1)
|
||||||
else:
|
else:
|
||||||
@ -114,6 +104,7 @@ class SessionController:
|
|||||||
logg.error('invalid input "{}"'.format(data_in_str))
|
logg.error('invalid input "{}"'.format(data_in_str))
|
||||||
raise ClientInputError()
|
raise ClientInputError()
|
||||||
|
|
||||||
|
logg.info('recv {} bytes'.format(len(data)))
|
||||||
return (srvs, data,)
|
return (srvs, data,)
|
||||||
|
|
||||||
|
|
||||||
@ -126,4 +117,3 @@ class SessionController:
|
|||||||
logg.debug('{} bytes sent'.format(len(v)))
|
logg.debug('{} bytes sent'.format(len(v)))
|
||||||
except BrokenPipeError:
|
except BrokenPipeError:
|
||||||
logg.debug('they just hung up. how rude.')
|
logg.debug('they just hung up. how rude.')
|
||||||
srvs.close()
|
|
||||||
|
@ -1,126 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import logging
|
|
||||||
import os
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
from chainlib.settings import ChainSettings
|
|
||||||
from chainqueue.settings import *
|
|
||||||
|
|
||||||
logg = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class ChaindSettings(ChainSettings):
|
|
||||||
|
|
||||||
def __init__(settings, include_sync=False, include_queue=False):
|
|
||||||
super(ChaindSettings, settings).__init__()
|
|
||||||
settings.include_sync = include_sync
|
|
||||||
settings.include_queue = include_queue
|
|
||||||
|
|
||||||
|
|
||||||
def dir_for(self, k):
|
|
||||||
return os.path.join(self.o['SESSION_PATH'], k)
|
|
||||||
|
|
||||||
|
|
||||||
def process_session(settings, config):
|
|
||||||
session_id = config.get('SESSION_ID')
|
|
||||||
|
|
||||||
base_dir = os.getcwd()
|
|
||||||
data_dir = config.get('SESSION_DATA_PATH')
|
|
||||||
if data_dir == None:
|
|
||||||
data_dir = os.path.join(base_dir, '.chaind', 'chaind', settings.get('CHAIND_BACKEND'))
|
|
||||||
data_engine_dir = os.path.join(data_dir, config.get('CHAIND_ENGINE'))
|
|
||||||
os.makedirs(data_engine_dir, exist_ok=True)
|
|
||||||
|
|
||||||
# check if existing session
|
|
||||||
if session_id == None:
|
|
||||||
fp = os.path.join(data_engine_dir, 'default')
|
|
||||||
try:
|
|
||||||
os.stat(fp)
|
|
||||||
fp = os.path.realpath(fp)
|
|
||||||
except FileNotFoundError:
|
|
||||||
fp = None
|
|
||||||
if fp != None:
|
|
||||||
session_id = os.path.basename(fp)
|
|
||||||
|
|
||||||
make_default = False
|
|
||||||
if session_id == None:
|
|
||||||
session_id = str(uuid.uuid4())
|
|
||||||
make_default = True
|
|
||||||
|
|
||||||
chain_spec = settings.get('CHAIN_SPEC')
|
|
||||||
network_id_str = str(chain_spec.network_id())
|
|
||||||
# create the session persistent dir
|
|
||||||
session_path = os.path.join(
|
|
||||||
data_engine_dir,
|
|
||||||
chain_spec.arch(),
|
|
||||||
chain_spec.fork(),
|
|
||||||
network_id_str,
|
|
||||||
session_id,
|
|
||||||
)
|
|
||||||
if make_default:
|
|
||||||
fp = os.path.join(data_engine_dir, 'default')
|
|
||||||
os.symlink(session_path, fp)
|
|
||||||
|
|
||||||
data_path = session_path
|
|
||||||
os.makedirs(data_path, exist_ok=True)
|
|
||||||
|
|
||||||
# create volatile dir
|
|
||||||
uid = os.getuid()
|
|
||||||
runtime_path = config.get('SESSION_RUNTIME_PATH')
|
|
||||||
if runtime_path == None:
|
|
||||||
runtime_path = os.path.join('/run', 'user', str(uid), 'chaind', settings.get('CHAIND_BACKEND'))
|
|
||||||
runtime_path = os.path.join(
|
|
||||||
runtime_path,
|
|
||||||
config.get('CHAIND_ENGINE'),
|
|
||||||
chain_spec.arch(),
|
|
||||||
chain_spec.fork(),
|
|
||||||
str(chain_spec.network_id()),
|
|
||||||
session_id,
|
|
||||||
)
|
|
||||||
os.makedirs(runtime_path, exist_ok=True)
|
|
||||||
|
|
||||||
settings.set('SESSION_RUNTIME_PATH', runtime_path)
|
|
||||||
settings.set('SESSION_PATH', session_path)
|
|
||||||
settings.set('SESSION_DATA_PATH', data_path)
|
|
||||||
settings.set('SESSION_ID', session_id)
|
|
||||||
|
|
||||||
return settings
|
|
||||||
|
|
||||||
|
|
||||||
def process_socket(settings, config):
|
|
||||||
socket_path = config.get('SESSION_SOCKET_PATH')
|
|
||||||
if socket_path == None:
|
|
||||||
socket_path = os.path.join(settings.get('SESSION_RUNTIME_PATH'), 'chaind.sock')
|
|
||||||
settings.set('SESSION_SOCKET_PATH', socket_path)
|
|
||||||
return settings
|
|
||||||
|
|
||||||
|
|
||||||
def process_dispatch(settings, config):
|
|
||||||
settings.set('SESSION_DISPATCH_DELAY', 0.01)
|
|
||||||
return settings
|
|
||||||
|
|
||||||
|
|
||||||
def process_token(settings, config):
|
|
||||||
settings.set('TOKEN_MODULE', config.get('TOKEN_MODULE'))
|
|
||||||
return settings
|
|
||||||
|
|
||||||
|
|
||||||
def process_backend(settings, config):
|
|
||||||
settings.set('CHAIND_BACKEND', config.get('STATE_BACKEND')) #backend)
|
|
||||||
return settings
|
|
||||||
|
|
||||||
|
|
||||||
def process_queue(settings, config):
|
|
||||||
if config.get('STATE_PATH') == None:
|
|
||||||
queue_state_dir = settings.dir_for('queue')
|
|
||||||
config.add(queue_state_dir, 'STATE_PATH', False)
|
|
||||||
logg.debug('setting queue state path {}'.format(queue_state_dir))
|
|
||||||
|
|
||||||
settings = process_queue_tx(settings, config)
|
|
||||||
settings = process_queue_paths(settings, config)
|
|
||||||
if config.get('STATE_BACKEND') == 'fs':
|
|
||||||
settings = process_queue_backend_fs(settings, config)
|
|
||||||
settings = process_queue_store(settings, config)
|
|
||||||
|
|
||||||
return settings
|
|
@ -1,4 +1,5 @@
|
|||||||
# standard imports
|
# standard imports
|
||||||
|
import unittest
|
||||||
import hashlib
|
import hashlib
|
||||||
import tempfile
|
import tempfile
|
||||||
|
|
||||||
@ -7,11 +8,9 @@ from chainqueue.cache import CacheTokenTx
|
|||||||
from chainlib.status import Status as TxStatus
|
from chainlib.status import Status as TxStatus
|
||||||
from chainlib.chain import ChainSpec
|
from chainlib.chain import ChainSpec
|
||||||
from chainlib.error import RPCException
|
from chainlib.error import RPCException
|
||||||
from chainlib.tx import (
|
|
||||||
Tx,
|
# local imports
|
||||||
TxResult,
|
from chaind.adapters.fs import ChaindFsAdapter
|
||||||
)
|
|
||||||
from chainlib.block import Block
|
|
||||||
|
|
||||||
|
|
||||||
class MockCacheAdapter(CacheTokenTx):
|
class MockCacheAdapter(CacheTokenTx):
|
||||||
@ -34,22 +33,22 @@ class MockDispatcher:
|
|||||||
|
|
||||||
|
|
||||||
def send(self, v):
|
def send(self, v):
|
||||||
if v in self.fails:
|
if v not in self.fails:
|
||||||
raise RPCException('{} is in fails'.format(v))
|
raise RPCException('{} is in fails'.format(v))
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class MockTx(Tx):
|
class MockTx:
|
||||||
|
|
||||||
def __init__(self, tx_hash, status=TxStatus.SUCCESS):
|
def __init__(self, tx_hash, status=TxStatus.SUCCESS):
|
||||||
result = TxResult()
|
self.hash = tx_hash
|
||||||
result.status = status
|
self.status = status
|
||||||
super(MockTx, self).__init__(result=result)
|
|
||||||
self.set_hash(tx_hash)
|
|
||||||
|
|
||||||
|
|
||||||
class MockBlock(Block):
|
class TestChaindFsBase(unittest.TestCase):
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
self.chain_spec = ChainSpec('foo', 'bar', 42, 'baz')
|
||||||
|
self.path = tempfile.mkdtemp()
|
||||||
|
self.adapter = ChaindFsAdapter(self.chain_spec, self.path, self.cache_adapter, self.dispatcher)
|
||||||
|
|
||||||
def __init__(self, number):
|
|
||||||
super(MockBlock, self).__init__()
|
|
||||||
self.number = number
|
|
||||||
|
@ -1,31 +0,0 @@
|
|||||||
# standard imports
|
|
||||||
import unittest
|
|
||||||
import tempfile
|
|
||||||
import logging
|
|
||||||
|
|
||||||
# external imports
|
|
||||||
from chainlib.chain import ChainSpec
|
|
||||||
|
|
||||||
# local imports
|
|
||||||
from chaind.adapters.fs import ChaindFsAdapter
|
|
||||||
|
|
||||||
logging.STATETRACE = 5
|
|
||||||
logg = logging.getLogger(__name__)
|
|
||||||
logg.setLevel(logging.STATETRACE)
|
|
||||||
|
|
||||||
|
|
||||||
class TestChaindFsBase(unittest.TestCase):
|
|
||||||
|
|
||||||
def setUp(self):
|
|
||||||
self.chain_spec = ChainSpec('foo', 'bar', 42, 'baz')
|
|
||||||
self.path = tempfile.mkdtemp()
|
|
||||||
self.adapter = ChaindFsAdapter(self.chain_spec, self.path, self.cache_adapter, self.dispatcher, event_callback=self.log_state)
|
|
||||||
|
|
||||||
|
|
||||||
def log_state(self, k, from_state, to_state):
|
|
||||||
logg.log(logging.STATETRACE, 'state change {}: {} -> {}'.format(
|
|
||||||
k,
|
|
||||||
from_state,
|
|
||||||
to_state,
|
|
||||||
)
|
|
||||||
)
|
|
@ -1,6 +1,6 @@
|
|||||||
chainlib~=0.3.0
|
chainlib~=0.1.0
|
||||||
chainqueue~=0.2.0
|
chainqueue~=0.1.1
|
||||||
chainsyncer~=0.5.0
|
chainsyncer~=0.3.5
|
||||||
confini~=0.6.1
|
confini~=0.6.0
|
||||||
funga~=0.5.2
|
funga~=0.5.2
|
||||||
pyxdg~=0.26
|
pyxdg~=0.26
|
||||||
|
80
scripts/migrate.py
Normal file
80
scripts/migrate.py
Normal file
@ -0,0 +1,80 @@
|
|||||||
|
#!/usr/bin/python
|
||||||
|
import os
|
||||||
|
import argparse
|
||||||
|
import logging
|
||||||
|
|
||||||
|
# external imports
|
||||||
|
import alembic
|
||||||
|
from alembic.config import Config as AlembicConfig
|
||||||
|
import confini
|
||||||
|
import chainqueue.db
|
||||||
|
import chainsyncer.db
|
||||||
|
|
||||||
|
# local imports
|
||||||
|
from chaind import Environment
|
||||||
|
|
||||||
|
logging.basicConfig(level=logging.WARNING)
|
||||||
|
logg = logging.getLogger()
|
||||||
|
|
||||||
|
rootdir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
||||||
|
dbdir = os.path.join(rootdir, 'chaind', 'db')
|
||||||
|
configdir = os.path.join(rootdir, 'chaind', 'data', 'config')
|
||||||
|
default_migrations_dir = os.path.join(dbdir, 'migrations')
|
||||||
|
|
||||||
|
env = Environment(env=os.environ)
|
||||||
|
|
||||||
|
argparser = argparse.ArgumentParser()
|
||||||
|
argparser.add_argument('-c', type=str, default=env.config_dir, help='config directory')
|
||||||
|
argparser.add_argument('--env-prefix', default=os.environ.get('CONFINI_ENV_PREFIX'), dest='env_prefix', type=str, help='environment prefix for variables to overwrite configuration')
|
||||||
|
argparser.add_argument('--data-dir', dest='data_dir', type=str, help='data directory')
|
||||||
|
argparser.add_argument('--migrations-dir', dest='migrations_dir', default=default_migrations_dir, type=str, help='path to alembic migrations directory')
|
||||||
|
argparser.add_argument('--reset', action='store_true', help='reset exsting database')
|
||||||
|
argparser.add_argument('-v', action='store_true', help='be verbose')
|
||||||
|
argparser.add_argument('-vv', action='store_true', help='be more verbose')
|
||||||
|
args = argparser.parse_args()
|
||||||
|
|
||||||
|
if args.vv:
|
||||||
|
logging.getLogger().setLevel(logging.DEBUG)
|
||||||
|
elif args.v:
|
||||||
|
logging.getLogger().setLevel(logging.INFO)
|
||||||
|
|
||||||
|
# process config
|
||||||
|
logg.debug('loading config from {}'.format(args.c))
|
||||||
|
config = confini.Config(configdir, args.env_prefix, override_dirs=[args.c])
|
||||||
|
config.process()
|
||||||
|
args_override = {
|
||||||
|
'SESSION_DATA_DIR': getattr(args, 'data_dir'),
|
||||||
|
}
|
||||||
|
config.dict_override(args_override, 'cli args')
|
||||||
|
|
||||||
|
if config.get('DATABASE_ENGINE') == 'sqlite':
|
||||||
|
config.add(os.path.join(config.get('SESSION_DATA_DIR'), config.get('DATABASE_NAME') + '.sqlite'), 'DATABASE_NAME', True)
|
||||||
|
|
||||||
|
config.censor('PASSWORD', 'DATABASE')
|
||||||
|
|
||||||
|
logg.debug('config loaded:\n{}'.format(config))
|
||||||
|
config.add(os.path.join(args.migrations_dir, config.get('DATABASE_ENGINE')), '_MIGRATIONS_DIR', True)
|
||||||
|
if not os.path.isdir(config.get('_MIGRATIONS_DIR')):
|
||||||
|
logg.debug('migrations dir for engine {} not found, reverting to default'.format(config.get('DATABASE_ENGINE')))
|
||||||
|
config.add(os.path.join(args.migrations_dir, 'default'), '_MIGRATIONS_DIR', True)
|
||||||
|
|
||||||
|
os.makedirs(config.get('SESSION_DATA_DIR'), exist_ok=True)
|
||||||
|
|
||||||
|
dsn = chainqueue.db.dsn_from_config(config)
|
||||||
|
|
||||||
|
def main():
|
||||||
|
logg.info('using migrations dir {}'.format(config.get('_MIGRATIONS_DIR')))
|
||||||
|
logg.info('using db {}'.format(dsn))
|
||||||
|
ac = AlembicConfig(os.path.join(config.get('_MIGRATIONS_DIR'), 'alembic.ini'))
|
||||||
|
ac.set_main_option('sqlalchemy.url', dsn)
|
||||||
|
ac.set_main_option('script_location', config.get('_MIGRATIONS_DIR'))
|
||||||
|
|
||||||
|
if args.reset:
|
||||||
|
logg.debug('reset is set, purging existing content')
|
||||||
|
alembic.command.downgrade(ac, 'base')
|
||||||
|
|
||||||
|
alembic.command.upgrade(ac, 'head')
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
17
setup.cfg
17
setup.cfg
@ -1,37 +1,36 @@
|
|||||||
[metadata]
|
[metadata]
|
||||||
name = chaind
|
name = chaind
|
||||||
version = 0.3.2
|
version = 0.1.0
|
||||||
description = Base package for chain queue service
|
description = Base package for chain queue servicek
|
||||||
author = Louis Holbrook
|
author = Louis Holbrook
|
||||||
author_email = dev@holbrook.no
|
author_email = dev@holbrook.no
|
||||||
url = https://git.defalsify.org/chaind.eth
|
url = https://gitlab.com/chaintool/chaind
|
||||||
keywords =
|
keywords =
|
||||||
blockchain
|
blockchain
|
||||||
cryptocurrency
|
cryptocurrency
|
||||||
dlt
|
p2p
|
||||||
classifiers =
|
classifiers =
|
||||||
Programming Language :: Python :: 3
|
Programming Language :: Python :: 3
|
||||||
Operating System :: OS Independent
|
Operating System :: OS Independent
|
||||||
Development Status :: 3 - Alpha
|
Development Status :: 3 - Alpha
|
||||||
Environment :: Console
|
Environment :: Console
|
||||||
Intended Audience :: Developers
|
Intended Audience :: Developers
|
||||||
License :: OSI Approved :: GNU Affero General Public License v3 or later (AGPLv3+)
|
License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)
|
||||||
Topic :: Internet
|
Topic :: Internet
|
||||||
# Topic :: Blockchain :: EVM
|
# Topic :: Blockchain :: EVM
|
||||||
license = OSI Approved :: GNU Affero General Public License v3 or later (AGPLv3+)
|
license = GPL3
|
||||||
licence_files =
|
licence_files =
|
||||||
LICENSE
|
LICENSE
|
||||||
|
|
||||||
[options]
|
[options]
|
||||||
python_requires = >= 3.7
|
python_requires = >= 3.6
|
||||||
include_package_data = True
|
include_package_data = True
|
||||||
packages =
|
packages =
|
||||||
chaind
|
chaind
|
||||||
|
# chaind.sql
|
||||||
# chaind.runnable
|
# chaind.runnable
|
||||||
chaind.adapters
|
chaind.adapters
|
||||||
chaind.unittest
|
chaind.unittest
|
||||||
chaind.data
|
|
||||||
chaind.cli
|
|
||||||
|
|
||||||
#[options.entry_points]
|
#[options.entry_points]
|
||||||
#console_scripts =
|
#console_scripts =
|
||||||
|
@ -14,16 +14,16 @@ from chaind.filter import StateFilter
|
|||||||
# test imports
|
# test imports
|
||||||
from chaind.unittest.common import (
|
from chaind.unittest.common import (
|
||||||
MockTx,
|
MockTx,
|
||||||
MockBlock,
|
|
||||||
MockCacheAdapter,
|
MockCacheAdapter,
|
||||||
MockDispatcher,
|
TestChaindFsBase,
|
||||||
)
|
)
|
||||||
from chaind.unittest.fs import TestChaindFsBase
|
|
||||||
|
|
||||||
logging.basicConfig(level=logging.DEBUG)
|
logging.basicConfig(level=logging.DEBUG)
|
||||||
logg = logging.getLogger()
|
logg = logging.getLogger()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class TestChaindFs(TestChaindFsBase):
|
class TestChaindFs(TestChaindFsBase):
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
@ -43,15 +43,12 @@ class TestChaindFs(TestChaindFsBase):
|
|||||||
self.assertEqual(data, v)
|
self.assertEqual(data, v)
|
||||||
|
|
||||||
|
|
||||||
def test_fs_fail(self):
|
def test_fs_defer(self):
|
||||||
data = os.urandom(128).hex()
|
data = os.urandom(128).hex()
|
||||||
hsh = self.adapter.put(data)
|
hsh = self.adapter.put(data)
|
||||||
self.dispatcher.add_fail(data)
|
self.dispatcher.add_fail(hsh)
|
||||||
|
self.adapter.dispatch(hsh)
|
||||||
r = self.adapter.dispatch(hsh)
|
txs = self.adapter.deferred()
|
||||||
self.assertFalse(r)
|
|
||||||
|
|
||||||
txs = self.adapter.failed()
|
|
||||||
self.assertEqual(len(txs), 1)
|
self.assertEqual(len(txs), 1)
|
||||||
|
|
||||||
|
|
||||||
@ -75,10 +72,9 @@ class TestChaindFs(TestChaindFsBase):
|
|||||||
data = os.urandom(128).hex()
|
data = os.urandom(128).hex()
|
||||||
hsh = self.adapter.put(data)
|
hsh = self.adapter.put(data)
|
||||||
|
|
||||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
fltr = StateFilter(self.adapter)
|
||||||
tx = MockTx(hsh)
|
tx = MockTx(hsh)
|
||||||
block = MockBlock(42)
|
fltr.filter(None, None, tx)
|
||||||
fltr.filter(None, block, tx)
|
|
||||||
|
|
||||||
|
|
||||||
def test_fs_filter_fail(self):
|
def test_fs_filter_fail(self):
|
||||||
@ -87,30 +83,9 @@ class TestChaindFs(TestChaindFsBase):
|
|||||||
data = os.urandom(128).hex()
|
data = os.urandom(128).hex()
|
||||||
hsh = self.adapter.put(data)
|
hsh = self.adapter.put(data)
|
||||||
|
|
||||||
fltr = StateFilter(self.chain_spec, self.path, MockCacheAdapter)
|
fltr = StateFilter(self.adapter)
|
||||||
tx = MockTx(hsh, TxStatus.ERROR)
|
tx = MockTx(hsh, TxStatus.ERROR)
|
||||||
block = MockBlock(42)
|
fltr.filter(None, None, tx)
|
||||||
fltr.filter(None, block, tx)
|
|
||||||
|
|
||||||
|
|
||||||
def test_upcoming(self):
|
|
||||||
drv = QueueDriver(self.adapter)
|
|
||||||
|
|
||||||
txs = []
|
|
||||||
for i in range(10):
|
|
||||||
data = os.urandom(128).hex()
|
|
||||||
hsh = self.adapter.put(data)
|
|
||||||
txs.append(hsh)
|
|
||||||
self.adapter.enqueue(hsh)
|
|
||||||
|
|
||||||
r = self.adapter.upcoming(limit=5)
|
|
||||||
self.assertEqual(len(r), 5)
|
|
||||||
|
|
||||||
r = self.adapter.dispatch(txs[0])
|
|
||||||
self.assertTrue(r)
|
|
||||||
|
|
||||||
r = self.adapter.upcoming(limit=5)
|
|
||||||
self.assertEqual(len(r), 4)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
|
Loading…
Reference in New Issue
Block a user