2021-02-21 16:41:37 +01:00
# standard imports
import os
import sys
import json
import logging
import argparse
import uuid
import datetime
import time
from glob import glob
# third-party imports
import redis
import confini
import celery
from hexathon import (
add_0x ,
strip_0x ,
)
from chainlib . eth . address import to_checksum
from cic_types . models . person import Person
from cic_eth . api . api_task import Api
2021-02-22 21:00:18 +01:00
from chainlib . chain import ChainSpec
2021-02-21 16:41:37 +01:00
from cic_types . processor import generate_metadata_pointer
logging . basicConfig ( level = logging . WARNING )
logg = logging . getLogger ( )
default_config_dir = ' /usr/local/etc/cic '
argparser = argparse . ArgumentParser ( )
argparser . add_argument ( ' -c ' , type = str , default = default_config_dir , help = ' config file ' )
argparser . add_argument ( ' -i ' , ' --chain-spec ' , dest = ' i ' , type = str , help = ' Chain specification string ' )
argparser . add_argument ( ' --redis-host ' , dest = ' redis_host ' , type = str , help = ' redis host to use for task submission ' )
argparser . add_argument ( ' --redis-port ' , dest = ' redis_port ' , type = int , help = ' redis host to use for task submission ' )
argparser . add_argument ( ' --redis-db ' , dest = ' redis_db ' , type = int , help = ' redis db to use for task submission and callback ' )
argparser . add_argument ( ' --redis-host-callback ' , dest = ' redis_host_callback ' , default = ' localhost ' , type = str , help = ' redis host to use for callback ' )
argparser . add_argument ( ' --redis-port-callback ' , dest = ' redis_port_callback ' , default = 6379 , type = int , help = ' redis port to use for callback ' )
2021-04-05 17:07:09 +02:00
argparser . add_argument ( ' --batch-size ' , dest = ' batch_size ' , default = 100 , type = int , help = ' burst size of sending transactions to node ' ) # batch size should be slightly below cumulative gas limit worth, eg 80000 gas txs with 8000000 limit is a bit less than 100 batch size
2021-02-21 16:41:37 +01:00
argparser . add_argument ( ' --batch-delay ' , dest = ' batch_delay ' , default = 2 , type = int , help = ' seconds delay between batches ' )
2021-03-07 19:01:44 +01:00
argparser . add_argument ( ' --timeout ' , default = 60.0 , type = float , help = ' Callback timeout ' )
2021-02-21 16:41:37 +01:00
argparser . add_argument ( ' -q ' , type = str , default = ' cic-eth ' , help = ' Task queue ' )
argparser . add_argument ( ' -v ' , action = ' store_true ' , help = ' Be verbose ' )
argparser . add_argument ( ' -vv ' , action = ' store_true ' , help = ' Be more verbose ' )
argparser . add_argument ( ' user_dir ' , type = str , help = ' path to users export dir tree ' )
args = argparser . parse_args ( )
if args . v :
logg . setLevel ( logging . INFO )
elif args . vv :
logg . setLevel ( logging . DEBUG )
config_dir = args . c
config = confini . Config ( config_dir , os . environ . get ( ' CONFINI_ENV_PREFIX ' ) )
config . process ( )
args_override = {
' CIC_CHAIN_SPEC ' : getattr ( args , ' i ' ) ,
' REDIS_HOST ' : getattr ( args , ' redis_host ' ) ,
' REDIS_PORT ' : getattr ( args , ' redis_port ' ) ,
' REDIS_DB ' : getattr ( args , ' redis_db ' ) ,
}
config . dict_override ( args_override , ' cli ' )
celery_app = celery . Celery ( broker = config . get ( ' CELERY_BROKER_URL ' ) , backend = config . get ( ' CELERY_RESULT_URL ' ) )
redis_host = config . get ( ' REDIS_HOST ' )
redis_port = config . get ( ' REDIS_PORT ' )
redis_db = config . get ( ' REDIS_DB ' )
r = redis . Redis ( redis_host , redis_port , redis_db )
ps = r . pubsub ( )
user_new_dir = os . path . join ( args . user_dir , ' new ' )
os . makedirs ( user_new_dir )
meta_dir = os . path . join ( args . user_dir , ' meta ' )
os . makedirs ( meta_dir )
user_old_dir = os . path . join ( args . user_dir , ' old ' )
os . stat ( user_old_dir )
chain_spec = ChainSpec . from_chain_str ( config . get ( ' CIC_CHAIN_SPEC ' ) )
chain_str = str ( chain_spec )
batch_size = args . batch_size
batch_delay = args . batch_delay
def register_eth ( i , u ) :
redis_channel = str ( uuid . uuid4 ( ) )
ps . subscribe ( redis_channel )
2021-03-01 21:15:17 +01:00
#ps.get_message()
2021-02-21 16:41:37 +01:00
api = Api (
config . get ( ' CIC_CHAIN_SPEC ' ) ,
queue = args . q ,
callback_param = ' {} : {} : {} : {} ' . format ( args . redis_host_callback , args . redis_port_callback , redis_db , redis_channel ) ,
callback_task = ' cic_eth.callbacks.redis.redis ' ,
callback_queue = args . q ,
)
t = api . create_account ( register = True )
2021-03-07 19:01:44 +01:00
logg . debug ( ' register {} -> {} ' . format ( u , t ) )
2021-02-21 16:41:37 +01:00
2021-03-01 21:15:17 +01:00
while True :
ps . get_message ( )
m = ps . get_message ( timeout = args . timeout )
address = None
2021-03-06 18:55:51 +01:00
if m == None :
logg . debug ( ' message timeout ' )
return
2021-03-01 21:15:17 +01:00
if m [ ' type ' ] == ' subscribe ' :
logg . debug ( ' skipping subscribe message ' )
continue
try :
r = json . loads ( m [ ' data ' ] )
address = r [ ' result ' ]
break
2021-03-07 19:01:44 +01:00
except Exception as e :
2021-03-01 21:15:17 +01:00
if m == None :
2021-03-07 19:01:44 +01:00
logg . critical ( ' empty response from redis callback (did the service crash?) {} ' . format ( e ) )
2021-03-01 21:15:17 +01:00
else :
2021-03-07 19:01:44 +01:00
logg . critical ( ' unexpected response from redis callback: {} {} ' . format ( m , e ) )
2021-03-01 21:15:17 +01:00
sys . exit ( 1 )
logg . debug ( ' [ {} ] register eth {} {} ' . format ( i , u , address ) )
2021-02-21 16:41:37 +01:00
return address
if __name__ == ' __main__ ' :
#fi = open(os.path.join(user_out_dir, 'addresses.csv'), 'a')
i = 0
j = 0
for x in os . walk ( user_old_dir ) :
for y in x [ 2 ] :
if y [ len ( y ) - 5 : ] != ' .json ' :
continue
filepath = os . path . join ( x [ 0 ] , y )
f = open ( filepath , ' r ' )
try :
o = json . load ( f )
except json . decoder . JSONDecodeError as e :
f . close ( )
logg . error ( ' load error for {} : {} ' . format ( y , e ) )
continue
f . close ( )
u = Person . deserialize ( o )
new_address = register_eth ( i , u )
if u . identities . get ( ' evm ' ) == None :
u . identities [ ' evm ' ] = { }
2021-02-22 21:00:18 +01:00
sub_chain_str = ' {} : {} ' . format ( chain_spec . common_name ( ) , chain_spec . network_id ( ) )
u . identities [ ' evm ' ] [ sub_chain_str ] = [ new_address ]
2021-02-21 16:41:37 +01:00
new_address_clean = strip_0x ( new_address )
filepath = os . path . join (
user_new_dir ,
new_address_clean [ : 2 ] . upper ( ) ,
new_address_clean [ 2 : 4 ] . upper ( ) ,
new_address_clean . upper ( ) + ' .json ' ,
)
os . makedirs ( os . path . dirname ( filepath ) , exist_ok = True )
o = u . serialize ( )
f = open ( filepath , ' w ' )
f . write ( json . dumps ( o ) )
f . close ( )
#old_address = to_checksum(add_0x(y[:len(y)-5]))
#fi.write('{},{}\n'.format(new_address, old_address))
meta_key = generate_metadata_pointer ( bytes . fromhex ( new_address_clean ) , ' cic.person ' )
meta_filepath = os . path . join ( meta_dir , ' {} .json ' . format ( new_address_clean . upper ( ) ) )
os . symlink ( os . path . realpath ( filepath ) , meta_filepath )
i + = 1
sys . stdout . write ( ' imported {} {} ' . format ( i , u ) . ljust ( 200 ) + " \r " )
j + = 1
if j == batch_size :
time . sleep ( batch_delay )
j = 0
#fi.close()