You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1248 lines
47 KiB
1248 lines
47 KiB
# Copyright (c) 2016-2018, Neil Booth
|
|
#
|
|
# All rights reserved.
|
|
#
|
|
# See the file "LICENCE" for information about the copyright
|
|
# and warranty status of this software.
|
|
|
|
'''Classes for local RPC server and remote client TCP/SSL servers.'''
|
|
|
|
import asyncio
|
|
import codecs
|
|
import datetime
|
|
import itertools
|
|
import json
|
|
import os
|
|
import ssl
|
|
import time
|
|
from collections import defaultdict
|
|
from functools import partial
|
|
|
|
from aiorpcx import ServerSession, JSONRPCAutoDetect, RPCError
|
|
|
|
import electrumx
|
|
import electrumx.lib.text as text
|
|
import electrumx.lib.util as util
|
|
from electrumx.lib.hash import (sha256, hash_to_hex_str, hex_str_to_hash,
|
|
HASHX_LEN)
|
|
from electrumx.lib.peer import Peer
|
|
from electrumx.server.daemon import DaemonError
|
|
|
|
|
|
BAD_REQUEST = 1
|
|
DAEMON_ERROR = 2
|
|
|
|
|
|
def scripthash_to_hashX(scripthash):
|
|
try:
|
|
bin_hash = hex_str_to_hash(scripthash)
|
|
if len(bin_hash) == 32:
|
|
return bin_hash[:HASHX_LEN]
|
|
except Exception:
|
|
pass
|
|
raise RPCError(BAD_REQUEST, f'{scripthash} is not a valid script hash')
|
|
|
|
|
|
def non_negative_integer(value):
|
|
'''Return param value it is or can be converted to a non-negative
|
|
integer, otherwise raise an RPCError.'''
|
|
try:
|
|
value = int(value)
|
|
if value >= 0:
|
|
return value
|
|
except ValueError:
|
|
pass
|
|
raise RPCError(BAD_REQUEST,
|
|
f'{value} should be a non-negative integer')
|
|
|
|
|
|
class Semaphores(object):
|
|
'''For aiorpcX's semaphore handling.'''
|
|
|
|
def __init__(self, semaphores):
|
|
self.semaphores = semaphores
|
|
self.acquired = []
|
|
|
|
async def __aenter__(self):
|
|
for semaphore in self.semaphores:
|
|
await semaphore.acquire()
|
|
self.acquired.append(semaphore)
|
|
|
|
async def __aexit__(self, exc_type, exc_value, traceback):
|
|
for semaphore in self.acquired:
|
|
semaphore.release()
|
|
|
|
|
|
class SessionGroup(object):
|
|
|
|
def __init__(self, gid):
|
|
self.gid = gid
|
|
# Concurrency per group
|
|
self.semaphore = asyncio.Semaphore(20)
|
|
|
|
|
|
class SessionManager(object):
|
|
'''Holds global state about all sessions.'''
|
|
|
|
CATCHING_UP, LISTENING, PAUSED, SHUTTING_DOWN = range(4)
|
|
|
|
def __init__(self, env, controller):
|
|
self.env = env
|
|
self.controller = controller
|
|
self.logger = util.class_logger(__name__, self.__class__.__name__)
|
|
self.servers = {}
|
|
self.sessions = set()
|
|
self.max_sessions = env.max_sessions
|
|
self.low_watermark = self.max_sessions * 19 // 20
|
|
self.max_subs = env.max_subs
|
|
self.next_log_sessions = 0
|
|
self.cur_group = SessionGroup(0)
|
|
self.state = self.CATCHING_UP
|
|
self.txs_sent = 0
|
|
self.start_time = time.time()
|
|
# Cache some idea of room to avoid recounting on each subscription
|
|
self.subs_room = 0
|
|
# Event triggered when electrumx is listening for incoming requests.
|
|
self.server_listening = asyncio.Event()
|
|
# Set up the RPC request handlers
|
|
cmds = ('add_peer daemon_url disconnect getinfo groups log peers '
|
|
'reorg sessions stop'.split())
|
|
self.rpc_handlers = {cmd: getattr(self, 'rpc_' + cmd) for cmd in cmds}
|
|
|
|
async def _start_server(self, kind, *args, **kw_args):
|
|
loop = asyncio.get_event_loop()
|
|
if kind == 'RPC':
|
|
protocol_class = LocalRPC
|
|
else:
|
|
protocol_class = self.env.coin.SESSIONCLS
|
|
protocol_factory = partial(protocol_class, self, self.controller, kind)
|
|
server = loop.create_server(protocol_factory, *args, **kw_args)
|
|
|
|
host, port = args[:2]
|
|
try:
|
|
self.servers[kind] = await server
|
|
except Exception as e:
|
|
self.logger.error('{} server failed to listen on {}:{:d} :{}'
|
|
.format(kind, host, port, e))
|
|
else:
|
|
self.logger.info('{} server listening on {}:{:d}'
|
|
.format(kind, host, port))
|
|
|
|
async def _start_external_servers(self):
|
|
'''Start listening on TCP and SSL ports, but only if the respective
|
|
port was given in the environment.
|
|
'''
|
|
env = self.env
|
|
host = env.cs_host(for_rpc=False)
|
|
if env.tcp_port is not None:
|
|
await self._start_server('TCP', host, env.tcp_port)
|
|
if env.ssl_port is not None:
|
|
sslc = ssl.SSLContext(ssl.PROTOCOL_TLS)
|
|
sslc.load_cert_chain(env.ssl_certfile, keyfile=env.ssl_keyfile)
|
|
await self._start_server('SSL', host, env.ssl_port, ssl=sslc)
|
|
# Change state
|
|
self.state = self.LISTENING
|
|
self.server_listening.set()
|
|
|
|
def _close_servers(self, kinds):
|
|
'''Close the servers of the given kinds (TCP etc.).'''
|
|
if kinds:
|
|
self.logger.info('closing down {} listening servers'
|
|
.format(', '.join(kinds)))
|
|
for kind in kinds:
|
|
server = self.servers.pop(kind, None)
|
|
if server:
|
|
server.close()
|
|
|
|
def _group_map(self):
|
|
group_map = defaultdict(list)
|
|
for session in self.sessions:
|
|
group_map[session.group].append(session)
|
|
return group_map
|
|
|
|
def _sub_count(self):
|
|
return sum(s.sub_count() for s in self.sessions)
|
|
|
|
def _lookup_session(self, session_id):
|
|
try:
|
|
session_id = int(session_id)
|
|
except Exception:
|
|
pass
|
|
else:
|
|
for session in self.sessions:
|
|
if session.session_id == session_id:
|
|
return session
|
|
return None
|
|
|
|
def _for_each_session(self, session_ids, operation):
|
|
if not isinstance(session_ids, list):
|
|
raise RPCError(BAD_REQUEST, 'expected a list of session IDs')
|
|
|
|
result = []
|
|
for session_id in session_ids:
|
|
session = self._lookup_session(session_id)
|
|
if session:
|
|
result.append(operation(session))
|
|
else:
|
|
result.append('unknown session: {}'.format(session_id))
|
|
return result
|
|
|
|
def _close_session(self, session):
|
|
'''Close the session's transport.'''
|
|
session.close()
|
|
return 'disconnected {:d}'.format(session.session_id)
|
|
|
|
def _clear_stale_sessions(self):
|
|
'''Cut off sessions that haven't done anything for 10 minutes.'''
|
|
now = time.time()
|
|
stale_cutoff = now - self.env.session_timeout
|
|
|
|
stale = []
|
|
for session in self.sessions:
|
|
if session.is_closing():
|
|
session.abort()
|
|
elif session.last_recv < stale_cutoff:
|
|
self._close_session(session)
|
|
stale.append(session.session_id)
|
|
if stale:
|
|
self.logger.info('closing stale connections {}'.format(stale))
|
|
|
|
# Consolidate small groups
|
|
bw_limit = self.env.bandwidth_limit
|
|
group_map = self._group_map()
|
|
groups = [group for group, sessions in group_map.items()
|
|
if len(sessions) <= 5 and
|
|
sum(s.bw_charge for s in sessions) < bw_limit]
|
|
if len(groups) > 1:
|
|
new_group = groups[-1]
|
|
for group in groups:
|
|
for session in group_map[group]:
|
|
session.group = new_group
|
|
|
|
def _getinfo(self):
|
|
'''A one-line summary of server state.'''
|
|
group_map = self._group_map()
|
|
daemon = self.controller.daemon
|
|
bp = self.controller.bp
|
|
peer_mgr = self.controller.peer_mgr
|
|
return {
|
|
'version': electrumx.version,
|
|
'daemon': daemon.logged_url(),
|
|
'daemon_height': daemon.cached_height(),
|
|
'db_height': bp.db_height,
|
|
'closing': len([s for s in self.sessions if s.is_closing()]),
|
|
'errors': sum(s.rpc.errors for s in self.sessions),
|
|
'groups': len(group_map),
|
|
'logged': len([s for s in self.sessions if s.log_me]),
|
|
'paused': sum(s.paused for s in self.sessions),
|
|
'pid': os.getpid(),
|
|
'peers': peer_mgr.info(),
|
|
'requests': sum(s.count_pending_items() for s in self.sessions),
|
|
'sessions': self._session_count(),
|
|
'subs': self._sub_count(),
|
|
'txs_sent': self.txs_sent,
|
|
'uptime': util.formatted_time(time.time() - self.start_time),
|
|
}
|
|
|
|
def _session_data(self, for_log):
|
|
'''Returned to the RPC 'sessions' call.'''
|
|
now = time.time()
|
|
sessions = sorted(self.sessions, key=lambda s: s.start_time)
|
|
return [(session.session_id,
|
|
session.flags(),
|
|
session.peer_address_str(for_log=for_log),
|
|
session.client,
|
|
session.protocol_version_string(),
|
|
session.count_pending_items(),
|
|
session.txs_sent,
|
|
session.sub_count(),
|
|
session.recv_count, session.recv_size,
|
|
session.send_count, session.send_size,
|
|
now - session.start_time)
|
|
for session in sessions]
|
|
|
|
def _group_data(self):
|
|
'''Returned to the RPC 'groups' call.'''
|
|
result = []
|
|
group_map = self._group_map()
|
|
for group, sessions in group_map.items():
|
|
result.append([group.gid,
|
|
len(sessions),
|
|
sum(s.bw_charge for s in sessions),
|
|
sum(s.count_pending_items() for s in sessions),
|
|
sum(s.txs_sent for s in sessions),
|
|
sum(s.sub_count() for s in sessions),
|
|
sum(s.recv_count for s in sessions),
|
|
sum(s.recv_size for s in sessions),
|
|
sum(s.send_count for s in sessions),
|
|
sum(s.send_size for s in sessions),
|
|
])
|
|
return result
|
|
|
|
# --- LocalRPC command handlers
|
|
|
|
def rpc_add_peer(self, real_name):
|
|
'''Add a peer.
|
|
|
|
real_name: a real name, as would appear on IRC
|
|
'''
|
|
peer = Peer.from_real_name(real_name, 'RPC')
|
|
self.controller.peer_mgr.add_peers([peer])
|
|
return "peer '{}' added".format(real_name)
|
|
|
|
def rpc_disconnect(self, session_ids):
|
|
'''Disconnect sesssions.
|
|
|
|
session_ids: array of session IDs
|
|
'''
|
|
return self._for_each_session(session_ids, self._close_session)
|
|
|
|
def rpc_log(self, session_ids):
|
|
'''Toggle logging of sesssions.
|
|
|
|
session_ids: array of session IDs
|
|
'''
|
|
def toggle_logging(session):
|
|
'''Toggle logging of the session.'''
|
|
session.toggle_logging()
|
|
return 'log {:d}: {}'.format(session.session_id, session.log_me)
|
|
|
|
return self._for_each_session(session_ids, toggle_logging)
|
|
|
|
def rpc_daemon_url(self, daemon_url=None):
|
|
'''Replace the daemon URL.'''
|
|
daemon_url = daemon_url or self.env.daemon_url
|
|
daemon = self.controller.daemon
|
|
try:
|
|
daemon.set_urls(self.env.coin.daemon_urls(daemon_url))
|
|
except Exception as e:
|
|
raise RPCError(BAD_REQUEST, f'an error occured: {e}')
|
|
return 'now using daemon at {}'.format(daemon.logged_url())
|
|
|
|
def rpc_stop(self):
|
|
'''Shut down the server cleanly.'''
|
|
loop = asyncio.get_event_loop()
|
|
loop.call_soon(self.controller.shutdown_event.set)
|
|
return 'stopping'
|
|
|
|
def rpc_getinfo(self):
|
|
'''Return summary information about the server process.'''
|
|
return self._getinfo()
|
|
|
|
def rpc_groups(self):
|
|
'''Return statistics about the session groups.'''
|
|
return self._group_data()
|
|
|
|
def rpc_peers(self):
|
|
'''Return a list of data about server peers.'''
|
|
return self.controller.peer_mgr.rpc_data()
|
|
|
|
def rpc_sessions(self):
|
|
'''Return statistics about connected sessions.'''
|
|
return self._session_data(for_log=False)
|
|
|
|
def rpc_reorg(self, count=3):
|
|
'''Force a reorg of the given number of blocks.
|
|
|
|
count: number of blocks to reorg (default 3)
|
|
'''
|
|
count = non_negative_integer(count)
|
|
if not self.controller.bp.force_chain_reorg(count):
|
|
raise RPCError(BAD_REQUEST, 'still catching up with daemon')
|
|
return 'scheduled a reorg of {:,d} blocks'.format(count)
|
|
|
|
# --- External Interface
|
|
|
|
async def start_serving(self):
|
|
'''Start TCP and SSL servers.'''
|
|
self.logger.info('max session count: {:,d}'.format(self.max_sessions))
|
|
self.logger.info('session timeout: {:,d} seconds'
|
|
.format(self.env.session_timeout))
|
|
self.logger.info('session bandwidth limit {:,d} bytes'
|
|
.format(self.env.bandwidth_limit))
|
|
self.logger.info('max response size {:,d} bytes'
|
|
.format(self.env.max_send))
|
|
self.logger.info('max subscriptions across all sessions: {:,d}'
|
|
.format(self.max_subs))
|
|
self.logger.info('max subscriptions per session: {:,d}'
|
|
.format(self.env.max_session_subs))
|
|
if self.env.drop_client is not None:
|
|
self.logger.info('drop clients matching: {}'
|
|
.format(self.env.drop_client.pattern))
|
|
await self._start_external_servers()
|
|
|
|
async def start_rpc_server(self):
|
|
if self.env.rpc_port is not None:
|
|
await self._start_server('RPC', self.env.cs_host(for_rpc=True),
|
|
self.env.rpc_port)
|
|
|
|
async def shutdown(self):
|
|
'''Close servers and sessions.'''
|
|
self.state = self.SHUTTING_DOWN
|
|
self._close_servers(list(self.servers.keys()))
|
|
for session in self.sessions:
|
|
session.abort()
|
|
for session in list(self.sessions):
|
|
await session.wait_closed()
|
|
|
|
def session_count(self):
|
|
'''The number of connections that we've sent something to.'''
|
|
return len(self.sessions)
|
|
|
|
def notify(self, height, touched):
|
|
# Height notifications are synchronous. Those sessions with
|
|
# touched addresses are scheduled for asynchronous completion
|
|
create_task = self.controller.create_task
|
|
for session in self.sessions:
|
|
if isinstance(session, LocalRPC):
|
|
continue
|
|
session_touched = session.notify(height, touched)
|
|
if session_touched is not None:
|
|
create_task(session.notify_async(session_touched))
|
|
|
|
async def housekeeping(self):
|
|
'''Regular housekeeping checks.'''
|
|
n = 0
|
|
while True:
|
|
n += 1
|
|
await asyncio.sleep(15)
|
|
if n % 10 == 0:
|
|
self._clear_stale_sessions()
|
|
|
|
# Start listening for incoming connections if paused and
|
|
# session count has fallen
|
|
if (self.state == self.PAUSED and
|
|
len(self.sessions) <= self.low_watermark):
|
|
await self._start_external_servers()
|
|
|
|
# Periodically log sessions
|
|
if self.env.log_sessions and time.time() > self.next_log_sessions:
|
|
if self.next_log_sessions:
|
|
data = self._session_data(for_log=True)
|
|
for line in text.sessions_lines(data):
|
|
self.logger.info(line)
|
|
self.logger.info(json.dumps(self._getinfo()))
|
|
self.next_log_sessions = time.time() + self.env.log_sessions
|
|
|
|
def add_session(self, session):
|
|
self.sessions.add(session)
|
|
if (len(self.sessions) >= self.max_sessions
|
|
and self.state == self.LISTENING):
|
|
self.state = self.PAUSED
|
|
session.logger.info('maximum sessions {:,d} reached, stopping new '
|
|
'connections until count drops to {:,d}'
|
|
.format(self.max_sessions, self.low_watermark))
|
|
self._close_servers(['TCP', 'SSL'])
|
|
gid = int(session.start_time - self.start_time) // 900
|
|
if self.cur_group.gid != gid:
|
|
self.cur_group = SessionGroup(gid)
|
|
return self.cur_group
|
|
|
|
def remove_session(self, session):
|
|
'''Remove a session from our sessions list if there.'''
|
|
self.sessions.remove(session)
|
|
|
|
def new_subscription(self):
|
|
if self.subs_room <= 0:
|
|
self.subs_room = self.max_subs - self._sub_count()
|
|
if self.subs_room <= 0:
|
|
raise RPCError(BAD_REQUEST, f'server subscription limit '
|
|
f'{self.max_subs:,d} reached')
|
|
self.subs_room -= 1
|
|
|
|
|
|
class SessionBase(ServerSession):
|
|
'''Base class of ElectrumX JSON sessions.
|
|
|
|
Each session runs its tasks in asynchronous parallelism with other
|
|
sessions.
|
|
'''
|
|
|
|
MAX_CHUNK_SIZE = 2016
|
|
session_counter = itertools.count()
|
|
|
|
def __init__(self, session_mgr, controller, kind):
|
|
super().__init__(rpc_protocol=JSONRPCAutoDetect)
|
|
self.logger = util.class_logger(__name__, self.__class__.__name__)
|
|
self.session_mgr = session_mgr
|
|
self.controller = controller
|
|
self.kind = kind # 'RPC', 'TCP' etc.
|
|
self.bp = controller.bp
|
|
self.env = controller.env
|
|
self.coin = self.env.coin
|
|
self.daemon = self.bp.daemon
|
|
self.client = 'unknown'
|
|
self.anon_logs = self.env.anon_logs
|
|
self.txs_sent = 0
|
|
self.log_me = False
|
|
self.bw_limit = self.env.bandwidth_limit
|
|
self._orig_mr = self.rpc.message_received
|
|
|
|
def peer_address_str(self, *, for_log=True):
|
|
'''Returns the peer's IP address and port as a human-readable
|
|
string, respecting anon logs if the output is for a log.'''
|
|
if for_log and self.anon_logs:
|
|
return 'xx.xx.xx.xx:xx'
|
|
return super().peer_address_str()
|
|
|
|
def message_received(self, message):
|
|
self.logger.info(f'processing {message}')
|
|
self._orig_mr(message)
|
|
|
|
def toggle_logging(self):
|
|
self.log_me = not self.log_me
|
|
if self.log_me:
|
|
self.rpc.message_received = self.message_received
|
|
else:
|
|
self.rpc.message_received = self._orig_mr
|
|
|
|
def flags(self):
|
|
'''Status flags.'''
|
|
status = self.kind[0]
|
|
if self.is_closing():
|
|
status += 'C'
|
|
if self.log_me:
|
|
status += 'L'
|
|
status += str(self.concurrency.max_concurrent)
|
|
return status
|
|
|
|
def connection_made(self, transport):
|
|
'''Handle an incoming client connection.'''
|
|
super().connection_made(transport)
|
|
self.session_id = next(self.session_counter)
|
|
context = {'conn_id': f'{self.session_id}'}
|
|
self.logger = util.ConnectionLogger(self.logger, context)
|
|
self.rpc.logger = self.logger
|
|
self.group = self.session_mgr.add_session(self)
|
|
self.logger.info(f'{self.kind} {self.peer_address_str()}, '
|
|
f'{self.session_mgr.session_count():,d} total')
|
|
|
|
def connection_lost(self, exc):
|
|
'''Handle client disconnection.'''
|
|
super().connection_lost(exc)
|
|
self.session_mgr.remove_session(self)
|
|
msg = ''
|
|
if self.paused:
|
|
msg += ' whilst paused'
|
|
if self.concurrency.max_concurrent != self.max_concurrent:
|
|
msg += ' whilst throttled'
|
|
if self.send_size >= 1024*1024:
|
|
msg += ('. Sent {:,d} bytes in {:,d} messages'
|
|
.format(self.send_size, self.send_count))
|
|
if msg:
|
|
msg = 'disconnected' + msg
|
|
self.logger.info(msg)
|
|
|
|
def count_pending_items(self):
|
|
return self.rpc.pending_requests
|
|
|
|
def semaphore(self):
|
|
return Semaphores([self.concurrency.semaphore, self.group.semaphore])
|
|
|
|
def sub_count(self):
|
|
return 0
|
|
|
|
|
|
class ElectrumX(SessionBase):
|
|
'''A TCP server that handles incoming Electrum connections.'''
|
|
|
|
PROTOCOL_MIN = (1, 1)
|
|
PROTOCOL_MAX = (1, 4)
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super().__init__(*args, **kwargs)
|
|
self.subscribe_headers = False
|
|
self.subscribe_headers_raw = False
|
|
self.notified_height = None
|
|
self.max_response_size = self.env.max_send
|
|
self.max_subs = self.env.max_session_subs
|
|
self.hashX_subs = {}
|
|
self.sv_seen = False
|
|
self.mempool_statuses = {}
|
|
self.set_protocol_handlers(self.PROTOCOL_MIN)
|
|
|
|
@classmethod
|
|
def protocol_min_max_strings(cls):
|
|
return [util.version_string(ver)
|
|
for ver in (cls.PROTOCOL_MIN, cls.PROTOCOL_MAX)]
|
|
|
|
@classmethod
|
|
def server_features(cls, env):
|
|
'''Return the server features dictionary.'''
|
|
min_str, max_str = cls.protocol_min_max_strings()
|
|
return {
|
|
'hosts': env.hosts_dict(),
|
|
'pruning': None,
|
|
'server_version': electrumx.version,
|
|
'protocol_min': min_str,
|
|
'protocol_max': max_str,
|
|
'genesis_hash': env.coin.GENESIS_HASH,
|
|
'hash_function': 'sha256',
|
|
}
|
|
|
|
@classmethod
|
|
def server_version_args(cls):
|
|
'''The arguments to a server.version RPC call to a peer.'''
|
|
return [electrumx.version, cls.protocol_min_max_strings()]
|
|
|
|
def protocol_version_string(self):
|
|
return util.version_string(self.protocol_tuple)
|
|
|
|
def sub_count(self):
|
|
return len(self.hashX_subs)
|
|
|
|
async def notify_async(self, our_touched):
|
|
changed = {}
|
|
|
|
for hashX in our_touched:
|
|
alias = self.hashX_subs[hashX]
|
|
status = await self.address_status(hashX)
|
|
changed[alias] = status
|
|
|
|
# Check mempool hashXs - the status is a function of the
|
|
# confirmed state of other transactions. Note: we cannot
|
|
# iterate over mempool_statuses as it changes size.
|
|
for hashX in set(self.mempool_statuses):
|
|
old_status = self.mempool_statuses[hashX]
|
|
status = await self.address_status(hashX)
|
|
if status != old_status:
|
|
alias = self.hashX_subs[hashX]
|
|
changed[alias] = status
|
|
|
|
for alias, status in changed.items():
|
|
if len(alias) == 64:
|
|
method = 'blockchain.scripthash.subscribe'
|
|
else:
|
|
method = 'blockchain.address.subscribe'
|
|
self.send_notification(method, (alias, status))
|
|
|
|
if changed:
|
|
es = '' if len(changed) == 1 else 'es'
|
|
self.logger.info('notified of {:,d} address{}'
|
|
.format(len(changed), es))
|
|
|
|
def notify(self, height, touched):
|
|
'''Notify the client about changes to touched addresses (from mempool
|
|
updates or new blocks) and height.
|
|
|
|
Return the set of addresses the session needs to be
|
|
asyncronously notified about. This can be empty if there are
|
|
possible mempool status updates.
|
|
|
|
Returns None if nothing needs to be notified asynchronously.
|
|
'''
|
|
height_changed = height != self.notified_height
|
|
if height_changed:
|
|
self.notified_height = height
|
|
if self.subscribe_headers:
|
|
args = (self.subscribe_headers_result(height), )
|
|
self.send_notification('blockchain.headers.subscribe', args)
|
|
|
|
our_touched = touched.intersection(self.hashX_subs)
|
|
if our_touched or (height_changed and self.mempool_statuses):
|
|
return our_touched
|
|
|
|
return None
|
|
|
|
def height(self):
|
|
'''Return the current flushed database height.'''
|
|
return self.bp.db_height
|
|
|
|
def assert_boolean(self, value):
|
|
'''Return param value it is boolean otherwise raise an RPCError.'''
|
|
if value in (False, True):
|
|
return value
|
|
raise RPCError(BAD_REQUEST, f'{value} should be a boolean value')
|
|
|
|
def subscribe_headers_result(self, height):
|
|
'''The result of a header subscription for the given height.'''
|
|
if self.subscribe_headers_raw:
|
|
raw_header = self.controller.raw_header(height)
|
|
return {'hex': raw_header.hex(), 'height': height}
|
|
return self.controller.electrum_header(height)
|
|
|
|
def _headers_subscribe(self, raw):
|
|
'''Subscribe to get headers of new blocks.'''
|
|
self.subscribe_headers = True
|
|
self.subscribe_headers_raw = self.assert_boolean(raw)
|
|
self.notified_height = self.height()
|
|
return self.subscribe_headers_result(self.height())
|
|
|
|
def headers_subscribe(self):
|
|
'''Subscribe to get raw headers of new blocks.'''
|
|
return self._headers_subscribe(True)
|
|
|
|
def headers_subscribe_True(self, raw=True):
|
|
'''Subscribe to get headers of new blocks.'''
|
|
return self._headers_subscribe(raw)
|
|
|
|
def headers_subscribe_False(self, raw=False):
|
|
'''Subscribe to get headers of new blocks.'''
|
|
return self._headers_subscribe(raw)
|
|
|
|
async def add_peer(self, features):
|
|
'''Add a peer (but only if the peer resolves to the source).'''
|
|
peer_mgr = self.controller.peer_mgr
|
|
return await peer_mgr.on_add_peer(features, self.peer_address())
|
|
|
|
def peers_subscribe(self):
|
|
'''Return the server peers as a list of (ip, host, details) tuples.'''
|
|
return self.controller.peer_mgr.on_peers_subscribe(self.is_tor())
|
|
|
|
async def address_status(self, hashX):
|
|
'''Returns an address status.
|
|
|
|
Status is a hex string, but must be None if there is no history.
|
|
'''
|
|
# Note history is ordered and mempool unordered in electrum-server
|
|
# For mempool, height is -1 if unconfirmed txins, otherwise 0
|
|
history = await self.controller.get_history(hashX)
|
|
mempool = await self.controller.mempool_transactions(hashX)
|
|
|
|
status = ''.join('{}:{:d}:'.format(hash_to_hex_str(tx_hash), height)
|
|
for tx_hash, height in history)
|
|
status += ''.join('{}:{:d}:'.format(hex_hash, -unconfirmed)
|
|
for hex_hash, tx_fee, unconfirmed in mempool)
|
|
if status:
|
|
status = sha256(status.encode()).hex()
|
|
else:
|
|
status = None
|
|
|
|
if mempool:
|
|
self.mempool_statuses[hashX] = status
|
|
else:
|
|
self.mempool_statuses.pop(hashX, None)
|
|
|
|
return status
|
|
|
|
async def hashX_listunspent(self, hashX):
|
|
'''Return the list of UTXOs of a script hash, including mempool
|
|
effects.'''
|
|
utxos = await self.controller.get_utxos(hashX)
|
|
utxos = sorted(utxos)
|
|
utxos.extend(self.controller.mempool.get_utxos(hashX))
|
|
spends = await self.controller.mempool.potential_spends(hashX)
|
|
|
|
return [{'tx_hash': hash_to_hex_str(utxo.tx_hash),
|
|
'tx_pos': utxo.tx_pos,
|
|
'height': utxo.height, 'value': utxo.value}
|
|
for utxo in utxos
|
|
if (utxo.tx_hash, utxo.tx_pos) not in spends]
|
|
|
|
async def hashX_subscribe(self, hashX, alias):
|
|
# First check our limit.
|
|
if len(self.hashX_subs) >= self.max_subs:
|
|
raise RPCError(BAD_REQUEST, 'your address subscription limit '
|
|
f'{self.max_subs:,d} reached')
|
|
|
|
# Now let the controller check its limit
|
|
self.session_mgr.new_subscription()
|
|
self.hashX_subs[hashX] = alias
|
|
return await self.address_status(hashX)
|
|
|
|
def address_to_hashX(self, address):
|
|
try:
|
|
return self.coin.address_to_hashX(address)
|
|
except Exception:
|
|
pass
|
|
raise RPCError(BAD_REQUEST, f'{address} is not a valid address')
|
|
|
|
async def address_get_balance(self, address):
|
|
'''Return the confirmed and unconfirmed balance of an address.'''
|
|
hashX = self.address_to_hashX(address)
|
|
return await self.get_balance(hashX)
|
|
|
|
async def address_get_history(self, address):
|
|
'''Return the confirmed and unconfirmed history of an address.'''
|
|
hashX = self.address_to_hashX(address)
|
|
return await self.confirmed_and_unconfirmed_history(hashX)
|
|
|
|
async def address_get_mempool(self, address):
|
|
'''Return the mempool transactions touching an address.'''
|
|
hashX = self.address_to_hashX(address)
|
|
return await self.unconfirmed_history(hashX)
|
|
|
|
async def address_listunspent(self, address):
|
|
'''Return the list of UTXOs of an address.'''
|
|
hashX = self.address_to_hashX(address)
|
|
return await self.hashX_listunspent(hashX)
|
|
|
|
async def address_subscribe(self, address):
|
|
'''Subscribe to an address.
|
|
|
|
address: the address to subscribe to'''
|
|
hashX = self.address_to_hashX(address)
|
|
return await self.hashX_subscribe(hashX, address)
|
|
|
|
async def get_balance(self, hashX):
|
|
utxos = await self.controller.get_utxos(hashX)
|
|
confirmed = sum(utxo.value for utxo in utxos)
|
|
unconfirmed = self.controller.mempool_value(hashX)
|
|
return {'confirmed': confirmed, 'unconfirmed': unconfirmed}
|
|
|
|
async def scripthash_get_balance(self, scripthash):
|
|
'''Return the confirmed and unconfirmed balance of a scripthash.'''
|
|
hashX = scripthash_to_hashX(scripthash)
|
|
return await self.get_balance(hashX)
|
|
|
|
async def unconfirmed_history(self, hashX):
|
|
# Note unconfirmed history is unordered in electrum-server
|
|
# Height is -1 if unconfirmed txins, otherwise 0
|
|
mempool = await self.controller.mempool_transactions(hashX)
|
|
return [{'tx_hash': tx_hash, 'height': -unconfirmed, 'fee': fee}
|
|
for tx_hash, fee, unconfirmed in mempool]
|
|
|
|
async def confirmed_and_unconfirmed_history(self, hashX):
|
|
# Note history is ordered but unconfirmed is unordered in e-s
|
|
history = await self.controller.get_history(hashX)
|
|
conf = [{'tx_hash': hash_to_hex_str(tx_hash), 'height': height}
|
|
for tx_hash, height in history]
|
|
return conf + await self.unconfirmed_history(hashX)
|
|
|
|
async def scripthash_get_history(self, scripthash):
|
|
'''Return the confirmed and unconfirmed history of a scripthash.'''
|
|
hashX = scripthash_to_hashX(scripthash)
|
|
return await self.confirmed_and_unconfirmed_history(hashX)
|
|
|
|
async def scripthash_get_mempool(self, scripthash):
|
|
'''Return the mempool transactions touching a scripthash.'''
|
|
hashX = scripthash_to_hashX(scripthash)
|
|
return await self.unconfirmed_history(hashX)
|
|
|
|
async def scripthash_listunspent(self, scripthash):
|
|
'''Return the list of UTXOs of a scripthash.'''
|
|
hashX = scripthash_to_hashX(scripthash)
|
|
return await self.hashX_listunspent(hashX)
|
|
|
|
async def scripthash_subscribe(self, scripthash):
|
|
'''Subscribe to a script hash.
|
|
|
|
scripthash: the SHA256 hash of the script to subscribe to'''
|
|
hashX = scripthash_to_hashX(scripthash)
|
|
return await self.hashX_subscribe(hashX, scripthash)
|
|
|
|
def _merkle_proof(self, cp_height, height):
|
|
max_height = self.height()
|
|
if not height <= cp_height <= max_height:
|
|
raise RPCError(BAD_REQUEST,
|
|
f'require header height {height:,d} <= '
|
|
f'cp_height {cp_height:,d} <= '
|
|
f'chain height {max_height:,d}')
|
|
branch, root = self.bp.header_mc.branch_and_root(cp_height + 1, height)
|
|
return {
|
|
'branch': [hash_to_hex_str(elt) for elt in branch],
|
|
'root': hash_to_hex_str(root),
|
|
}
|
|
|
|
def block_header(self, height, cp_height=0):
|
|
'''Return a raw block header as a hexadecimal string, or as a
|
|
dictionary with a merkle proof.'''
|
|
height = non_negative_integer(height)
|
|
cp_height = non_negative_integer(cp_height)
|
|
raw_header_hex = self.controller.raw_header(height).hex()
|
|
if cp_height == 0:
|
|
return raw_header_hex
|
|
result = {'header': raw_header_hex}
|
|
result.update(self._merkle_proof(cp_height, height))
|
|
return result
|
|
|
|
def block_header_13(self, height):
|
|
'''Return a raw block header as a hexadecimal string.
|
|
|
|
height: the header's height'''
|
|
return self.block_header(height)
|
|
|
|
def block_headers(self, start_height, count, cp_height=0):
|
|
'''Return count concatenated block headers as hex for the main chain;
|
|
starting at start_height.
|
|
|
|
start_height and count must be non-negative integers. At most
|
|
MAX_CHUNK_SIZE headers will be returned.
|
|
'''
|
|
start_height = non_negative_integer(start_height)
|
|
count = non_negative_integer(count)
|
|
cp_height = non_negative_integer(cp_height)
|
|
|
|
max_size = self.MAX_CHUNK_SIZE
|
|
count = min(count, max_size)
|
|
headers, count = self.bp.read_headers(start_height, count)
|
|
result = {'hex': headers.hex(), 'count': count, 'max': max_size}
|
|
if count and cp_height:
|
|
last_height = start_height + count - 1
|
|
result.update(self._merkle_proof(cp_height, last_height))
|
|
return result
|
|
|
|
def block_headers_12(self, start_height, count):
|
|
return self.block_headers(start_height, count)
|
|
|
|
def block_get_chunk(self, index):
|
|
'''Return a chunk of block headers as a hexadecimal string.
|
|
|
|
index: the chunk index'''
|
|
index = non_negative_integer(index)
|
|
chunk_size = self.coin.CHUNK_SIZE
|
|
start_height = index * chunk_size
|
|
headers, count = self.bp.read_headers(start_height, chunk_size)
|
|
return headers.hex()
|
|
|
|
def block_get_header(self, height):
|
|
'''The deserialized header at a given height.
|
|
|
|
height: the header's height'''
|
|
height = non_negative_integer(height)
|
|
return self.controller.electrum_header(height)
|
|
|
|
def is_tor(self):
|
|
'''Try to detect if the connection is to a tor hidden service we are
|
|
running.'''
|
|
peername = self.controller.peer_mgr.proxy_peername()
|
|
if not peername:
|
|
return False
|
|
peer_address = self.peer_address()
|
|
return peer_address and peer_address[0] == peername[0]
|
|
|
|
async def replaced_banner(self, banner):
|
|
network_info = await self.controller.daemon_request('getnetworkinfo')
|
|
ni_version = network_info['version']
|
|
major, minor = divmod(ni_version, 1000000)
|
|
minor, revision = divmod(minor, 10000)
|
|
revision //= 100
|
|
daemon_version = '{:d}.{:d}.{:d}'.format(major, minor, revision)
|
|
for pair in [
|
|
('$SERVER_VERSION', electrumx.version_short),
|
|
('$SERVER_SUBVERSION', electrumx.version),
|
|
('$DAEMON_VERSION', daemon_version),
|
|
('$DAEMON_SUBVERSION', network_info['subversion']),
|
|
('$DONATION_ADDRESS', self.env.donation_address),
|
|
]:
|
|
banner = banner.replace(*pair)
|
|
return banner
|
|
|
|
def donation_address(self):
|
|
'''Return the donation address as a string, empty if there is none.'''
|
|
return self.env.donation_address
|
|
|
|
async def banner(self):
|
|
'''Return the server banner text.'''
|
|
banner = 'Welcome to Electrum!'
|
|
|
|
if self.is_tor():
|
|
banner_file = self.env.tor_banner_file
|
|
else:
|
|
banner_file = self.env.banner_file
|
|
if banner_file:
|
|
try:
|
|
with codecs.open(banner_file, 'r', 'utf-8') as f:
|
|
banner = f.read()
|
|
except Exception as e:
|
|
self.logger.error(f'reading banner file {banner_file}: {e}')
|
|
else:
|
|
banner = await self.replaced_banner(banner)
|
|
|
|
return banner
|
|
|
|
def mempool_get_fee_histogram(self):
|
|
'''Memory pool fee histogram.'''
|
|
return self.controller.mempool.get_fee_histogram()
|
|
|
|
async def relayfee(self):
|
|
'''The minimum fee a low-priority tx must pay in order to be accepted
|
|
to the daemon's memory pool.'''
|
|
return await self.controller.daemon_request('relayfee')
|
|
|
|
async def estimatefee(self, number):
|
|
'''The estimated transaction fee per kilobyte to be paid for a
|
|
transaction to be included within a certain number of blocks.
|
|
|
|
number: the number of blocks
|
|
'''
|
|
number = non_negative_integer(number)
|
|
return await self.controller.daemon_request('estimatefee', [number])
|
|
|
|
def ping(self):
|
|
'''Serves as a connection keep-alive mechanism and for the client to
|
|
confirm the server is still responding.
|
|
'''
|
|
return None
|
|
|
|
def server_version(self, client_name='', protocol_version=None):
|
|
'''Returns the server version as a string.
|
|
|
|
client_name: a string identifying the client
|
|
protocol_version: the protocol version spoken by the client
|
|
'''
|
|
if self.sv_seen and self.protocol_tuple >= (1, 4):
|
|
raise RPCError(BAD_REQUEST, f'server.version already sent')
|
|
self.sv_seen = True
|
|
|
|
if client_name:
|
|
client_name = str(client_name)
|
|
if self.env.drop_client is not None and \
|
|
self.env.drop_client.match(client_name):
|
|
self.close_after_send = True
|
|
raise RPCError(BAD_REQUEST,
|
|
f'unsupported client: {client_name}')
|
|
self.client = client_name[:17]
|
|
|
|
# Find the highest common protocol version. Disconnect if
|
|
# that protocol version in unsupported.
|
|
ptuple, client_min = util.protocol_version(
|
|
protocol_version, self.PROTOCOL_MIN, self.PROTOCOL_MAX)
|
|
if ptuple is None:
|
|
if client_min > self.PROTOCOL_MIN:
|
|
self.logger.info(f'client requested future protocol version '
|
|
f'{util.version_string(client_min)} '
|
|
f'- is your software out of date?')
|
|
self.close_after_send = True
|
|
raise RPCError(BAD_REQUEST,
|
|
f'unsupported protocol version: {protocol_version}')
|
|
self.set_protocol_handlers(ptuple)
|
|
|
|
return (electrumx.version, self.protocol_version_string())
|
|
|
|
async def transaction_broadcast(self, raw_tx):
|
|
'''Broadcast a raw transaction to the network.
|
|
|
|
raw_tx: the raw transaction as a hexadecimal string'''
|
|
# This returns errors as JSON RPC errors, as is natural
|
|
try:
|
|
tx_hash = await self.daemon.sendrawtransaction([raw_tx])
|
|
self.txs_sent += 1
|
|
self.session_mgr.txs_sent += 1
|
|
self.logger.info('sent tx: {}'.format(tx_hash))
|
|
return tx_hash
|
|
except DaemonError as e:
|
|
error, = e.args
|
|
message = error['message']
|
|
self.logger.info('sendrawtransaction: {}'.format(message))
|
|
raise RPCError(BAD_REQUEST, 'the transaction was rejected by '
|
|
f'network rules.\n\n{message}\n[{raw_tx}]')
|
|
|
|
def set_protocol_handlers(self, ptuple):
|
|
self.protocol_tuple = ptuple
|
|
|
|
controller = self.controller
|
|
handlers = {
|
|
'blockchain.block.get_chunk': self.block_get_chunk,
|
|
'blockchain.block.get_header': self.block_get_header,
|
|
'blockchain.estimatefee': self.estimatefee,
|
|
'blockchain.relayfee': self.relayfee,
|
|
'blockchain.scripthash.get_balance': self.scripthash_get_balance,
|
|
'blockchain.scripthash.get_history': self.scripthash_get_history,
|
|
'blockchain.scripthash.get_mempool': self.scripthash_get_mempool,
|
|
'blockchain.scripthash.listunspent': self.scripthash_listunspent,
|
|
'blockchain.scripthash.subscribe': self.scripthash_subscribe,
|
|
'blockchain.transaction.broadcast': self.transaction_broadcast,
|
|
'blockchain.transaction.get': controller.transaction_get,
|
|
'blockchain.transaction.get_merkle':
|
|
controller.transaction_get_merkle,
|
|
'server.add_peer': self.add_peer,
|
|
'server.banner': self.banner,
|
|
'server.donation_address': self.donation_address,
|
|
'server.features': partial(self.server_features, self.env),
|
|
'server.peers.subscribe': self.peers_subscribe,
|
|
'server.version': self.server_version,
|
|
}
|
|
|
|
if ptuple >= (1, 2):
|
|
# New handler as of 1.2
|
|
handlers.update({
|
|
'mempool.get_fee_histogram': self.mempool_get_fee_histogram,
|
|
'blockchain.block.headers': self.block_headers_12,
|
|
'server.ping': self.ping,
|
|
})
|
|
|
|
if ptuple >= (1, 4):
|
|
handlers.update({
|
|
'blockchain.block.header': self.block_header,
|
|
'blockchain.block.headers': self.block_headers,
|
|
'blockchain.headers.subscribe': self.headers_subscribe,
|
|
})
|
|
elif ptuple >= (1, 3):
|
|
handlers.update({
|
|
'blockchain.block.header': self.block_header_13,
|
|
'blockchain.headers.subscribe': self.headers_subscribe_True,
|
|
})
|
|
else:
|
|
handlers.update({
|
|
'blockchain.headers.subscribe': self.headers_subscribe_False,
|
|
'blockchain.address.get_balance': self.address_get_balance,
|
|
'blockchain.address.get_history': self.address_get_history,
|
|
'blockchain.address.get_mempool': self.address_get_mempool,
|
|
'blockchain.address.listunspent': self.address_listunspent,
|
|
'blockchain.address.subscribe': self.address_subscribe,
|
|
})
|
|
|
|
self.electrumx_handlers = handlers
|
|
|
|
def request_handler(self, method):
|
|
'''Return the async handler for the given request method.'''
|
|
return self.electrumx_handlers.get(method)
|
|
|
|
|
|
class LocalRPC(SessionBase):
|
|
'''A local TCP RPC server session.'''
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super().__init__(*args, **kwargs)
|
|
self.client = 'RPC'
|
|
self.max_response_size = 0
|
|
|
|
def protocol_version_string(self):
|
|
return 'RPC'
|
|
|
|
def request_handler(self, method):
|
|
'''Return the async handler for the given request method.'''
|
|
return self.session_mgr.rpc_handlers.get(method)
|
|
|
|
|
|
class DashElectrumX(ElectrumX):
|
|
'''A TCP server that handles incoming Electrum Dash connections.'''
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super().__init__(*args, **kwargs)
|
|
self.mns = set()
|
|
|
|
def set_protocol_handlers(self, ptuple):
|
|
super().set_protocol_handlers(ptuple)
|
|
self.electrumx_handlers.update({
|
|
'masternode.announce.broadcast':
|
|
self.masternode_announce_broadcast,
|
|
'masternode.subscribe': self.masternode_subscribe,
|
|
'masternode.list': self.masternode_list
|
|
})
|
|
|
|
async def notify_masternodes_async(self):
|
|
for masternode in self.mns:
|
|
status = await self.daemon.masternode_list(['status', masternode])
|
|
self.send_notification('masternode.subscribe',
|
|
[masternode, status.get(masternode)])
|
|
|
|
def notify(self, height, touched):
|
|
'''Notify the client about changes in masternode list.'''
|
|
result = super().notify(height, touched)
|
|
self.controller.create_task(self.notify_masternodes_async())
|
|
return result
|
|
|
|
# Masternode command handlers
|
|
async def masternode_announce_broadcast(self, signmnb):
|
|
'''Pass through the masternode announce message to be broadcast
|
|
by the daemon.
|
|
|
|
signmnb: signed masternode broadcast message.'''
|
|
try:
|
|
return await self.daemon.masternode_broadcast(['relay', signmnb])
|
|
except DaemonError as e:
|
|
error, = e.args
|
|
message = error['message']
|
|
self.logger.info('masternode_broadcast: {}'.format(message))
|
|
raise RPCError(BAD_REQUEST, 'the masternode broadcast was '
|
|
f'rejected.\n\n{message}\n[{signmnb}]')
|
|
|
|
async def masternode_subscribe(self, collateral):
|
|
'''Returns the status of masternode.
|
|
|
|
collateral: masternode collateral.
|
|
'''
|
|
result = await self.daemon.masternode_list(['status', collateral])
|
|
if result is not None:
|
|
self.mns.add(collateral)
|
|
return result.get(collateral)
|
|
return None
|
|
|
|
async def masternode_list(self, payees):
|
|
'''
|
|
Returns the list of masternodes.
|
|
|
|
payees: a list of masternode payee addresses.
|
|
'''
|
|
if not isinstance(payees, list):
|
|
raise RPCError(BAD_REQUEST, 'expected a list of payees')
|
|
|
|
result = []
|
|
|
|
def get_masternode_payment_queue(mns):
|
|
'''Returns the calculated position in the payment queue for all the
|
|
valid masterernodes in the given mns list.
|
|
|
|
mns: a list of masternodes information.
|
|
'''
|
|
now = int(datetime.datetime.utcnow().strftime("%s"))
|
|
mn_queue = []
|
|
|
|
# Only ENABLED masternodes are considered for the list.
|
|
for line in mns:
|
|
mnstat = mns[line].split()
|
|
if mnstat[0] == 'ENABLED':
|
|
# if last paid time == 0
|
|
if int(mnstat[5]) == 0:
|
|
# use active seconds
|
|
mnstat.append(int(mnstat[4]))
|
|
else:
|
|
# now minus last paid
|
|
delta = now - int(mnstat[5])
|
|
# if > active seconds, use active seconds
|
|
if delta >= int(mnstat[4]):
|
|
mnstat.append(int(mnstat[4]))
|
|
# use active seconds
|
|
else:
|
|
mnstat.append(delta)
|
|
mn_queue.append(mnstat)
|
|
mn_queue = sorted(mn_queue, key=lambda x: x[8], reverse=True)
|
|
return mn_queue
|
|
|
|
def get_payment_position(payment_queue, address):
|
|
'''
|
|
Returns the position of the payment list for the given address.
|
|
|
|
payment_queue: position in the payment queue for the masternode.
|
|
address: masternode payee address.
|
|
'''
|
|
position = -1
|
|
for pos, mn in enumerate(payment_queue, start=1):
|
|
if mn[2] == address:
|
|
position = pos
|
|
break
|
|
return position
|
|
|
|
# Accordingly with the masternode payment queue, a custom list
|
|
# with the masternode information including the payment
|
|
# position is returned.
|
|
if (self.controller.cache_mn_height != self.height()
|
|
or not self.controller.mn_cache):
|
|
self.controller.cache_mn_height = self.height()
|
|
self.controller.mn_cache.clear()
|
|
full_mn_list = await self.daemon.masternode_list(['full'])
|
|
mn_payment_queue = get_masternode_payment_queue(full_mn_list)
|
|
mn_payment_count = len(mn_payment_queue)
|
|
mn_list = []
|
|
for key, value in full_mn_list.items():
|
|
mn_data = value.split()
|
|
mn_info = {}
|
|
mn_info['vin'] = key
|
|
mn_info['status'] = mn_data[0]
|
|
mn_info['protocol'] = mn_data[1]
|
|
mn_info['payee'] = mn_data[2]
|
|
mn_info['lastseen'] = mn_data[3]
|
|
mn_info['activeseconds'] = mn_data[4]
|
|
mn_info['lastpaidtime'] = mn_data[5]
|
|
mn_info['lastpaidblock'] = mn_data[6]
|
|
mn_info['ip'] = mn_data[7]
|
|
mn_info['paymentposition'] = get_payment_position(
|
|
mn_payment_queue, mn_info['payee'])
|
|
mn_info['inselection'] = (
|
|
mn_info['paymentposition'] < mn_payment_count // 10)
|
|
balance = await self.address_get_balance(mn_info['payee'])
|
|
mn_info['balance'] = (sum(balance.values())
|
|
/ self.coin.VALUE_PER_COIN)
|
|
mn_list.append(mn_info)
|
|
self.controller.mn_cache = mn_list
|
|
|
|
# If payees is an empty list the whole masternode list is returned
|
|
if payees:
|
|
result = [mn for mn in self.controller.mn_cache
|
|
for address in payees if mn['payee'] == address]
|
|
else:
|
|
result = self.controller.mn_cache
|
|
|
|
return result
|
|
|