mirror of
https://github.com/bitcoin/bitcoin.git
synced 2025-01-27 03:33:27 -03:00
411ba32af2
bc9283c441
[test] Add functional test to test early key response behaviour in BIP 324 (stratospher)ffe6a56d75
[test] Check whether v2 TestNode performs downgrading (stratospher)ba737358a3
[test] Add functional tests to test v2 P2P behaviour (stratospher)4115cf9956
[test] Ignore BIP324 decoy messages (stratospher)8c054aa04d
[test] Allow inbound and outbound connections supporting v2 P2P protocol (stratospher)382894c3ac
[test] Reconnect using v1 P2P when v2 P2P terminates due to magic byte mismatch (stratospher)a94e350ac0
[test] Build v2 P2P messages (stratospher)bb7bffed79
[test] Use lock for sending P2P messages in test framework (stratospher)5b91fb14ab
[test] Read v2 P2P messages (stratospher)05bddb20f5
[test] Perform initial v2 handshake (stratospher)a049d1bd08
[test] Introduce EncryptedP2PState object in P2PConnection (stratospher)b89fa59e71
[test] Construct class to handle v2 P2P protocol functions (stratospher)8d6c848a48
[test] Move MAGIC_BYTES to messages.py (stratospher)595ad4b168
[test/crypto] Add ECDH (stratospher)4487b80517
[rpc/net] Allow v2 p2p support in addconnection (stratospher) Pull request description: This PR introduces support for v2 P2P encryption(BIP 324) in the existing functional test framework and adds functional tests for the same. ### commits overview 1. introduces a new class `EncryptedP2PState` to store the keys, functions for performing the initial v2 handshake and encryption/decryption. 3. this class is used by `P2PConnection` in inbound/outbound connections to perform the initial v2 handshake before the v1 version handshake. Only after the initial v2 handshake is performed do application layer P2P messages(version, verack etc..) get exchanged. (in a v2 connection) - `v2_state` is the object of class `EncryptedP2PState` in `P2PConnection` used to store its keys, session-id etc. - a node [advertising](https://github.com/stratospher/blogosphere/blob/main/integration_test_bip324.md#advertising-to-support-v2-p2p) support for v2 P2P is different from a node actually [supporting v2 P2P](https://github.com/stratospher/blogosphere/blob/main/integration_test_bip324.md#supporting-v2-p2p) (differ when false advertisement of services occur) - introduce a boolean variable `supports_v2_p2p` in `P2PConnection` to denote if it supports v2 P2P. - introduce a boolean variable `advertises_v2_p2p` to denote whether `P2PConnection` which mimics peer behaviour advertises V2 P2P support. Default option is `False`. - In the test framework, you can create Inbound and Outbound connections to `TestNode` 1. During **Inbound Connections**, `P2PConnection` is the initiator [`TestNode` <--------- `P2PConnection`] - Case 1: - if the `TestNode` advertises/signals v2 P2P support (means `self.nodes[i]` set up with `"-v2transport=1"`), different behaviour will be exhibited based on whether: 1. `P2PConnection` supports v2 P2P 2. `P2PConnection` does not support v2 P2P - In a real world scenario, the initiator node would intrinsically know if they support v2 P2P based on whatever code they choose to run. However, in the test scenario where we mimic peer behaviour, we have no way of knowing if `P2PConnection` should support v2 P2P or not. So `supports_v2_p2p` boolean variable is used as an option to enable support for v2 P2P in `P2PConnection`. - Since the `TestNode` advertises v2 P2P support (using "-v2transport=1"), our initiator `P2PConnection` would send: 1. (if the `P2PConnection` supports v2 P2P) ellswift + garbage bytes to initiate the connection 2. (if the `P2PConnection` does not support v2 P2P) version message to initiate the connection - Case 2: - if the `TestNode` doesn't signal v2 P2P support; `P2PConnection` being the initiator would send version message to initiate a connection. 2. During **Outbound Connections** [TestNode --------> P2PConnection] - initiator `TestNode` would send: - (if the `P2PConnection` advertises v2 P2P) ellswift + garbage bytes to initiate the connection - (if the `P2PConnection` advertises v2 P2P) version message to initiate the connection - Suppose `P2PConnection` advertises v2 P2P support when it actually doesn't support v2 P2P (false advertisement scenario) - `TestNode` sends ellswift + garbage bytes - `P2PConnection` receives but can't process it and disconnects. - `TestNode` then tries using v1 P2P and sends version message - `P2PConnection` receives/processes this successfully and they communicate on v1 P2P 4. the encrypted P2P messages follow a different format - 3 byte length + 1-13 byte message_type + payload + 16 byte MAC 5. includes support for testing decoy messages and v2 connection downgrade(using false advertisement - when a v2 node makes an outbound connection to a node which doesn't support v2 but is advertised as v2 by some malicious intermediary) ### run the tests * functional test - `test/functional/p2p_v2_encrypted.py` `test/functional/p2p_v2_earlykeyresponse.py` I'm also super grateful to @ dhruv for his really valuable feedback on this branch. Also written a more elaborate explanation here - https://github.com/stratospher/blogosphere/blob/main/integration_test_bip324.md ACKs for top commit: naumenkogs: ACKbc9283c441
mzumsande: Code Review ACKbc9283c441
theStack: Code-review ACKbc9283c441
glozow: ACKbc9283c441
Tree-SHA512: 9b54ed27e925e1775e0e0d35e959cdbf2a9a1aab7bcf5d027e66f8b59780bdd0458a7a4311ddc7dd67657a4a2a2cd5034ead75524420d58a83f642a8304c9811
960 lines
43 KiB
Python
Executable file
960 lines
43 KiB
Python
Executable file
#!/usr/bin/env python3
|
|
# Copyright (c) 2017-2022 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
"""Class for bitcoind node under test"""
|
|
|
|
import contextlib
|
|
import decimal
|
|
import errno
|
|
from enum import Enum
|
|
import http.client
|
|
import json
|
|
import logging
|
|
import os
|
|
import platform
|
|
import re
|
|
import subprocess
|
|
import tempfile
|
|
import time
|
|
import urllib.parse
|
|
import collections
|
|
import shlex
|
|
from pathlib import Path
|
|
|
|
from .authproxy import (
|
|
JSONRPCException,
|
|
serialization_fallback,
|
|
)
|
|
from .descriptors import descsum_create
|
|
from .messages import NODE_P2P_V2
|
|
from .p2p import P2P_SERVICES, P2P_SUBVERSION
|
|
from .util import (
|
|
MAX_NODES,
|
|
assert_equal,
|
|
append_config,
|
|
delete_cookie_file,
|
|
get_auth_cookie,
|
|
get_rpc_proxy,
|
|
rpc_url,
|
|
wait_until_helper_internal,
|
|
p2p_port,
|
|
)
|
|
|
|
BITCOIND_PROC_WAIT_TIMEOUT = 60
|
|
|
|
|
|
class FailedToStartError(Exception):
|
|
"""Raised when a node fails to start correctly."""
|
|
|
|
|
|
class ErrorMatch(Enum):
|
|
FULL_TEXT = 1
|
|
FULL_REGEX = 2
|
|
PARTIAL_REGEX = 3
|
|
|
|
|
|
class TestNode():
|
|
"""A class for representing a bitcoind node under test.
|
|
|
|
This class contains:
|
|
|
|
- state about the node (whether it's running, etc)
|
|
- a Python subprocess.Popen object representing the running process
|
|
- an RPC connection to the node
|
|
- one or more P2P connections to the node
|
|
|
|
|
|
To make things easier for the test writer, any unrecognised messages will
|
|
be dispatched to the RPC connection."""
|
|
|
|
def __init__(self, i, datadir_path, *, chain, rpchost, timewait, timeout_factor, bitcoind, bitcoin_cli, coverage_dir, cwd, extra_conf=None, extra_args=None, use_cli=False, start_perf=False, use_valgrind=False, version=None, descriptors=False, v2transport=False):
|
|
"""
|
|
Kwargs:
|
|
start_perf (bool): If True, begin profiling the node with `perf` as soon as
|
|
the node starts.
|
|
"""
|
|
|
|
self.index = i
|
|
self.p2p_conn_index = 1
|
|
self.datadir_path = datadir_path
|
|
self.bitcoinconf = self.datadir_path / "bitcoin.conf"
|
|
self.stdout_dir = self.datadir_path / "stdout"
|
|
self.stderr_dir = self.datadir_path / "stderr"
|
|
self.chain = chain
|
|
self.rpchost = rpchost
|
|
self.rpc_timeout = timewait
|
|
self.binary = bitcoind
|
|
self.coverage_dir = coverage_dir
|
|
self.cwd = cwd
|
|
self.descriptors = descriptors
|
|
if extra_conf is not None:
|
|
append_config(self.datadir_path, extra_conf)
|
|
# Most callers will just need to add extra args to the standard list below.
|
|
# For those callers that need more flexibility, they can just set the args property directly.
|
|
# Note that common args are set in the config file (see initialize_datadir)
|
|
self.extra_args = extra_args
|
|
self.version = version
|
|
# Configuration for logging is set as command-line args rather than in the bitcoin.conf file.
|
|
# This means that starting a bitcoind using the temp dir to debug a failed test won't
|
|
# spam debug.log.
|
|
self.args = [
|
|
self.binary,
|
|
f"-datadir={self.datadir_path}",
|
|
"-logtimemicros",
|
|
"-debug",
|
|
"-debugexclude=libevent",
|
|
"-debugexclude=leveldb",
|
|
"-debugexclude=rand",
|
|
"-uacomment=testnode%d" % i,
|
|
]
|
|
if self.descriptors is None:
|
|
self.args.append("-disablewallet")
|
|
|
|
# Use valgrind, expect for previous release binaries
|
|
if use_valgrind and version is None:
|
|
default_suppressions_file = Path(__file__).parents[3] / "contrib" / "valgrind.supp"
|
|
suppressions_file = os.getenv("VALGRIND_SUPPRESSIONS_FILE",
|
|
default_suppressions_file)
|
|
self.args = ["valgrind", "--suppressions={}".format(suppressions_file),
|
|
"--gen-suppressions=all", "--exit-on-first-error=yes",
|
|
"--error-exitcode=1", "--quiet"] + self.args
|
|
|
|
if self.version_is_at_least(190000):
|
|
self.args.append("-logthreadnames")
|
|
if self.version_is_at_least(219900):
|
|
self.args.append("-logsourcelocations")
|
|
if self.version_is_at_least(239000):
|
|
self.args.append("-loglevel=trace")
|
|
|
|
# Default behavior from global -v2transport flag is added to args to persist it over restarts.
|
|
# May be overwritten in individual tests, using extra_args.
|
|
self.default_to_v2 = v2transport
|
|
if self.default_to_v2:
|
|
self.args.append("-v2transport=1")
|
|
|
|
self.cli = TestNodeCLI(bitcoin_cli, self.datadir_path)
|
|
self.use_cli = use_cli
|
|
self.start_perf = start_perf
|
|
|
|
self.running = False
|
|
self.process = None
|
|
self.rpc_connected = False
|
|
self.rpc = None
|
|
self.url = None
|
|
self.log = logging.getLogger('TestFramework.node%d' % i)
|
|
self.cleanup_on_exit = True # Whether to kill the node when this object goes away
|
|
# Cache perf subprocesses here by their data output filename.
|
|
self.perf_subprocesses = {}
|
|
|
|
self.p2ps = []
|
|
self.timeout_factor = timeout_factor
|
|
|
|
self.mocktime = None
|
|
|
|
AddressKeyPair = collections.namedtuple('AddressKeyPair', ['address', 'key'])
|
|
PRIV_KEYS = [
|
|
# address , privkey
|
|
AddressKeyPair('mjTkW3DjgyZck4KbiRusZsqTgaYTxdSz6z', 'cVpF924EspNh8KjYsfhgY96mmxvT6DgdWiTYMtMjuM74hJaU5psW'),
|
|
AddressKeyPair('msX6jQXvxiNhx3Q62PKeLPrhrqZQdSimTg', 'cUxsWyKyZ9MAQTaAhUQWJmBbSvHMwSmuv59KgxQV7oZQU3PXN3KE'),
|
|
AddressKeyPair('mnonCMyH9TmAsSj3M59DsbH8H63U3RKoFP', 'cTrh7dkEAeJd6b3MRX9bZK8eRmNqVCMH3LSUkE3dSFDyzjU38QxK'),
|
|
AddressKeyPair('mqJupas8Dt2uestQDvV2NH3RU8uZh2dqQR', 'cVuKKa7gbehEQvVq717hYcbE9Dqmq7KEBKqWgWrYBa2CKKrhtRim'),
|
|
AddressKeyPair('msYac7Rvd5ywm6pEmkjyxhbCDKqWsVeYws', 'cQDCBuKcjanpXDpCqacNSjYfxeQj8G6CAtH1Dsk3cXyqLNC4RPuh'),
|
|
AddressKeyPair('n2rnuUnwLgXqf9kk2kjvVm8R5BZK1yxQBi', 'cQakmfPSLSqKHyMFGwAqKHgWUiofJCagVGhiB4KCainaeCSxeyYq'),
|
|
AddressKeyPair('myzuPxRwsf3vvGzEuzPfK9Nf2RfwauwYe6', 'cQMpDLJwA8DBe9NcQbdoSb1BhmFxVjWD5gRyrLZCtpuF9Zi3a9RK'),
|
|
AddressKeyPair('mumwTaMtbxEPUswmLBBN3vM9oGRtGBrys8', 'cSXmRKXVcoouhNNVpcNKFfxsTsToY5pvB9DVsFksF1ENunTzRKsy'),
|
|
AddressKeyPair('mpV7aGShMkJCZgbW7F6iZgrvuPHjZjH9qg', 'cSoXt6tm3pqy43UMabY6eUTmR3eSUYFtB2iNQDGgb3VUnRsQys2k'),
|
|
AddressKeyPair('mq4fBNdckGtvY2mijd9am7DRsbRB4KjUkf', 'cN55daf1HotwBAgAKWVgDcoppmUNDtQSfb7XLutTLeAgVc3u8hik'),
|
|
AddressKeyPair('mpFAHDjX7KregM3rVotdXzQmkbwtbQEnZ6', 'cT7qK7g1wkYEMvKowd2ZrX1E5f6JQ7TM246UfqbCiyF7kZhorpX3'),
|
|
AddressKeyPair('mzRe8QZMfGi58KyWCse2exxEFry2sfF2Y7', 'cPiRWE8KMjTRxH1MWkPerhfoHFn5iHPWVK5aPqjW8NxmdwenFinJ'),
|
|
]
|
|
|
|
def get_deterministic_priv_key(self):
|
|
"""Return a deterministic priv key in base58, that only depends on the node's index"""
|
|
assert len(self.PRIV_KEYS) == MAX_NODES
|
|
return self.PRIV_KEYS[self.index]
|
|
|
|
def _node_msg(self, msg: str) -> str:
|
|
"""Return a modified msg that identifies this node by its index as a debugging aid."""
|
|
return "[node %d] %s" % (self.index, msg)
|
|
|
|
def _raise_assertion_error(self, msg: str):
|
|
"""Raise an AssertionError with msg modified to identify this node."""
|
|
raise AssertionError(self._node_msg(msg))
|
|
|
|
def __del__(self):
|
|
# Ensure that we don't leave any bitcoind processes lying around after
|
|
# the test ends
|
|
if self.process and self.cleanup_on_exit:
|
|
# Should only happen on test failure
|
|
# Avoid using logger, as that may have already been shutdown when
|
|
# this destructor is called.
|
|
print(self._node_msg("Cleaning up leftover process"))
|
|
self.process.kill()
|
|
|
|
def __getattr__(self, name):
|
|
"""Dispatches any unrecognised messages to the RPC connection or a CLI instance."""
|
|
if self.use_cli:
|
|
return getattr(RPCOverloadWrapper(self.cli, True, self.descriptors), name)
|
|
else:
|
|
assert self.rpc_connected and self.rpc is not None, self._node_msg("Error: no RPC connection")
|
|
return getattr(RPCOverloadWrapper(self.rpc, descriptors=self.descriptors), name)
|
|
|
|
def start(self, extra_args=None, *, cwd=None, stdout=None, stderr=None, env=None, **kwargs):
|
|
"""Start the node."""
|
|
if extra_args is None:
|
|
extra_args = self.extra_args
|
|
|
|
self.use_v2transport = "-v2transport=1" in extra_args or (self.default_to_v2 and "-v2transport=0" not in extra_args)
|
|
|
|
# Add a new stdout and stderr file each time bitcoind is started
|
|
if stderr is None:
|
|
stderr = tempfile.NamedTemporaryFile(dir=self.stderr_dir, delete=False)
|
|
if stdout is None:
|
|
stdout = tempfile.NamedTemporaryFile(dir=self.stdout_dir, delete=False)
|
|
self.stderr = stderr
|
|
self.stdout = stdout
|
|
|
|
if cwd is None:
|
|
cwd = self.cwd
|
|
|
|
# Delete any existing cookie file -- if such a file exists (eg due to
|
|
# unclean shutdown), it will get overwritten anyway by bitcoind, and
|
|
# potentially interfere with our attempt to authenticate
|
|
delete_cookie_file(self.datadir_path, self.chain)
|
|
|
|
# add environment variable LIBC_FATAL_STDERR_=1 so that libc errors are written to stderr and not the terminal
|
|
subp_env = dict(os.environ, LIBC_FATAL_STDERR_="1")
|
|
if env is not None:
|
|
subp_env.update(env)
|
|
|
|
self.process = subprocess.Popen(self.args + extra_args, env=subp_env, stdout=stdout, stderr=stderr, cwd=cwd, **kwargs)
|
|
|
|
self.running = True
|
|
self.log.debug("bitcoind started, waiting for RPC to come up")
|
|
|
|
if self.start_perf:
|
|
self._start_perf()
|
|
|
|
def wait_for_rpc_connection(self):
|
|
"""Sets up an RPC connection to the bitcoind process. Returns False if unable to connect."""
|
|
# Poll at a rate of four times per second
|
|
poll_per_s = 4
|
|
for _ in range(poll_per_s * self.rpc_timeout):
|
|
if self.process.poll() is not None:
|
|
# Attach abrupt shutdown error/s to the exception message
|
|
self.stderr.seek(0)
|
|
str_error = ''.join(line.decode('utf-8') for line in self.stderr)
|
|
str_error += "************************\n" if str_error else ''
|
|
|
|
raise FailedToStartError(self._node_msg(
|
|
f'bitcoind exited with status {self.process.returncode} during initialization. {str_error}'))
|
|
try:
|
|
rpc = get_rpc_proxy(
|
|
rpc_url(self.datadir_path, self.index, self.chain, self.rpchost),
|
|
self.index,
|
|
timeout=self.rpc_timeout // 2, # Shorter timeout to allow for one retry in case of ETIMEDOUT
|
|
coveragedir=self.coverage_dir,
|
|
)
|
|
rpc.getblockcount()
|
|
# If the call to getblockcount() succeeds then the RPC connection is up
|
|
if self.version_is_at_least(190000):
|
|
# getmempoolinfo.loaded is available since commit
|
|
# bb8ae2c (version 0.19.0)
|
|
self.wait_until(lambda: rpc.getmempoolinfo()['loaded'])
|
|
# Wait for the node to finish reindex, block import, and
|
|
# loading the mempool. Usually importing happens fast or
|
|
# even "immediate" when the node is started. However, there
|
|
# is no guarantee and sometimes ImportBlocks might finish
|
|
# later. This is going to cause intermittent test failures,
|
|
# because generally the tests assume the node is fully
|
|
# ready after being started.
|
|
#
|
|
# For example, the node will reject block messages from p2p
|
|
# when it is still importing with the error "Unexpected
|
|
# block message received"
|
|
#
|
|
# The wait is done here to make tests as robust as possible
|
|
# and prevent racy tests and intermittent failures as much
|
|
# as possible. Some tests might not need this, but the
|
|
# overhead is trivial, and the added guarantees are worth
|
|
# the minimal performance cost.
|
|
self.log.debug("RPC successfully started")
|
|
if self.use_cli:
|
|
return
|
|
self.rpc = rpc
|
|
self.rpc_connected = True
|
|
self.url = self.rpc.rpc_url
|
|
return
|
|
except JSONRPCException as e: # Initialization phase
|
|
# -28 RPC in warmup
|
|
# -342 Service unavailable, RPC server started but is shutting down due to error
|
|
if e.error['code'] != -28 and e.error['code'] != -342:
|
|
raise # unknown JSON RPC exception
|
|
except ConnectionResetError:
|
|
# This might happen when the RPC server is in warmup, but shut down before the call to getblockcount
|
|
# succeeds. Try again to properly raise the FailedToStartError
|
|
pass
|
|
except OSError as e:
|
|
if e.errno == errno.ETIMEDOUT:
|
|
pass # Treat identical to ConnectionResetError
|
|
elif e.errno == errno.ECONNREFUSED:
|
|
pass # Port not yet open?
|
|
else:
|
|
raise # unknown OS error
|
|
except ValueError as e: # cookie file not found and no rpcuser or rpcpassword; bitcoind is still starting
|
|
if "No RPC credentials" not in str(e):
|
|
raise
|
|
time.sleep(1.0 / poll_per_s)
|
|
self._raise_assertion_error("Unable to connect to bitcoind after {}s".format(self.rpc_timeout))
|
|
|
|
def wait_for_cookie_credentials(self):
|
|
"""Ensures auth cookie credentials can be read, e.g. for testing CLI with -rpcwait before RPC connection is up."""
|
|
self.log.debug("Waiting for cookie credentials")
|
|
# Poll at a rate of four times per second.
|
|
poll_per_s = 4
|
|
for _ in range(poll_per_s * self.rpc_timeout):
|
|
try:
|
|
get_auth_cookie(self.datadir_path, self.chain)
|
|
self.log.debug("Cookie credentials successfully retrieved")
|
|
return
|
|
except ValueError: # cookie file not found and no rpcuser or rpcpassword; bitcoind is still starting
|
|
pass # so we continue polling until RPC credentials are retrieved
|
|
time.sleep(1.0 / poll_per_s)
|
|
self._raise_assertion_error("Unable to retrieve cookie credentials after {}s".format(self.rpc_timeout))
|
|
|
|
def generate(self, nblocks, maxtries=1000000, **kwargs):
|
|
self.log.debug("TestNode.generate() dispatches `generate` call to `generatetoaddress`")
|
|
return self.generatetoaddress(nblocks=nblocks, address=self.get_deterministic_priv_key().address, maxtries=maxtries, **kwargs)
|
|
|
|
def generateblock(self, *args, invalid_call, **kwargs):
|
|
assert not invalid_call
|
|
return self.__getattr__('generateblock')(*args, **kwargs)
|
|
|
|
def generatetoaddress(self, *args, invalid_call, **kwargs):
|
|
assert not invalid_call
|
|
return self.__getattr__('generatetoaddress')(*args, **kwargs)
|
|
|
|
def generatetodescriptor(self, *args, invalid_call, **kwargs):
|
|
assert not invalid_call
|
|
return self.__getattr__('generatetodescriptor')(*args, **kwargs)
|
|
|
|
def setmocktime(self, timestamp):
|
|
"""Wrapper for setmocktime RPC, sets self.mocktime"""
|
|
if timestamp == 0:
|
|
# setmocktime(0) resets to system time.
|
|
self.mocktime = None
|
|
else:
|
|
self.mocktime = timestamp
|
|
return self.__getattr__('setmocktime')(timestamp)
|
|
|
|
def get_wallet_rpc(self, wallet_name):
|
|
if self.use_cli:
|
|
return RPCOverloadWrapper(self.cli("-rpcwallet={}".format(wallet_name)), True, self.descriptors)
|
|
else:
|
|
assert self.rpc_connected and self.rpc, self._node_msg("RPC not connected")
|
|
wallet_path = "wallet/{}".format(urllib.parse.quote(wallet_name))
|
|
return RPCOverloadWrapper(self.rpc / wallet_path, descriptors=self.descriptors)
|
|
|
|
def version_is_at_least(self, ver):
|
|
return self.version is None or self.version >= ver
|
|
|
|
def stop_node(self, expected_stderr='', *, wait=0, wait_until_stopped=True):
|
|
"""Stop the node."""
|
|
if not self.running:
|
|
return
|
|
self.log.debug("Stopping node")
|
|
try:
|
|
# Do not use wait argument when testing older nodes, e.g. in wallet_backwards_compatibility.py
|
|
if self.version_is_at_least(180000):
|
|
self.stop(wait=wait)
|
|
else:
|
|
self.stop()
|
|
except http.client.CannotSendRequest:
|
|
self.log.exception("Unable to stop node.")
|
|
|
|
# If there are any running perf processes, stop them.
|
|
for profile_name in tuple(self.perf_subprocesses.keys()):
|
|
self._stop_perf(profile_name)
|
|
|
|
del self.p2ps[:]
|
|
|
|
assert (not expected_stderr) or wait_until_stopped # Must wait to check stderr
|
|
if wait_until_stopped:
|
|
self.wait_until_stopped(expected_stderr=expected_stderr)
|
|
|
|
def is_node_stopped(self, *, expected_stderr="", expected_ret_code=0):
|
|
"""Checks whether the node has stopped.
|
|
|
|
Returns True if the node has stopped. False otherwise.
|
|
This method is responsible for freeing resources (self.process)."""
|
|
if not self.running:
|
|
return True
|
|
return_code = self.process.poll()
|
|
if return_code is None:
|
|
return False
|
|
|
|
# process has stopped. Assert that it didn't return an error code.
|
|
assert return_code == expected_ret_code, self._node_msg(
|
|
f"Node returned unexpected exit code ({return_code}) vs ({expected_ret_code}) when stopping")
|
|
# Check that stderr is as expected
|
|
self.stderr.seek(0)
|
|
stderr = self.stderr.read().decode('utf-8').strip()
|
|
if stderr != expected_stderr:
|
|
raise AssertionError("Unexpected stderr {} != {}".format(stderr, expected_stderr))
|
|
|
|
self.stdout.close()
|
|
self.stderr.close()
|
|
|
|
self.running = False
|
|
self.process = None
|
|
self.rpc_connected = False
|
|
self.rpc = None
|
|
self.log.debug("Node stopped")
|
|
return True
|
|
|
|
def wait_until_stopped(self, *, timeout=BITCOIND_PROC_WAIT_TIMEOUT, expect_error=False, **kwargs):
|
|
expected_ret_code = 1 if expect_error else 0 # Whether node shutdown return EXIT_FAILURE or EXIT_SUCCESS
|
|
self.wait_until(lambda: self.is_node_stopped(expected_ret_code=expected_ret_code, **kwargs), timeout=timeout)
|
|
|
|
def replace_in_config(self, replacements):
|
|
"""
|
|
Perform replacements in the configuration file.
|
|
The substitutions are passed as a list of search-replace-tuples, e.g.
|
|
[("old", "new"), ("foo", "bar"), ...]
|
|
"""
|
|
with open(self.bitcoinconf, 'r', encoding='utf8') as conf:
|
|
conf_data = conf.read()
|
|
for replacement in replacements:
|
|
assert_equal(len(replacement), 2)
|
|
old, new = replacement[0], replacement[1]
|
|
conf_data = conf_data.replace(old, new)
|
|
with open(self.bitcoinconf, 'w', encoding='utf8') as conf:
|
|
conf.write(conf_data)
|
|
|
|
@property
|
|
def chain_path(self) -> Path:
|
|
return self.datadir_path / self.chain
|
|
|
|
@property
|
|
def debug_log_path(self) -> Path:
|
|
return self.chain_path / 'debug.log'
|
|
|
|
@property
|
|
def blocks_path(self) -> Path:
|
|
return self.chain_path / "blocks"
|
|
|
|
@property
|
|
def wallets_path(self) -> Path:
|
|
return self.chain_path / "wallets"
|
|
|
|
def debug_log_size(self, **kwargs) -> int:
|
|
with open(self.debug_log_path, **kwargs) as dl:
|
|
dl.seek(0, 2)
|
|
return dl.tell()
|
|
|
|
@contextlib.contextmanager
|
|
def assert_debug_log(self, expected_msgs, unexpected_msgs=None, timeout=2):
|
|
if unexpected_msgs is None:
|
|
unexpected_msgs = []
|
|
assert_equal(type(expected_msgs), list)
|
|
assert_equal(type(unexpected_msgs), list)
|
|
|
|
time_end = time.time() + timeout * self.timeout_factor
|
|
prev_size = self.debug_log_size(encoding="utf-8") # Must use same encoding that is used to read() below
|
|
|
|
yield
|
|
|
|
while True:
|
|
found = True
|
|
with open(self.debug_log_path, encoding="utf-8", errors="replace") as dl:
|
|
dl.seek(prev_size)
|
|
log = dl.read()
|
|
print_log = " - " + "\n - ".join(log.splitlines())
|
|
for unexpected_msg in unexpected_msgs:
|
|
if re.search(re.escape(unexpected_msg), log, flags=re.MULTILINE):
|
|
self._raise_assertion_error('Unexpected message "{}" partially matches log:\n\n{}\n\n'.format(unexpected_msg, print_log))
|
|
for expected_msg in expected_msgs:
|
|
if re.search(re.escape(expected_msg), log, flags=re.MULTILINE) is None:
|
|
found = False
|
|
if found:
|
|
return
|
|
if time.time() >= time_end:
|
|
break
|
|
time.sleep(0.05)
|
|
self._raise_assertion_error('Expected messages "{}" does not partially match log:\n\n{}\n\n'.format(str(expected_msgs), print_log))
|
|
|
|
@contextlib.contextmanager
|
|
def wait_for_debug_log(self, expected_msgs, timeout=60):
|
|
"""
|
|
Block until we see a particular debug log message fragment or until we exceed the timeout.
|
|
Return:
|
|
the number of log lines we encountered when matching
|
|
"""
|
|
time_end = time.time() + timeout * self.timeout_factor
|
|
prev_size = self.debug_log_size(mode="rb") # Must use same mode that is used to read() below
|
|
|
|
yield
|
|
|
|
while True:
|
|
found = True
|
|
with open(self.debug_log_path, "rb") as dl:
|
|
dl.seek(prev_size)
|
|
log = dl.read()
|
|
|
|
for expected_msg in expected_msgs:
|
|
if expected_msg not in log:
|
|
found = False
|
|
|
|
if found:
|
|
return
|
|
|
|
if time.time() >= time_end:
|
|
print_log = " - " + "\n - ".join(log.decode("utf8", errors="replace").splitlines())
|
|
break
|
|
|
|
# No sleep here because we want to detect the message fragment as fast as
|
|
# possible.
|
|
|
|
self._raise_assertion_error(
|
|
'Expected messages "{}" does not partially match log:\n\n{}\n\n'.format(
|
|
str(expected_msgs), print_log))
|
|
|
|
@contextlib.contextmanager
|
|
def wait_for_new_peer(self, timeout=5):
|
|
"""
|
|
Wait until the node is connected to at least one new peer. We detect this
|
|
by watching for an increased highest peer id, using the `getpeerinfo` RPC call.
|
|
Note that the simpler approach of only accounting for the number of peers
|
|
suffers from race conditions, as disconnects from unrelated previous peers
|
|
could happen anytime in-between.
|
|
"""
|
|
def get_highest_peer_id():
|
|
peer_info = self.getpeerinfo()
|
|
return peer_info[-1]["id"] if peer_info else -1
|
|
|
|
initial_peer_id = get_highest_peer_id()
|
|
yield
|
|
self.wait_until(lambda: get_highest_peer_id() > initial_peer_id, timeout=timeout)
|
|
|
|
@contextlib.contextmanager
|
|
def profile_with_perf(self, profile_name: str):
|
|
"""
|
|
Context manager that allows easy profiling of node activity using `perf`.
|
|
|
|
See `test/functional/README.md` for details on perf usage.
|
|
|
|
Args:
|
|
profile_name: This string will be appended to the
|
|
profile data filename generated by perf.
|
|
"""
|
|
subp = self._start_perf(profile_name)
|
|
|
|
yield
|
|
|
|
if subp:
|
|
self._stop_perf(profile_name)
|
|
|
|
def _start_perf(self, profile_name=None):
|
|
"""Start a perf process to profile this node.
|
|
|
|
Returns the subprocess running perf."""
|
|
subp = None
|
|
|
|
def test_success(cmd):
|
|
return subprocess.call(
|
|
# shell=True required for pipe use below
|
|
cmd, shell=True,
|
|
stderr=subprocess.DEVNULL, stdout=subprocess.DEVNULL) == 0
|
|
|
|
if platform.system() != 'Linux':
|
|
self.log.warning("Can't profile with perf; only available on Linux platforms")
|
|
return None
|
|
|
|
if not test_success('which perf'):
|
|
self.log.warning("Can't profile with perf; must install perf-tools")
|
|
return None
|
|
|
|
if not test_success('readelf -S {} | grep .debug_str'.format(shlex.quote(self.binary))):
|
|
self.log.warning(
|
|
"perf output won't be very useful without debug symbols compiled into bitcoind")
|
|
|
|
output_path = tempfile.NamedTemporaryFile(
|
|
dir=self.datadir_path,
|
|
prefix="{}.perf.data.".format(profile_name or 'test'),
|
|
delete=False,
|
|
).name
|
|
|
|
cmd = [
|
|
'perf', 'record',
|
|
'-g', # Record the callgraph.
|
|
'--call-graph', 'dwarf', # Compatibility for gcc's --fomit-frame-pointer.
|
|
'-F', '101', # Sampling frequency in Hz.
|
|
'-p', str(self.process.pid),
|
|
'-o', output_path,
|
|
]
|
|
subp = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
self.perf_subprocesses[profile_name] = subp
|
|
|
|
return subp
|
|
|
|
def _stop_perf(self, profile_name):
|
|
"""Stop (and pop) a perf subprocess."""
|
|
subp = self.perf_subprocesses.pop(profile_name)
|
|
output_path = subp.args[subp.args.index('-o') + 1]
|
|
|
|
subp.terminate()
|
|
subp.wait(timeout=10)
|
|
|
|
stderr = subp.stderr.read().decode()
|
|
if 'Consider tweaking /proc/sys/kernel/perf_event_paranoid' in stderr:
|
|
self.log.warning(
|
|
"perf couldn't collect data! Try "
|
|
"'sudo sysctl -w kernel.perf_event_paranoid=-1'")
|
|
else:
|
|
report_cmd = "perf report -i {}".format(output_path)
|
|
self.log.info("See perf output by running '{}'".format(report_cmd))
|
|
|
|
def assert_start_raises_init_error(self, extra_args=None, expected_msg=None, match=ErrorMatch.FULL_TEXT, *args, **kwargs):
|
|
"""Attempt to start the node and expect it to raise an error.
|
|
|
|
extra_args: extra arguments to pass through to bitcoind
|
|
expected_msg: regex that stderr should match when bitcoind fails
|
|
|
|
Will throw if bitcoind starts without an error.
|
|
Will throw if an expected_msg is provided and it does not match bitcoind's stdout."""
|
|
assert not self.running
|
|
with tempfile.NamedTemporaryFile(dir=self.stderr_dir, delete=False) as log_stderr, \
|
|
tempfile.NamedTemporaryFile(dir=self.stdout_dir, delete=False) as log_stdout:
|
|
try:
|
|
self.start(extra_args, stdout=log_stdout, stderr=log_stderr, *args, **kwargs)
|
|
ret = self.process.wait(timeout=self.rpc_timeout)
|
|
self.log.debug(self._node_msg(f'bitcoind exited with status {ret} during initialization'))
|
|
assert ret != 0 # Exit code must indicate failure
|
|
self.running = False
|
|
self.process = None
|
|
# Check stderr for expected message
|
|
if expected_msg is not None:
|
|
log_stderr.seek(0)
|
|
stderr = log_stderr.read().decode('utf-8').strip()
|
|
if match == ErrorMatch.PARTIAL_REGEX:
|
|
if re.search(expected_msg, stderr, flags=re.MULTILINE) is None:
|
|
self._raise_assertion_error(
|
|
'Expected message "{}" does not partially match stderr:\n"{}"'.format(expected_msg, stderr))
|
|
elif match == ErrorMatch.FULL_REGEX:
|
|
if re.fullmatch(expected_msg, stderr) is None:
|
|
self._raise_assertion_error(
|
|
'Expected message "{}" does not fully match stderr:\n"{}"'.format(expected_msg, stderr))
|
|
elif match == ErrorMatch.FULL_TEXT:
|
|
if expected_msg != stderr:
|
|
self._raise_assertion_error(
|
|
'Expected message "{}" does not fully match stderr:\n"{}"'.format(expected_msg, stderr))
|
|
except subprocess.TimeoutExpired:
|
|
self.process.kill()
|
|
self.running = False
|
|
self.process = None
|
|
assert_msg = f'bitcoind should have exited within {self.rpc_timeout}s '
|
|
if expected_msg is None:
|
|
assert_msg += "with an error"
|
|
else:
|
|
assert_msg += "with expected error " + expected_msg
|
|
self._raise_assertion_error(assert_msg)
|
|
|
|
def add_p2p_connection(self, p2p_conn, *, wait_for_verack=True, send_version=True, supports_v2_p2p=False, **kwargs):
|
|
"""Add an inbound p2p connection to the node.
|
|
|
|
This method adds the p2p connection to the self.p2ps list and also
|
|
returns the connection to the caller.
|
|
|
|
When self.use_v2transport is True, TestNode advertises NODE_P2P_V2 service flag
|
|
|
|
An inbound connection is made from TestNode <------ P2PConnection
|
|
- if TestNode doesn't advertise NODE_P2P_V2 service, P2PConnection sends version message and v1 P2P is followed
|
|
- if TestNode advertises NODE_P2P_V2 service, (and if P2PConnections supports v2 P2P)
|
|
P2PConnection sends ellswift bytes and v2 P2P is followed
|
|
"""
|
|
if 'dstport' not in kwargs:
|
|
kwargs['dstport'] = p2p_port(self.index)
|
|
if 'dstaddr' not in kwargs:
|
|
kwargs['dstaddr'] = '127.0.0.1'
|
|
|
|
p2p_conn.p2p_connected_to_node = True
|
|
if self.use_v2transport:
|
|
kwargs['services'] = kwargs.get('services', P2P_SERVICES) | NODE_P2P_V2
|
|
supports_v2_p2p = self.use_v2transport and supports_v2_p2p
|
|
p2p_conn.peer_connect(**kwargs, send_version=send_version, net=self.chain, timeout_factor=self.timeout_factor, supports_v2_p2p=supports_v2_p2p)()
|
|
|
|
self.p2ps.append(p2p_conn)
|
|
p2p_conn.wait_until(lambda: p2p_conn.is_connected, check_connected=False)
|
|
if send_version:
|
|
p2p_conn.wait_until(lambda: not p2p_conn.on_connection_send_msg)
|
|
if wait_for_verack:
|
|
# Wait for the node to send us the version and verack
|
|
p2p_conn.wait_for_verack()
|
|
# At this point we have sent our version message and received the version and verack, however the full node
|
|
# has not yet received the verack from us (in reply to their version). So, the connection is not yet fully
|
|
# established (fSuccessfullyConnected).
|
|
#
|
|
# This shouldn't lead to any issues when sending messages, since the verack will be in-flight before the
|
|
# message we send. However, it might lead to races where we are expecting to receive a message. E.g. a
|
|
# transaction that will be added to the mempool as soon as we return here.
|
|
#
|
|
# So syncing here is redundant when we only want to send a message, but the cost is low (a few milliseconds)
|
|
# in comparison to the upside of making tests less fragile and unexpected intermittent errors less likely.
|
|
p2p_conn.sync_with_ping()
|
|
|
|
# Consistency check that the node received our user agent string.
|
|
# Find our connection in getpeerinfo by our address:port and theirs, as this combination is unique.
|
|
sockname = p2p_conn._transport.get_extra_info("socket").getsockname()
|
|
our_addr_and_port = f"{sockname[0]}:{sockname[1]}"
|
|
dst_addr_and_port = f"{p2p_conn.dstaddr}:{p2p_conn.dstport}"
|
|
info = [peer for peer in self.getpeerinfo() if peer["addr"] == our_addr_and_port and peer["addrbind"] == dst_addr_and_port]
|
|
assert_equal(len(info), 1)
|
|
assert_equal(info[0]["subver"], P2P_SUBVERSION)
|
|
|
|
return p2p_conn
|
|
|
|
def add_outbound_p2p_connection(self, p2p_conn, *, wait_for_verack=True, p2p_idx, connection_type="outbound-full-relay", supports_v2_p2p=False, advertise_v2_p2p=False, **kwargs):
|
|
"""Add an outbound p2p connection from node. Must be an
|
|
"outbound-full-relay", "block-relay-only", "addr-fetch" or "feeler" connection.
|
|
|
|
This method adds the p2p connection to the self.p2ps list and returns
|
|
the connection to the caller.
|
|
|
|
p2p_idx must be different for simultaneously connected peers. When reusing it for the next peer
|
|
after disconnecting the previous one, it is necessary to wait for the disconnect to finish to avoid
|
|
a race condition.
|
|
|
|
Parameters:
|
|
supports_v2_p2p: whether p2p_conn supports v2 P2P or not
|
|
advertise_v2_p2p: whether p2p_conn is advertised to support v2 P2P or not
|
|
|
|
An outbound connection is made from TestNode -------> P2PConnection
|
|
- if P2PConnection doesn't advertise_v2_p2p, TestNode sends version message and v1 P2P is followed
|
|
- if P2PConnection both supports_v2_p2p and advertise_v2_p2p, TestNode sends ellswift bytes and v2 P2P is followed
|
|
- if P2PConnection doesn't supports_v2_p2p but advertise_v2_p2p,
|
|
TestNode sends ellswift bytes and P2PConnection disconnects,
|
|
TestNode reconnects by sending version message and v1 P2P is followed
|
|
"""
|
|
|
|
def addconnection_callback(address, port):
|
|
self.log.debug("Connecting to %s:%d %s" % (address, port, connection_type))
|
|
self.addconnection('%s:%d' % (address, port), connection_type, advertise_v2_p2p)
|
|
|
|
p2p_conn.p2p_connected_to_node = False
|
|
if advertise_v2_p2p:
|
|
kwargs['services'] = kwargs.get('services', P2P_SERVICES) | NODE_P2P_V2
|
|
assert self.use_v2transport # only a v2 TestNode could make a v2 outbound connection
|
|
|
|
# if P2PConnection is advertised to support v2 P2P when it doesn't actually support v2 P2P,
|
|
# reconnection needs to be attempted using v1 P2P by sending version message
|
|
reconnect = advertise_v2_p2p and not supports_v2_p2p
|
|
# P2PConnection needs to be advertised to support v2 P2P so that ellswift bytes are sent instead of msg_version
|
|
supports_v2_p2p = supports_v2_p2p and advertise_v2_p2p
|
|
p2p_conn.peer_accept_connection(connect_cb=addconnection_callback, connect_id=p2p_idx + 1, net=self.chain, timeout_factor=self.timeout_factor, supports_v2_p2p=supports_v2_p2p, reconnect=reconnect, **kwargs)()
|
|
|
|
if reconnect:
|
|
p2p_conn.wait_for_reconnect()
|
|
|
|
if connection_type == "feeler":
|
|
# feeler connections are closed as soon as the node receives a `version` message
|
|
p2p_conn.wait_until(lambda: p2p_conn.message_count["version"] == 1, check_connected=False)
|
|
p2p_conn.wait_until(lambda: not p2p_conn.is_connected, check_connected=False)
|
|
else:
|
|
p2p_conn.wait_for_connect()
|
|
self.p2ps.append(p2p_conn)
|
|
|
|
p2p_conn.wait_until(lambda: not p2p_conn.on_connection_send_msg)
|
|
if wait_for_verack:
|
|
p2p_conn.wait_for_verack()
|
|
p2p_conn.sync_with_ping()
|
|
|
|
return p2p_conn
|
|
|
|
def num_test_p2p_connections(self):
|
|
"""Return number of test framework p2p connections to the node."""
|
|
return len([peer for peer in self.getpeerinfo() if peer['subver'] == P2P_SUBVERSION])
|
|
|
|
def disconnect_p2ps(self):
|
|
"""Close all p2p connections to the node.
|
|
Use only after each p2p has sent a version message to ensure the wait works."""
|
|
for p in self.p2ps:
|
|
p.peer_disconnect()
|
|
del self.p2ps[:]
|
|
|
|
self.wait_until(lambda: self.num_test_p2p_connections() == 0)
|
|
|
|
def bumpmocktime(self, seconds):
|
|
"""Fast forward using setmocktime to self.mocktime + seconds. Requires setmocktime to have
|
|
been called at some point in the past."""
|
|
assert self.mocktime
|
|
self.mocktime += seconds
|
|
self.setmocktime(self.mocktime)
|
|
|
|
def wait_until(self, test_function, timeout=60):
|
|
return wait_until_helper_internal(test_function, timeout=timeout, timeout_factor=self.timeout_factor)
|
|
|
|
|
|
class TestNodeCLIAttr:
|
|
def __init__(self, cli, command):
|
|
self.cli = cli
|
|
self.command = command
|
|
|
|
def __call__(self, *args, **kwargs):
|
|
return self.cli.send_cli(self.command, *args, **kwargs)
|
|
|
|
def get_request(self, *args, **kwargs):
|
|
return lambda: self(*args, **kwargs)
|
|
|
|
|
|
def arg_to_cli(arg):
|
|
if isinstance(arg, bool):
|
|
return str(arg).lower()
|
|
elif arg is None:
|
|
return 'null'
|
|
elif isinstance(arg, dict) or isinstance(arg, list):
|
|
return json.dumps(arg, default=serialization_fallback)
|
|
else:
|
|
return str(arg)
|
|
|
|
|
|
class TestNodeCLI():
|
|
"""Interface to bitcoin-cli for an individual node"""
|
|
def __init__(self, binary, datadir):
|
|
self.options = []
|
|
self.binary = binary
|
|
self.datadir = datadir
|
|
self.input = None
|
|
self.log = logging.getLogger('TestFramework.bitcoincli')
|
|
|
|
def __call__(self, *options, input=None):
|
|
# TestNodeCLI is callable with bitcoin-cli command-line options
|
|
cli = TestNodeCLI(self.binary, self.datadir)
|
|
cli.options = [str(o) for o in options]
|
|
cli.input = input
|
|
return cli
|
|
|
|
def __getattr__(self, command):
|
|
return TestNodeCLIAttr(self, command)
|
|
|
|
def batch(self, requests):
|
|
results = []
|
|
for request in requests:
|
|
try:
|
|
results.append(dict(result=request()))
|
|
except JSONRPCException as e:
|
|
results.append(dict(error=e))
|
|
return results
|
|
|
|
def send_cli(self, clicommand=None, *args, **kwargs):
|
|
"""Run bitcoin-cli command. Deserializes returned string as python object."""
|
|
pos_args = [arg_to_cli(arg) for arg in args]
|
|
named_args = [str(key) + "=" + arg_to_cli(value) for (key, value) in kwargs.items()]
|
|
p_args = [self.binary, f"-datadir={self.datadir}"] + self.options
|
|
if named_args:
|
|
p_args += ["-named"]
|
|
if clicommand is not None:
|
|
p_args += [clicommand]
|
|
p_args += pos_args + named_args
|
|
self.log.debug("Running bitcoin-cli {}".format(p_args[2:]))
|
|
process = subprocess.Popen(p_args, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True)
|
|
cli_stdout, cli_stderr = process.communicate(input=self.input)
|
|
returncode = process.poll()
|
|
if returncode:
|
|
match = re.match(r'error code: ([-0-9]+)\nerror message:\n(.*)', cli_stderr)
|
|
if match:
|
|
code, message = match.groups()
|
|
raise JSONRPCException(dict(code=int(code), message=message))
|
|
# Ignore cli_stdout, raise with cli_stderr
|
|
raise subprocess.CalledProcessError(returncode, self.binary, output=cli_stderr)
|
|
try:
|
|
return json.loads(cli_stdout, parse_float=decimal.Decimal)
|
|
except (json.JSONDecodeError, decimal.InvalidOperation):
|
|
return cli_stdout.rstrip("\n")
|
|
|
|
class RPCOverloadWrapper():
|
|
def __init__(self, rpc, cli=False, descriptors=False):
|
|
self.rpc = rpc
|
|
self.is_cli = cli
|
|
self.descriptors = descriptors
|
|
|
|
def __getattr__(self, name):
|
|
return getattr(self.rpc, name)
|
|
|
|
def createwallet_passthrough(self, *args, **kwargs):
|
|
return self.__getattr__("createwallet")(*args, **kwargs)
|
|
|
|
def createwallet(self, wallet_name, disable_private_keys=None, blank=None, passphrase='', avoid_reuse=None, descriptors=None, load_on_startup=None, external_signer=None):
|
|
if descriptors is None:
|
|
descriptors = self.descriptors
|
|
return self.__getattr__('createwallet')(wallet_name, disable_private_keys, blank, passphrase, avoid_reuse, descriptors, load_on_startup, external_signer)
|
|
|
|
def importprivkey(self, privkey, label=None, rescan=None):
|
|
wallet_info = self.getwalletinfo()
|
|
if 'descriptors' not in wallet_info or ('descriptors' in wallet_info and not wallet_info['descriptors']):
|
|
return self.__getattr__('importprivkey')(privkey, label, rescan)
|
|
desc = descsum_create('combo(' + privkey + ')')
|
|
req = [{
|
|
'desc': desc,
|
|
'timestamp': 0 if rescan else 'now',
|
|
'label': label if label else ''
|
|
}]
|
|
import_res = self.importdescriptors(req)
|
|
if not import_res[0]['success']:
|
|
raise JSONRPCException(import_res[0]['error'])
|
|
|
|
def addmultisigaddress(self, nrequired, keys, label=None, address_type=None):
|
|
wallet_info = self.getwalletinfo()
|
|
if 'descriptors' not in wallet_info or ('descriptors' in wallet_info and not wallet_info['descriptors']):
|
|
return self.__getattr__('addmultisigaddress')(nrequired, keys, label, address_type)
|
|
cms = self.createmultisig(nrequired, keys, address_type)
|
|
req = [{
|
|
'desc': cms['descriptor'],
|
|
'timestamp': 0,
|
|
'label': label if label else ''
|
|
}]
|
|
import_res = self.importdescriptors(req)
|
|
if not import_res[0]['success']:
|
|
raise JSONRPCException(import_res[0]['error'])
|
|
return cms
|
|
|
|
def importpubkey(self, pubkey, label=None, rescan=None):
|
|
wallet_info = self.getwalletinfo()
|
|
if 'descriptors' not in wallet_info or ('descriptors' in wallet_info and not wallet_info['descriptors']):
|
|
return self.__getattr__('importpubkey')(pubkey, label, rescan)
|
|
desc = descsum_create('combo(' + pubkey + ')')
|
|
req = [{
|
|
'desc': desc,
|
|
'timestamp': 0 if rescan else 'now',
|
|
'label': label if label else ''
|
|
}]
|
|
import_res = self.importdescriptors(req)
|
|
if not import_res[0]['success']:
|
|
raise JSONRPCException(import_res[0]['error'])
|
|
|
|
def importaddress(self, address, label=None, rescan=None, p2sh=None):
|
|
wallet_info = self.getwalletinfo()
|
|
if 'descriptors' not in wallet_info or ('descriptors' in wallet_info and not wallet_info['descriptors']):
|
|
return self.__getattr__('importaddress')(address, label, rescan, p2sh)
|
|
is_hex = False
|
|
try:
|
|
int(address ,16)
|
|
is_hex = True
|
|
desc = descsum_create('raw(' + address + ')')
|
|
except Exception:
|
|
desc = descsum_create('addr(' + address + ')')
|
|
reqs = [{
|
|
'desc': desc,
|
|
'timestamp': 0 if rescan else 'now',
|
|
'label': label if label else ''
|
|
}]
|
|
if is_hex and p2sh:
|
|
reqs.append({
|
|
'desc': descsum_create('p2sh(raw(' + address + '))'),
|
|
'timestamp': 0 if rescan else 'now',
|
|
'label': label if label else ''
|
|
})
|
|
import_res = self.importdescriptors(reqs)
|
|
for res in import_res:
|
|
if not res['success']:
|
|
raise JSONRPCException(res['error'])
|