jeffro256 40eb82873e
cryptonote_core: cache input verification results directly in mempool
This replaces `ver_rct_non_semantics_simple_cached()` with an API that offloads
the responsibility of tracking input verification successes to the caller. The
main caller of this function in the codebase, `cryptonote::Blockchain()` instead
keeps track of the verification results for transaction in the mempool by
storing a "verification ID" in the mempool metadata table (with `txpool_tx_meta_t`).
This has several benefits, including:

* When the mempool is large (>8192 txs), we no longer experience cache misses and unnecessarily re-verify ring signatures. This greatly improves block propagation time for FCMP++ blocks under load
* For the same reason, reorg handling can be sped up by storing verification IDs of transactions popped from the chain
* Speeds up re-validating every mempool transaction on fork change (monerod revalidates the whole tx-pool on HFs #10142)
* Caches results for every single type of Monero transaction, not just latest RCT type
* Cache persists over a node restart
* Uses 512KiB less RAM (8192*2*32B)
* No additional storage or DB migration required since `txpool_tx_meta_t` already had padding allocated
* Moves more verification logic out of `cryptonote::Blockchain`

Furthermore, this opens the door to future multi-threaded block verification
speed-ups. Right now, transactions' input proof verification is limited to one
transaction at a time. However, one can imagine a scenario with verification IDs
where input proofs are optimistically multi-threaded in advance of block
processing. Then, even though ring member fetching and verification is
single-threaded inside of `cryptonote::Blockchain::check_tx_inputs()`, the
single thread can skip the CPU-intensive cryptographic code if the verification
ID allows it.

Also changes the default log category in `tx_verification_utils.cpp` from "blockchain" to "verify".
2026-01-02 15:47:19 -06:00

459 lines
19 KiB
Python
Executable File

#!/usr/bin/env python3
# Copyright (c) 2018-2024, The Monero Project
#
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without modification, are
# permitted provided that the following conditions are met:
#
# 1. Redistributions of source code must retain the above copyright notice, this list of
# conditions and the following disclaimer.
#
# 2. Redistributions in binary form must reproduce the above copyright notice, this list
# of conditions and the following disclaimer in the documentation and/or other
# materials provided with the distribution.
#
# 3. Neither the name of the copyright holder nor the names of its contributors may be
# used to endorse or promote products derived from this software without specific
# prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY
# EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
# THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
# STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF
# THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
import time
"""Test daemon P2P
"""
from framework.daemon import Daemon
from framework.wallet import Wallet
def average(a):
return sum(a) / len(a)
def median(a):
a = sorted(a)
i0 = len(a)//2
if len(a) % 2 == 0:
return average(a[i0-1:i0+1])
else:
return a[i0]
class P2PTest():
def run_test(self):
self.reset()
self.create()
self.mine(80)
self.test_p2p_reorg()
txid = self.test_p2p_tx_propagation()
self.test_p2p_block_propagation_shared(txid)
txid = self.test_p2p_tx_propagation()
self.test_p2p_block_propagation_new(txid)
self.bench_p2p_heavy_block_propagation_speed()
def reset(self):
print('Resetting blockchain')
daemon = Daemon()
res = daemon.get_height()
daemon.pop_blocks(res.height - 1)
daemon.flush_txpool()
def create(self):
print('Creating wallet')
seed = 'velvet lymph giddy number token physics poetry unquoted nibs useful sabotage limits benches lifestyle eden nitrogen anvil fewest avoid batch vials washing fences goat unquoted'
self.wallet = Wallet(idx = 4)
# close the wallet if any, will throw if none is loaded
try: self.wallet.close_wallet()
except: pass
res = self.wallet.restore_deterministic_wallet(seed = seed)
def mine(self, blocks):
assert blocks >= 1
print("Generating", blocks, 'blocks')
daemon = Daemon(idx = 2)
# generate blocks
res_generateblocks = daemon.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', blocks)
def test_p2p_reorg(self):
print('Testing P2P reorg')
daemon2 = Daemon(idx = 2)
daemon3 = Daemon(idx = 3)
# give sync some time
time.sleep(1)
res = daemon2.get_info()
height = res.height
assert height > 0
top_block_hash = res.top_block_hash
assert len(top_block_hash) == 64
res = daemon3.get_info()
assert res.height == height
assert res.top_block_hash == top_block_hash
# disconnect daemons and mine separately on both
daemon2.out_peers(0)
daemon3.out_peers(0)
res = daemon2.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 2)
res = daemon3.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 3)
res = daemon2.get_info()
assert res.height == height + 2
daemon2_top_block_hash = res.top_block_hash
assert daemon2_top_block_hash != top_block_hash
res = daemon3.get_info()
assert res.height == height + 3
daemon3_top_block_hash = res.top_block_hash
assert daemon3_top_block_hash != top_block_hash
assert daemon3_top_block_hash != daemon2_top_block_hash
# reconnect, daemon2 will now switch to daemon3's chain
daemon2.out_peers(8)
daemon3.out_peers(8)
time.sleep(10)
res = daemon2.get_info()
assert res.height == height + 3
assert res.top_block_hash == daemon3_top_block_hash
# disconect, mine on daemon2 again more than daemon3
daemon2.out_peers(0)
daemon3.out_peers(0)
res = daemon2.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 3)
res = daemon3.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 2)
res = daemon2.get_info()
assert res.height == height + 6
daemon2_top_block_hash = res.top_block_hash
assert daemon2_top_block_hash != top_block_hash
res = daemon3.get_info()
assert res.height == height + 5
daemon3_top_block_hash = res.top_block_hash
assert daemon3_top_block_hash != top_block_hash
assert daemon3_top_block_hash != daemon2_top_block_hash
# reconnect, daemon3 will now switch to daemon2's chain
daemon2.out_peers(8)
daemon3.out_peers(8)
time.sleep(5)
res = daemon3.get_info()
assert res.height == height + 6
assert res.top_block_hash == daemon2_top_block_hash
# disconnect and mine a lot on daemon3
daemon2.out_peers(0)
daemon3.out_peers(0)
res = daemon3.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 500)
# reconnect and wait for sync
daemon2.out_peers(8)
daemon3.out_peers(8)
deadline = time.monotonic() + 240
result = None
while result is None:
res2 = daemon2.get_info()
res3 = daemon3.get_info()
if res2.top_block_hash == res3.top_block_hash:
result = True
elif time.monotonic() >= deadline:
result = False
else:
time.sleep(.25)
assert result, 'Sync timed out'
def test_p2p_tx_propagation(self):
print('Testing P2P tx propagation')
daemons = (Daemon(idx=2), Daemon(idx=3))
for daemon in daemons:
res = daemon.get_transaction_pool_hashes()
assert len(res.get('tx_hashes', [])) == 0
self.wallet.refresh()
res = self.wallet.get_balance()
dst = {'address': '42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 'amount': 1000000000000}
res = self.wallet.transfer([dst])
assert len(res.tx_hash) == 32*2
txid = res.tx_hash
# Due to Dandelion++, the network propagates transactions with a
# random delay, so poll for the transaction with a timeout. The delay
# should almost never exceed a maximum of 13s (~1/billion samples).
# Set the timeout slightly higher than the maximum delay to account
# for transmission and processing time.
timeout = 13.5
pending_daemons = set(daemons)
expected_hashes = [txid]
wait_cutoff = time.monotonic() + timeout
while True:
done = []
for daemon in pending_daemons:
res = daemon.get_transaction_pool_hashes()
hashes = res.get('tx_hashes')
if hashes:
assert hashes == expected_hashes
done.append(daemon)
pending_daemons.difference_update(done)
if len(pending_daemons) == 0:
break
max_delay = wait_cutoff - time.monotonic()
if max_delay <= 0:
break
time.sleep(min(.2, max_delay))
npending = len(pending_daemons)
assert npending == 0, '%d daemons pending' % npending
return txid
def test_p2p_block_propagation_shared(self, mempool_txid):
print('Testing P2P block propagation with shared TX')
daemon2 = Daemon(idx = 2)
daemon3 = Daemon(idx = 3)
# check precondition: txid in daemon2's and daemon3's mempool
res = daemon2.get_transaction_pool_hashes()
assert mempool_txid in res.get('tx_hashes', [])
res = daemon3.get_transaction_pool_hashes()
assert mempool_txid in res.get('tx_hashes', [])
# mine block on daemon2
res = daemon2.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 1)
block_height = res.height
# wait until both are synced, or 5 seconds, whichever is first.
# the timeout time should be very, very low here since block propagation, unlike tx propagation, is designed
# to be as fast as possible.
# and since both daemons already have the tx in their mempools, a notification of a new fluffy block, which is
# pushed out immediately upon being mined, should result in an instant addition of that block to the chain,
# without any round trips.
# this test should only fail if you're running it on a potato where PoW verification + check_tx_inputs() takes
# more than 5 second.
deadline = time.monotonic() + 5
result = None
while result is None:
res2 = daemon2.get_info()
res3 = daemon3.get_info()
if res2.top_block_hash == res3.top_block_hash:
result = True
elif time.monotonic() > deadline:
result = False
else:
time.sleep(.25)
assert result, "Shared tx block propagation timed out"
# check the tx is moved to both daemons's blockchains at top block
for daemon in [daemon2, daemon3]:
res = daemon.get_transaction_pool_hashes()
assert not 'tx_hashes' in res or len(res.tx_hashes) == 0
res = daemon.get_transactions([mempool_txid])
assert len(res.get('txs', [])) == 1
tx_details = res.txs[0]
assert ('in_pool' not in tx_details) or (not tx_details.in_pool)
assert tx_details.block_height == block_height
def test_p2p_block_propagation_new(self, mempool_txid):
# there's a big problem with this testcase in that there's not yet a way to prevent daemon's from syncing
# mempools only, but still allow block propagation. so there's a good chance that the transaction will be synced
# between daemons between when daemon2's mempool is flushed and when daemon3 mines a new block. in this
# scenario, this testcase basically just degenerates into test_p2p_block_propagation_shared(). however, if this
# one ever fails but test_p2p_block_propagation_shared() passes, then we might have actually caught a problem
# with block propagation when one of the daemons is missing a tx(s)
print('Testing P2P block propagation with (possibly) new TX')
daemon2 = Daemon(idx = 2)
daemon3 = Daemon(idx = 3)
# check precondition: txid in daemon3's mempool
res = daemon3.get_transaction_pool_hashes()
assert mempool_txid in res.get('tx_hashes', [])
# flush daemon2 mempool
daemon2.flush_txpool()
# mine block on daemon3
res = daemon3.generateblocks('42ey1afDFnn4886T7196doS9GPMzexD9gXpsZJDwVjeRVdFCSoHnv7KPbBeGpzJBzHRCAs9UxqeoyFQMYbqSWYTfJJQAWDm', 1)
block_height = res.height
# wait until both are synced, or 5 seconds, whichever is first.
# the timeout time should be very, very low here since block propagation, unlike tx propagation, is designed
# to be as fast as possible. however, it might have to be raised if the daemon actually does make a round trip
# to request a missing tx in a fluffy block
deadline = time.monotonic() + 5
result = None
while result is None:
res2 = daemon2.get_info()
res3 = daemon3.get_info()
if res2.top_block_hash == res3.top_block_hash:
result = True
elif time.monotonic() > deadline:
result = False
else:
time.sleep(.25)
assert result, "New tx block propagation timed out"
# check the tx is moved to both daemons's blockchains at top block
for daemon in [daemon2, daemon3]:
res = daemon.get_transaction_pool_hashes()
assert not 'tx_hashes' in res or len(res.tx_hashes) == 0
res = daemon.get_transactions([mempool_txid])
assert len(res.get('txs', [])) == 1
tx_details = res.txs[0]
assert ('in_pool' not in tx_details) or (not tx_details.in_pool)
assert tx_details.block_height == block_height
def bench_p2p_heavy_block_propagation_speed(self):
ENABLED = False
if not ENABLED:
print('SKIPPING benchmark of P2P heavy block propagation')
return
print('Benchmarking P2P heavy block propagation')
daemon2 = Daemon(idx = 2)
daemon3 = Daemon(idx = 3)
start_height = daemon2.get_height().height
current_height = start_height
daemon2_address = '{}:{}'.format(daemon2.host, daemon2.port)
print(' Setup: creating new wallet')
wallet = Wallet()
try: wallet.close_wallet()
except: pass
wallet.create_wallet()
wallet.auto_refresh(enable = False)
wallet.set_daemon(daemon2_address)
assert wallet.get_transfers() == {}
main_address = wallet.get_address().address
CURRENT_RING_SIZE = 16
min_height = CURRENT_RING_SIZE + 1 + 60
if start_height < min_height:
print(' Setup: mining to mixable RingCT height: {}'.format(min_height))
n_to_mine = min_height - start_height
daemon2.generateblocks(main_address, n_to_mine)
current_height += n_to_mine
print(' Setup: spamming self-send transactions into mempool to increase size')
update_unlocked_inputs = lambda: \
[x for x in wallet.incoming_transfers().get('transfers', []) if not x.spent
and x.unlocked
and x.amount > 2 * last_fee]
MAX_TX_OUTPUTS = 16
MEMPOOL_TX_TARGET = 2 * 8192 # 2x previous ver_rct_non_semantics_simple_cached() cache size
n_mempool_txs = 0
unlocked_inputs = update_unlocked_inputs()
last_fee = 10000000000 # 0.01 XMR to start off with is an over-estimation for a 1/16 in the penalty-free zone
print_progress = lambda action: \
print(' Progress: {}/{} ({:.1f}%) txs in mempool, {} usable inputs, {} blocks mined, just {} {}'.format(
n_mempool_txs, MEMPOOL_TX_TARGET, n_mempool_txs/MEMPOOL_TX_TARGET*100, len(unlocked_inputs),
current_height - start_height, action, ' ' * 10
), end='\r')
print_progress('started')
while n_mempool_txs < MEMPOOL_TX_TARGET:
try:
if len(unlocked_inputs) == 0:
daemon2.generateblocks(main_address, 1)
wallet.refresh()
current_height += 1
wallet_height = wallet.get_height().height
assert wallet_height == current_height, wallet_height
n_mempool_txs = len(daemon2.get_transaction_pool_hashes().get('tx_hashes', []))
res = wallet.incoming_transfers()
unlocked_inputs = update_unlocked_inputs()
last_action = 'mined'
else:
inp = unlocked_inputs.pop()
res = wallet.sweep_single(main_address, outputs = MAX_TX_OUTPUTS - 1, key_image = inp.key_image)
assert res.spent_key_images.key_images == [inp.key_image]
last_fee = res.fee
n_mempool_txs += 1
last_action = 'swept'
except AssertionError as ae:
print() # Clear carriage return
if 'Transaction sanity check failed' not in str(ae):
raise
# The RingCT output distribution gets so skewed in this test that the wallet
# thinks something is wrong with decoy selection. To recover, try mining a block on
# the next action.
print(' WARNING: caught transaction sanity check, stepping forward chain to try to fix')
unlocked_inputs = []
last_action = 'caught sanity'
print_progress(last_action)
print()
print(' Setup: wait for daemons to reach equilibrium on mempool contents')
assert n_mempool_txs > 0
sync_start = time.time()
sync_deadline = sync_start + 120
while True:
mempool_hashes_2 = daemon2.get_transaction_pool_hashes().get('tx_hashes', [])
assert len(mempool_hashes_2) == n_mempool_txs # Txs were submitted to daemon 2
mempool_hashes_3 = daemon3.get_transaction_pool_hashes().get('tx_hashes', [])
print(' {}/{} mempool txs propagated'.format(len(mempool_hashes_2), len(mempool_hashes_3)), end='\r')
if sorted(mempool_hashes_2) == sorted(mempool_hashes_3):
break
elif time.time() > sync_deadline:
raise RuntimeError('daemons did not sync mempools within deadline')
time.sleep(0.25)
print()
print(' Bench: mine and propagate blocks until mempool is empty of profitable txs')
timings = []
while True:
time1 = time.time()
daemon2.generateblocks(main_address, 1)
current_height += 1
time2 = time.time()
while daemon3.get_height().height != current_height:
time.sleep(0.01)
time3 = time.time()
new_n_mempool_txs = len(daemon2.get_transaction_pool_hashes().get('tx_hashes', []))
assert new_n_mempool_txs <= n_mempool_txs
n_mined_txs = n_mempool_txs - new_n_mempool_txs
elapsed_mining = time2 - time1
elapsed_prop = time3 - time2
timings.append((n_mined_txs, elapsed_mining, elapsed_prop))
n_mempool_txs = new_n_mempool_txs
print(' * Mined {} txs in {:.2f}s, propagated in {:.2f}s'.format(*(timings[-1])))
if n_mempool_txs == 0 or n_mined_txs == 0:
break
print(' Analysis of {}-tx mempool handling:'.format(MEMPOOL_TX_TARGET))
avg_mining = average([x[1] for x in timings])
median_mining = median([x[1] for x in timings])
avg_prop = average([x[2] for x in timings])
median_prop = median([x[2] for x in timings])
print(' Average mining time: {:.2f}'.format(avg_mining))
print(' Median mining time: {:.2f}'.format(median_mining))
print(' Average block propagation time: {:.2f}'.format(avg_prop))
print(' Median block propagation time: {:.2f}'.format(median_prop))
if __name__ == '__main__':
P2PTest().run_test()