54#include <boost/multi_index/hashed_index.hpp>
55#include <boost/multi_index/member.hpp>
56#include <boost/multi_index/ordered_index.hpp>
57#include <boost/multi_index_container.hpp>
115 "Max protocol message length must be greater than largest "
116 "possible INV message");
173 std::chrono::seconds(2),
174 std::chrono::seconds(2),
175 std::chrono::seconds(60),
182 std::chrono::seconds(2),
183 std::chrono::seconds(2),
184 std::chrono::seconds(60),
217 "MAX_BLOCKTXN_DEPTH too high");
276 std::chrono::seconds{1},
277 "INVENTORY_RELAY_MAX too low");
328 std::unique_ptr<PartiallyDownloadedBlock> partialBlock;
333 std::chrono::seconds timeAdded;
335 StalledTxId(
TxId txid_, std::chrono::seconds timeAdded_)
336 : txid(txid_), timeAdded(timeAdded_){};
342using StalledTxIdSet = boost::multi_index_container<
344 boost::multi_index::indexed_by<
346 boost::multi_index::hashed_unique<
347 boost::multi_index::tag<by_txid>,
348 boost::multi_index::member<StalledTxId, TxId, &StalledTxId::txid>,
351 boost::multi_index::ordered_non_unique<
352 boost::multi_index::tag<by_time>,
353 boost::multi_index::member<StalledTxId, std::chrono::seconds,
354 &StalledTxId::timeAdded>>>>;
391 std::atomic<ServiceFlags> m_their_services{
NODE_NONE};
394 Mutex m_misbehavior_mutex;
399 bool m_should_discourage
GUARDED_BY(m_misbehavior_mutex){
false};
402 Mutex m_block_inv_mutex;
408 std::vector<BlockHash> m_blocks_for_inv_relay
GUARDED_BY(m_block_inv_mutex);
414 std::vector<BlockHash>
415 m_blocks_for_headers_relay
GUARDED_BY(m_block_inv_mutex);
426 std::atomic<int> m_starting_height{-1};
429 std::atomic<uint64_t> m_ping_nonce_sent{0};
431 std::atomic<std::chrono::microseconds> m_ping_start{0us};
433 std::atomic<bool> m_ping_queued{
false};
443 std::chrono::microseconds m_next_send_feefilter
456 bool m_relay_txs
GUARDED_BY(m_bloom_filter_mutex){
false};
461 std::unique_ptr<CBloomFilter>
477 GUARDED_BY(m_tx_inventory_mutex){50000, 0.000001};
483 std::set<TxId> m_tx_inventory_to_send
GUARDED_BY(m_tx_inventory_mutex);
489 bool m_send_mempool
GUARDED_BY(m_tx_inventory_mutex){
false};
491 std::atomic<std::chrono::seconds> m_last_mempool_req{0s};
496 std::chrono::microseconds
497 m_next_inv_send_time
GUARDED_BY(m_tx_inventory_mutex){0};
503 std::atomic<Amount> m_fee_filter_received{
Amount::zero()};
509 m_avalanche_stalled_txids
GUARDED_BY(m_tx_inventory_mutex);
517 LOCK(m_tx_relay_mutex);
519 m_tx_relay = std::make_unique<Peer::TxRelay>();
520 return m_tx_relay.get();
524 return WITH_LOCK(m_tx_relay_mutex,
return m_tx_relay.get());
526 const TxRelay *GetTxRelay() const
528 return WITH_LOCK(m_tx_relay_mutex,
return m_tx_relay.get());
533 std::set<avalanche::ProofId>
534 m_proof_inventory_to_send
GUARDED_BY(m_proof_inventory_mutex);
537 GUARDED_BY(m_proof_inventory_mutex){10000, 0.000001};
544 std::chrono::microseconds m_next_inv_send_time{0};
548 std::atomic<std::chrono::seconds> lastSharedProofsUpdate{0s};
549 std::atomic<bool> compactproofs_requested{
false};
556 const std::unique_ptr<ProofRelay> m_proof_relay;
561 std::vector<CAddress>
573 std::unique_ptr<CRollingBloomFilter>
591 std::atomic_bool m_addr_relay_enabled{
false};
595 mutable Mutex m_addr_send_times_mutex;
597 std::chrono::microseconds
598 m_next_addr_send
GUARDED_BY(m_addr_send_times_mutex){0};
600 std::chrono::microseconds
601 m_next_local_addr_send
GUARDED_BY(m_addr_send_times_mutex){0};
606 std::atomic_bool m_wants_addrv2{
false};
610 mutable Mutex m_addr_token_bucket_mutex;
615 double m_addr_token_bucket
GUARDED_BY(m_addr_token_bucket_mutex){1.0};
617 std::chrono::microseconds
619 GetTime<std::chrono::microseconds>()};
621 std::atomic<uint64_t> m_addr_rate_limited{0};
626 std::atomic<uint64_t> m_addr_processed{0};
632 bool m_inv_triggered_getheaders_before_sync
636 Mutex m_getdata_requests_mutex;
638 std::deque<CInv> m_getdata_requests
GUARDED_BY(m_getdata_requests_mutex);
645 Mutex m_headers_sync_mutex;
650 std::unique_ptr<HeadersSyncState>
655 std::atomic<bool> m_sent_sendheaders{
false};
658 std::chrono::microseconds m_headers_sync_timeout
669 : m_id(id), m_our_services{our_services},
670 m_proof_relay(fRelayProofs ?
std::make_unique<ProofRelay>()
674 mutable Mutex m_tx_relay_mutex;
677 std::unique_ptr<TxRelay> m_tx_relay
GUARDED_BY(m_tx_relay_mutex);
680using PeerRef = std::shared_ptr<Peer>;
698 bool fSyncStarted{
false};
701 std::chrono::microseconds m_stalling_since{0us};
702 std::list<QueuedBlock> vBlocksInFlight;
705 std::chrono::microseconds m_downloading_since{0us};
707 bool fPreferredDownload{
false};
712 bool m_requested_hb_cmpctblocks{
false};
714 bool m_provides_cmpctblocks{
false};
742 struct ChainSyncTimeoutState {
745 std::chrono::seconds m_timeout{0s};
749 bool m_sent_getheaders{
false};
752 bool m_protect{
false};
755 ChainSyncTimeoutState m_chain_sync;
758 int64_t m_last_block_announcement{0};
761 const bool m_is_inbound;
763 CNodeState(
bool is_inbound) : m_is_inbound(is_inbound) {}
774 const std::shared_ptr<const CBlock> &pblock,
782 bool fInitialDownload)
override
788 const std::shared_ptr<const CBlock> &pblock)
override
797 !m_headers_presync_mutex);
799 std::atomic<bool> &interrupt)
override
801 !m_recent_confirmed_transactions_mutex,
802 !m_most_recent_block_mutex, !cs_proofrequest,
803 !m_headers_presync_mutex, g_msgproc_mutex);
806 !m_recent_confirmed_transactions_mutex,
807 !m_most_recent_block_mutex, !cs_proofrequest,
813 std::optional<std::string>
820 void RelayTransaction(const
TxId &txid) override
822 void RelayProof(const
avalanche::ProofId &proofid) override
824 void SetBestHeight(
int height)
override { m_best_height = height; };
827 Misbehaving(*
Assert(GetPeerRef(peer_id)),
"");
830 const std::string &msg_type,
DataStream &vRecv,
831 const std::chrono::microseconds time_received,
832 const std::atomic<bool> &interruptMsgProc)
override
834 !m_recent_confirmed_transactions_mutex,
835 !m_most_recent_block_mutex, !cs_proofrequest,
836 !m_headers_presync_mutex, g_msgproc_mutex);
838 int64_t time_in_seconds)
override;
845 void ConsiderEviction(
CNode &pto, Peer &peer,
846 std::chrono::seconds time_in_seconds)
853 void EvictExtraOutboundPeers(std::chrono::seconds now)
860 void ReattemptInitialBroadcast(
CScheduler &scheduler)
866 void UpdateAvalancheStatistics()
const;
871 void AvalanchePeriodicNetworking(
CScheduler &scheduler)
const;
889 void Misbehaving(Peer &peer,
const std::string &message);
901 void MaybePunishNodeForBlock(
NodeId nodeid,
903 bool via_compact_block,
904 const std::string &message =
"")
912 const
std::
string &message = "")
924 bool MaybeDiscourageAndDisconnect(
CNode &pnode, Peer &peer);
942 bool maybe_add_extra_compact_tx)
945 struct PackageToValidate {
947 const std::vector<NodeId> m_senders;
952 : m_txns{parent, child}, m_senders{parent_sender, child_sender} {}
955 Assume(m_txns.size() == 2);
957 "parent %s (sender=%d) + child %s (sender=%d)",
958 m_txns.front()->GetId().ToString(), m_senders.front(),
959 m_txns.back()->GetId().ToString(), m_senders.back());
968 void ProcessPackageResult(
const PackageToValidate &package_to_validate,
978 std::optional<PackageToValidate> Find1P1CPackage(
const CTransactionRef &ptx,
1005 bool ProcessOrphanTx(
const Config &config, Peer &peer)
1018 void ProcessHeadersMessage(
const Config &config,
CNode &pfrom, Peer &peer,
1019 std::vector<CBlockHeader> &&headers,
1020 bool via_compact_block)
1030 bool CheckHeadersPoW(
const std::vector<CBlockHeader> &headers,
1040 void HandleUnconnectingHeaders(
CNode &pfrom, Peer &peer,
1041 const std::vector<CBlockHeader> &headers)
1045 CheckHeadersAreContinuous(
const std::vector<CBlockHeader> &headers)
const;
1065 bool IsContinuationOfLowWorkHeadersSync(Peer &peer,
CNode &pfrom,
1066 std::vector<CBlockHeader> &headers)
1068 !m_headers_presync_mutex, g_msgproc_mutex);
1082 bool TryLowWorkHeadersSync(Peer &peer,
CNode &pfrom,
1084 std::vector<CBlockHeader> &headers)
1086 !m_headers_presync_mutex, g_msgproc_mutex);
1092 bool IsAncestorOfBestHeaderOrTip(
const CBlockIndex *header)
1106 void HeadersDirectFetchBlocks(
const Config &config,
CNode &pfrom,
1109 void UpdatePeerStateForReceivedHeaders(
CNode &pfrom, Peer &peer,
1111 bool received_new_header,
1112 bool may_have_more_headers)
1115 void SendBlockTransactions(
CNode &pfrom, Peer &peer,
const CBlock &block,
1124 std::chrono::microseconds current_time)
1134 std::chrono::microseconds current_time,
bool preferred)
1139 m_connman.PushMessage(&
node, std::move(msg));
1141 template <
typename... Args>
1142 void MakeAndPushMessage(
CNode &
node, std::string msg_type,
1143 Args &&...args)
const {
1145 std::forward<Args>(args)...));
1149 void PushNodeVersion(
const Config &config,
CNode &pnode,
const Peer &peer);
1157 void MaybeSendPing(
CNode &node_to, Peer &peer,
1158 std::chrono::microseconds now);
1161 void MaybeSendAddr(
CNode &
node, Peer &peer,
1162 std::chrono::microseconds current_time)
1169 void MaybeSendSendHeaders(
CNode &
node, Peer &peer)
1173 void MaybeSendFeefilter(
CNode &
node, Peer &peer,
1174 std::chrono::microseconds current_time)
1186 void RelayAddress(
NodeId originator,
const CAddress &addr,
bool fReachable)
1207 Mutex cs_proofrequest;
1212 std::atomic<int> m_best_height{-1};
1217 const Options m_opts;
1219 bool RejectIncomingTxs(
const CNode &peer)
const;
1231 mutable Mutex m_peer_mutex;
1238 std::map<NodeId, PeerRef> m_peer_map
GUARDED_BY(m_peer_mutex);
1247 const CNodeState *State(
NodeId pnode)
const
1252 std::atomic<std::chrono::microseconds> m_next_inv_to_inbounds{0us};
1259 m_last_block_inv_triggering_headers_sync
GUARDED_BY(g_msgproc_mutex){};
1267 std::map<BlockHash, std::pair<NodeId, bool>>
1277 std::atomic<std::chrono::seconds> m_block_stalling_timeout{
1291 bool AlreadyHaveTx(
const TxId &txid,
bool include_reconsiderable)
1293 !m_recent_confirmed_transactions_mutex);
1357 mutable Mutex m_recent_confirmed_transactions_mutex;
1359 GUARDED_BY(m_recent_confirmed_transactions_mutex){24'000, 0.000'001};
1368 std::chrono::microseconds
1369 NextInvToInbounds(std::chrono::microseconds now,
1370 std::chrono::seconds average_interval)
1375 mutable Mutex m_most_recent_block_mutex;
1376 std::shared_ptr<const CBlock>
1377 m_most_recent_block
GUARDED_BY(m_most_recent_block_mutex);
1378 std::shared_ptr<const CBlockHeaderAndShortTxIDs>
1379 m_most_recent_compact_block
GUARDED_BY(m_most_recent_block_mutex);
1381 std::unique_ptr<const std::map<TxId, CTransactionRef>>
1382 m_most_recent_block_txs
GUARDED_BY(m_most_recent_block_mutex);
1387 Mutex m_headers_presync_mutex;
1398 using HeadersPresyncStats =
1399 std::pair<arith_uint256, std::optional<std::pair<int64_t, uint32_t>>>;
1401 std::map<NodeId, HeadersPresyncStats>
1402 m_headers_presync_stats
GUARDED_BY(m_headers_presync_mutex){};
1406 std::atomic_bool m_headers_presync_should_signal{
false};
1414 bool IsBlockRequested(
const BlockHash &hash)
1418 bool IsBlockRequestedFromOutbound(
const BlockHash &hash)
1429 void RemoveBlockRequest(
const BlockHash &hash,
1430 std::optional<NodeId> from_peer)
1439 bool BlockRequested(
const Config &config,
NodeId nodeid,
1441 std::list<QueuedBlock>::iterator **pit =
nullptr)
1450 void FindNextBlocksToDownload(const Peer &peer,
unsigned int count,
1456 void TryDownloadingHistoricalBlocks(
1457 const Peer &peer,
unsigned int count,
1491 const Peer &peer, CNodeState *state,
1493 int nWindowEnd, const
CChain *activeChain =
nullptr,
1494 NodeId *nodeStaller =
nullptr)
1504 std::atomic<
std::chrono::seconds> m_last_tip_update{0s};
1511 const std::chrono::seconds mempool_req,
1512 const std::chrono::seconds now)
1517 void ProcessGetData(
const Config &config,
CNode &pfrom, Peer &peer,
1518 const std::atomic<bool> &interruptMsgProc)
1520 peer.m_getdata_requests_mutex,
1526 const std::shared_ptr<const CBlock> &block,
1527 bool force_processing,
bool min_pow_checked);
1535 void MaybeSetPeerAsAnnouncingHeaderAndIDs(
NodeId nodeid)
1554 std::vector<CTransactionRef>
1555 vExtraTxnForCompact
GUARDED_BY(g_msgproc_mutex);
1557 size_t vExtraTxnForCompactIt
GUARDED_BY(g_msgproc_mutex) = 0;
1562 void ProcessBlockAvailability(
NodeId nodeid)
1577 bool BlockRequestAllowed(const
CBlockIndex *pindex)
1579 bool AlreadyHaveBlock(const
BlockHash &block_hash)
1581 bool AlreadyHaveProof(const
avalanche::ProofId &proofid);
1582 void ProcessGetBlockData(const
Config &config,
CNode &pfrom, Peer &peer,
1605 bool PrepareBlockFilterRequest(
CNode &
node, Peer &peer,
1607 uint32_t start_height,
1609 uint32_t max_height_diff,
1651 uint32_t GetAvalancheVoteForBlock(const
BlockHash &hash) const
1662 const
TxId &
id) const
1664 !m_recent_confirmed_transactions_mutex);
1673 bool SetupAddressRelay(const
CNode &
node, Peer &peer)
1676 void AddAddressKnown(Peer &peer, const
CAddress &addr)
1678 void PushAddress(Peer &peer, const
CAddress &addr)
1686 bool ReceivedAvalancheProof(
CNode &
node, Peer &peer,
1692 const
std::chrono::seconds now)
1695 bool isPreferredDownloadPeer(const
CNode &pfrom);
1698const CNodeState *PeerManagerImpl::State(
NodeId pnode) const
1700 std::map<NodeId, CNodeState>::const_iterator it = m_node_states.find(pnode);
1701 if (it == m_node_states.end()) {
1708CNodeState *PeerManagerImpl::State(
NodeId pnode)
1710 return const_cast<CNodeState *
>(std::as_const(*this).State(pnode));
1718static bool IsAddrCompatible(
const Peer &peer,
const CAddress &addr) {
1722void PeerManagerImpl::AddAddressKnown(Peer &peer,
const CAddress &addr) {
1723 assert(peer.m_addr_known);
1724 peer.m_addr_known->insert(addr.
GetKey());
1727void PeerManagerImpl::PushAddress(Peer &peer,
const CAddress &addr) {
1731 assert(peer.m_addr_known);
1732 if (addr.
IsValid() && !peer.m_addr_known->contains(addr.
GetKey()) &&
1733 IsAddrCompatible(peer, addr)) {
1734 if (peer.m_addrs_to_send.size() >= m_opts.max_addr_to_send) {
1735 peer.m_addrs_to_send[m_rng.randrange(peer.m_addrs_to_send.size())] =
1738 peer.m_addrs_to_send.push_back(addr);
1743static void AddKnownTx(Peer &peer,
const TxId &txid) {
1744 auto tx_relay = peer.GetTxRelay();
1749 LOCK(tx_relay->m_tx_inventory_mutex);
1750 tx_relay->m_tx_inventory_known_filter.insert(txid);
1754 if (peer.m_proof_relay !=
nullptr) {
1755 LOCK(peer.m_proof_relay->m_proof_inventory_mutex);
1756 peer.m_proof_relay->m_proof_inventory_known_filter.insert(proofid);
1760bool PeerManagerImpl::isPreferredDownloadPeer(
const CNode &pfrom) {
1762 const CNodeState *state = State(pfrom.
GetId());
1763 return state && state->fPreferredDownload;
1766static bool CanServeBlocks(
const Peer &peer) {
1774static bool IsLimitedPeer(
const Peer &peer) {
1779std::chrono::microseconds
1780PeerManagerImpl::NextInvToInbounds(std::chrono::microseconds now,
1781 std::chrono::seconds average_interval) {
1782 if (m_next_inv_to_inbounds.load() < now) {
1787 m_next_inv_to_inbounds =
1788 now + m_rng.rand_exp_duration(average_interval);
1790 return m_next_inv_to_inbounds;
1793bool PeerManagerImpl::IsBlockRequested(
const BlockHash &hash) {
1794 return mapBlocksInFlight.count(hash);
1797bool PeerManagerImpl::IsBlockRequestedFromOutbound(
const BlockHash &hash) {
1798 for (
auto range = mapBlocksInFlight.equal_range(hash);
1799 range.first != range.second; range.first++) {
1800 auto [nodeid, block_it] = range.first->second;
1801 CNodeState &nodestate = *
Assert(State(nodeid));
1802 if (!nodestate.m_is_inbound) {
1810void PeerManagerImpl::RemoveBlockRequest(
const BlockHash &hash,
1811 std::optional<NodeId> from_peer) {
1812 auto range = mapBlocksInFlight.equal_range(hash);
1813 if (range.first == range.second) {
1821 while (range.first != range.second) {
1822 auto [node_id, list_it] = range.first->second;
1824 if (from_peer && *from_peer != node_id) {
1829 CNodeState &state = *
Assert(State(node_id));
1831 if (state.vBlocksInFlight.begin() == list_it) {
1834 state.m_downloading_since =
1835 std::max(state.m_downloading_since,
1836 GetTime<std::chrono::microseconds>());
1838 state.vBlocksInFlight.erase(list_it);
1840 if (state.vBlocksInFlight.empty()) {
1842 m_peers_downloading_from--;
1844 state.m_stalling_since = 0us;
1846 range.first = mapBlocksInFlight.erase(range.first);
1850bool PeerManagerImpl::BlockRequested(
const Config &config,
NodeId nodeid,
1852 std::list<QueuedBlock>::iterator **pit) {
1855 CNodeState *state = State(nodeid);
1856 assert(state !=
nullptr);
1861 for (
auto range = mapBlocksInFlight.equal_range(hash);
1862 range.first != range.second; range.first++) {
1863 if (range.first->second.first == nodeid) {
1865 *pit = &range.first->second.second;
1872 RemoveBlockRequest(hash, nodeid);
1874 std::list<QueuedBlock>::iterator it = state->vBlocksInFlight.insert(
1875 state->vBlocksInFlight.end(),
1876 {&block, std::unique_ptr<PartiallyDownloadedBlock>(
1877 pit ? new PartiallyDownloadedBlock(config, &m_mempool)
1879 if (state->vBlocksInFlight.size() == 1) {
1881 state->m_downloading_since = GetTime<std::chrono::microseconds>();
1882 m_peers_downloading_from++;
1885 auto itInFlight = mapBlocksInFlight.insert(
1886 std::make_pair(hash, std::make_pair(nodeid, it)));
1889 *pit = &itInFlight->second.second;
1895void PeerManagerImpl::MaybeSetPeerAsAnnouncingHeaderAndIDs(
NodeId nodeid) {
1901 if (m_opts.ignore_incoming_txs) {
1905 CNodeState *nodestate = State(nodeid);
1910 if (!nodestate->m_provides_cmpctblocks) {
1913 int num_outbound_hb_peers = 0;
1914 for (std::list<NodeId>::iterator it = lNodesAnnouncingHeaderAndIDs.begin();
1915 it != lNodesAnnouncingHeaderAndIDs.end(); it++) {
1916 if (*it == nodeid) {
1917 lNodesAnnouncingHeaderAndIDs.erase(it);
1918 lNodesAnnouncingHeaderAndIDs.push_back(nodeid);
1921 CNodeState *state = State(*it);
1922 if (state !=
nullptr && !state->m_is_inbound) {
1923 ++num_outbound_hb_peers;
1926 if (nodestate->m_is_inbound) {
1929 if (lNodesAnnouncingHeaderAndIDs.size() >= 3 &&
1930 num_outbound_hb_peers == 1) {
1931 CNodeState *remove_node =
1932 State(lNodesAnnouncingHeaderAndIDs.front());
1933 if (remove_node !=
nullptr && !remove_node->m_is_inbound) {
1936 std::swap(lNodesAnnouncingHeaderAndIDs.front(),
1937 *std::next(lNodesAnnouncingHeaderAndIDs.begin()));
1944 if (lNodesAnnouncingHeaderAndIDs.size() >= 3) {
1948 lNodesAnnouncingHeaderAndIDs.front(), [
this](
CNode *pnodeStop) {
1949 MakeAndPushMessage(*pnodeStop, NetMsgType::SENDCMPCT,
1951 CMPCTBLOCKS_VERSION);
1954 pnodeStop->m_bip152_highbandwidth_to = false;
1957 lNodesAnnouncingHeaderAndIDs.pop_front();
1964 lNodesAnnouncingHeaderAndIDs.push_back(pfrom->
GetId());
1969bool PeerManagerImpl::TipMayBeStale() {
1972 if (m_last_tip_update.load() == 0s) {
1973 m_last_tip_update = GetTime<std::chrono::seconds>();
1975 return m_last_tip_update.load() <
1976 GetTime<std::chrono::seconds>() -
1979 mapBlocksInFlight.empty();
1982bool PeerManagerImpl::CanDirectFetch() {
1988static bool PeerHasHeader(CNodeState *state,
const CBlockIndex *pindex)
1990 if (state->pindexBestKnownBlock &&
1991 pindex == state->pindexBestKnownBlock->GetAncestor(pindex->nHeight)) {
1994 if (state->pindexBestHeaderSent &&
1995 pindex == state->pindexBestHeaderSent->GetAncestor(pindex->nHeight)) {
2001void PeerManagerImpl::ProcessBlockAvailability(
NodeId nodeid) {
2002 CNodeState *state = State(nodeid);
2003 assert(state !=
nullptr);
2005 if (!state->hashLastUnknownBlock.IsNull()) {
2009 if (state->pindexBestKnownBlock ==
nullptr ||
2010 pindex->
nChainWork >= state->pindexBestKnownBlock->nChainWork) {
2011 state->pindexBestKnownBlock = pindex;
2013 state->hashLastUnknownBlock.SetNull();
2018void PeerManagerImpl::UpdateBlockAvailability(
NodeId nodeid,
2020 CNodeState *state = State(nodeid);
2021 assert(state !=
nullptr);
2023 ProcessBlockAvailability(nodeid);
2028 if (state->pindexBestKnownBlock ==
nullptr ||
2029 pindex->
nChainWork >= state->pindexBestKnownBlock->nChainWork) {
2030 state->pindexBestKnownBlock = pindex;
2035 state->hashLastUnknownBlock = hash;
2041void PeerManagerImpl::FindNextBlocksToDownload(
2042 const Peer &peer,
unsigned int count,
2043 std::vector<const CBlockIndex *> &vBlocks,
NodeId &nodeStaller) {
2048 vBlocks.reserve(vBlocks.size() +
count);
2049 CNodeState *state = State(peer.m_id);
2050 assert(state !=
nullptr);
2053 ProcessBlockAvailability(peer.m_id);
2055 if (state->pindexBestKnownBlock ==
nullptr ||
2056 state->pindexBestKnownBlock->nChainWork <
2058 state->pindexBestKnownBlock->nChainWork <
2068 const CBlockIndex *snap_base{m_chainman.GetSnapshotBaseBlock()};
2069 if (snap_base && state->pindexBestKnownBlock->GetAncestor(
2070 snap_base->nHeight) != snap_base) {
2072 "Not downloading blocks from peer=%d, which doesn't have the "
2073 "snapshot block in its best chain.\n",
2082 if (state->pindexLastCommonBlock ==
nullptr ||
2084 state->pindexLastCommonBlock->nHeight < snap_base->nHeight)) {
2085 state->pindexLastCommonBlock =
2087 .
ActiveChain()[std::min(state->pindexBestKnownBlock->nHeight,
2094 state->pindexLastCommonBlock, state->pindexBestKnownBlock);
2095 if (state->pindexLastCommonBlock == state->pindexBestKnownBlock) {
2099 const CBlockIndex *pindexWalk = state->pindexLastCommonBlock;
2107 FindNextBlocks(vBlocks, peer, state, pindexWalk,
count, nWindowEnd,
2111void PeerManagerImpl::TryDownloadingHistoricalBlocks(
2112 const Peer &peer,
unsigned int count,
2113 std::vector<const CBlockIndex *> &vBlocks,
const CBlockIndex *from_tip,
2118 if (vBlocks.size() >=
count) {
2122 vBlocks.reserve(
count);
2123 CNodeState *state =
Assert(State(peer.m_id));
2125 if (state->pindexBestKnownBlock ==
nullptr ||
2126 state->pindexBestKnownBlock->GetAncestor(target_block->
nHeight) !=
2141 FindNextBlocks(vBlocks, peer, state, from_tip,
count,
2146void PeerManagerImpl::FindNextBlocks(std::vector<const CBlockIndex *> &vBlocks,
2147 const Peer &peer, CNodeState *state,
2149 unsigned int count,
int nWindowEnd,
2150 const CChain *activeChain,
2152 std::vector<const CBlockIndex *> vToFetch;
2154 std::min<int>(state->pindexBestKnownBlock->nHeight, nWindowEnd + 1);
2156 while (pindexWalk->
nHeight < nMaxHeight) {
2161 int nToFetch = std::min(nMaxHeight - pindexWalk->
nHeight,
2162 std::max<int>(
count - vBlocks.size(), 128));
2163 vToFetch.resize(nToFetch);
2164 pindexWalk = state->pindexBestKnownBlock->
GetAncestor(
2165 pindexWalk->
nHeight + nToFetch);
2166 vToFetch[nToFetch - 1] = pindexWalk;
2167 for (
unsigned int i = nToFetch - 1; i > 0; i--) {
2168 vToFetch[i - 1] = vToFetch[i]->
pprev;
2181 if (pindex->nStatus.hasData() ||
2182 (activeChain && activeChain->
Contains(pindex))) {
2184 state->pindexLastCommonBlock = pindex;
2186 }
else if (!IsBlockRequested(pindex->
GetBlockHash())) {
2188 if (pindex->
nHeight > nWindowEnd) {
2190 if (vBlocks.size() == 0 && waitingfor != peer.m_id) {
2194 *nodeStaller = waitingfor;
2199 vBlocks.push_back(pindex);
2200 if (vBlocks.size() ==
count) {
2203 }
else if (waitingfor == -1) {
2215template <
class InvId>
2219 return !
node.HasPermission(
2232template <
class InvId>
2233static std::chrono::microseconds
2237 std::chrono::microseconds current_time,
bool preferred) {
2238 auto delay = std::chrono::microseconds{0};
2250 return current_time + delay;
2253void PeerManagerImpl::PushNodeVersion(
const Config &config,
CNode &pnode,
2255 uint64_t my_services{peer.m_our_services};
2256 const int64_t nTime{
count_seconds(GetTime<std::chrono::seconds>())};
2258 const int nNodeStartingHeight{m_best_height};
2269 const bool tx_relay{!RejectIncomingTxs(pnode)};
2278 nNodeStartingHeight, tx_relay, extraEntropy);
2282 "send version message: version %d, blocks=%d, them=%s, "
2283 "txrelay=%d, peer=%d\n",
2288 "send version message: version %d, blocks=%d, "
2289 "txrelay=%d, peer=%d\n",
2294void PeerManagerImpl::AddTxAnnouncement(
2296 std::chrono::microseconds current_time) {
2304 const bool preferred = isPreferredDownloadPeer(
node);
2306 current_time, preferred);
2308 m_txrequest.ReceivedInv(
node.GetId(), txid, preferred, reqtime);
2311void PeerManagerImpl::AddProofAnnouncement(
2313 std::chrono::microseconds current_time,
bool preferred) {
2324 m_proofrequest.ReceivedInv(
node.GetId(), proofid, preferred, reqtime);
2327void PeerManagerImpl::UpdateLastBlockAnnounceTime(
NodeId node,
2328 int64_t time_in_seconds) {
2330 CNodeState *state = State(
node);
2332 state->m_last_block_announcement = time_in_seconds;
2336void PeerManagerImpl::InitializeNode(
const Config &config,
CNode &
node,
2341 m_node_states.emplace_hint(m_node_states.end(),
2342 std::piecewise_construct,
2343 std::forward_as_tuple(nodeid),
2344 std::forward_as_tuple(
node.IsInboundConn()));
2345 assert(m_txrequest.Count(nodeid) == 0);
2353 PeerRef peer = std::make_shared<Peer>(nodeid, our_services, !!m_avalanche);
2356 m_peer_map.emplace_hint(m_peer_map.end(), nodeid, peer);
2358 if (!
node.IsInboundConn()) {
2359 PushNodeVersion(config,
node, *peer);
2363void PeerManagerImpl::ReattemptInitialBroadcast(
CScheduler &scheduler) {
2366 for (
const TxId &txid : unbroadcast_txids) {
2368 if (m_mempool.
exists(txid)) {
2369 RelayTransaction(txid);
2380 auto unbroadcasted_proofids =
2384 auto it = unbroadcasted_proofids.begin();
2385 while (it != unbroadcasted_proofids.end()) {
2388 if (!pm.isBoundToPeer(*it)) {
2389 pm.removeUnbroadcastProof(*it);
2390 it = unbroadcasted_proofids.erase(it);
2397 return unbroadcasted_proofids;
2401 for (
const auto &proofid : unbroadcasted_proofids) {
2402 RelayProof(proofid);
2409 const auto reattemptBroadcastInterval =
2411 scheduler.
scheduleFromNow([&] { ReattemptInitialBroadcast(scheduler); },
2412 reattemptBroadcastInterval);
2415void PeerManagerImpl::UpdateAvalancheStatistics()
const {
2421void PeerManagerImpl::AvalanchePeriodicNetworking(
CScheduler &scheduler)
const {
2422 const auto now = GetTime<std::chrono::seconds>();
2423 std::vector<NodeId> avanode_ids;
2424 bool fQuorumEstablished;
2425 bool fShouldRequestMoreNodes;
2435 fShouldRequestMoreNodes =
2443 avanode_ids.push_back(pnode->GetId());
2446 PeerRef peer = GetPeerRef(pnode->
GetId());
2447 if (peer ==
nullptr) {
2451 if (peer->m_proof_relay &&
2452 now > (peer->m_proof_relay->lastSharedProofsUpdate.load() +
2454 peer->m_proof_relay->sharedProofs = {};
2458 if (avanode_ids.empty()) {
2466 for (
NodeId avanodeId : avanode_ids) {
2467 const bool sentGetavaaddr =
2470 MakeAndPushMessage(*pavanode, NetMsgType::GETAVAADDR);
2471 PeerRef peer = GetPeerRef(avanodeId);
2472 WITH_LOCK(peer->m_addr_token_bucket_mutex,
2473 peer->m_addr_token_bucket +=
2474 m_opts.max_addr_to_send);
2482 if (sentGetavaaddr && fQuorumEstablished && !fShouldRequestMoreNodes) {
2497 avanode_ids.resize(std::min<size_t>(avanode_ids.size(), 3));
2500 for (
NodeId nodeid : avanode_ids) {
2503 PeerRef peer = GetPeerRef(nodeid);
2504 if (peer->m_proof_relay) {
2506 peer->m_proof_relay->compactproofs_requested =
true;
2516 const auto avalanchePeriodicNetworkingInterval =
2518 scheduler.
scheduleFromNow([&] { AvalanchePeriodicNetworking(scheduler); },
2519 avalanchePeriodicNetworkingInterval);
2522void PeerManagerImpl::FinalizeNode(
const Config &config,
const CNode &
node) {
2532 PeerRef peer = RemovePeer(nodeid);
2535 m_peer_map.erase(nodeid);
2537 CNodeState *state = State(nodeid);
2538 assert(state !=
nullptr);
2540 if (state->fSyncStarted) {
2544 for (
const QueuedBlock &entry : state->vBlocksInFlight) {
2546 mapBlocksInFlight.equal_range(entry.pindex->GetBlockHash());
2547 while (range.first != range.second) {
2548 auto [node_id, list_it] = range.first->second;
2549 if (node_id != nodeid) {
2552 range.first = mapBlocksInFlight.erase(range.first);
2559 m_txrequest.DisconnectedPeer(nodeid);
2560 m_num_preferred_download_peers -= state->fPreferredDownload;
2561 m_peers_downloading_from -= (!state->vBlocksInFlight.empty());
2562 assert(m_peers_downloading_from >= 0);
2563 m_outbound_peers_with_protect_from_disconnect -=
2564 state->m_chain_sync.m_protect;
2565 assert(m_outbound_peers_with_protect_from_disconnect >= 0);
2567 m_node_states.erase(nodeid);
2569 if (m_node_states.empty()) {
2571 assert(mapBlocksInFlight.empty());
2572 assert(m_num_preferred_download_peers == 0);
2573 assert(m_peers_downloading_from == 0);
2574 assert(m_outbound_peers_with_protect_from_disconnect == 0);
2575 assert(m_txrequest.Size() == 0);
2577 return orphanage.Size();
2582 if (
node.fSuccessfullyConnected && !
node.IsBlockOnlyConn() &&
2583 !
node.IsInboundConn()) {
2590 LOCK(m_headers_presync_mutex);
2591 m_headers_presync_stats.erase(nodeid);
2594 WITH_LOCK(cs_proofrequest, m_proofrequest.DisconnectedPeer(nodeid));
2599PeerRef PeerManagerImpl::GetPeerRef(
NodeId id)
const {
2601 auto it = m_peer_map.find(
id);
2602 return it != m_peer_map.end() ? it->second :
nullptr;
2605PeerRef PeerManagerImpl::RemovePeer(
NodeId id) {
2608 auto it = m_peer_map.find(
id);
2609 if (it != m_peer_map.end()) {
2610 ret = std::move(it->second);
2611 m_peer_map.erase(it);
2616bool PeerManagerImpl::GetNodeStateStats(
NodeId nodeid,
2620 const CNodeState *state = State(nodeid);
2621 if (state ==
nullptr) {
2625 ? state->pindexBestKnownBlock->nHeight
2628 ? state->pindexLastCommonBlock->nHeight
2630 for (
const QueuedBlock &queue : state->vBlocksInFlight) {
2637 PeerRef peer = GetPeerRef(nodeid);
2638 if (peer ==
nullptr) {
2650 auto ping_wait{0us};
2651 if ((0 != peer->m_ping_nonce_sent) &&
2652 (0 != peer->m_ping_start.load().count())) {
2654 GetTime<std::chrono::microseconds>() - peer->m_ping_start.load();
2657 if (
auto tx_relay = peer->GetTxRelay()) {
2659 return tx_relay->m_relay_txs);
2671 LOCK(peer->m_headers_sync_mutex);
2672 if (peer->m_headers_sync) {
2680void PeerManagerImpl::AddToCompactExtraTransactions(
const CTransactionRef &tx) {
2681 if (m_opts.max_extra_txs <= 0) {
2685 if (!vExtraTxnForCompact.size()) {
2686 vExtraTxnForCompact.resize(m_opts.max_extra_txs);
2689 vExtraTxnForCompact[vExtraTxnForCompactIt] = tx;
2690 vExtraTxnForCompactIt = (vExtraTxnForCompactIt + 1) % m_opts.max_extra_txs;
2693void PeerManagerImpl::Misbehaving(Peer &peer,
const std::string &message) {
2694 LOCK(peer.m_misbehavior_mutex);
2696 const std::string message_prefixed =
2697 message.empty() ?
"" : (
": " + message);
2698 peer.m_should_discourage =
true;
2703void PeerManagerImpl::MaybePunishNodeForBlock(
NodeId nodeid,
2705 bool via_compact_block,
2706 const std::string &message) {
2707 PeerRef peer{GetPeerRef(nodeid)};
2718 if (!via_compact_block) {
2720 Misbehaving(*peer, message);
2727 CNodeState *node_state = State(nodeid);
2728 if (node_state ==
nullptr) {
2735 if (!via_compact_block && !node_state->m_is_inbound) {
2737 Misbehaving(*peer, message);
2747 Misbehaving(*peer, message);
2753 Misbehaving(*peer, message);
2759 if (message !=
"") {
2764void PeerManagerImpl::MaybePunishNodeForTx(
NodeId nodeid,
2766 const std::string &message) {
2767 PeerRef peer{GetPeerRef(nodeid)};
2774 Misbehaving(*peer, message);
2792 if (message !=
"") {
2797bool PeerManagerImpl::BlockRequestAllowed(
const CBlockIndex *pindex) {
2803 (m_chainman.m_best_header !=
nullptr) &&
2804 (m_chainman.m_best_header->GetBlockTime() - pindex->
GetBlockTime() <
2807 *m_chainman.m_best_header, *pindex, *m_chainman.m_best_header,
2811std::optional<std::string>
2812PeerManagerImpl::FetchBlock(
const Config &config,
NodeId peer_id,
2815 return "Loading blocks ...";
2821 CNodeState *state = State(peer_id);
2822 if (state ==
nullptr) {
2823 return "Peer does not exist";
2827 RemoveBlockRequest(block_index.
GetBlockHash(), std::nullopt);
2830 if (!BlockRequested(config, peer_id, block_index)) {
2831 return "Already requested from this peer";
2840 this->MakeAndPushMessage(*node, NetMsgType::GETDATA, invs);
2843 return "Node not fully connected";
2848 return std::nullopt;
2851std::unique_ptr<PeerManager>
2855 return std::make_unique<PeerManagerImpl>(connman, addrman, banman, chainman,
2864 : m_rng{opts.deterministic_rng},
2866 m_chainparams(chainman.GetParams()), m_connman(connman),
2867 m_addrman(addrman), m_banman(banman), m_chainman(chainman),
2868 m_mempool(pool), m_avalanche(
avalanche), m_opts{opts} {}
2870void PeerManagerImpl::StartScheduledTasks(
CScheduler &scheduler) {
2877 "peer eviction timer should be less than stale tip check timer");
2880 this->CheckForStaleTipAndEvictPeers();
2886 const auto reattemptBroadcastInterval =
2888 scheduler.
scheduleFromNow([&] { ReattemptInitialBroadcast(scheduler); },
2889 reattemptBroadcastInterval);
2894 UpdateAvalancheStatistics();
2900 const auto avalanchePeriodicNetworkingInterval =
2902 scheduler.
scheduleFromNow([&] { AvalanchePeriodicNetworking(scheduler); },
2903 avalanchePeriodicNetworkingInterval);
2912void PeerManagerImpl::BlockConnected(
2913 ChainstateRole role,
const std::shared_ptr<const CBlock> &pblock,
2917 m_last_tip_update = GetTime<std::chrono::seconds>();
2921 auto stalling_timeout = m_block_stalling_timeout.load();
2924 const auto new_timeout =
2925 std::max(std::chrono::duration_cast<std::chrono::seconds>(
2926 stalling_timeout * 0.85),
2928 if (m_block_stalling_timeout.compare_exchange_strong(stalling_timeout,
2948 LOCK(m_recent_confirmed_transactions_mutex);
2950 m_recent_confirmed_transactions.insert(ptx->GetId());
2955 for (
const auto &ptx : pblock->vtx) {
2956 m_txrequest.ForgetInvId(ptx->GetId());
2961void PeerManagerImpl::BlockDisconnected(
2962 const std::shared_ptr<const CBlock> &block,
const CBlockIndex *pindex) {
2971 LOCK(m_recent_confirmed_transactions_mutex);
2972 m_recent_confirmed_transactions.reset();
2979void PeerManagerImpl::NewPoWValidBlock(
2980 const CBlockIndex *pindex,
const std::shared_ptr<const CBlock> &pblock) {
2981 std::shared_ptr<const CBlockHeaderAndShortTxIDs> pcmpctblock =
2982 std::make_shared<const CBlockHeaderAndShortTxIDs>(
2987 if (pindex->
nHeight <= m_highest_fast_announce) {
2990 m_highest_fast_announce = pindex->
nHeight;
2993 const std::shared_future<CSerializedNetMsg> lazy_ser{
2994 std::async(std::launch::deferred, [&] {
2999 auto most_recent_block_txs =
3000 std::make_unique<std::map<TxId, CTransactionRef>>();
3001 for (
const auto &tx : pblock->vtx) {
3002 most_recent_block_txs->emplace(tx->GetId(), tx);
3005 LOCK(m_most_recent_block_mutex);
3006 m_most_recent_block_hash = hashBlock;
3007 m_most_recent_block = pblock;
3008 m_most_recent_compact_block = pcmpctblock;
3009 m_most_recent_block_txs = std::move(most_recent_block_txs);
3013 [
this, pindex, &lazy_ser, &hashBlock](
CNode *pnode)
3021 ProcessBlockAvailability(pnode->
GetId());
3022 CNodeState &state = *State(pnode->
GetId());
3026 if (state.m_requested_hb_cmpctblocks &&
3027 !PeerHasHeader(&state, pindex) &&
3028 PeerHasHeader(&state, pindex->
pprev)) {
3030 "%s sending header-and-ids %s to peer=%d\n",
3031 "PeerManager::NewPoWValidBlock",
3032 hashBlock.ToString(), pnode->
GetId());
3035 PushMessage(*pnode, ser_cmpctblock.Copy());
3036 state.pindexBestHeaderSent = pindex;
3045void PeerManagerImpl::UpdatedBlockTip(
const CBlockIndex *pindexNew,
3047 bool fInitialDownload) {
3048 SetBestHeight(pindexNew->
nHeight);
3052 if (fInitialDownload) {
3057 std::vector<BlockHash> vHashes;
3059 while (pindexToAnnounce != pindexFork) {
3061 pindexToAnnounce = pindexToAnnounce->
pprev;
3071 for (
auto &it : m_peer_map) {
3072 Peer &peer = *it.second;
3073 LOCK(peer.m_block_inv_mutex);
3075 peer.m_blocks_for_headers_relay.push_back(hash);
3087void PeerManagerImpl::BlockChecked(
const CBlock &block,
3092 std::map<BlockHash, std::pair<NodeId, bool>>::iterator it =
3093 mapBlockSource.find(hash);
3097 if (state.
IsInvalid() && it != mapBlockSource.end() &&
3098 State(it->second.first)) {
3099 MaybePunishNodeForBlock(it->second.first, state,
3100 !it->second.second);
3109 mapBlocksInFlight.count(hash) == mapBlocksInFlight.size()) {
3110 if (it != mapBlockSource.end()) {
3111 MaybeSetPeerAsAnnouncingHeaderAndIDs(it->second.first);
3115 if (it != mapBlockSource.end()) {
3116 mapBlockSource.erase(it);
3125bool PeerManagerImpl::AlreadyHaveTx(
const TxId &txid,
3126 bool include_reconsiderable) {
3128 hashRecentRejectsChainTip) {
3133 hashRecentRejectsChainTip =
3135 m_recent_rejects.reset();
3136 m_recent_rejects_package_reconsiderable.reset();
3140 return orphanage.HaveTx(txid);
3146 return conflicting.HaveTx(txid);
3151 if (include_reconsiderable &&
3152 m_recent_rejects_package_reconsiderable.contains(txid)) {
3157 LOCK(m_recent_confirmed_transactions_mutex);
3158 if (m_recent_confirmed_transactions.contains(txid)) {
3163 return m_recent_rejects.contains(txid) || m_mempool.
exists(txid);
3166bool PeerManagerImpl::AlreadyHaveBlock(
const BlockHash &block_hash) {
3171 if (!
Assume(m_avalanche)) {
3176 if (localProof && localProof->getId() == proofid) {
3185void PeerManagerImpl::SendPings() {
3187 for (
auto &it : m_peer_map) {
3188 it.second->m_ping_queued =
true;
3192void PeerManagerImpl::RelayTransaction(
const TxId &txid) {
3194 for (
auto &it : m_peer_map) {
3195 Peer &peer = *it.second;
3196 auto tx_relay = peer.GetTxRelay();
3200 LOCK(tx_relay->m_tx_inventory_mutex);
3206 if (tx_relay->m_next_inv_send_time == 0s) {
3210 if (!tx_relay->m_tx_inventory_known_filter.contains(txid) ||
3211 tx_relay->m_avalanche_stalled_txids.count(txid) > 0) {
3212 tx_relay->m_tx_inventory_to_send.insert(txid);
3219 for (
auto &it : m_peer_map) {
3220 Peer &peer = *it.second;
3222 if (!peer.m_proof_relay) {
3225 LOCK(peer.m_proof_relay->m_proof_inventory_mutex);
3226 if (!peer.m_proof_relay->m_proof_inventory_known_filter.contains(
3228 peer.m_proof_relay->m_proof_inventory_to_send.insert(proofid);
3233void PeerManagerImpl::RelayAddress(
NodeId originator,
const CAddress &addr,
3249 const auto current_time{GetTime<std::chrono::seconds>()};
3252 const uint64_t time_addr{
3253 (
static_cast<uint64_t
>(
count_seconds(current_time)) + hash_addr) /
3263 unsigned int nRelayNodes = (fReachable || (hasher.Finalize() & 1)) ? 2 : 1;
3264 std::array<std::pair<uint64_t, Peer *>, 2> best{
3265 {{0,
nullptr}, {0,
nullptr}}};
3266 assert(nRelayNodes <= best.size());
3270 for (
auto &[
id, peer] : m_peer_map) {
3271 if (peer->m_addr_relay_enabled &&
id != originator &&
3272 IsAddrCompatible(*peer, addr)) {
3274 for (
unsigned int i = 0; i < nRelayNodes; i++) {
3275 if (hashKey > best[i].first) {
3276 std::copy(best.begin() + i, best.begin() + nRelayNodes - 1,
3277 best.begin() + i + 1);
3278 best[i] = std::make_pair(hashKey, peer.get());
3285 for (
unsigned int i = 0; i < nRelayNodes && best[i].first != 0; i++) {
3286 PushAddress(*best[i].second, addr);
3290void PeerManagerImpl::ProcessGetBlockData(
const Config &config,
CNode &pfrom,
3291 Peer &peer,
const CInv &inv) {
3294 std::shared_ptr<const CBlock> a_recent_block;
3295 std::shared_ptr<const CBlockHeaderAndShortTxIDs> a_recent_compact_block;
3297 LOCK(m_most_recent_block_mutex);
3298 a_recent_block = m_most_recent_block;
3299 a_recent_compact_block = m_most_recent_compact_block;
3302 bool need_activate_chain =
false;
3316 need_activate_chain =
true;
3320 if (need_activate_chain) {
3323 state, a_recent_block, m_avalanche)) {
3331 bool can_direct_fetch{
false};
3339 if (!BlockRequestAllowed(pindex)) {
3341 "%s: ignoring request from peer=%i for old "
3342 "block that isn't in the main chain\n",
3343 __func__, pfrom.
GetId());
3349 (((m_chainman.m_best_header !=
nullptr) &&
3350 (m_chainman.m_best_header->GetBlockTime() -
3358 "historical block serving limit reached, disconnect peer=%d\n",
3371 (tip->nHeight - pindex->
nHeight >
3374 "Ignore block request below NODE_NETWORK_LIMITED "
3375 "threshold, disconnect peer=%d\n",
3385 if (!pindex->nStatus.hasData()) {
3388 can_direct_fetch = CanDirectFetch();
3392 std::shared_ptr<const CBlock> pblock;
3393 auto handle_block_read_error = [&]() {
3395 return m_chainman.
m_blockman.IsBlockPruned(*pindex))) {
3397 "Block was pruned before it could be read, disconnect "
3401 LogError(
"Cannot load block from disk, disconnect peer=%d\n",
3407 if (a_recent_block && a_recent_block->GetHash() == pindex->
GetBlockHash()) {
3408 pblock = a_recent_block;
3412 std::vector<uint8_t> block_data;
3414 handle_block_read_error();
3421 std::shared_ptr<CBlock> pblockRead = std::make_shared<CBlock>();
3423 handle_block_read_error();
3426 pblock = pblockRead;
3432 bool sendMerkleBlock =
false;
3434 if (
auto tx_relay = peer.GetTxRelay()) {
3435 LOCK(tx_relay->m_bloom_filter_mutex);
3436 if (tx_relay->m_bloom_filter) {
3437 sendMerkleBlock =
true;
3442 if (sendMerkleBlock) {
3453 typedef std::pair<size_t, uint256> PairType;
3456 *pblock->vtx[pair.first]);
3466 if (can_direct_fetch &&
3468 if (a_recent_compact_block &&
3469 a_recent_compact_block->header.GetHash() ==
3472 *a_recent_compact_block);
3486 LOCK(peer.m_block_inv_mutex);
3489 if (hash == peer.m_continuation_block) {
3493 std::vector<CInv> vInv;
3496 peer.m_continuation_block =
BlockHash();
3502PeerManagerImpl::FindTxForGetData(
const Peer &peer,
const TxId &txid,
3503 const std::chrono::seconds mempool_req,
3504 const std::chrono::seconds now) {
3505 auto txinfo = m_mempool.
info(txid);
3510 if ((mempool_req.count() && txinfo.m_time <= mempool_req) ||
3512 return std::move(txinfo.tx);
3521 Assume(peer.GetTxRelay())->m_recently_announced_invs.contains(txid);
3522 if (recent && txinfo.tx) {
3523 return std::move(txinfo.tx);
3528 LOCK(m_most_recent_block_mutex);
3529 if (m_most_recent_block_txs !=
nullptr) {
3530 auto it = m_most_recent_block_txs->find(txid);
3531 if (it != m_most_recent_block_txs->end()) {
3544PeerManagerImpl::FindProofForGetData(
const Peer &peer,
3546 const std::chrono::seconds now) {
3549 bool send_unconditionally =
3575 if (send_unconditionally) {
3580 if (peer.m_proof_relay->m_recently_announced_proofs.contains(proofid)) {
3587void PeerManagerImpl::ProcessGetData(
3589 const std::atomic<bool> &interruptMsgProc) {
3592 auto tx_relay = peer.GetTxRelay();
3594 std::deque<CInv>::iterator it = peer.m_getdata_requests.begin();
3595 std::vector<CInv> vNotFound;
3597 const auto now{GetTime<std::chrono::seconds>()};
3599 const auto mempool_req = tx_relay !=
nullptr
3600 ? tx_relay->m_last_mempool_req.load()
3601 : std::chrono::seconds::min();
3606 while (it != peer.m_getdata_requests.end()) {
3607 if (interruptMsgProc) {
3616 const CInv &inv = *it;
3618 if (it->IsMsgProof()) {
3620 vNotFound.push_back(inv);
3625 auto proof = FindProofForGetData(peer, proofid, now);
3632 vNotFound.push_back(inv);
3639 if (it->IsMsgTx()) {
3640 if (tx_relay ==
nullptr) {
3654 std::vector<TxId> parent_ids_to_add;
3657 auto txiter = m_mempool.
GetIter(tx->GetId());
3659 auto &pentry = *txiter;
3661 (*pentry)->GetMemPoolParentsConst();
3662 parent_ids_to_add.reserve(parents.size());
3663 for (
const auto &parent : parents) {
3664 if (parent.get()->GetTime() >
3666 parent_ids_to_add.push_back(
3667 parent.get()->GetTx().GetId());
3672 for (
const TxId &parent_txid : parent_ids_to_add) {
3675 if (
WITH_LOCK(tx_relay->m_tx_inventory_mutex,
3676 return !tx_relay->m_tx_inventory_known_filter
3677 .contains(parent_txid))) {
3678 tx_relay->m_recently_announced_invs.insert(parent_txid);
3682 vNotFound.push_back(inv);
3695 if (it != peer.m_getdata_requests.end() && !pfrom.
fPauseSend) {
3696 const CInv &inv = *it++;
3698 ProcessGetBlockData(config, pfrom, peer, inv);
3704 peer.m_getdata_requests.erase(peer.m_getdata_requests.begin(), it);
3706 if (!vNotFound.empty()) {
3723void PeerManagerImpl::SendBlockTransactions(
3727 for (
size_t i = 0; i < req.
indices.size(); i++) {
3729 Misbehaving(peer,
"getblocktxn with out-of-bounds tx indices");
3738bool PeerManagerImpl::CheckHeadersPoW(
const std::vector<CBlockHeader> &headers,
3743 Misbehaving(peer,
"header with invalid proof of work");
3748 if (!CheckHeadersAreContinuous(headers)) {
3749 Misbehaving(peer,
"non-continuous headers sequence");
3762 near_chaintip_work =
3775void PeerManagerImpl::HandleUnconnectingHeaders(
3776 CNode &pfrom, Peer &peer,
const std::vector<CBlockHeader> &headers) {
3780 if (MaybeSendGetHeaders(pfrom,
GetLocator(best_header), peer)) {
3783 "received header %s: missing prev block %s, sending getheaders "
3784 "(%d) to end (peer=%d)\n",
3786 headers[0].hashPrevBlock.ToString(), best_header->nHeight,
3794 UpdateBlockAvailability(pfrom.
GetId(), headers.back().GetHash()));
3797bool PeerManagerImpl::CheckHeadersAreContinuous(
3798 const std::vector<CBlockHeader> &headers)
const {
3801 if (!hashLastBlock.
IsNull() && header.hashPrevBlock != hashLastBlock) {
3804 hashLastBlock = header.GetHash();
3809bool PeerManagerImpl::IsContinuationOfLowWorkHeadersSync(
3810 Peer &peer,
CNode &pfrom, std::vector<CBlockHeader> &headers) {
3811 if (peer.m_headers_sync) {
3812 auto result = peer.m_headers_sync->ProcessNextHeaders(
3816 if (result.success) {
3817 peer.m_last_getheaders_timestamp = {};
3819 if (result.request_more) {
3820 auto locator = peer.m_headers_sync->NextHeadersRequestLocator();
3823 Assume(!locator.vHave.empty());
3827 if (!locator.vHave.empty()) {
3830 bool sent_getheaders =
3831 MaybeSendGetHeaders(pfrom, locator, peer);
3834 locator.vHave.front().ToString(), pfrom.
GetId());
3839 peer.m_headers_sync.reset(
nullptr);
3844 LOCK(m_headers_presync_mutex);
3845 m_headers_presync_stats.erase(pfrom.
GetId());
3848 HeadersPresyncStats stats;
3849 stats.first = peer.m_headers_sync->GetPresyncWork();
3850 if (peer.m_headers_sync->GetState() ==
3852 stats.second = {peer.m_headers_sync->GetPresyncHeight(),
3853 peer.m_headers_sync->GetPresyncTime()};
3857 LOCK(m_headers_presync_mutex);
3858 m_headers_presync_stats[pfrom.
GetId()] = stats;
3860 m_headers_presync_stats.find(m_headers_presync_bestpeer);
3861 bool best_updated =
false;
3862 if (best_it == m_headers_presync_stats.end()) {
3867 const HeadersPresyncStats *stat_best{
nullptr};
3868 for (
const auto &[_peer, _stat] : m_headers_presync_stats) {
3869 if (!stat_best || _stat > *stat_best) {
3874 m_headers_presync_bestpeer = peer_best;
3875 best_updated = (peer_best == pfrom.
GetId());
3876 }
else if (best_it->first == pfrom.
GetId() ||
3877 stats > best_it->second) {
3880 m_headers_presync_bestpeer = pfrom.
GetId();
3881 best_updated =
true;
3883 if (best_updated && stats.second.has_value()) {
3886 m_headers_presync_should_signal =
true;
3890 if (result.success) {
3893 headers.swap(result.pow_validated_headers);
3896 return result.success;
3904bool PeerManagerImpl::TryLowWorkHeadersSync(
3906 std::vector<CBlockHeader> &headers) {
3913 arith_uint256 minimum_chain_work = GetAntiDoSWorkThreshold();
3917 if (total_work < minimum_chain_work) {
3931 LOCK(peer.m_headers_sync_mutex);
3932 peer.m_headers_sync.reset(
3934 chain_start_header, minimum_chain_work));
3939 (void)IsContinuationOfLowWorkHeadersSync(peer, pfrom, headers);
3942 "Ignoring low-work chain (height=%u) from peer=%d\n",
3943 chain_start_header->
nHeight + headers.size(),
3955bool PeerManagerImpl::IsAncestorOfBestHeaderOrTip(
const CBlockIndex *header) {
3956 return header !=
nullptr &&
3957 ((m_chainman.m_best_header !=
nullptr &&
3959 m_chainman.m_best_header->GetAncestor(header->
nHeight)) ||
3963bool PeerManagerImpl::MaybeSendGetHeaders(
CNode &pfrom,
3970 if (current_time - peer.m_last_getheaders_timestamp >
3973 peer.m_last_getheaders_timestamp = current_time;
3985void PeerManagerImpl::HeadersDirectFetchBlocks(
const Config &config,
3989 CNodeState *nodestate = State(pfrom.
GetId());
3993 std::vector<const CBlockIndex *> vToFetch;
3999 if (!pindexWalk->nStatus.hasData() &&
4002 vToFetch.push_back(pindexWalk);
4004 pindexWalk = pindexWalk->
pprev;
4015 std::vector<CInv> vGetData;
4018 if (nodestate->vBlocksInFlight.size() >=
4024 BlockRequested(config, pfrom.
GetId(), *pindex);
4028 if (vGetData.size() > 1) {
4030 "Downloading blocks toward %s (%d) via headers "
4035 if (vGetData.size() > 0) {
4036 if (!m_opts.ignore_incoming_txs &&
4037 nodestate->m_provides_cmpctblocks && vGetData.size() == 1 &&
4038 mapBlocksInFlight.size() == 1 &&
4055void PeerManagerImpl::UpdatePeerStateForReceivedHeaders(
4057 bool received_new_header,
bool may_have_more_headers) {
4060 CNodeState *nodestate = State(pfrom.
GetId());
4068 if (received_new_header &&
4070 nodestate->m_last_block_announcement =
GetTime();
4078 if (nodestate->pindexBestKnownBlock &&
4079 nodestate->pindexBestKnownBlock->nChainWork <
4090 LogPrintf(
"Disconnecting outbound peer %d -- headers "
4091 "chain has insufficient work\n",
4105 nodestate->pindexBestKnownBlock !=
nullptr) {
4106 if (m_outbound_peers_with_protect_from_disconnect <
4108 nodestate->pindexBestKnownBlock->nChainWork >=
4110 !nodestate->m_chain_sync.m_protect) {
4113 nodestate->m_chain_sync.m_protect =
true;
4114 ++m_outbound_peers_with_protect_from_disconnect;
4119void PeerManagerImpl::ProcessHeadersMessage(
const Config &config,
CNode &pfrom,
4121 std::vector<CBlockHeader> &&headers,
4122 bool via_compact_block) {
4123 size_t nCount = headers.size();
4131 LOCK(peer.m_headers_sync_mutex);
4132 if (peer.m_headers_sync) {
4133 peer.m_headers_sync.reset(
nullptr);
4134 LOCK(m_headers_presync_mutex);
4135 m_headers_presync_stats.erase(pfrom.
GetId());
4140 peer.m_last_getheaders_timestamp = {};
4148 if (!CheckHeadersPoW(headers, m_chainparams.
GetConsensus(), peer)) {
4163 bool already_validated_work =
false;
4166 bool have_headers_sync =
false;
4168 LOCK(peer.m_headers_sync_mutex);
4170 already_validated_work =
4171 IsContinuationOfLowWorkHeadersSync(peer, pfrom, headers);
4183 if (headers.empty()) {
4187 have_headers_sync = !!peer.m_headers_sync;
4193 headers[0].hashPrevBlock))};
4194 bool headers_connect_blockindex{chain_start_header !=
nullptr};
4196 if (!headers_connect_blockindex) {
4200 HandleUnconnectingHeaders(pfrom, peer, headers);
4208 peer.m_last_getheaders_timestamp = {};
4217 last_received_header =
4219 if (IsAncestorOfBestHeaderOrTip(last_received_header)) {
4220 already_validated_work =
true;
4228 already_validated_work =
true;
4234 if (!already_validated_work &&
4235 TryLowWorkHeadersSync(peer, pfrom, chain_start_header, headers)) {
4247 bool received_new_header{last_received_header ==
nullptr};
4252 state, &pindexLast)) {
4254 MaybePunishNodeForBlock(pfrom.
GetId(), state, via_compact_block,
4255 "invalid header received");
4265 if (MaybeSendGetHeaders(pfrom,
GetLocator(pindexLast), peer)) {
4268 "more getheaders (%d) to end to peer=%d (startheight:%d)\n",
4269 pindexLast->
nHeight, pfrom.
GetId(), peer.m_starting_height);
4273 UpdatePeerStateForReceivedHeaders(pfrom, peer, *pindexLast,
4274 received_new_header,
4278 HeadersDirectFetchBlocks(config, pfrom, *pindexLast);
4281void PeerManagerImpl::ProcessInvalidTx(
NodeId nodeid,
4284 bool maybe_add_extra_compact_tx) {
4289 const TxId &txid = ptx->GetId();
4309 m_recent_rejects_package_reconsiderable.insert(txid);
4311 m_recent_rejects.insert(txid);
4313 m_txrequest.ForgetInvId(txid);
4316 AddToCompactExtraTransactions(ptx);
4319 MaybePunishNodeForTx(nodeid, state);
4325 return orphanage.EraseTx(txid);
4339 m_txrequest.ForgetInvId(tx->GetId());
4345 orphanage.
EraseTx(tx->GetId());
4350 "AcceptToMemoryPool: peer=%d: accepted %s (poolsz %u txn, %u kB)\n",
4351 nodeid, tx->GetId().ToString(), m_mempool.
size(),
4354 RelayTransaction(tx->GetId());
4357void PeerManagerImpl::ProcessPackageResult(
4358 const PackageToValidate &package_to_validate,
4364 const auto &
package = package_to_validate.m_txns;
4365 const auto &senders = package_to_validate.m_senders;
4368 m_recent_rejects_package_reconsiderable.insert(
GetPackageHash(package));
4372 if (!
Assume(package.size() == 2)) {
4378 auto package_iter = package.rbegin();
4379 auto senders_iter = senders.rbegin();
4380 while (package_iter != package.rend()) {
4381 const auto &tx = *package_iter;
4382 const NodeId nodeid = *senders_iter;
4383 const auto it_result{package_result.
m_tx_results.find(tx->GetId())};
4387 const auto &tx_result = it_result->second;
4388 switch (tx_result.m_result_type) {
4390 ProcessValidTx(nodeid, tx);
4400 ProcessInvalidTx(nodeid, tx, tx_result.m_state,
4417std::optional<PeerManagerImpl::PackageToValidate>
4423 const auto &parent_txid{ptx->GetId()};
4425 Assume(m_recent_rejects_package_reconsiderable.contains(parent_txid));
4431 const auto cpfp_candidates_same_peer{
4437 for (
const auto &child : cpfp_candidates_same_peer) {
4438 Package maybe_cpfp_package{ptx, child};
4439 if (!m_recent_rejects_package_reconsiderable.contains(
4441 return PeerManagerImpl::PackageToValidate{ptx, child, nodeid,
4455 const auto cpfp_candidates_different_peer{
4465 std::vector<size_t> tx_indices(cpfp_candidates_different_peer.size());
4466 std::iota(tx_indices.begin(), tx_indices.end(), 0);
4467 Shuffle(tx_indices.begin(), tx_indices.end(), m_rng);
4469 for (
const auto index : tx_indices) {
4472 const auto [child_tx, child_sender] =
4473 cpfp_candidates_different_peer.at(index);
4474 Package maybe_cpfp_package{ptx, child_tx};
4475 if (!m_recent_rejects_package_reconsiderable.contains(
4477 return PeerManagerImpl::PackageToValidate{ptx, child_tx, nodeid,
4481 return std::nullopt;
4484bool PeerManagerImpl::ProcessOrphanTx(
const Config &config, Peer &peer) {
4490 return orphanage.GetTxToReconsider(peer.m_id);
4495 const TxId &orphanTxId = porphanTx->GetId();
4500 ProcessValidTx(peer.m_id, porphanTx);
4506 " invalid orphan tx %s from peer=%d. %s\n",
4513 ProcessInvalidTx(peer.m_id, porphanTx, state,
4524bool PeerManagerImpl::PrepareBlockFilterRequest(
4526 const BlockHash &stop_hash, uint32_t max_height_diff,
4528 const bool supported_filter_type =
4531 if (!supported_filter_type) {
4533 "peer %d requested unsupported block filter type: %d\n",
4534 node.GetId(),
static_cast<uint8_t
>(filter_type));
4535 node.fDisconnect =
true;
4545 if (!stop_index || !BlockRequestAllowed(stop_index)) {
4548 node.fDisconnect =
true;
4553 uint32_t stop_height = stop_index->
nHeight;
4554 if (start_height > stop_height) {
4557 "peer %d sent invalid getcfilters/getcfheaders with "
4559 "start height %d and stop height %d\n",
4560 node.GetId(), start_height, stop_height);
4561 node.fDisconnect =
true;
4564 if (stop_height - start_height >= max_height_diff) {
4566 "peer %d requested too many cfilters/cfheaders: %d / %d\n",
4567 node.GetId(), stop_height - start_height + 1, max_height_diff);
4568 node.fDisconnect =
true;
4573 if (!filter_index) {
4582void PeerManagerImpl::ProcessGetCFilters(
CNode &
node, Peer &peer,
4584 uint8_t filter_type_ser;
4585 uint32_t start_height;
4588 vRecv >> filter_type_ser >> start_height >> stop_hash;
4595 if (!PrepareBlockFilterRequest(
node, peer, filter_type, start_height,
4601 std::vector<BlockFilter> filters;
4604 "Failed to find block filter in index: filter_type=%s, "
4605 "start_height=%d, stop_hash=%s\n",
4611 for (
const auto &filter : filters) {
4616void PeerManagerImpl::ProcessGetCFHeaders(
CNode &
node, Peer &peer,
4618 uint8_t filter_type_ser;
4619 uint32_t start_height;
4622 vRecv >> filter_type_ser >> start_height >> stop_hash;
4629 if (!PrepareBlockFilterRequest(
node, peer, filter_type, start_height,
4636 if (start_height > 0) {
4638 stop_index->
GetAncestor(
static_cast<int>(start_height - 1));
4641 "Failed to find block filter header in index: "
4642 "filter_type=%s, block_hash=%s\n",
4649 std::vector<uint256> filter_hashes;
4653 "Failed to find block filter hashes in index: filter_type=%s, "
4654 "start_height=%d, stop_hash=%s\n",
4661 stop_index->
GetBlockHash(), prev_header, filter_hashes);
4664void PeerManagerImpl::ProcessGetCFCheckPt(
CNode &
node, Peer &peer,
4666 uint8_t filter_type_ser;
4669 vRecv >> filter_type_ser >> stop_hash;
4676 if (!PrepareBlockFilterRequest(
4677 node, peer, filter_type, 0, stop_hash,
4678 std::numeric_limits<uint32_t>::max(),
4679 stop_index, filter_index)) {
4687 for (
int i = headers.size() - 1; i >= 0; i--) {
4693 "Failed to find block filter header in index: "
4694 "filter_type=%s, block_hash=%s\n",
4717PeerManagerImpl::GetAvalancheVoteForBlock(
const BlockHash &hash)
const {
4728 if (pindex->nStatus.isInvalid()) {
4733 if (pindex->nStatus.isOnParkedChain()) {
4741 if (pindex == pindexFork) {
4746 if (pindexFork != pindexTip) {
4751 if (!pindex->nStatus.hasData()) {
4762 const TxId &
id)
const {
4764 if (
WITH_LOCK(m_recent_confirmed_transactions_mutex,
4765 return m_recent_confirmed_transactions.contains(
id))) {
4774 if (m_recent_rejects.contains(
id)) {
4786 if (
auto iter = m_mempool.
GetIter(
id)) {
4787 mempool_tx = (**iter)->GetSharedTx();
4792 return conflicting.HaveTx(id);
4799 return orphanage.HaveTx(id);
4865 const std::shared_ptr<const CBlock> &block,
4866 bool force_processing,
4867 bool min_pow_checked) {
4868 bool new_block{
false};
4870 &new_block, m_avalanche);
4872 node.m_last_block_time = GetTime<std::chrono::seconds>();
4877 RemoveBlockRequest(block->GetHash(), std::nullopt);
4880 mapBlockSource.erase(block->GetHash());
4884void PeerManagerImpl::ProcessMessage(
4885 const Config &config,
CNode &pfrom,
const std::string &msg_type,
4886 DataStream &vRecv,
const std::chrono::microseconds time_received,
4887 const std::atomic<bool> &interruptMsgProc) {
4893 PeerRef peer = GetPeerRef(pfrom.
GetId());
4894 if (peer ==
nullptr) {
4900 "Avalanche is not initialized, ignoring %s message\n",
4915 uint64_t nNonce = 1;
4918 std::string cleanSubVer;
4919 int starting_height = -1;
4921 uint64_t nExtraEntropy = 1;
4923 vRecv >> nVersion >> Using<CustomUintFormatter<8>>(nServices) >> nTime;
4936 "peer=%d does not offer the expected services "
4937 "(%08x offered, %08x expected); disconnecting\n",
4938 pfrom.
GetId(), nServices,
4948 "peer=%d does not offer the avalanche service; disconnecting\n",
4957 "peer=%d using obsolete version %i; disconnecting\n",
4958 pfrom.
GetId(), nVersion);
4963 if (!vRecv.
empty()) {
4972 if (!vRecv.
empty()) {
4973 std::string strSubVer;
4977 if (!vRecv.
empty()) {
4978 vRecv >> starting_height;
4980 if (!vRecv.
empty()) {
4983 if (!vRecv.
empty()) {
4984 vRecv >> nExtraEntropy;
4988 LogPrintf(
"connected to self at %s, disconnecting\n",
5001 PushNodeVersion(config, pfrom, *peer);
5005 const int greatest_common_version =
5017 peer->m_their_services = nServices;
5021 pfrom.cleanSubVer = cleanSubVer;
5023 peer->m_starting_height = starting_height;
5031 (fRelay || (peer->m_our_services &
NODE_BLOOM))) {
5032 auto *
const tx_relay = peer->SetTxRelay();
5034 LOCK(tx_relay->m_bloom_filter_mutex);
5036 tx_relay->m_relay_txs = fRelay;
5049 CNodeState *state = State(pfrom.
GetId());
5050 state->fPreferredDownload =
5054 m_num_preferred_download_peers += state->fPreferredDownload;
5060 bool send_getaddr{
false};
5062 send_getaddr = SetupAddressRelay(pfrom, *peer);
5073 peer->m_getaddr_sent =
true;
5077 WITH_LOCK(peer->m_addr_token_bucket_mutex,
5078 peer->m_addr_token_bucket += m_opts.max_addr_to_send);
5099 std::string remoteAddr;
5105 "receive version message: [%s] %s: version %d, blocks=%d, "
5106 "us=%s, txrelay=%d, peer=%d%s\n",
5108 peer->m_starting_height, addrMe.
ToString(), fRelay,
5109 pfrom.
GetId(), remoteAddr);
5111 int64_t currentTime =
GetTime();
5112 int64_t nTimeOffset = nTime - currentTime;
5117 Misbehaving(*peer,
"Ignoring invalid timestamp in version message");
5127 "feeler connection completed peer=%d; disconnecting\n",
5136 Misbehaving(*peer,
"non-version message before version handshake");
5143 "ignoring redundant verack message from peer=%d\n",
5150 "New outbound peer connected: version: %d, blocks=%d, "
5152 pfrom.
nVersion.load(), peer->m_starting_height, pfrom.
GetId(),
5173 AddKnownProof(*peer, localProof->getId());
5177 peer->m_proof_relay->m_recently_announced_proofs.insert(
5178 localProof->getId());
5183 if (
auto tx_relay = peer->GetTxRelay()) {
5192 return tx_relay->m_tx_inventory_to_send.empty() &&
5193 tx_relay->m_next_inv_send_time == 0s));
5202 Misbehaving(*peer,
"non-verack message before version handshake");
5207 const auto ser_params{
5216 std::vector<CAddress> vAddr;
5220 if (!SetupAddressRelay(pfrom, *peer)) {
5226 if (vAddr.size() > m_opts.max_addr_to_send) {
5227 Misbehaving(*peer,
strprintf(
"%s message size = %u", msg_type,
5233 std::vector<CAddress> vAddrOk;
5234 const auto current_a_time{Now<NodeSeconds>()};
5237 const auto current_time = GetTime<std::chrono::microseconds>();
5239 LOCK(peer->m_addr_token_bucket_mutex);
5242 const auto time_diff =
5243 std::max(current_time - peer->m_addr_token_timestamp, 0us);
5244 const double increment =
5246 peer->m_addr_token_bucket =
5247 std::min<double>(peer->m_addr_token_bucket + increment,
5251 peer->m_addr_token_timestamp = current_time;
5253 const bool rate_limited =
5255 uint64_t num_proc = 0;
5256 uint64_t num_rate_limit = 0;
5257 Shuffle(vAddr.begin(), vAddr.end(), m_rng);
5259 if (interruptMsgProc) {
5264 LOCK(peer->m_addr_token_bucket_mutex);
5266 if (peer->m_addr_token_bucket < 1.0) {
5272 peer->m_addr_token_bucket -= 1.0;
5285 addr.
nTime > current_a_time + 10min) {
5286 addr.
nTime = current_a_time - 5 * 24h;
5288 AddAddressKnown(*peer, addr);
5297 if (addr.
nTime > current_a_time - 10min && !peer->m_getaddr_sent &&
5300 RelayAddress(pfrom.
GetId(), addr, fReachable);
5304 vAddrOk.push_back(addr);
5307 peer->m_addr_processed += num_proc;
5308 peer->m_addr_rate_limited += num_rate_limit;
5310 "Received addr: %u addresses (%u processed, %u rate-limited) "
5312 vAddr.size(), num_proc, num_rate_limit, pfrom.
GetId());
5314 m_addrman.
Add(vAddrOk, pfrom.
addr, 2h);
5315 if (vAddr.size() < 1000) {
5316 peer->m_getaddr_sent =
false;
5323 "addrfetch connection completed peer=%d; disconnecting\n",
5331 peer->m_wants_addrv2 =
true;
5336 peer->m_prefers_headers =
true;
5341 bool sendcmpct_hb{
false};
5342 uint64_t sendcmpct_version{0};
5343 vRecv >> sendcmpct_hb >> sendcmpct_version;
5350 CNodeState *nodestate = State(pfrom.
GetId());
5351 nodestate->m_provides_cmpctblocks =
true;
5352 nodestate->m_requested_hb_cmpctblocks = sendcmpct_hb;
5361 std::vector<CInv> vInv;
5364 Misbehaving(*peer,
strprintf(
"inv message size = %u", vInv.size()));
5368 const bool reject_tx_invs{RejectIncomingTxs(pfrom)};
5370 const auto current_time{GetTime<std::chrono::microseconds>()};
5371 std::optional<BlockHash> best_block;
5373 auto logInv = [&](
const CInv &inv,
bool fAlreadyHave) {
5375 fAlreadyHave ?
"have" :
"new", pfrom.
GetId());
5378 for (
CInv &inv : vInv) {
5379 if (interruptMsgProc) {
5391 const bool fAlreadyHave = AlreadyHaveBlock(
BlockHash(inv.
hash));
5392 logInv(inv, fAlreadyHave);
5395 UpdateBlockAvailability(pfrom.
GetId(), hash);
5397 !IsBlockRequested(hash)) {
5404 best_block = std::move(hash);
5415 const bool fAlreadyHave = AlreadyHaveProof(proofid);
5416 logInv(inv, fAlreadyHave);
5417 AddKnownProof(*peer, proofid);
5419 if (!fAlreadyHave && m_avalanche &&
5421 const bool preferred = isPreferredDownloadPeer(pfrom);
5423 LOCK(cs_proofrequest);
5424 AddProofAnnouncement(pfrom, proofid, current_time,
5433 const bool fAlreadyHave =
5434 AlreadyHaveTx(txid,
true);
5435 logInv(inv, fAlreadyHave);
5437 AddKnownTx(*peer, txid);
5438 if (reject_tx_invs) {
5440 "transaction (%s) inv sent in violation of "
5441 "protocol, disconnecting peer=%d\n",
5445 }
else if (!fAlreadyHave &&
5447 AddTxAnnouncement(pfrom, txid, current_time);
5454 "Unknown inv type \"%s\" received from peer=%d\n",
5471 if (state.fSyncStarted ||
5472 (!peer->m_inv_triggered_getheaders_before_sync &&
5473 *best_block != m_last_block_inv_triggering_headers_sync)) {
5474 if (MaybeSendGetHeaders(
5475 pfrom,
GetLocator(m_chainman.m_best_header), *peer)) {
5477 m_chainman.m_best_header->nHeight,
5478 best_block->ToString(), pfrom.
GetId());
5480 if (!state.fSyncStarted) {
5481 peer->m_inv_triggered_getheaders_before_sync =
true;
5485 m_last_block_inv_triggering_headers_sync = *best_block;
5494 std::vector<CInv> vInv;
5498 strprintf(
"getdata message size = %u", vInv.size()));
5503 vInv.size(), pfrom.
GetId());
5505 if (vInv.size() > 0) {
5511 LOCK(peer->m_getdata_requests_mutex);
5512 peer->m_getdata_requests.insert(peer->m_getdata_requests.end(),
5513 vInv.begin(), vInv.end());
5514 ProcessGetData(config, pfrom, *peer, interruptMsgProc);
5523 vRecv >> locator >> hashStop;
5527 "getblocks locator size %lld > %d, disconnect peer=%d\n",
5541 std::shared_ptr<const CBlock> a_recent_block;
5543 LOCK(m_most_recent_block_mutex);
5544 a_recent_block = m_most_recent_block;
5548 state, a_recent_block, m_avalanche)) {
5566 (pindex ? pindex->
nHeight : -1),
5569 for (; pindex; pindex = m_chainman.
ActiveChain().Next(pindex)) {
5578 const int nPrunedBlocksLikelyToHave =
5582 (!pindex->nStatus.hasData() ||
5584 nPrunedBlocksLikelyToHave)) {
5587 " getblocks stopping, pruned or too old block at %d %s\n",
5592 peer->m_block_inv_mutex,
5593 peer->m_blocks_for_inv_relay.push_back(pindex->
GetBlockHash()));
5594 if (--nLimit <= 0) {
5600 peer->m_continuation_block = pindex->GetBlockHash();
5612 std::shared_ptr<const CBlock> recent_block;
5614 LOCK(m_most_recent_block_mutex);
5615 if (m_most_recent_block_hash == req.
blockhash) {
5616 recent_block = m_most_recent_block;
5621 SendBlockTransactions(pfrom, *peer, *recent_block, req);
5631 if (!pindex || !pindex->nStatus.hasData()) {
5634 "Peer %d sent us a getblocktxn for a block we don't have\n",
5645 if (!block_pos.IsNull()) {
5653 SendBlockTransactions(pfrom, *peer, block, req);
5665 "Peer %d sent us a getblocktxn for a block > %i deep\n",
5670 WITH_LOCK(peer->m_getdata_requests_mutex,
5671 peer->m_getdata_requests.push_back(inv));
5680 vRecv >> locator >> hashStop;
5684 "getheaders locator size %lld > %d, disconnect peer=%d\n",
5693 "Ignoring getheaders from peer=%d while importing/reindexing\n",
5707 if (m_chainman.
ActiveTip() ==
nullptr ||
5712 "Ignoring getheaders from peer=%d because active chain "
5713 "has too little work; sending empty response\n",
5718 std::vector<CBlock>());
5722 CNodeState *nodestate = State(pfrom.
GetId());
5731 if (!BlockRequestAllowed(pindex)) {
5733 "%s: ignoring request from peer=%i for old block "
5734 "header that isn't in the main chain\n",
5735 __func__, pfrom.
GetId());
5749 std::vector<CBlock> vHeaders;
5752 (pindex ? pindex->
nHeight : -1),
5755 for (; pindex; pindex = m_chainman.
ActiveChain().Next(pindex)) {
5757 if (--nLimit <= 0 || pindex->GetBlockHash() == hashStop) {
5774 nodestate->pindexBestHeaderSent =
5781 if (RejectIncomingTxs(pfrom)) {
5783 "transaction sent in violation of protocol peer=%d\n",
5799 const CTransaction &tx = *ptx;
5800 const TxId &txid = tx.GetId();
5801 AddKnownTx(*peer, txid);
5806 m_txrequest.ReceivedResponse(pfrom.
GetId(), txid);
5808 if (AlreadyHaveTx(txid,
true)) {
5814 if (!m_mempool.
exists(tx.GetId())) {
5816 "Not relaying non-mempool transaction %s from "
5817 "forcerelay peer=%d\n",
5818 tx.GetId().ToString(), pfrom.
GetId());
5820 LogPrintf(
"Force relaying tx %s from peer=%d\n",
5821 tx.GetId().ToString(), pfrom.
GetId());
5822 RelayTransaction(tx.GetId());
5826 if (m_recent_rejects_package_reconsiderable.contains(txid)) {
5834 "found tx %s in reconsiderable rejects, looking for "
5835 "child in orphanage\n",
5837 if (
auto package_to_validate{
5838 Find1P1CPackage(ptx, pfrom.
GetId())}) {
5841 package_to_validate->m_txns,
5844 "package evaluation for %s: %s (%s)\n",
5845 package_to_validate->ToString(),
5847 ?
"package accepted"
5848 :
"package rejected",
5850 ProcessPackageResult(package_to_validate.value(),
5879 ProcessValidTx(pfrom.
GetId(), ptx);
5885 bool fRejectedParents =
false;
5889 std::vector<TxId> unique_parents;
5890 unique_parents.reserve(tx.vin.size());
5891 for (
const CTxIn &txin : tx.vin) {
5894 unique_parents.push_back(txin.prevout.GetTxId());
5896 std::sort(unique_parents.begin(), unique_parents.end());
5897 unique_parents.erase(
5898 std::unique(unique_parents.begin(), unique_parents.end()),
5899 unique_parents.end());
5907 std::optional<TxId> rejected_parent_reconsiderable;
5908 for (
const TxId &parent_txid : unique_parents) {
5909 if (m_recent_rejects.contains(parent_txid)) {
5910 fRejectedParents =
true;
5914 if (m_recent_rejects_package_reconsiderable.contains(
5916 !m_mempool.
exists(parent_txid)) {
5921 if (rejected_parent_reconsiderable.has_value()) {
5922 fRejectedParents =
true;
5925 rejected_parent_reconsiderable = parent_txid;
5928 if (!fRejectedParents) {
5929 const auto current_time{
5930 GetTime<std::chrono::microseconds>()};
5932 for (
const TxId &parent_txid : unique_parents) {
5934 AddKnownTx(*peer, parent_txid);
5938 if (!AlreadyHaveTx(parent_txid,
5940 AddTxAnnouncement(pfrom, parent_txid, current_time);
5946 if (
unsigned int nEvicted =
5950 if (orphanage.AddTx(ptx,
5952 AddToCompactExtraTransactions(ptx);
5955 m_opts.max_orphan_txs, m_rng);
5958 "orphanage overflow, removed %u tx\n",
5964 m_txrequest.ForgetInvId(tx.GetId());
5968 "not keeping orphan with rejected parents %s\n",
5969 tx.GetId().ToString());
5972 m_recent_rejects.insert(tx.GetId());
5973 m_txrequest.ForgetInvId(tx.GetId());
5977 ProcessInvalidTx(pfrom.
GetId(), ptx, state,
5987 "tx %s failed but reconsiderable, looking for child in "
5990 if (
auto package_to_validate{
5991 Find1P1CPackage(ptx, pfrom.
GetId())}) {
5994 package_to_validate->m_txns,
false)};
5996 "package evaluation for %s: %s (%s)\n",
5997 package_to_validate->ToString(),
5999 ?
"package accepted"
6000 :
"package rejected",
6002 ProcessPackageResult(package_to_validate.value(),
6011 m_txrequest.ForgetInvId(tx.GetId());
6013 unsigned int nEvicted{0};
6020 m_opts.max_conflicting_txs, m_rng);
6025 "conflicting pool overflow, removed %u tx\n",
6038 "Unexpected cmpctblock message received from peer %d\n",
6045 vRecv >> cmpctblock;
6046 }
catch (std::ios_base::failure &e) {
6048 Misbehaving(*peer,
"cmpctblock-bad-indexes");
6052 bool received_new_header =
false;
6065 MaybeSendGetHeaders(
6066 pfrom,
GetLocator(m_chainman.m_best_header), *peer);
6072 GetAntiDoSWorkThreshold()) {
6076 "Ignoring low-work compact block from peer %d\n",
6082 received_new_header =
true;
6092 MaybePunishNodeForBlock(pfrom.
GetId(), state,
6094 "invalid header via cmpctblock");
6099 if (received_new_header) {
6100 LogInfo(
"Saw new cmpctblock header hash=%s peer=%d\n",
6101 blockhash.ToString(), pfrom.
GetId());
6108 bool fProcessBLOCKTXN =
false;
6114 bool fRevertToHeaderProcessing =
false;
6118 std::shared_ptr<CBlock> pblock = std::make_shared<CBlock>();
6119 bool fBlockReconstructed =
false;
6127 CNodeState *nodestate = State(pfrom.
GetId());
6131 if (received_new_header &&
6134 nodestate->m_last_block_announcement =
GetTime();
6137 if (pindex->nStatus.hasData()) {
6144 size_t already_in_flight =
6145 std::distance(range_flight.first, range_flight.second);
6146 bool requested_block_from_this_peer{
false};
6150 bool first_in_flight =
6151 already_in_flight == 0 ||
6152 (range_flight.first->second.first == pfrom.
GetId());
6154 while (range_flight.first != range_flight.second) {
6155 if (range_flight.first->second.first == pfrom.
GetId()) {
6156 requested_block_from_this_peer =
true;
6159 range_flight.first++;
6168 if (requested_block_from_this_peer) {
6172 std::vector<CInv> vInv(1);
6181 if (!already_in_flight && !CanDirectFetch()) {
6189 nodestate->vBlocksInFlight.size() <
6191 requested_block_from_this_peer) {
6192 std::list<QueuedBlock>::iterator *queuedBlockIt =
nullptr;
6193 if (!BlockRequested(config, pfrom.
GetId(), *pindex,
6195 if (!(*queuedBlockIt)->partialBlock) {
6197 ->partialBlock.reset(
6204 "we were already syncing!\n");
6210 *(*queuedBlockIt)->partialBlock;
6212 partialBlock.
InitData(cmpctblock, vExtraTxnForCompact);
6218 Misbehaving(*peer,
"invalid compact block");
6221 if (first_in_flight) {
6224 std::vector<CInv> vInv(1);
6237 for (
size_t i = 0; i < cmpctblock.
BlockTxCount(); i++) {
6248 fProcessBLOCKTXN =
true;
6249 }
else if (first_in_flight) {
6256 IsBlockRequestedFromOutbound(blockhash) ||
6279 tempBlock.InitData(cmpctblock, vExtraTxnForCompact);
6284 std::vector<CTransactionRef> dummy;
6285 status = tempBlock.FillBlock(*pblock, dummy);
6287 fBlockReconstructed =
true;
6291 if (requested_block_from_this_peer) {
6295 std::vector<CInv> vInv(1);
6302 fRevertToHeaderProcessing =
true;
6307 if (fProcessBLOCKTXN) {
6309 blockTxnMsg, time_received, interruptMsgProc);
6312 if (fRevertToHeaderProcessing) {
6318 return ProcessHeadersMessage(config, pfrom, *peer,
6323 if (fBlockReconstructed) {
6328 mapBlockSource.emplace(pblock->GetHash(),
6329 std::make_pair(pfrom.
GetId(),
false));
6340 ProcessBlock(config, pfrom, pblock,
true,
6349 RemoveBlockRequest(pblock->GetHash(), std::nullopt);
6359 "Unexpected blocktxn message received from peer %d\n",
6367 std::shared_ptr<CBlock> pblock = std::make_shared<CBlock>();
6368 bool fBlockRead =
false;
6372 auto range_flight = mapBlocksInFlight.equal_range(resp.
blockhash);
6373 size_t already_in_flight =
6374 std::distance(range_flight.first, range_flight.second);
6375 bool requested_block_from_this_peer{
false};
6379 bool first_in_flight =
6380 already_in_flight == 0 ||
6381 (range_flight.first->second.first == pfrom.
GetId());
6383 while (range_flight.first != range_flight.second) {
6384 auto [node_id, block_it] = range_flight.first->second;
6385 if (node_id == pfrom.
GetId() && block_it->partialBlock) {
6386 requested_block_from_this_peer =
true;
6389 range_flight.first++;
6392 if (!requested_block_from_this_peer) {
6394 "Peer %d sent us block transactions for block "
6395 "we weren't expecting\n",
6401 *range_flight.first->second.second->partialBlock;
6409 "invalid compact block/non-matching block transactions");
6412 if (first_in_flight) {
6414 std::vector<CInv> invs;
6421 "Peer %d sent us a compact block but it failed to "
6422 "reconstruct, waiting on first download to complete\n",
6455 std::make_pair(pfrom.
GetId(),
false));
6466 ProcessBlock(config, pfrom, pblock,
true,
6476 "Unexpected headers message received from peer %d\n",
6481 std::vector<CBlockHeader> headers;
6488 strprintf(
"too-many-headers: headers message size = %u",
6492 headers.resize(nCount);
6493 for (
unsigned int n = 0; n < nCount; n++) {
6494 vRecv >> headers[n];
6499 ProcessHeadersMessage(config, pfrom, *peer, std::move(headers),
6505 if (m_headers_presync_should_signal.exchange(
false)) {
6506 HeadersPresyncStats stats;
6508 LOCK(m_headers_presync_mutex);
6510 m_headers_presync_stats.find(m_headers_presync_bestpeer);
6511 if (it != m_headers_presync_stats.end()) {
6517 stats.first, stats.second->first, stats.second->second);
6528 "Unexpected block message received from peer %d\n",
6533 std::shared_ptr<CBlock> pblock = std::make_shared<CBlock>();
6537 pblock->GetHash().ToString(), pfrom.
GetId());
6542 pblock->hashPrevBlock))};
6546 "Received mutated block from peer=%d\n", peer->m_id);
6547 Misbehaving(*peer,
"mutated block");
6549 RemoveBlockRequest(pblock->GetHash(), peer->m_id));
6559 const BlockHash hash = pblock->GetHash();
6560 bool min_pow_checked =
false;
6565 forceProcessing = IsBlockRequested(hash);
6566 RemoveBlockRequest(hash, pfrom.
GetId());
6570 mapBlockSource.emplace(hash, std::make_pair(pfrom.
GetId(),
true));
6576 GetAntiDoSWorkThreshold()) {
6577 min_pow_checked =
true;
6580 ProcessBlock(config, pfrom, pblock, forceProcessing, min_pow_checked);
6590 if (pfrom.m_avalanche_pubkey.has_value()) {
6593 "Ignoring avahello from peer %d: already in our node set\n",
6599 vRecv >> delegation;
6606 if (!delegation.
verify(state, pubkey)) {
6607 Misbehaving(*peer,
"invalid-delegation");
6610 pfrom.m_avalanche_pubkey = std::move(pubkey);
6613 sighasher << delegation.
getId();
6621 if (!(*pfrom.m_avalanche_pubkey)
6622 .VerifySchnorr(sighasher.GetHash(),
sig)) {
6623 Misbehaving(*peer,
"invalid-avahello-signature");
6630 if (!AlreadyHaveProof(proofid)) {
6631 const bool preferred = isPreferredDownloadPeer(pfrom);
6632 LOCK(cs_proofrequest);
6633 AddProofAnnouncement(pfrom, proofid,
6634 GetTime<std::chrono::microseconds>(),
6652 WITH_LOCK(peer->m_addr_token_bucket_mutex,
6653 peer->m_addr_token_bucket += m_opts.max_addr_to_send);
6657 peer->m_proof_relay->compactproofs_requested =
true;
6668 const auto now = Now<SteadyMilliseconds>();
6674 last_poll + std::chrono::milliseconds(m_opts.avalanche_cooldown)) {
6676 "Ignoring repeated avapoll from peer %d: cooldown not "
6691 strprintf(
"too-many-ava-poll: poll message size = %u", nCount));
6695 std::vector<avalanche::Vote> votes;
6696 votes.reserve(nCount);
6698 bool fPreconsensus{
false};
6699 bool fStakingPreconsensus{
false};
6704 fStakingPreconsensus =
6708 for (
unsigned int n = 0; n < nCount; n++) {
6716 if (!quorum_established) {
6717 votes.emplace_back(vote, inv.
hash);
6724 if (fPreconsensus) {
6726 GetAvalancheVoteForTx(*m_avalanche,
TxId(inv.
hash));
6738 if (fStakingPreconsensus) {
6745 "poll inv type %d unknown from peer=%d\n",
6750 votes.emplace_back(vote, inv.
hash);
6776 if (!pfrom.m_avalanche_pubkey.has_value() ||
6777 !(*pfrom.m_avalanche_pubkey)
6778 .VerifySchnorr(verifier.GetHash(),
sig)) {
6779 Misbehaving(*peer,
"invalid-ava-response-signature");
6784 auto now = GetTime<std::chrono::seconds>();
6786 std::vector<avalanche::VoteItemUpdate> updates;
6787 bool disconnect{
false};
6790 disconnect, error)) {
6792 Misbehaving(*peer, error);
6810 "Repeated failure to register votes from peer %d: %s\n",
6811 pfrom.
GetId(), error);
6814 Misbehaving(*peer, error);
6827 auto logVoteUpdate = [](
const auto &voteUpdate,
6828 const std::string &voteItemTypeStr,
6829 const auto &voteItemId) {
6830 std::string voteOutcome;
6831 bool alwaysPrint =
false;
6832 switch (voteUpdate.getStatus()) {
6834 voteOutcome =
"invalidated";
6838 voteOutcome =
"rejected";
6841 voteOutcome =
"accepted";
6844 voteOutcome =
"finalized";
6847 alwaysPrint = voteItemTypeStr !=
"tx";
6850 voteOutcome =
"stalled";
6859 alwaysPrint &= (voteItemTypeStr !=
"contender");
6862 LogPrintf(
"Avalanche %s %s %s\n", voteOutcome, voteItemTypeStr,
6863 voteItemId.ToString());
6867 voteItemTypeStr, voteItemId.ToString());
6871 bool shouldActivateBestChain =
false;
6873 bool fPreconsensus{
false};
6874 bool fStakingPreconsensus{
false};
6879 fStakingPreconsensus =
6883 for (
const auto &u : updates) {
6888 if (
auto pitem = std::get_if<const avalanche::ProofRef>(&item)) {
6892 logVoteUpdate(u,
"proof", proofid);
6894 auto rejectionMode =
6896 auto nextCooldownTimePoint = GetTime<std::chrono::seconds>();
6897 switch (u.getStatus()) {
6913 return pm.rejectProof(proofid,
6917 "ERROR: Failed to reject proof: %s\n",
6923 nextCooldownTimePoint += std::chrono::seconds(
6924 m_opts.avalanche_peer_replacement_cooldown);
6930 avalanche::PeerManager::
6931 RegistrationMode::FORCE_ACCEPT);
6934 [&](const avalanche::Peer &peer) {
6935 pm.updateNextPossibleConflictTime(
6937 nextCooldownTimePoint);
6938 if (u.getStatus() ==
6939 avalanche::VoteStatus::
6941 pm.setFinalized(peer.peerid);
6949 "ERROR: Failed to accept proof: %s\n",
6956 auto getBlockFromIndex = [
this](
const CBlockIndex *pindex) {
6959 std::shared_ptr<const CBlock> pblock =
WITH_LOCK(
6960 m_most_recent_block_mutex,
return m_most_recent_block);
6962 if (!pblock || pblock->GetHash() != pindex->
GetBlockHash()) {
6963 std::shared_ptr<CBlock> pblockRead =
6964 std::make_shared<CBlock>();
6967 assert(!
"cannot load block from disk");
6969 pblock = pblockRead;
6974 if (
auto pitem = std::get_if<const CBlockIndex *>(&item)) {
6977 shouldActivateBestChain =
true;
6981 switch (u.getStatus()) {
6986 LogPrintf(
"ERROR: Database error: %s\n",
6995 LogPrintf(
"ERROR: Database error: %s\n",
7000 auto pblock = getBlockFromIndex(pindex);
7016 std::unique_ptr<node::CBlockTemplate> blockTemplate;
7020 chainstate.UnparkBlock(pindex);
7022 const bool newlyFinalized =
7023 !chainstate.IsBlockAvalancheFinalized(pindex) &&
7024 chainstate.AvalancheFinalizeBlock(pindex,
7029 if (fPreconsensus && newlyFinalized) {
7030 auto pblock = getBlockFromIndex(pindex);
7044 std::unordered_set<TxId, SaltedTxIdHasher>
7045 confirmedTxIdsInNonFinalizedBlocks;
7047 block !=
nullptr && block != pindex;
7048 block = block->
pprev) {
7050 getBlockFromIndex(block);
7052 for (
const auto &tx :
7053 currentBlock->vtx) {
7054 confirmedTxIdsInNonFinalizedBlocks
7055 .insert(tx->GetId());
7063 confirmedTxIdsInNonFinalizedBlocks);
7075 config, chainstate, &m_mempool,
7077 blockAssembler.pblocktemplate.reset(
7080 if (blockAssembler.pblocktemplate) {
7081 blockAssembler.addTxs(m_mempool);
7082 blockTemplate = std::move(
7083 blockAssembler.pblocktemplate);
7089 if (blockTemplate) {
7095 for (
const auto &templateEntry :
7109 if (fStakingPreconsensus) {
7111 std::get_if<const avalanche::StakeContenderId>(&item)) {
7113 logVoteUpdate(u,
"contender", contenderId);
7115 switch (u.getStatus()) {
7136 if (!fPreconsensus) {
7140 if (
auto pitem = std::get_if<const CTransactionRef>(&item)) {
7144 const TxId &txid = tx->GetId();
7145 const auto status{u.getStatus()};
7150 logVoteUpdate(u,
"tx", txid);
7161 if (m_mempool.
exists(txid)) {
7165 std::vector<CTransactionRef> conflictingTxs =
7171 if (conflictingTxs.size() > 0) {
7182 for (
const auto &conflictingTx :
7185 conflictingTx->GetId());
7204 m_recent_rejects.insert(txid);
7211 std::make_shared<const std::vector<Coin>>(
7227 return conflicting.HaveTx(txid);
7230 std::vector<CTransactionRef>
7231 mempool_conflicting_txs;
7232 for (
const auto &txin : tx->vin) {
7237 mempool_conflicting_txs.push_back(
7238 std::move(conflict));
7247 [&txid, &mempool_conflicting_txs](
7252 if (mempool_conflicting_txs.size() >
7255 mempool_conflicting_txs[0],
7264 auto it = m_mempool.
GetIter(txid);
7265 if (!it.has_value()) {
7268 "Error: finalized tx (%s) is not in the "
7274 std::vector<TxId> finalizedTxIds;
7275 m_mempool.setAvalancheFinalized(
7280 for (
const auto &finalized_txid : finalizedTxIds) {
7285 logVoteUpdate(u,
"tx", finalized_txid);
7293 std::vector<CTransactionRef>
7296 for (
const auto &conflictingTx :
7298 m_recent_rejects.insert(
7299 conflictingTx->GetId());
7301 conflictingTx->GetId());
7327 m_txrequest.ForgetInvId(txid);
7333 for (
auto &it : m_peer_map) {
7334 auto tx_relay = (*it.second).GetTxRelay();
7339 LOCK(tx_relay->m_tx_inventory_mutex);
7346 auto &stalled_by_time =
7347 tx_relay->m_avalanche_stalled_txids
7349 if (stalled_by_time.size() >=
7351 stalled_by_time.erase(
7352 stalled_by_time.begin()->timeAdded);
7355 tx_relay->m_avalanche_stalled_txids.insert(
7366 if (shouldActivateBestChain) {
7369 state,
nullptr, m_avalanche)) {
7384 ReceivedAvalancheProof(pfrom, *peer, proof);
7393 if (peer->m_proof_relay ==
nullptr) {
7397 peer->m_proof_relay->lastSharedProofsUpdate =
7398 GetTime<std::chrono::seconds>();
7400 peer->m_proof_relay->sharedProofs =
7406 peer->m_proof_relay->sharedProofs);
7416 if (peer->m_proof_relay ==
nullptr) {
7421 if (!peer->m_proof_relay->compactproofs_requested) {
7425 peer->m_proof_relay->compactproofs_requested =
false;
7429 vRecv >> compactProofs;
7430 }
catch (std::ios_base::failure &e) {
7432 Misbehaving(*peer,
"avaproofs-bad-indexes");
7438 if (!ReceivedAvalancheProof(pfrom, *peer, prefilledProof.proof)) {
7468 auto shortIdProcessor =
7472 if (shortIdProcessor.hasOutOfBoundIndex()) {
7475 Misbehaving(*peer,
"avaproofs-bad-indexes");
7478 if (!shortIdProcessor.isEvenlyDistributed()) {
7483 std::vector<std::pair<avalanche::ProofId, bool>> remoteProofsStatus;
7490 shortIdProcessor.matchKnownItem(shortid, peer.
proof);
7497 remoteProofsStatus.emplace_back(peer.
getProofId(),
7508 for (
size_t i = 0; i < compactProofs.
size(); i++) {
7509 if (shortIdProcessor.getItem(i) ==
nullptr) {
7525 return pfrom.m_avalanche_pubkey.has_value())) {
7528 for (
const auto &[proofid, present] : remoteProofsStatus) {
7538 if (peer->m_proof_relay ==
nullptr) {
7545 auto requestedIndiceIt = proofreq.
indices.begin();
7546 uint32_t treeIndice = 0;
7547 peer->m_proof_relay->sharedProofs.forEachLeaf([&](
const auto &proof) {
7548 if (requestedIndiceIt == proofreq.
indices.end()) {
7553 if (treeIndice++ == *requestedIndiceIt) {
7555 requestedIndiceIt++;
7561 peer->m_proof_relay->sharedProofs = {};
7574 "Ignoring \"getaddr\" from %s connection. peer=%d\n",
7581 Assume(SetupAddressRelay(pfrom, *peer));
7585 if (peer->m_getaddr_recvd) {
7590 peer->m_getaddr_recvd =
true;
7592 peer->m_addrs_to_send.clear();
7593 std::vector<CAddress> vAddr;
7594 const size_t maxAddrToSend = m_opts.max_addr_to_send;
7602 for (
const CAddress &addr : vAddr) {
7603 PushAddress(*peer, addr);
7609 auto now = GetTime<std::chrono::seconds>();
7619 if (!SetupAddressRelay(pfrom, *peer)) {
7621 "Ignoring getavaaddr message from %s peer=%d\n",
7626 auto availabilityScoreComparator = [](
const CNode *lhs,
7629 double scoreRhs = rhs->getAvailabilityScore();
7631 if (scoreLhs != scoreRhs) {
7632 return scoreLhs > scoreRhs;
7641 std::set<
const CNode *,
decltype(availabilityScoreComparator)> avaNodes(
7642 availabilityScoreComparator);
7649 avaNodes.insert(pnode);
7650 if (avaNodes.size() > m_opts.max_addr_to_send) {
7651 avaNodes.erase(std::prev(avaNodes.end()));
7655 peer->m_addrs_to_send.clear();
7656 for (
const CNode *pnode : avaNodes) {
7657 PushAddress(*peer, pnode->
addr);
7668 "mempool request with bloom filters disabled, "
7669 "disconnect peer=%d\n",
7680 "mempool request with bandwidth limit reached, "
7681 "disconnect peer=%d\n",
7688 if (
auto tx_relay = peer->GetTxRelay()) {
7689 LOCK(tx_relay->m_tx_inventory_mutex);
7690 tx_relay->m_send_mempool =
true;
7719 const auto ping_end = time_received;
7722 bool bPingFinished =
false;
7723 std::string sProblem;
7725 if (nAvail >=
sizeof(nonce)) {
7730 if (peer->m_ping_nonce_sent != 0) {
7731 if (nonce == peer->m_ping_nonce_sent) {
7734 bPingFinished =
true;
7735 const auto ping_time = ping_end - peer->m_ping_start.load();
7736 if (ping_time.count() >= 0) {
7741 sProblem =
"Timing mishap";
7745 sProblem =
"Nonce mismatch";
7749 bPingFinished =
true;
7750 sProblem =
"Nonce zero";
7754 sProblem =
"Unsolicited pong without ping";
7759 bPingFinished =
true;
7760 sProblem =
"Short payload";
7763 if (!(sProblem.empty())) {
7765 "pong peer=%d: %s, %x expected, %x received, %u bytes\n",
7766 pfrom.
GetId(), sProblem, peer->m_ping_nonce_sent, nonce,
7769 if (bPingFinished) {
7770 peer->m_ping_nonce_sent = 0;
7778 "filterload received despite not offering bloom services "
7779 "from peer=%d; disconnecting\n",
7789 Misbehaving(*peer,
"too-large bloom filter");
7790 }
else if (
auto tx_relay = peer->GetTxRelay()) {
7792 LOCK(tx_relay->m_bloom_filter_mutex);
7793 tx_relay->m_bloom_filter.reset(
new CBloomFilter(filter));
7794 tx_relay->m_relay_txs =
true;
7804 "filteradd received despite not offering bloom services "
7805 "from peer=%d; disconnecting\n",
7810 std::vector<uint8_t> vData;
7819 }
else if (
auto tx_relay = peer->GetTxRelay()) {
7820 LOCK(tx_relay->m_bloom_filter_mutex);
7821 if (tx_relay->m_bloom_filter) {
7822 tx_relay->m_bloom_filter->insert(vData);
7830 Misbehaving(*peer,
"bad filteradd message");
7838 "filterclear received despite not offering bloom services "
7839 "from peer=%d; disconnecting\n",
7844 auto tx_relay = peer->GetTxRelay();
7850 LOCK(tx_relay->m_bloom_filter_mutex);
7851 tx_relay->m_bloom_filter =
nullptr;
7852 tx_relay->m_relay_txs =
true;
7861 vRecv >> newFeeFilter;
7863 if (
auto tx_relay = peer->GetTxRelay()) {
7864 tx_relay->m_fee_filter_received = newFeeFilter;
7873 ProcessGetCFilters(pfrom, *peer, vRecv);
7878 ProcessGetCFHeaders(pfrom, *peer, vRecv);
7883 ProcessGetCFCheckPt(pfrom, *peer, vRecv);
7888 std::vector<CInv> vInv;
7894 for (
CInv &inv : vInv) {
7900 m_txrequest.ReceivedResponse(pfrom.
GetId(),
TxId(inv.
hash));
7907 LOCK(cs_proofrequest);
7908 m_proofrequest.ReceivedResponse(
7922bool PeerManagerImpl::MaybeDiscourageAndDisconnect(
CNode &pnode, Peer &peer) {
7924 LOCK(peer.m_misbehavior_mutex);
7927 if (!peer.m_should_discourage) {
7931 peer.m_should_discourage =
false;
7937 LogPrintf(
"Warning: not punishing noban peer %d!\n", peer.m_id);
7943 LogPrintf(
"Warning: not punishing manually connected peer %d!\n",
7952 "Warning: disconnecting but not discouraging %s peer %d!\n",
7969bool PeerManagerImpl::ProcessMessages(
const Config &config,
CNode *pfrom,
7970 std::atomic<bool> &interruptMsgProc) {
7982 PeerRef peer = GetPeerRef(pfrom->
GetId());
7983 if (peer ==
nullptr) {
7988 LOCK(peer->m_getdata_requests_mutex);
7989 if (!peer->m_getdata_requests.empty()) {
7990 ProcessGetData(config, *pfrom, *peer, interruptMsgProc);
7994 const bool processed_orphan = ProcessOrphanTx(config, *peer);
8000 if (processed_orphan) {
8007 LOCK(peer->m_getdata_requests_mutex);
8008 if (!peer->m_getdata_requests.empty()) {
8025 bool fMoreWork = poll_result->second;
8029 msg.m_recv.size(), msg.m_recv.data());
8031 if (m_opts.capture_messages) {
8037 if (!msg.m_valid_netmagic) {
8039 "PROCESSMESSAGE: INVALID MESSAGESTART %s peer=%d\n",
8053 if (!msg.m_valid_header) {
8061 if (!msg.m_valid_checksum) {
8073 ProcessMessage(config, *pfrom, msg.m_type, vRecv, msg.m_time,
8075 if (interruptMsgProc) {
8080 LOCK(peer->m_getdata_requests_mutex);
8081 if (!peer->m_getdata_requests.empty()) {
8090 return orphanage.HaveTxToReconsider(peer->m_id);
8094 }
catch (
const std::exception &e) {
8097 e.what(),
typeid(e).name());
8106void PeerManagerImpl::ConsiderEviction(
CNode &pto, Peer &peer,
8107 std::chrono::seconds time_in_seconds) {
8110 CNodeState &state = *State(pto.
GetId());
8113 state.fSyncStarted) {
8120 if (state.pindexBestKnownBlock !=
nullptr &&
8121 state.pindexBestKnownBlock->nChainWork >=
8123 if (state.m_chain_sync.m_timeout != 0s) {
8124 state.m_chain_sync.m_timeout = 0s;
8125 state.m_chain_sync.m_work_header =
nullptr;
8126 state.m_chain_sync.m_sent_getheaders =
false;
8128 }
else if (state.m_chain_sync.m_timeout == 0s ||
8129 (state.m_chain_sync.m_work_header !=
nullptr &&
8130 state.pindexBestKnownBlock !=
nullptr &&
8131 state.pindexBestKnownBlock->nChainWork >=
8132 state.m_chain_sync.m_work_header->nChainWork)) {
8138 state.m_chain_sync.m_work_header = m_chainman.
ActiveChain().
Tip();
8139 state.m_chain_sync.m_sent_getheaders =
false;
8140 }
else if (state.m_chain_sync.m_timeout > 0s &&
8141 time_in_seconds > state.m_chain_sync.m_timeout) {
8146 if (state.m_chain_sync.m_sent_getheaders) {
8149 "Disconnecting outbound peer %d for old chain, best known "
8152 state.pindexBestKnownBlock !=
nullptr
8153 ? state.pindexBestKnownBlock->GetBlockHash().ToString()
8157 assert(state.m_chain_sync.m_work_header);
8162 MaybeSendGetHeaders(
8163 pto,
GetLocator(state.m_chain_sync.m_work_header->pprev),
8167 "sending getheaders to outbound peer=%d to verify chain "
8168 "work (current best known block:%s, benchmark blockhash: "
8171 state.pindexBestKnownBlock !=
nullptr
8172 ? state.pindexBestKnownBlock->GetBlockHash().ToString()
8174 state.m_chain_sync.m_work_header->GetBlockHash()
8176 state.m_chain_sync.m_sent_getheaders =
true;
8183 state.m_chain_sync.m_timeout =
8190void PeerManagerImpl::EvictExtraOutboundPeers(std::chrono::seconds now) {
8199 std::pair<NodeId, std::chrono::seconds> youngest_peer{-1, 0},
8200 next_youngest_peer{-1, 0};
8206 if (pnode->
GetId() > youngest_peer.first) {
8207 next_youngest_peer = youngest_peer;
8208 youngest_peer.first = pnode->GetId();
8209 youngest_peer.second = pnode->m_last_block_time;
8213 NodeId to_disconnect = youngest_peer.first;
8214 if (youngest_peer.second > next_youngest_peer.second) {
8217 to_disconnect = next_youngest_peer.first;
8229 CNodeState *node_state = State(pnode->
GetId());
8230 if (node_state ==
nullptr ||
8232 node_state->vBlocksInFlight.empty())) {
8235 "disconnecting extra block-relay-only peer=%d "
8236 "(last block received at time %d)\n",
8243 "keeping block-relay-only peer=%d chosen for eviction "
8244 "(connect time: %d, blocks_in_flight: %d)\n",
8246 node_state->vBlocksInFlight.size());
8262 int64_t oldest_block_announcement = std::numeric_limits<int64_t>::max();
8273 CNodeState *state = State(pnode->
GetId());
8274 if (state ==
nullptr) {
8279 if (state->m_chain_sync.m_protect) {
8282 if (state->m_last_block_announcement < oldest_block_announcement ||
8283 (state->m_last_block_announcement == oldest_block_announcement &&
8284 pnode->
GetId() > worst_peer)) {
8285 worst_peer = pnode->
GetId();
8286 oldest_block_announcement = state->m_last_block_announcement;
8290 if (worst_peer == -1) {
8294 bool disconnected = m_connman.
ForNode(
8302 CNodeState &state = *State(pnode->
GetId());
8304 state.vBlocksInFlight.empty()) {
8306 "disconnecting extra outbound peer=%d (last block "
8307 "announcement received at time %d)\n",
8308 pnode->
GetId(), oldest_block_announcement);
8313 "keeping outbound peer=%d chosen for eviction "
8314 "(connect time: %d, blocks_in_flight: %d)\n",
8316 state.vBlocksInFlight.size());
8331void PeerManagerImpl::CheckForStaleTipAndEvictPeers() {
8334 auto now{GetTime<std::chrono::seconds>()};
8336 EvictExtraOutboundPeers(now);
8338 if (now > m_stale_tip_check_time) {
8344 LogPrintf(
"Potential stale tip detected, will try using extra "
8345 "outbound peer (last tip update: %d seconds ago)\n",
8354 if (!m_initial_sync_finished && CanDirectFetch()) {
8356 m_initial_sync_finished =
true;
8360void PeerManagerImpl::MaybeSendPing(
CNode &node_to, Peer &peer,
8361 std::chrono::microseconds now) {
8363 node_to, std::chrono::duration_cast<std::chrono::seconds>(now)) &&
8364 peer.m_ping_nonce_sent &&
8375 bool pingSend =
false;
8377 if (peer.m_ping_queued) {
8382 if (peer.m_ping_nonce_sent == 0 &&
8392 }
while (nonce == 0);
8393 peer.m_ping_queued =
false;
8394 peer.m_ping_start = now;
8396 peer.m_ping_nonce_sent = nonce;
8401 peer.m_ping_nonce_sent = 0;
8407void PeerManagerImpl::MaybeSendAddr(
CNode &
node, Peer &peer,
8408 std::chrono::microseconds current_time) {
8410 if (!peer.m_addr_relay_enabled) {
8414 LOCK(peer.m_addr_send_times_mutex);
8416 peer.m_next_local_addr_send < current_time) {
8423 if (peer.m_next_local_addr_send != 0us) {
8424 peer.m_addr_known->reset();
8427 CAddress local_addr{*local_service, peer.m_our_services,
8428 Now<NodeSeconds>()};
8429 PushAddress(peer, local_addr);
8431 peer.m_next_local_addr_send =
8437 if (current_time <= peer.m_next_addr_send) {
8441 peer.m_next_addr_send =
8444 const size_t max_addr_to_send = m_opts.max_addr_to_send;
8445 if (!
Assume(peer.m_addrs_to_send.size() <= max_addr_to_send)) {
8448 peer.m_addrs_to_send.resize(max_addr_to_send);
8453 auto addr_already_known =
8456 bool ret = peer.m_addr_known->contains(addr.
GetKey());
8458 peer.m_addr_known->insert(addr.
GetKey());
8462 peer.m_addrs_to_send.erase(std::remove_if(peer.m_addrs_to_send.begin(),
8463 peer.m_addrs_to_send.end(),
8464 addr_already_known),
8465 peer.m_addrs_to_send.end());
8468 if (peer.m_addrs_to_send.empty()) {
8472 const char *msg_type;
8474 if (peer.m_wants_addrv2) {
8484 peer.m_addrs_to_send));
8485 peer.m_addrs_to_send.clear();
8488 if (peer.m_addrs_to_send.capacity() > 40) {
8489 peer.m_addrs_to_send.shrink_to_fit();
8493void PeerManagerImpl::MaybeSendSendHeaders(
CNode &
node, Peer &peer) {
8498 if (!peer.m_sent_sendheaders &&
8501 CNodeState &state = *State(
node.GetId());
8502 if (state.pindexBestKnownBlock !=
nullptr &&
8503 state.pindexBestKnownBlock->nChainWork >
8510 peer.m_sent_sendheaders =
true;
8515void PeerManagerImpl::MaybeSendFeefilter(
8516 CNode &pto, Peer &peer, std::chrono::microseconds current_time) {
8517 if (m_opts.ignore_incoming_txs) {
8541 static const Amount MAX_FILTER{m_fee_filter_rounder.round(
MAX_MONEY)};
8542 if (peer.m_fee_filter_sent == MAX_FILTER) {
8545 peer.m_next_send_feefilter = 0us;
8548 if (current_time > peer.m_next_send_feefilter) {
8549 Amount filterToSend = m_fee_filter_rounder.round(currentFilter);
8553 if (filterToSend != peer.m_fee_filter_sent) {
8555 peer.m_fee_filter_sent = filterToSend;
8557 peer.m_next_send_feefilter =
8565 peer.m_next_send_feefilter &&
8566 (currentFilter < 3 * peer.m_fee_filter_sent / 4 ||
8567 currentFilter > 4 * peer.m_fee_filter_sent / 3)) {
8568 peer.m_next_send_feefilter =
8576class CompareInvMempoolOrder {
8580 explicit CompareInvMempoolOrder(
CTxMemPool *_mempool) : mp(_mempool) {}
8582 bool operator()(std::set<TxId>::iterator a, std::set<TxId>::iterator b) {
8592bool PeerManagerImpl::RejectIncomingTxs(
const CNode &peer)
const {
8601 if (m_opts.ignore_incoming_txs &&
8608bool PeerManagerImpl::SetupAddressRelay(
const CNode &
node, Peer &peer) {
8612 if (
node.IsBlockOnlyConn()) {
8616 if (!peer.m_addr_relay_enabled.exchange(
true)) {
8620 peer.m_addr_known = std::make_unique<CRollingBloomFilter>(5000, 0.001);
8626bool PeerManagerImpl::SendMessages(
const Config &config,
CNode *pto) {
8629 PeerRef peer = GetPeerRef(pto->
GetId());
8638 if (MaybeDiscourageAndDisconnect(*pto, *peer)) {
8647 const auto current_time{GetTime<std::chrono::microseconds>()};
8652 "addrfetch connection timeout; disconnecting peer=%d\n",
8658 MaybeSendPing(*pto, *peer, current_time);
8665 bool sync_blocks_and_headers_from_peer =
false;
8667 MaybeSendAddr(*pto, *peer, current_time);
8669 MaybeSendSendHeaders(*pto, *peer);
8674 CNodeState &state = *State(pto->
GetId());
8677 if (m_chainman.m_best_header ==
nullptr) {
8684 if (state.fPreferredDownload) {
8685 sync_blocks_and_headers_from_peer =
true;
8696 if (m_num_preferred_download_peers == 0 ||
8697 mapBlocksInFlight.empty()) {
8698 sync_blocks_and_headers_from_peer =
true;
8702 if (!state.fSyncStarted && CanServeBlocks(*peer) &&
8706 if ((nSyncStarted == 0 && sync_blocks_and_headers_from_peer) ||
8708 const CBlockIndex *pindexStart = m_chainman.m_best_header;
8717 if (pindexStart->
pprev) {
8718 pindexStart = pindexStart->
pprev;
8720 if (MaybeSendGetHeaders(*pto,
GetLocator(pindexStart), *peer)) {
8723 "initial getheaders (%d) to peer=%d (startheight:%d)\n",
8725 peer->m_starting_height);
8727 state.fSyncStarted =
true;
8728 peer->m_headers_sync_timeout =
8733 std::chrono::microseconds{
8735 Ticks<std::chrono::seconds>(
8737 m_chainman.m_best_header->Time()) /
8754 LOCK(peer->m_block_inv_mutex);
8755 std::vector<CBlock> vHeaders;
8757 ((!peer->m_prefers_headers &&
8758 (!state.m_requested_hb_cmpctblocks ||
8759 peer->m_blocks_for_headers_relay.size() > 1)) ||
8760 peer->m_blocks_for_headers_relay.size() >
8765 ProcessBlockAvailability(pto->
GetId());
8767 if (!fRevertToInv) {
8768 bool fFoundStartingHeader =
false;
8772 for (
const BlockHash &hash : peer->m_blocks_for_headers_relay) {
8778 fRevertToInv =
true;
8781 if (pBestIndex !=
nullptr && pindex->
pprev != pBestIndex) {
8792 fRevertToInv =
true;
8795 pBestIndex = pindex;
8796 if (fFoundStartingHeader) {
8799 }
else if (PeerHasHeader(&state, pindex)) {
8802 }
else if (pindex->
pprev ==
nullptr ||
8803 PeerHasHeader(&state, pindex->
pprev)) {
8806 fFoundStartingHeader =
true;
8811 fRevertToInv =
true;
8816 if (!fRevertToInv && !vHeaders.empty()) {
8817 if (vHeaders.size() == 1 && state.m_requested_hb_cmpctblocks) {
8822 "%s sending header-and-ids %s to peer=%d\n",
8823 __func__, vHeaders.front().GetHash().ToString(),
8826 std::optional<CSerializedNetMsg> cached_cmpctblock_msg;
8828 LOCK(m_most_recent_block_mutex);
8829 if (m_most_recent_block_hash ==
8831 cached_cmpctblock_msg =
8833 *m_most_recent_compact_block);
8836 if (cached_cmpctblock_msg.has_value()) {
8838 std::move(cached_cmpctblock_msg.value()));
8842 block, *pBestIndex)};
8849 state.pindexBestHeaderSent = pBestIndex;
8850 }
else if (peer->m_prefers_headers) {
8851 if (vHeaders.size() > 1) {
8853 "%s: %u headers, range (%s, %s), to peer=%d\n",
8854 __func__, vHeaders.size(),
8855 vHeaders.front().GetHash().ToString(),
8856 vHeaders.back().GetHash().ToString(),
8860 "%s: sending header %s to peer=%d\n", __func__,
8861 vHeaders.front().GetHash().ToString(),
8865 state.pindexBestHeaderSent = pBestIndex;
8867 fRevertToInv =
true;
8874 if (!peer->m_blocks_for_headers_relay.empty()) {
8876 peer->m_blocks_for_headers_relay.back();
8887 "Announcing block %s not on main chain (tip=%s)\n",
8896 if (!PeerHasHeader(&state, pindex)) {
8897 peer->m_blocks_for_inv_relay.push_back(hashToAnnounce);
8899 "%s: sending inv peer=%d hash=%s\n", __func__,
8904 peer->m_blocks_for_headers_relay.clear();
8911 std::vector<CInv> vInv;
8912 auto addInvAndMaybeFlush = [&](uint32_t type,
const uint256 &hash) {
8913 vInv.emplace_back(type, hash);
8924 LOCK(peer->m_block_inv_mutex);
8926 vInv.reserve(std::max<size_t>(peer->m_blocks_for_inv_relay.size(),
8932 for (
const BlockHash &hash : peer->m_blocks_for_inv_relay) {
8935 peer->m_blocks_for_inv_relay.clear();
8938 auto computeNextInvSendTime =
8939 [&](std::chrono::microseconds &next)
8943 if (next < current_time) {
8944 fSendTrickle =
true;
8946 next = NextInvToInbounds(
8951 next = current_time;
8955 return fSendTrickle;
8959 if (peer->m_proof_relay !=
nullptr) {
8960 LOCK(peer->m_proof_relay->m_proof_inventory_mutex);
8962 if (computeNextInvSendTime(
8963 peer->m_proof_relay->m_next_inv_send_time)) {
8965 peer->m_proof_relay->m_proof_inventory_to_send.begin();
8967 peer->m_proof_relay->m_proof_inventory_to_send.end()) {
8970 it = peer->m_proof_relay->m_proof_inventory_to_send.erase(
8973 if (peer->m_proof_relay->m_proof_inventory_known_filter
8974 .contains(proofid)) {
8978 peer->m_proof_relay->m_proof_inventory_known_filter.insert(
8981 peer->m_proof_relay->m_recently_announced_proofs.insert(
8987 if (
auto tx_relay = peer->GetTxRelay()) {
8988 LOCK(tx_relay->m_tx_inventory_mutex);
8990 const bool fSendTrickle =
8991 computeNextInvSendTime(tx_relay->m_next_inv_send_time);
8996 LOCK(tx_relay->m_bloom_filter_mutex);
8997 if (!tx_relay->m_relay_txs) {
8998 tx_relay->m_tx_inventory_to_send.clear();
9003 if (fSendTrickle && tx_relay->m_send_mempool) {
9004 auto vtxinfo = m_mempool.
infoAll();
9005 tx_relay->m_send_mempool =
false;
9007 tx_relay->m_fee_filter_received.load()};
9009 LOCK(tx_relay->m_bloom_filter_mutex);
9011 for (
const auto &txinfo : vtxinfo) {
9012 const TxId &txid = txinfo.tx->GetId();
9013 tx_relay->m_tx_inventory_to_send.erase(txid);
9016 if (txinfo.fee < filterrate.GetFee(txinfo.vsize)) {
9019 if (tx_relay->m_bloom_filter &&
9020 !tx_relay->m_bloom_filter->IsRelevantAndUpdate(
9024 tx_relay->m_tx_inventory_known_filter.insert(txid);
9027 addInvAndMaybeFlush(
MSG_TX, txid);
9029 tx_relay->m_last_mempool_req =
9030 std::chrono::duration_cast<std::chrono::seconds>(
9037 std::vector<std::set<TxId>::iterator> vInvTx;
9038 vInvTx.reserve(tx_relay->m_tx_inventory_to_send.size());
9039 for (std::set<TxId>::iterator it =
9040 tx_relay->m_tx_inventory_to_send.begin();
9041 it != tx_relay->m_tx_inventory_to_send.end(); it++) {
9042 vInvTx.push_back(it);
9045 tx_relay->m_fee_filter_received.load()};
9050 CompareInvMempoolOrder compareInvMempoolOrder(&m_mempool);
9051 std::make_heap(vInvTx.begin(), vInvTx.end(),
9052 compareInvMempoolOrder);
9056 unsigned int nRelayedTransactions = 0;
9057 LOCK(tx_relay->m_bloom_filter_mutex);
9058 while (!vInvTx.empty() &&
9063 std::pop_heap(vInvTx.begin(), vInvTx.end(),
9064 compareInvMempoolOrder);
9065 std::set<TxId>::iterator it = vInvTx.back();
9067 const TxId txid = *it;
9069 tx_relay->m_tx_inventory_to_send.erase(it);
9071 if (tx_relay->m_tx_inventory_known_filter.contains(txid) &&
9072 tx_relay->m_avalanche_stalled_txids.count(txid) == 0) {
9076 auto txinfo = m_mempool.
info(txid);
9082 if (txinfo.fee < filterrate.GetFee(txinfo.vsize)) {
9085 if (tx_relay->m_bloom_filter &&
9086 !tx_relay->m_bloom_filter->IsRelevantAndUpdate(
9091 tx_relay->m_recently_announced_invs.insert(txid);
9092 addInvAndMaybeFlush(
MSG_TX, txid);
9093 nRelayedTransactions++;
9094 tx_relay->m_tx_inventory_known_filter.insert(txid);
9095 tx_relay->m_avalanche_stalled_txids.erase(txid);
9101 if (!vInv.empty()) {
9108 CNodeState &state = *State(pto->
GetId());
9111 auto stalling_timeout = m_block_stalling_timeout.load();
9112 if (state.m_stalling_since.count() &&
9113 state.m_stalling_since < current_time - stalling_timeout) {
9118 LogPrintf(
"Peer=%d is stalling block download, disconnecting\n",
9123 const auto new_timeout =
9125 if (stalling_timeout != new_timeout &&
9126 m_block_stalling_timeout.compare_exchange_strong(
9127 stalling_timeout, new_timeout)) {
9130 "Increased stalling timeout temporarily to %d seconds\n",
9142 if (state.vBlocksInFlight.size() > 0) {
9143 QueuedBlock &queuedBlock = state.vBlocksInFlight.front();
9144 int nOtherPeersWithValidatedDownloads =
9145 m_peers_downloading_from - 1;
9147 state.m_downloading_since +
9148 std::chrono::seconds{consensusParams.nPowTargetSpacing} *
9151 nOtherPeersWithValidatedDownloads)) {
9152 LogPrintf(
"Timeout downloading block %s from peer=%d, "
9154 queuedBlock.pindex->GetBlockHash().ToString(),
9162 if (state.fSyncStarted &&
9163 peer->m_headers_sync_timeout < std::chrono::microseconds::max()) {
9166 if (current_time > peer->m_headers_sync_timeout &&
9167 nSyncStarted == 1 &&
9168 (m_num_preferred_download_peers -
9169 state.fPreferredDownload >=
9178 LogPrintf(
"Timeout downloading headers from peer=%d, "
9184 LogPrintf(
"Timeout downloading headers from noban "
9185 "peer=%d, not disconnecting\n",
9191 state.fSyncStarted =
false;
9193 peer->m_headers_sync_timeout = 0us;
9199 peer->m_headers_sync_timeout = std::chrono::microseconds::max();
9205 ConsiderEviction(*pto, *peer, GetTime<std::chrono::seconds>());
9208 std::vector<CInv> vGetData;
9216 CNodeState &state = *State(pto->
GetId());
9218 if (CanServeBlocks(*peer) &&
9219 ((sync_blocks_and_headers_from_peer && !IsLimitedPeer(*peer)) ||
9222 std::vector<const CBlockIndex *> vToDownload;
9224 auto get_inflight_budget = [&state]() {
9227 static_cast<int>(state.vBlocksInFlight.size()));
9233 FindNextBlocksToDownload(*peer, get_inflight_budget(), vToDownload,
9236 !IsLimitedPeer(*peer)) {
9242 m_chainman.GetSnapshotBaseBlock());
9244 TryDownloadingHistoricalBlocks(
9245 *peer, get_inflight_budget(), vToDownload, from_tip,
9246 Assert(m_chainman.GetSnapshotBaseBlock()));
9250 BlockRequested(config, pto->
GetId(), *pindex);
9255 if (state.vBlocksInFlight.empty() && staller != -1) {
9256 if (State(staller)->m_stalling_since == 0us) {
9257 State(staller)->m_stalling_since = current_time;
9264 auto addGetDataAndMaybeFlush = [&](uint32_t type,
const uint256 &hash) {
9265 CInv inv(type, hash);
9268 vGetData.push_back(std::move(inv));
9279 LOCK(cs_proofrequest);
9280 std::vector<std::pair<NodeId, avalanche::ProofId>> expired;
9282 m_proofrequest.GetRequestable(pto->
GetId(), current_time, &expired);
9283 for (
const auto &entry : expired) {
9285 "timeout of inflight proof %s from peer=%d\n",
9286 entry.second.ToString(), entry.first);
9288 for (
const auto &proofid : requestable) {
9289 if (!AlreadyHaveProof(proofid)) {
9291 m_proofrequest.RequestedData(
9292 pto->
GetId(), proofid,
9299 m_proofrequest.ForgetInvId(proofid);
9309 std::vector<std::pair<NodeId, TxId>> expired;
9311 m_txrequest.GetRequestable(pto->
GetId(), current_time, &expired);
9312 for (
const auto &entry : expired) {
9314 entry.second.ToString(), entry.first);
9316 for (
const TxId &txid : requestable) {
9320 if (!AlreadyHaveTx(txid,
false)) {
9321 addGetDataAndMaybeFlush(
MSG_TX, txid);
9322 m_txrequest.RequestedData(
9329 m_txrequest.ForgetInvId(txid);
9333 if (!vGetData.empty()) {
9338 MaybeSendFeefilter(*pto, *peer, current_time);
9342bool PeerManagerImpl::ReceivedAvalancheProof(
CNode &
node, Peer &peer,
9344 assert(proof !=
nullptr);
9348 AddKnownProof(peer, proofid);
9360 return node.m_avalanche_pubkey.has_value());
9361 auto saveProofIfStaker = [
this, isStaker](
const CNode &
node,
9363 const NodeId nodeid) ->
bool {
9375 LOCK(cs_proofrequest);
9376 m_proofrequest.ReceivedResponse(nodeid, proofid);
9378 if (AlreadyHaveProof(proofid)) {
9379 m_proofrequest.ForgetInvId(proofid);
9380 saveProofIfStaker(
node, proofid, nodeid);
9390 return pm.registerProof(proof, state);
9392 WITH_LOCK(cs_proofrequest, m_proofrequest.ForgetInvId(proofid));
9393 RelayProof(proofid);
9395 node.m_last_proof_time = GetTime<std::chrono::seconds>();
9398 nodeid, proofid.ToString());
9420 "Not polling the avalanche proof (%s): peer=%d, proofid %s\n",
9421 state.
IsValid() ?
"not-worth-polling"
9423 nodeid, proofid.ToString());
9426 saveProofIfStaker(
node, proofid, nodeid);
bool MoneyRange(const Amount nValue)
static constexpr Amount MAX_MONEY
No amount larger than this (in satoshi) is valid.
enum ReadStatus_t ReadStatus
const std::string & BlockFilterTypeName(BlockFilterType filter_type)
Get the human-readable name for a filter type.
BlockFilterIndex * GetBlockFilterIndex(BlockFilterType filter_type)
Get a block filter index by type.
static constexpr int CFCHECKPT_INTERVAL
Interval between compact filter checkpoints.
@ CHAIN
Outputs do not overspend inputs, no double spends, coinbase output ok, no immature coinbase spends,...
@ TRANSACTIONS
Only first tx is coinbase, 2 <= coinbase input script length <= 100, transactions valid,...
@ SCRIPTS
Scripts & signatures ok.
@ TREE
All parent headers found, difficulty matches, timestamp >= median previous, checkpoint.
arith_uint256 GetBlockProof(const CBlockIndex &block)
CBlockLocator GetLocator(const CBlockIndex *index)
Get a locator for a block index entry.
int64_t GetBlockProofEquivalentTime(const CBlockIndex &to, const CBlockIndex &from, const CBlockIndex &tip, const Consensus::Params ¶ms)
Return the time it would take to redo the work difference between from and to, assuming the current h...
const CBlockIndex * LastCommonAncestor(const CBlockIndex *pa, const CBlockIndex *pb)
Find the last common ancestor two blocks have.
#define Assert(val)
Identity function.
#define Assume(val)
Assume is the identity function.
Stochastic address manager.
void Connected(const CService &addr, NodeSeconds time=Now< NodeSeconds >())
We have successfully connected to this peer.
void Good(const CService &addr, bool test_before_evict=true, NodeSeconds time=Now< NodeSeconds >())
Mark an entry as accessible, possibly moving it from "new" to "tried".
bool Add(const std::vector< CAddress > &vAddr, const CNetAddr &source, std::chrono::seconds time_penalty=0s)
Attempt to add one or more addresses to addrman's new table.
void SetServices(const CService &addr, ServiceFlags nServices)
Update an entry's service bits.
void Discourage(const CNetAddr &net_addr)
bool IsBanned(const CNetAddr &net_addr)
Return whether net_addr is banned.
bool IsDiscouraged(const CNetAddr &net_addr)
Return whether net_addr is discouraged.
BlockFilterIndex is used to store and retrieve block filters, hashes, and headers for a range of bloc...
bool LookupFilterRange(int start_height, const CBlockIndex *stop_index, std::vector< BlockFilter > &filters_out) const
Get a range of filters between two heights on a chain.
bool LookupFilterHashRange(int start_height, const CBlockIndex *stop_index, std::vector< uint256 > &hashes_out) const
Get a range of filter hashes between two heights on a chain.
bool LookupFilterHeader(const CBlockIndex *block_index, uint256 &header_out) EXCLUSIVE_LOCKS_REQUIRED(!m_cs_headers_cache)
Get a single filter header by block.
std::vector< CTransactionRef > txn
std::vector< uint32_t > indices
A CService with information about it as peer.
ServiceFlags nServices
Serialized as uint64_t in V1, and as CompactSize in V2.
static constexpr SerParams V1_NETWORK
NodeSeconds nTime
Always included in serialization, except in the network format on INIT_PROTO_VERSION.
static constexpr SerParams V2_NETWORK
size_t BlockTxCount() const
std::vector< CTransactionRef > vtx
The block chain is a tree shaped structure starting with the genesis block at the root,...
bool IsValid(enum BlockValidity nUpTo=BlockValidity::TRANSACTIONS) const EXCLUSIVE_LOCKS_REQUIRED(
Check whether this block index entry is valid up to the passed validity level.
CBlockIndex * pprev
pointer to the index of the predecessor of this block
CBlockHeader GetBlockHeader() const
arith_uint256 nChainWork
(memory only) Total amount of work (expected number of hashes) in the chain up to and including this ...
bool HaveNumChainTxs() const
Check whether this block and all previous blocks back to the genesis block or an assumeutxo snapshot ...
int64_t GetBlockTime() const
unsigned int nTx
Number of transactions in this block.
CBlockIndex * GetAncestor(int height)
Efficiently find an ancestor of this block.
BlockHash GetBlockHash() const
int nHeight
height of the entry in the chain. The genesis block has height 0
FlatFilePos GetBlockPos() const EXCLUSIVE_LOCKS_REQUIRED(
BloomFilter is a probabilistic filter which SPV clients provide so that we can filter the transaction...
bool IsWithinSizeConstraints() const
True if the size is <= MAX_BLOOM_FILTER_SIZE and the number of hash functions is <= MAX_HASH_FUNCS (c...
An in-memory indexed chain of blocks.
CBlockIndex * Tip() const
Returns the index entry for the tip of this chain, or nullptr if none.
CBlockIndex * Next(const CBlockIndex *pindex) const
Find the successor of a block in this chain, or nullptr if the given index is not found or is the tip...
int Height() const
Return the maximal height in the chain.
bool Contains(const CBlockIndex *pindex) const
Efficiently check whether a block is present in this chain.
CChainParams defines various tweakable parameters of a given instance of the Bitcoin system.
const CBlock & GenesisBlock() const
const Consensus::Params & GetConsensus() const
CCoinsView that adds a memory cache for transactions to another CCoinsView.
CCoinsView that brings transactions from a mempool into view.
void ForEachNode(const NodeFn &func)
bool OutboundTargetReached(bool historicalBlockServingLimit) const
check if the outbound target is reached.
bool ForNode(NodeId id, std::function< bool(CNode *pnode)> func)
bool GetNetworkActive() const
bool GetTryNewOutboundPeer() const
void SetTryNewOutboundPeer(bool flag)
int GetExtraBlockRelayCount() const
void WakeMessageHandler() EXCLUSIVE_LOCKS_REQUIRED(!mutexMsgProc)
void StartExtraBlockRelayPeers()
bool DisconnectNode(const std::string &node)
CSipHasher GetDeterministicRandomizer(uint64_t id) const
Get a unique deterministic randomizer.
int GetExtraFullOutboundCount() const
std::vector< CAddress > GetAddresses(size_t max_addresses, size_t max_pct, std::optional< Network > network) const
Return all or many randomly selected addresses, optionally by network.
bool CheckIncomingNonce(uint64_t nonce)
bool ShouldRunInactivityChecks(const CNode &node, std::chrono::seconds now) const
Return true if we should disconnect the peer for failing an inactivity check.
bool GetUseAddrmanOutgoing() const
Fee rate in satoshis per kilobyte: Amount / kB.
Amount GetFeePerK() const
Return the fee in satoshis for a size of 1000 bytes.
Inv(ventory) message data.
bool IsMsgCmpctBlk() const
std::string ToString() const
bool IsMsgStakeContender() const
bool IsMsgFilteredBlk() const
void TransactionInvalidated(const CTransactionRef &tx, std::shared_ptr< const std::vector< Coin > > spent_coins)
Used to create a Merkle proof (usually from a subset of transactions), which consists of a block head...
std::vector< std::pair< size_t, uint256 > > vMatchedTxn
Public only for unit testing and relay testing (not relayed).
bool IsRelayable() const
Whether this address should be relayed to other peers even if we can't reach it ourselves.
static constexpr SerParams V1
bool IsAddrV1Compatible() const
Check if the current object can be serialized in pre-ADDRv2/BIP155 format.
Transport protocol agnostic message container.
Information about a peer.
Mutex cs_avalanche_pubkey
bool IsFeelerConn() const
const std::chrono::seconds m_connected
Unix epoch time at peer connection.
bool ExpectServicesFromConn() const
std::atomic< int > nVersion
std::atomic_bool m_has_all_wanted_services
Whether this peer provides all services that we want.
bool IsInboundConn() const
bool HasPermission(NetPermissionFlags permission) const
bool IsOutboundOrBlockRelayConn() const
bool IsManualConn() const
std::atomic< int64_t > nTimeOffset
const std::string m_addr_name
std::string ConnectionTypeAsString() const
void SetCommonVersion(int greatest_common_version)
std::atomic< bool > m_bip152_highbandwidth_to
std::atomic_bool m_relays_txs
Whether we should relay transactions to this peer.
std::atomic< bool > m_bip152_highbandwidth_from
void PongReceived(std::chrono::microseconds ping_time)
A ping-pong round trip has completed successfully.
std::atomic_bool fSuccessfullyConnected
bool IsAddrFetchConn() const
uint64_t GetLocalNonce() const
void SetAddrLocal(const CService &addrLocalIn) EXCLUSIVE_LOCKS_REQUIRED(!m_addr_local_mutex)
May not be called more than once.
bool IsBlockOnlyConn() const
int GetCommonVersion() const
bool IsFullOutboundConn() const
uint64_t nRemoteHostNonce
Mutex m_subver_mutex
cleanSubVer is a sanitized string of the user agent byte array we read from the wire.
std::atomic_bool fPauseSend
std::chrono::seconds m_nextGetAvaAddr
uint64_t nRemoteExtraEntropy
std::optional< std::pair< CNetMessage, bool > > PollMessage() EXCLUSIVE_LOCKS_REQUIRED(!m_msg_process_queue_mutex)
Poll the next message from the processing queue of this connection.
uint64_t GetLocalExtraEntropy() const
SteadyMilliseconds m_last_poll
double getAvailabilityScore() const
std::atomic_bool m_bloom_filter_loaded
Whether this peer has loaded a bloom filter.
void updateAvailabilityScore(double decayFactor)
The availability score is calculated using an exponentially weighted average.
std::atomic< std::chrono::seconds > m_avalanche_last_message_fault
const bool m_inbound_onion
Whether this peer is an inbound onion, i.e.
std::atomic< int > m_avalanche_message_fault_counter
How much faulty messages did this node accumulate.
std::atomic< bool > m_avalanche_enabled
std::atomic< std::chrono::seconds > m_last_block_time
UNIX epoch time of the last block received from this peer that we had not yet seen (e....
std::atomic_bool fDisconnect
std::atomic< int > m_avalanche_message_fault_score
This score is incremented for every new faulty message received when m_avalanche_message_fault_counte...
std::atomic< std::chrono::seconds > m_last_tx_time
UNIX epoch time of the last transaction received from this peer that we had not yet seen (e....
void invsVoted(uint32_t count)
The node voted for count invs.
bool IsAvalancheOutboundConnection() const
An encapsulated public key.
RollingBloomFilter is a probabilistic "keep track of most recently inserted" set.
Simple class for background tasks that should be run periodically or once "after a while".
void scheduleEvery(Predicate p, std::chrono::milliseconds delta) EXCLUSIVE_LOCKS_REQUIRED(!newTaskMutex)
Repeat p until it return false.
void scheduleFromNow(Function f, std::chrono::milliseconds delta) EXCLUSIVE_LOCKS_REQUIRED(!newTaskMutex)
Call f once after the delta has passed.
A combination of a network address (CNetAddr) and a (TCP) port.
std::string ToString() const
std::vector< uint8_t > GetKey() const
uint64_t Finalize() const
Compute the 64-bit SipHash-2-4 of the data written so far.
CSipHasher & Write(uint64_t data)
Hash a 64-bit integer worth of data.
std::set< std::reference_wrapper< const CTxMemPoolEntryRef >, CompareIteratorById > Parents
CTxMemPool stores valid-according-to-the-current-best-chain transactions that may be included in the ...
void removeConflicts(const CTransaction &tx) EXCLUSIVE_LOCKS_REQUIRED(cs)
void RemoveUnbroadcastTx(const TxId &txid, const bool unchecked=false)
Removes a transaction from the unbroadcast set.
CFeeRate GetMinFee() const
The minimum fee to get into the mempool, which may itself not be enough for larger-sized transactions...
RecursiveMutex cs
This mutex needs to be locked when accessing mapTx or other members that are guarded by it.
void removeRecursive(const CTransaction &tx, MemPoolRemovalReason reason) EXCLUSIVE_LOCKS_REQUIRED(cs)
bool CompareTopologically(const TxId &txida, const TxId &txidb) const
TxMempoolInfo info(const TxId &txid) const
size_t DynamicMemoryUsage() const
bool setAvalancheFinalized(const CTxMemPoolEntryRef &tx, const Consensus::Params ¶ms, const CBlockIndex &active_chain_tip, std::vector< TxId > &finalizedTxIds) EXCLUSIVE_LOCKS_REQUIRED(bool isAvalancheFinalizedPreConsensus(const TxId &txid) const EXCLUSIVE_LOCKS_REQUIRED(cs)
std::vector< TxMempoolInfo > infoAll() const
CTransactionRef GetConflictTx(const COutPoint &prevout) const EXCLUSIVE_LOCKS_REQUIRED(cs)
Get the transaction in the pool that spends the same prevout.
bool exists(const TxId &txid) const
std::set< TxId > GetUnbroadcastTxs() const
Returns transactions in unbroadcast set.
auto withOrphanage(Callable &&func) const EXCLUSIVE_LOCKS_REQUIRED(!cs_orphanage)
const CFeeRate m_min_relay_feerate
auto withConflicting(Callable &&func) const EXCLUSIVE_LOCKS_REQUIRED(!cs_conflicting)
void removeForFinalizedBlock(const std::unordered_set< TxId, SaltedTxIdHasher > &confirmedTxIdsInNonFinalizedBlocks) EXCLUSIVE_LOCKS_REQUIRED(cs)
unsigned long size() const
std::optional< txiter > GetIter(const TxId &txid) const EXCLUSIVE_LOCKS_REQUIRED(cs)
Returns an iterator to the given txid, if found.
virtual void NewPoWValidBlock(const CBlockIndex *pindex, const std::shared_ptr< const CBlock > &block)
Notifies listeners that a block which builds directly on our current tip has been received and connec...
virtual void BlockConnected(ChainstateRole role, const std::shared_ptr< const CBlock > &block, const CBlockIndex *pindex)
Notifies listeners of a block being connected.
virtual void BlockChecked(const CBlock &, const BlockValidationState &)
Notifies listeners of a block validation result.
virtual void UpdatedBlockTip(const CBlockIndex *pindexNew, const CBlockIndex *pindexFork, bool fInitialDownload)
Notifies listeners when the block chain tip advances.
virtual void BlockDisconnected(const std::shared_ptr< const CBlock > &block, const CBlockIndex *pindex)
Notifies listeners of a block being disconnected.
Provides an interface for creating and interacting with one or two chainstates: an IBD chainstate gen...
SnapshotCompletionResult MaybeCompleteSnapshotValidation() EXCLUSIVE_LOCKS_REQUIRED(const CBlockIndex *GetSnapshotBaseBlock() const EXCLUSIVE_LOCKS_REQUIRED(Chainstate ActiveChainstate)() const
Once the background validation chainstate has reached the height which is the base of the UTXO snapsh...
const CBlockIndex * GetBackgroundSyncTip() const EXCLUSIVE_LOCKS_REQUIRED(GetMutex())
The tip of the background sync chain.
MempoolAcceptResult ProcessTransaction(const CTransactionRef &tx, bool test_accept=false) EXCLUSIVE_LOCKS_REQUIRED(cs_main)
Try to add a transaction to the memory pool.
bool IsInitialBlockDownload() const
Check whether we are doing an initial block download (synchronizing from disk or network)
bool ProcessNewBlock(const std::shared_ptr< const CBlock > &block, bool force_processing, bool min_pow_checked, bool *new_block, avalanche::Processor *const avalanche=nullptr) LOCKS_EXCLUDED(cs_main)
Process an incoming block.
RecursiveMutex & GetMutex() const LOCK_RETURNED(
Alias for cs_main.
CBlockIndex * ActiveTip() const EXCLUSIVE_LOCKS_REQUIRED(GetMutex())
bool BackgroundSyncInProgress() const EXCLUSIVE_LOCKS_REQUIRED(GetMutex())
The state of a background sync (for net processing)
bool ProcessNewBlockHeaders(const std::vector< CBlockHeader > &block, bool min_pow_checked, BlockValidationState &state, const CBlockIndex **ppindex=nullptr, const std::optional< CCheckpointData > &test_checkpoints=std::nullopt) LOCKS_EXCLUDED(cs_main)
Process incoming block headers.
const arith_uint256 & MinimumChainWork() const
CChain & ActiveChain() const EXCLUSIVE_LOCKS_REQUIRED(GetMutex())
void MaybeRebalanceCaches() EXCLUSIVE_LOCKS_REQUIRED(void ReportHeadersPresync(const arith_uint256 &work, int64_t height, int64_t timestamp)
Check to see if caches are out of balance and if so, call ResizeCoinsCaches() as needed.
node::BlockManager m_blockman
A single BlockManager instance is shared across each constructed chainstate to avoid duplicating bloc...
virtual uint64_t GetMaxBlockSize() const =0
Double ended buffer combining vector and stream-like interfaces.
void ignore(size_t num_ignore)
uint64_t rand64() noexcept
Generate a random 64-bit integer.
Reads data from an underlying stream, while hashing the read data.
A writer stream (for serialization) that computes a 256-bit hash.
size_t Count(NodeId peer) const
Count how many announcements a peer has (REQUESTED, CANDIDATE, and COMPLETED combined).
size_t CountInFlight(NodeId peer) const
Count how many REQUESTED announcements a peer has.
Interface for message handling.
static Mutex g_msgproc_mutex
Mutex for anything that is only accessed via the msg processing thread.
virtual bool ProcessMessages(const Config &config, CNode *pnode, std::atomic< bool > &interrupt) EXCLUSIVE_LOCKS_REQUIRED(g_msgproc_mutex)=0
Process protocol messages received from a given node.
virtual bool SendMessages(const Config &config, CNode *pnode) EXCLUSIVE_LOCKS_REQUIRED(g_msgproc_mutex)=0
Send queued protocol messages to a given node.
virtual void InitializeNode(const Config &config, CNode &node, ServiceFlags our_services)=0
Initialize a peer (setup state, queue any initial messages)
virtual void FinalizeNode(const Config &config, const CNode &node)=0
Handle removal of a peer (clear state)
static bool HasFlag(NetPermissionFlags flags, NetPermissionFlags f)
ReadStatus InitData(const CBlockHeaderAndShortTxIDs &cmpctblock, const std::vector< CTransactionRef > &extra_txn)
bool IsTxAvailable(size_t index) const
ReadStatus FillBlock(CBlock &block, const std::vector< CTransactionRef > &vtx_missing)
virtual std::optional< std::string > FetchBlock(const Config &config, NodeId peer_id, const CBlockIndex &block_index)=0
Attempt to manually fetch block from a given peer.
virtual void SendPings()=0
Send ping message to all peers.
static std::unique_ptr< PeerManager > make(CConnman &connman, AddrMan &addrman, BanMan *banman, ChainstateManager &chainman, CTxMemPool &pool, avalanche::Processor *const avalanche, Options opts)
virtual void StartScheduledTasks(CScheduler &scheduler)=0
Begin running background tasks, should only be called once.
virtual bool IgnoresIncomingTxs()=0
Whether this node ignores txs received over p2p.
virtual void ProcessMessage(const Config &config, CNode &pfrom, const std::string &msg_type, DataStream &vRecv, const std::chrono::microseconds time_received, const std::atomic< bool > &interruptMsgProc) EXCLUSIVE_LOCKS_REQUIRED(g_msgproc_mutex)=0
Process a single message from a peer.
virtual bool GetNodeStateStats(NodeId nodeid, CNodeStateStats &stats) const =0
Get statistics from node state.
virtual void UnitTestMisbehaving(const NodeId peer_id)=0
Public for unit testing.
virtual void UpdateLastBlockAnnounceTime(NodeId node, int64_t time_in_seconds)=0
This function is used for testing the stale tip eviction logic, see denialofservice_tests....
virtual void CheckForStaleTipAndEvictPeers()=0
Evict extra outbound peers.
static RCUPtr make(Args &&...args)
Construct a new object that is owned by the pointer.
I randrange(I range) noexcept
Generate a random integer in the range [0..range), with range > 0.
A Span is an object that can refer to a contiguous sequence of objects.
int EraseTx(const TxId &txid) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Erase a tx by txid.
void EraseForPeer(NodeId peer) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Erase all txs announced by a peer (eg, after that peer disconnects)
std::vector< CTransactionRef > GetChildrenFromSamePeer(const CTransactionRef &parent, NodeId nodeid) const EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Get all children that spend from this tx and were received from nodeid.
bool AddTx(const CTransactionRef &tx, NodeId peer) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Add a new transaction to the pool.
unsigned int LimitTxs(unsigned int max_txs, FastRandomContext &rng) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Limit the txs to the given maximum.
void EraseForBlock(const CBlock &block) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Erase all txs included in or invalidated by a new block.
std::vector< CTransactionRef > GetConflictTxs(const CTransactionRef &tx) const EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
void AddChildrenToWorkSet(const CTransaction &tx) EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Add any tx that list a particular tx as a parent into the from peer's work set.
std::vector< std::pair< CTransactionRef, NodeId > > GetChildrenFromDifferentPeer(const CTransactionRef &parent, NodeId nodeid) const EXCLUSIVE_LOCKS_REQUIRED(!m_mutex)
Get all children that spend from this tx but were not received from nodeid.
std::string GetRejectReason() const
std::string ToString() const
256-bit unsigned big integer.
const std::vector< PrefilledProof > & getPrefilledProofs() const
uint64_t getShortID(const ProofId &proofid) const
const std::vector< uint64_t > & getShortIDs() const
ProofId getProofId() const
bool verify(DelegationState &state, CPubKey &auth) const
const DelegationId & getId() const
const LimitedProofId & getLimitedProofId() const
bool shouldRequestMoreNodes()
Returns true if we encountered a lack of node since the last call.
bool exists(const ProofId &proofid) const
Return true if the (valid) proof exists, but only for non-dangling proofs.
bool forPeer(const ProofId &proofid, Callable &&func) const
bool addNode(NodeId nodeid, const ProofId &proofid)
Node API.
void removeUnbroadcastProof(const ProofId &proofid)
const ProofRadixTree & getShareableProofsSnapshot() const
bool isBoundToPeer(const ProofId &proofid) const
bool saveRemoteProof(const ProofId &proofid, const NodeId nodeid, const bool present)
void forEachPeer(Callable &&func) const
void setInvalid(const ProofId &proofid)
bool isInvalid(const ProofId &proofid) const
bool isImmature(const ProofId &proofid) const
auto getUnbroadcastProofs() const
bool isInConflictingPool(const ProofId &proofid) const
void sendResponse(CNode *pfrom, Response response) const
bool addToReconcile(const AnyVoteItem &item) EXCLUSIVE_LOCKS_REQUIRED(!cs_finalizedItems)
bool isStakingPreconsensusActivated(const CBlockIndex *pprev) const
int64_t getAvaproofsNodeCounter() const
bool sendHello(CNode *pfrom) EXCLUSIVE_LOCKS_REQUIRED(!cs_delayedAvahelloNodeIds)
Send a avahello message.
void setRecentlyFinalized(const uint256 &itemId) EXCLUSIVE_LOCKS_REQUIRED(!cs_finalizedItems)
bool isQuorumEstablished() LOCKS_EXCLUDED(cs_main) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager
void cleanupStakingRewards(const int minHeight) EXCLUSIVE_LOCKS_REQUIRED(!cs_stakingRewards
ProofRef getLocalProof() const
void acceptStakeContender(const StakeContenderId &contenderId) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager)
bool reconcileOrFinalize(const ProofRef &proof) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager
Wrapper around the addToReconcile for proofs that adds back the finalization flag to the peer if it i...
int getStakeContenderStatus(const StakeContenderId &contenderId) const EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager
Track votes on stake contenders.
void sendDelayedAvahello() EXCLUSIVE_LOCKS_REQUIRED(!cs_delayedAvahelloNodeIds)
void finalizeStakeContender(const StakeContenderId &contenderId) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager
bool isPreconsensusActivated(const CBlockIndex *pprev) const
auto withPeerManager(Callable &&func) const EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager)
bool registerVotes(NodeId nodeid, const Response &response, std::vector< VoteItemUpdate > &updates, bool &disconnect, std::string &error) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager
void rejectStakeContender(const StakeContenderId &contenderId) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager)
void avaproofsSent(NodeId nodeid) LOCKS_EXCLUDED(cs_main) EXCLUSIVE_LOCKS_REQUIRED(!cs_peerManager)
std::vector< uint32_t > indices
std::string ToString() const
std::string GetHex() const
Generate a new block, without valid proof-of-work.
bool ReadRawBlock(std::vector< uint8_t > &block, const FlatFilePos &pos) const
CBlockIndex * LookupBlockIndex(const BlockHash &hash) EXCLUSIVE_LOCKS_REQUIRED(cs_main)
bool LoadingBlocks() const
bool IsPruneMode() const
Whether running in -prune mode.
bool ReadBlock(CBlock &block, const FlatFilePos &pos) const
Functions for disk access for blocks.
static const uint256 ZERO
@ BLOCK_CHECKPOINT
the block failed to meet one of our checkpoints
@ BLOCK_HEADER_LOW_WORK
the block header may be on a too-little-work chain
@ BLOCK_INVALID_HEADER
invalid proof of work or time too old
@ BLOCK_CACHED_INVALID
this block was cached as being invalid and we didn't store the reason why
@ BLOCK_CONSENSUS
invalid by consensus rules (excluding any below reasons)
@ BLOCK_MISSING_PREV
We don't have the previous block the checked one is built on.
@ BLOCK_INVALID_PREV
A block this one builds on is invalid.
@ BLOCK_MUTATED
the block's data didn't match the data committed to by the PoW
@ BLOCK_TIME_FUTURE
block timestamp was > 2 hours in the future (or our clock is bad)
@ BLOCK_RESULT_UNSET
initial value. Block has not yet been rejected
@ TX_MISSING_INPUTS
transaction was missing some of its inputs
@ TX_CHILD_BEFORE_PARENT
This tx outputs are already spent in the mempool.
@ TX_MEMPOOL_POLICY
violated mempool's fee/size/descendant/etc limits
@ TX_PACKAGE_RECONSIDERABLE
fails some policy, but might be acceptable if submitted in a (different) package
@ TX_UNKNOWN
transaction was not validated because package failed
@ TX_PREMATURE_SPEND
transaction spends a coinbase too early, or violates locktime/sequence locks
@ TX_DUPLICATE
Tx already in mempool or in the chain.
@ TX_INPUTS_NOT_STANDARD
inputs failed policy rules
@ TX_CONFLICT
Tx conflicts with a finalized tx, i.e.
@ TX_NOT_STANDARD
otherwise didn't meet our local policy rules
@ TX_AVALANCHE_RECONSIDERABLE
fails some policy, but might be reconsidered by avalanche voting
@ TX_NO_MEMPOOL
this node does not have a mempool so can't validate the transaction
@ TX_RESULT_UNSET
initial value. Tx has not yet been rejected
@ TX_CONSENSUS
invalid by consensus rules
static size_t RecursiveDynamicUsage(const CScript &script)
RecursiveMutex cs_main
Mutex to guard access to validation specific variables, such as reading or changing the chainstate.
ChainstateRole
This enum describes the various roles a specific Chainstate instance can take.
std::array< uint8_t, CPubKey::SCHNORR_SIZE > SchnorrSig
a Schnorr signature
#define LogPrintLevel(category, level,...)
#define LogPrint(category,...)
#define LogDebug(category,...)
CSerializedNetMsg Make(std::string msg_type, Args &&...args)
const char * FILTERLOAD
The filterload message tells the receiving peer to filter all relayed transactions and requested merk...
const char * CFHEADERS
cfheaders is a response to a getcfheaders request containing a filter header and a vector of filter h...
const char * AVAPROOFSREQ
Request for missing avalanche proofs after an avaproofs message has been processed.
const char * CFILTER
cfilter is a response to a getcfilters request containing a single compact filter.
const char * BLOCK
The block message transmits a single serialized block.
const char * FILTERCLEAR
The filterclear message tells the receiving peer to remove a previously-set bloom filter.
const char * HEADERS
The headers message sends one or more block headers to a node which previously requested certain head...
const char * ADDRV2
The addrv2 message relays connection information for peers on the network just like the addr message,...
const char * SENDHEADERS
Indicates that a node prefers to receive new block announcements via a "headers" message rather than ...
const char * AVAPROOFS
The avaproofs message the proof short ids of all the valid proofs that we know.
const char * PONG
The pong message replies to a ping message, proving to the pinging node that the ponging node is stil...
const char * GETAVAPROOFS
The getavaproofs message requests an avaproofs message that provides the proof short ids of all the v...
const char * SENDCMPCT
Contains a 1-byte bool and 8-byte LE version number.
const char * GETADDR
The getaddr message requests an addr message from the receiving node, preferably one with lots of IP ...
const char * GETCFCHECKPT
getcfcheckpt requests evenly spaced compact filter headers, enabling parallelized download and valida...
const char * NOTFOUND
The notfound message is a reply to a getdata message which requested an object the receiving node doe...
const char * GETAVAADDR
The getavaaddr message requests an addr message from the receiving node, containing IP addresses of t...
const char * CMPCTBLOCK
Contains a CBlockHeaderAndShortTxIDs object - providing a header and list of "short txids".
const char * MEMPOOL
The mempool message requests the TXIDs of transactions that the receiving node has verified as valid ...
const char * GETCFILTERS
getcfilters requests compact filters for a range of blocks.
const char * TX
The tx message transmits a single transaction.
const char * AVAHELLO
Contains a delegation and a signature.
const char * FILTERADD
The filteradd message tells the receiving peer to add a single element to a previously-set bloom filt...
const char * ADDR
The addr (IP address) message relays connection information for peers on the network.
const char * VERSION
The version message provides information about the transmitting node to the receiving node at the beg...
const char * GETBLOCKS
The getblocks message requests an inv message that provides block header hashes starting from a parti...
const char * FEEFILTER
The feefilter message tells the receiving peer not to inv us any txs which do not meet the specified ...
const char * GETHEADERS
The getheaders message requests a headers message that provides block headers starting from a particu...
const char * AVARESPONSE
Contains an avalanche::Response.
const char * GETDATA
The getdata message requests one or more data objects from another node.
const char * VERACK
The verack message acknowledges a previously-received version message, informing the connecting node ...
const char * BLOCKTXN
Contains a BlockTransactions.
const char * GETCFHEADERS
getcfheaders requests a compact filter header and the filter hashes for a range of blocks,...
const char * SENDADDRV2
The sendaddrv2 message signals support for receiving ADDRV2 messages (BIP155).
const char * PING
The ping message is sent periodically to help confirm that the receiving peer is still connected.
const char * AVAPOLL
Contains an avalanche::Poll.
const char * MERKLEBLOCK
The merkleblock message is a reply to a getdata message which requested a block using the inventory t...
const char * AVAPROOF
Contains an avalanche::Proof.
const char * CFCHECKPT
cfcheckpt is a response to a getcfcheckpt request containing a vector of evenly spaced filter headers...
const char * GETBLOCKTXN
Contains a BlockTransactionsRequest Peer should respond with "blocktxn" message.
const char * INV
The inv message (inventory message) transmits one or more inventories of objects known to the transmi...
ShortIdProcessor< PrefilledProof, ShortIdProcessorPrefilledProofAdapter, ProofRefCompare > ProofShortIdProcessor
std::variant< const ProofRef, const CBlockIndex *, const StakeContenderId, const CTransactionRef > AnyVoteItem
RCUPtr< const Proof > ProofRef
Implement std::hash so RCUPtr can be used as a key for maps or sets.
std::optional< CService > GetLocalAddrForPeer(CNode &node)
Returns a local address that we should advertise to this peer.
std::function< void(const CAddress &addr, const std::string &msg_type, Span< const uint8_t > data, bool is_incoming)> CaptureMessage
Defaults to CaptureMessageToFile(), but can be overridden by unit tests.
std::string userAgent(const Config &config)
bool IsReachable(enum Network net)
bool SeenLocal(const CService &addr)
vote for a local address
static const unsigned int MAX_SUBVERSION_LENGTH
Maximum length of the user agent string in version message.
static constexpr std::chrono::minutes TIMEOUT_INTERVAL
Time after which to disconnect, after waiting for a ping response (or inactivity).
@ BypassProofRequestLimits
static constexpr auto HEADERS_RESPONSE_TIME
How long to wait for a peer to respond to a getheaders request.
static constexpr size_t MAX_ADDR_PROCESSING_TOKEN_BUCKET
The soft limit of the address processing token bucket (the regular MAX_ADDR_RATE_PER_SECOND based inc...
static constexpr size_t MAX_AVALANCHE_STALLED_TXIDS_PER_PEER
Maximum number of stalled avalanche txids to store per peer.
static const int MAX_BLOCKS_IN_TRANSIT_PER_PEER
Number of blocks that can be requested at any given time from a single peer.
static constexpr auto BLOCK_STALLING_TIMEOUT_DEFAULT
Default time during which a peer must stall block download progress before being disconnected.
static constexpr auto GETAVAADDR_INTERVAL
Minimum time between 2 successives getavaaddr messages from the same peer.
static constexpr auto AVG_FEEFILTER_BROADCAST_INTERVAL
Verify that INVENTORY_MAX_RECENT_RELAY is enough to cache everything typically relayed before uncondi...
static constexpr unsigned int INVENTORY_BROADCAST_MAX_PER_MB
Maximum number of inventory items to send per transmission.
static constexpr auto EXTRA_PEER_CHECK_INTERVAL
How frequently to check for extra outbound peers and disconnect.
static const unsigned int BLOCK_DOWNLOAD_WINDOW
Size of the "block download window": how far ahead of our current height do we fetch?...
static uint32_t getAvalancheVoteForProof(const avalanche::Processor &avalanche, const avalanche::ProofId &id)
Decide a response for an Avalanche poll about the given proof.
static constexpr int STALE_RELAY_AGE_LIMIT
Age after which a stale block will no longer be served if requested as protection against fingerprint...
static constexpr int HISTORICAL_BLOCK_AGE
Age after which a block is considered historical for purposes of rate limiting block relay.
static constexpr auto ROTATE_ADDR_RELAY_DEST_INTERVAL
Delay between rotating the peers we relay a particular address to.
static constexpr auto MINIMUM_CONNECT_TIME
Minimum time an outbound-peer-eviction candidate must be connected for, in order to evict.
static constexpr auto CHAIN_SYNC_TIMEOUT
Timeout for (unprotected) outbound peers to sync to our chainwork.
static const unsigned int NODE_NETWORK_LIMITED_MIN_BLOCKS
Minimum blocks required to signal NODE_NETWORK_LIMITED.
static constexpr auto AVG_LOCAL_ADDRESS_BROADCAST_INTERVAL
Average delay between local address broadcasts.
static const int MAX_BLOCKTXN_DEPTH
Maximum depth of blocks we're willing to respond to GETBLOCKTXN requests for.
static constexpr uint64_t CMPCTBLOCKS_VERSION
The compactblocks version we support.
bool IsAvalancheMessageType(const std::string &msg_type)
static constexpr int32_t MAX_OUTBOUND_PEERS_TO_PROTECT_FROM_DISCONNECT
Protect at least this many outbound peers from disconnection due to slow/behind headers chain.
static std::chrono::microseconds ComputeRequestTime(const CNode &node, const InvRequestTracker< InvId > &requestTracker, const DataRequestParameters &requestParams, std::chrono::microseconds current_time, bool preferred)
Compute the request time for this announcement, current time plus delays for:
static constexpr auto INBOUND_INVENTORY_BROADCAST_INTERVAL
Average delay between trickled inventory transmissions for inbound peers.
static constexpr DataRequestParameters TX_REQUEST_PARAMS
static constexpr auto MAX_FEEFILTER_CHANGE_DELAY
Maximum feefilter broadcast delay after significant change.
static constexpr uint32_t MAX_GETCFILTERS_SIZE
Maximum number of compact filters that may be requested with one getcfilters.
static constexpr auto HEADERS_DOWNLOAD_TIMEOUT_BASE
Headers download timeout.
static const unsigned int MAX_GETDATA_SZ
Limit to avoid sending big packets.
static constexpr double BLOCK_DOWNLOAD_TIMEOUT_BASE
Block download timeout base, expressed in multiples of the block interval (i.e.
static constexpr auto AVALANCHE_AVAPROOFS_TIMEOUT
If no proof was requested from a compact proof message after this timeout expired,...
static constexpr auto STALE_CHECK_INTERVAL
How frequently to check for stale tips.
static constexpr unsigned int INVENTORY_MAX_RECENT_RELAY
The number of most recently announced transactions a peer can request.
static constexpr auto UNCONDITIONAL_RELAY_DELAY
How long a transaction has to be in the mempool before it can unconditionally be relayed.
static constexpr auto AVG_ADDRESS_BROADCAST_INTERVAL
Average delay between peer address broadcasts.
static const unsigned int MAX_LOCATOR_SZ
The maximum number of entries in a locator.
static constexpr double BLOCK_DOWNLOAD_TIMEOUT_PER_PEER
Additional block download timeout per parallel downloading peer (i.e.
static constexpr double MAX_ADDR_RATE_PER_SECOND
The maximum rate of address records we're willing to process on average.
static constexpr auto PING_INTERVAL
Time between pings automatically sent out for latency probing and keepalive.
static const int MAX_CMPCTBLOCK_DEPTH
Maximum depth of blocks we're willing to serve as compact blocks to peers when requested.
static constexpr DataRequestParameters PROOF_REQUEST_PARAMS
static const unsigned int MAX_BLOCKS_TO_ANNOUNCE
Maximum number of headers to announce when relaying blocks with headers message.
static bool TooManyAnnouncements(const CNode &node, const InvRequestTracker< InvId > &requestTracker, const DataRequestParameters &requestParams)
static constexpr uint32_t MAX_GETCFHEADERS_SIZE
Maximum number of cf hashes that may be requested with one getcfheaders.
static constexpr auto BLOCK_STALLING_TIMEOUT_MAX
Maximum timeout for stalling block download.
static constexpr auto HEADERS_DOWNLOAD_TIMEOUT_PER_HEADER
static constexpr uint64_t RANDOMIZER_ID_ADDRESS_RELAY
SHA256("main address relay")[0:8].
static constexpr size_t MAX_PCT_ADDR_TO_SEND
the maximum percentage of addresses from our addrman to return in response to a getaddr message.
static const unsigned int MAX_INV_SZ
The maximum number of entries in an 'inv' protocol message.
static constexpr unsigned int INVENTORY_BROADCAST_PER_SECOND
Maximum rate of inventory items to send per second.
static constexpr size_t MAX_ADDR_TO_SEND
The maximum number of address records permitted in an ADDR message.
static const unsigned int MAX_CMPCTBLOCKS_INFLIGHT_PER_BLOCK
Maximum number of outstanding CMPCTBLOCK requests for the same block.
static const unsigned int MAX_HEADERS_RESULTS
Number of headers sent in one getheaders result.
bool IsProxy(const CNetAddr &addr)
static constexpr NodeId NO_NODE
Special NodeId that represent no node.
uint256 GetPackageHash(const Package &package)
std::vector< CTransactionRef > Package
A package is an ordered list of transactions.
static constexpr Amount DEFAULT_MIN_RELAY_TX_FEE_PER_KB(1000 *SATOSHI)
Default for -minrelaytxfee, minimum relay fee for transactions.
std::shared_ptr< const CTransaction > CTransactionRef
static constexpr size_t AVALANCHE_MAX_ELEMENT_POLL
Maximum item that can be polled at once.
void SetServiceFlagsIBDCache(bool state)
Set the current IBD status in order to figure out the desirable service flags.
ServiceFlags GetDesirableServiceFlags(ServiceFlags services)
Gets the set of service flags which are "desirable" for a given peer.
static const unsigned int MAX_PROTOCOL_MESSAGE_LENGTH
Maximum length of incoming protocol messages (Currently 2MB).
static bool HasAllDesirableServiceFlags(ServiceFlags services)
A shortcut for (services & GetDesirableServiceFlags(services)) == GetDesirableServiceFlags(services),...
@ MSG_AVA_STAKE_CONTENDER
@ MSG_CMPCT_BLOCK
Defined in BIP152.
ServiceFlags
nServices flags.
static bool MayHaveUsefulAddressDB(ServiceFlags services)
Checks if a peer with the given service flags may be capable of having a robust address-storage DB.
static const int SHORT_IDS_BLOCKS_VERSION
short-id-based block download starts with this version
static const int SENDHEADERS_VERSION
"sendheaders" command and announcing blocks with headers starts with this version
static const int PROTOCOL_VERSION
network protocol versioning
static const int FEEFILTER_VERSION
"feefilter" tells peers to filter invs to you by fee starts with this version
static const int MIN_PEER_PROTO_VERSION
disconnect from peers older than this proto version
static const int INVALID_CB_NO_BAN_VERSION
not banning for invalid compact blocks starts with this version
static const int BIP0031_VERSION
BIP 0031, pong message, is enabled for all versions AFTER this one.
void Shuffle(I first, I last, R &&rng)
More efficient than using std::shuffle on a FastRandomContext.
reverse_range< T > reverse_iterate(T &x)
static const unsigned int MAX_SCRIPT_ELEMENT_SIZE
void Unserialize(Stream &, V)=delete
#define LIMITED_STRING(obj, n)
static auto WithParams(const Params ¶ms, T &&t)
Return a wrapper around t that (de)serializes it with specified parameter params.
uint64_t ReadCompactSize(Stream &is, bool range_check=true)
Decode a CompactSize-encoded variable-length integer.
constexpr auto MakeUCharSpan(V &&v) -> decltype(UCharSpanCast(Span{std::forward< V >(v)}))
Like the Span constructor, but for (const) uint8_t member types only.
static const double AVALANCHE_STATISTICS_DECAY_FACTOR
Pre-computed decay factor for the avalanche statistics computation.
static constexpr std::chrono::minutes AVALANCHE_STATISTICS_REFRESH_PERIOD
Refresh period for the avalanche statistics computation.
std::string ToString(const T &t)
Locale-independent version of std::to_string.
static constexpr Amount zero() noexcept
A BlockHash is a unqiue identifier for a block.
Describes a place in the block chain to another node such that if the other node doesn't have the sam...
std::vector< BlockHash > vHave
std::chrono::microseconds m_ping_wait
Amount m_fee_filter_received
std::vector< int > vHeightInFlight
bool m_addr_relay_enabled
uint64_t m_addr_rate_limited
uint64_t m_addr_processed
ServiceFlags their_services
Parameters that influence chain consensus.
int64_t nPowTargetSpacing
std::chrono::seconds PowTargetSpacing() const
const std::chrono::seconds overloaded_peer_delay
How long to delay requesting data from overloaded peers (see max_peer_request_in_flight).
const size_t max_peer_announcements
Maximum number of inventories to consider for requesting, per peer.
const std::chrono::seconds nonpref_peer_delay
How long to delay requesting data from non-preferred peers.
const NetPermissionFlags bypass_request_limits_permissions
Permission flags a peer requires to bypass the request limits tracking limits and delay penalty.
const std::chrono::microseconds getdata_interval
How long to wait (in microseconds) before a data request from an additional peer.
const size_t max_peer_request_in_flight
Maximum number of in-flight data requests from a peer.
Validation result for a transaction evaluated by MemPoolAccept (single or package).
const ResultType m_result_type
Result type.
const TxValidationState m_state
Contains information about why the transaction failed.
@ MEMPOOL_ENTRY
Valid, transaction was already in the mempool.
@ VALID
Fully validated, valid.
static time_point now() noexcept
Return current system time or mocked time, if set.
std::chrono::time_point< NodeClock > time_point
Validation result for package mempool acceptance.
PackageValidationState m_state
std::map< TxId, MempoolAcceptResult > m_tx_results
Map from txid to finished MempoolAcceptResults.
This is a radix tree storing values identified by a unique key.
A TxId is the identifier of a transaction.
std::chrono::seconds registration_time
const ProofId & getProofId() const
StakeContenderIds are unique for each block to ensure that the peer polling for their acceptance has ...
#define AssertLockNotHeld(cs)
#define WITH_LOCK(cs, code)
Run code while locking a mutex.
#define EXCLUSIVE_LOCKS_REQUIRED(...)
#define LOCKS_EXCLUDED(...)
#define NO_THREAD_SAFETY_ANALYSIS
int64_t GetTime()
DEPRECATED Use either ClockType::now() or Now<TimePointType>() if a cast is needed.
constexpr int64_t count_microseconds(std::chrono::microseconds t)
constexpr int64_t count_seconds(std::chrono::seconds t)
std::chrono::time_point< NodeClock, std::chrono::seconds > NodeSeconds
double CountSecondsDouble(SecondsDouble t)
Helper to count the seconds in any std::chrono::duration type.
NodeClock::time_point GetAdjustedTime()
void AddTimeData(const CNetAddr &ip, int64_t nOffsetSample)
#define TRACE6(context, event, a, b, c, d, e, f)
@ AVALANCHE
Removed by avalanche vote.
std::string SanitizeString(std::string_view str, int rule)
Remove unsafe chars.
arith_uint256 CalculateHeadersWork(const std::vector< CBlockHeader > &headers)
Return the sum of the work on a given set of headers.
bool HasValidProofOfWork(const std::vector< CBlockHeader > &headers, const Consensus::Params &consensusParams)
Check with the proof of work on each blockheader matches the value in nBits.
PackageMempoolAcceptResult ProcessNewPackage(Chainstate &active_chainstate, CTxMemPool &pool, const Package &package, bool test_accept)
Validate (and maybe submit) a package to the mempool.
bool IsBlockMutated(const CBlock &block)
Check if a block has been mutated (with respect to its merkle root).
std::vector< Coin > GetSpentCoins(const CTransactionRef &ptx, const CCoinsViewCache &coins_view)
Get the coins spent by ptx from the coins_view.
static const unsigned int MIN_BLOCKS_TO_KEEP
Block files containing a block-height within MIN_BLOCKS_TO_KEEP of ActiveChain().Tip() will not be pr...
CMainSignals & GetMainSignals()