@@ -135,7 +135,7 @@ static constexpr double BLOCK_DOWNLOAD_TIMEOUT_PER_PEER = 0.5;
135
135
/* * Maximum number of headers to announce when relaying blocks with headers message.*/
136
136
static const unsigned int MAX_BLOCKS_TO_ANNOUNCE = 8 ;
137
137
/* * Maximum number of unconnecting headers announcements before DoS score */
138
- static const int MAX_UNCONNECTING_HEADERS = 10 ;
138
+ static const int MAX_NUM_UNCONNECTING_HEADERS_MSGS = 10 ;
139
139
/* * Minimum blocks required to signal NODE_NETWORK_LIMITED */
140
140
static const unsigned int NODE_NETWORK_LIMITED_MIN_BLOCKS = 288 ;
141
141
/* * Average delay between local address broadcasts */
@@ -393,13 +393,13 @@ struct Peer {
393
393
std::atomic<bool > m_sent_sendheaders{false };
394
394
395
395
/* * Length of current-streak of unconnecting headers announcements */
396
- int nUnconnectingHeaders GUARDED_BY (NetEventsInterface::g_msgproc_mutex){0 };
396
+ int m_num_unconnecting_headers_msgs GUARDED_BY (NetEventsInterface::g_msgproc_mutex){0 };
397
397
398
398
/* * When to potentially disconnect peer for stalling headers download */
399
399
std::chrono::microseconds m_headers_sync_timeout GUARDED_BY (NetEventsInterface::g_msgproc_mutex){0us};
400
400
401
401
/* * Whether this peer wants invs or headers (when possible) for block announcements */
402
- bool fPreferHeaders GUARDED_BY (NetEventsInterface::g_msgproc_mutex){false };
402
+ bool m_prefers_headers GUARDED_BY (NetEventsInterface::g_msgproc_mutex){false };
403
403
404
404
explicit Peer (NodeId id, ServiceFlags our_services)
405
405
: m_id{id}
@@ -2439,24 +2439,24 @@ arith_uint256 PeerManagerImpl::GetAntiDoSWorkThreshold()
2439
2439
*
2440
2440
* We'll send a getheaders message in response to try to connect the chain.
2441
2441
*
2442
- * The peer can send up to MAX_UNCONNECTING_HEADERS in a row that
2442
+ * The peer can send up to MAX_NUM_UNCONNECTING_HEADERS_MSGS in a row that
2443
2443
* don't connect before given DoS points.
2444
2444
*
2445
2445
* Once a headers message is received that is valid and does connect,
2446
- * nUnconnectingHeaders gets reset back to 0.
2446
+ * m_num_unconnecting_headers_msgs gets reset back to 0.
2447
2447
*/
2448
2448
void PeerManagerImpl::HandleFewUnconnectingHeaders (CNode& pfrom, Peer& peer,
2449
2449
const std::vector<CBlockHeader>& headers)
2450
2450
{
2451
- peer.nUnconnectingHeaders ++;
2451
+ peer.m_num_unconnecting_headers_msgs ++;
2452
2452
// Try to fill in the missing headers.
2453
2453
const CBlockIndex* best_header{WITH_LOCK (cs_main, return m_chainman.m_best_header )};
2454
2454
if (MaybeSendGetHeaders (pfrom, GetLocator (best_header), peer)) {
2455
- LogPrint (BCLog::NET, " received header %s: missing prev block %s, sending getheaders (%d) to end (peer=%d, nUnconnectingHeaders =%d)\n " ,
2455
+ LogPrint (BCLog::NET, " received header %s: missing prev block %s, sending getheaders (%d) to end (peer=%d, m_num_unconnecting_headers_msgs =%d)\n " ,
2456
2456
headers[0 ].GetHash ().ToString (),
2457
2457
headers[0 ].hashPrevBlock .ToString (),
2458
2458
best_header->nHeight ,
2459
- pfrom.GetId (), peer.nUnconnectingHeaders );
2459
+ pfrom.GetId (), peer.m_num_unconnecting_headers_msgs );
2460
2460
}
2461
2461
2462
2462
// Set hashLastUnknownBlock for this peer, so that if we
@@ -2466,8 +2466,8 @@ void PeerManagerImpl::HandleFewUnconnectingHeaders(CNode& pfrom, Peer& peer,
2466
2466
2467
2467
// The peer may just be broken, so periodically assign DoS points if this
2468
2468
// condition persists.
2469
- if (peer.nUnconnectingHeaders % MAX_UNCONNECTING_HEADERS == 0 ) {
2470
- Misbehaving (peer, 20 , strprintf (" %d non-connecting headers" , peer.nUnconnectingHeaders ));
2469
+ if (peer.m_num_unconnecting_headers_msgs % MAX_NUM_UNCONNECTING_HEADERS_MSGS == 0 ) {
2470
+ Misbehaving (peer, 20 , strprintf (" %d non-connecting headers" , peer.m_num_unconnecting_headers_msgs ));
2471
2471
}
2472
2472
}
2473
2473
@@ -2718,10 +2718,10 @@ void PeerManagerImpl::HeadersDirectFetchBlocks(CNode& pfrom, const Peer& peer, c
2718
2718
void PeerManagerImpl::UpdatePeerStateForReceivedHeaders (CNode& pfrom, Peer& peer,
2719
2719
const CBlockIndex& last_header, bool received_new_header, bool may_have_more_headers)
2720
2720
{
2721
- if (peer.nUnconnectingHeaders > 0 ) {
2722
- LogPrint (BCLog::NET, " peer=%d: resetting nUnconnectingHeaders (%d -> 0)\n " , pfrom.GetId (), peer.nUnconnectingHeaders );
2721
+ if (peer.m_num_unconnecting_headers_msgs > 0 ) {
2722
+ LogPrint (BCLog::NET, " peer=%d: resetting m_num_unconnecting_headers_msgs (%d -> 0)\n " , pfrom.GetId (), peer.m_num_unconnecting_headers_msgs );
2723
2723
}
2724
- peer.nUnconnectingHeaders = 0 ;
2724
+ peer.m_num_unconnecting_headers_msgs = 0 ;
2725
2725
2726
2726
LOCK (cs_main);
2727
2727
CNodeState *nodestate = State (pfrom.GetId ());
@@ -3453,7 +3453,7 @@ void PeerManagerImpl::ProcessMessage(CNode& pfrom, const std::string& msg_type,
3453
3453
}
3454
3454
3455
3455
if (msg_type == NetMsgType::SENDHEADERS) {
3456
- peer->fPreferHeaders = true ;
3456
+ peer->m_prefers_headers = true ;
3457
3457
return ;
3458
3458
}
3459
3459
@@ -5459,7 +5459,7 @@ bool PeerManagerImpl::SendMessages(CNode* pto)
5459
5459
// add all to the inv queue.
5460
5460
LOCK (peer->m_block_inv_mutex );
5461
5461
std::vector<CBlock> vHeaders;
5462
- bool fRevertToInv = ((!peer->fPreferHeaders &&
5462
+ bool fRevertToInv = ((!peer->m_prefers_headers &&
5463
5463
(!state.m_requested_hb_cmpctblocks || peer->m_blocks_for_headers_relay .size () > 1 )) ||
5464
5464
peer->m_blocks_for_headers_relay .size () > MAX_BLOCKS_TO_ANNOUNCE);
5465
5465
const CBlockIndex *pBestIndex = nullptr ; // last header queued for delivery
@@ -5536,7 +5536,7 @@ bool PeerManagerImpl::SendMessages(CNode* pto)
5536
5536
m_connman.PushMessage (pto, msgMaker.Make (NetMsgType::CMPCTBLOCK, cmpctblock));
5537
5537
}
5538
5538
state.pindexBestHeaderSent = pBestIndex;
5539
- } else if (peer->fPreferHeaders ) {
5539
+ } else if (peer->m_prefers_headers ) {
5540
5540
if (vHeaders.size () > 1 ) {
5541
5541
LogPrint (BCLog::NET, " %s: %u headers, range (%s, %s), to peer=%d\n " , __func__,
5542
5542
vHeaders.size (),
0 commit comments