From 19b9d1a62531a3c985ea490f57fe7c25b03fddcb Mon Sep 17 00:00:00 2001 From: zjumathcode Date: Mon, 1 Dec 2025 18:39:03 +0800 Subject: [PATCH] refactor: omit unnecessary reassignment Signed-off-by: zjumathcode --- module/chunks/chunkVerifier_test.go | 1 - module/component/component.go | 1 - module/component/component_manager_test.go | 2 -- module/executiondatasync/execution_data/downloader.go | 2 -- module/executiondatasync/provider/provider.go | 2 -- module/mempool/queue/heroQueue_test.go | 1 - module/mempool/queue/heroStore_test.go | 1 - .../state_synchronization/requester/unittest/unittest.go | 1 - module/trace/trace_test.go | 1 - module/util/util_test.go | 2 -- network/alsp/internal/cache_test.go | 7 ------- network/p2p/scoring/internal/subscriptionCache_test.go | 1 - network/p2p/test/fixtures.go | 1 - network/p2p/tracer/internal/rpc_sent_tracker_test.go | 1 - 14 files changed, 24 deletions(-) diff --git a/module/chunks/chunkVerifier_test.go b/module/chunks/chunkVerifier_test.go index c108c95eca8..18f63d2bea7 100644 --- a/module/chunks/chunkVerifier_test.go +++ b/module/chunks/chunkVerifier_test.go @@ -671,7 +671,6 @@ func generateEvents(t *testing.T, collection *flow.Collection, includeServiceEve // service events are also included as regular events if includeServiceEvent { for _, e := range serviceEventsList { - e := e event, err := convert.ServiceEvent(testChain, e) require.NoError(t, err) diff --git a/module/component/component.go b/module/component/component.go index 11b543f239f..a798a5aae92 100644 --- a/module/component/component.go +++ b/module/component/component.go @@ -264,7 +264,6 @@ func (c *ComponentManager) Start(parent irrecoverable.SignalerContext) { // launch workers for _, worker := range c.workers { - worker := worker go func() { defer workersDone.Done() var readyOnce sync.Once diff --git a/module/component/component_manager_test.go b/module/component/component_manager_test.go index 1431abdd4c0..4758dff4950 100644 --- a/module/component/component_manager_test.go +++ b/module/component/component_manager_test.go @@ -531,8 +531,6 @@ func (c *ComponentManagerMachine) ExecuteStateTransition(t *rapid.T) { } for i, workerId := range st.workerIDs { - i := i - workerId := workerId addTransition(func() { wst := st.workerTransitions[i] t.Logf("executing worker %v transition: %v\n", workerId, wst) diff --git a/module/executiondatasync/execution_data/downloader.go b/module/executiondatasync/execution_data/downloader.go index b0eae050519..a9e942127ea 100644 --- a/module/executiondatasync/execution_data/downloader.go +++ b/module/executiondatasync/execution_data/downloader.go @@ -105,8 +105,6 @@ func (d *downloader) Get(ctx context.Context, executionDataID flow.Identifier) ( var mu sync.Mutex for i, chunkDataID := range edRoot.ChunkExecutionDataIDs { - i := i - chunkDataID := chunkDataID g.Go(func() error { ced, cids, err := d.getChunkExecutionData( diff --git a/module/executiondatasync/provider/provider.go b/module/executiondatasync/provider/provider.go index 72551826634..76b73846543 100644 --- a/module/executiondatasync/provider/provider.go +++ b/module/executiondatasync/provider/provider.go @@ -160,8 +160,6 @@ func (p *ExecutionDataProvider) provide(ctx context.Context, blockHeight uint64, chunkDataIDs := make([]cid.Cid, len(executionData.ChunkExecutionDatas)) for i, chunkExecutionData := range executionData.ChunkExecutionDatas { - i := i - chunkExecutionData := chunkExecutionData g.Go(func() error { logger.Debug().Int("chunk_index", i).Msg("adding chunk execution data") diff --git a/module/mempool/queue/heroQueue_test.go b/module/mempool/queue/heroQueue_test.go index 0107c789a56..233e06b7f3b 100644 --- a/module/mempool/queue/heroQueue_test.go +++ b/module/mempool/queue/heroQueue_test.go @@ -73,7 +73,6 @@ func TestHeroQueue_Concurrent(t *testing.T) { entities := unittest.EntityListFixture(uint(sizeLimit)) // pushing entities concurrently. for _, e := range entities { - e := e // suppress loop variable go func() { require.True(t, q.Push(e.Identifier, e)) pushWG.Done() diff --git a/module/mempool/queue/heroStore_test.go b/module/mempool/queue/heroStore_test.go index 209faee37cd..35de1a41a12 100644 --- a/module/mempool/queue/heroStore_test.go +++ b/module/mempool/queue/heroStore_test.go @@ -56,7 +56,6 @@ func TestHeroStore_Concurrent(t *testing.T) { messages := unittest.EngineMessageFixtures(sizeLimit) // putting messages concurrently. for _, m := range messages { - m := m go func() { require.True(t, store.Put(m)) putWG.Done() diff --git a/module/state_synchronization/requester/unittest/unittest.go b/module/state_synchronization/requester/unittest/unittest.go index fd350ffd444..d403ad2781b 100644 --- a/module/state_synchronization/requester/unittest/unittest.go +++ b/module/state_synchronization/requester/unittest/unittest.go @@ -29,7 +29,6 @@ func MockBlobService(bs blockstore.Blockstore) *mocknetwork.BlobService { wg.Add(len(cids)) for _, c := range cids { - c := c go func() { defer wg.Done() diff --git a/module/trace/trace_test.go b/module/trace/trace_test.go index f1011589930..be5ca1f3f2a 100644 --- a/module/trace/trace_test.go +++ b/module/trace/trace_test.go @@ -43,7 +43,6 @@ func BenchmarkStartBlockSpan(b *testing.B) { {name: "cacheHit", n: 100}, {name: "cacheMiss", n: 100000}, } { - t := t b.Run(t.name, func(b *testing.B) { randomIDs := make([]flow.Identifier, 0, t.n) for i := 0; i < t.n; i++ { diff --git a/module/util/util_test.go b/module/util/util_test.go index 8d0f42ed1ed..f5baf79b6c6 100644 --- a/module/util/util_test.go +++ b/module/util/util_test.go @@ -138,8 +138,6 @@ func TestMergeChannels(t *testing.T) { channels := []chan int{make(chan int), make(chan int), make(chan int)} merged := util.MergeChannels(channels).(<-chan int) for i, ch := range channels { - i := i - ch := ch go func() { ch <- i close(ch) diff --git a/network/alsp/internal/cache_test.go b/network/alsp/internal/cache_test.go index 8a7e116e3bb..7355dfbff63 100644 --- a/network/alsp/internal/cache_test.go +++ b/network/alsp/internal/cache_test.go @@ -539,7 +539,6 @@ func TestSpamRecordCache_ConcurrentInitAndRemove(t *testing.T) { // initialize spam records concurrently for _, originID := range originIDsToAdd { - originID := originID // capture range variable go func() { defer wg.Done() penalty, err := cache.AdjustWithInit(originID, adjustFnNoOp) @@ -612,7 +611,6 @@ func TestSpamRecordCache_ConcurrentInitRemoveAdjust(t *testing.T) { // Initialize spam records concurrently for _, originID := range originIDsToAdd { - originID := originID // capture range variable go func() { defer wg.Done() penalty, err := cache.AdjustWithInit(originID, adjustFnNoOp) @@ -684,7 +682,6 @@ func TestSpamRecordCache_ConcurrentInitRemoveAndAdjust(t *testing.T) { // initialize spam records concurrently for _, originID := range originIDsToAdd { - originID := originID go func() { defer wg.Done() penalty, err := cache.AdjustWithInit(originID, adjustFnNoOp) @@ -695,7 +692,6 @@ func TestSpamRecordCache_ConcurrentInitRemoveAndAdjust(t *testing.T) { // remove spam records concurrently for _, originID := range originIDsToRemove { - originID := originID go func() { defer wg.Done() cache.Remove(originID) @@ -704,7 +700,6 @@ func TestSpamRecordCache_ConcurrentInitRemoveAndAdjust(t *testing.T) { // adjust spam records concurrently for _, originID := range originIDsToAdjust { - originID := originID go func() { defer wg.Done() _, err := cache.AdjustWithInit(originID, func(record *model.ProtocolSpamRecord) (*model.ProtocolSpamRecord, error) { @@ -773,7 +768,6 @@ func TestSpamRecordCache_ConcurrentIdentitiesAndOperations(t *testing.T) { // initialize spam records concurrently for _, originID := range originIDsToAdd { - originID := originID go func() { defer wg.Done() penalty, err := cache.AdjustWithInit(originID, adjustFnNoOp) @@ -787,7 +781,6 @@ func TestSpamRecordCache_ConcurrentIdentitiesAndOperations(t *testing.T) { // remove spam records concurrently for _, originID := range originIDsToRemove { - originID := originID go func() { defer wg.Done() require.True(t, cache.Remove(originID)) diff --git a/network/p2p/scoring/internal/subscriptionCache_test.go b/network/p2p/scoring/internal/subscriptionCache_test.go index 54b88707702..a984d0a466e 100644 --- a/network/p2p/scoring/internal/subscriptionCache_test.go +++ b/network/p2p/scoring/internal/subscriptionCache_test.go @@ -258,7 +258,6 @@ func TestSubscriptionCache_ConcurrentUpdate(t *testing.T) { // verify that all peers have all topics; concurrently allTopicsVerified := sync.WaitGroup{} for _, pid := range peerIds { - pid := pid allTopicsVerified.Add(1) go func() { defer allTopicsVerified.Done() diff --git a/network/p2p/test/fixtures.go b/network/p2p/test/fixtures.go index 88f3c7667cd..d0d2bce0291 100644 --- a/network/p2p/test/fixtures.go +++ b/network/p2p/test/fixtures.go @@ -752,7 +752,6 @@ func EnsureNoPubsubMessageExchange(t *testing.T, wg := &sync.WaitGroup{} for _, node := range from { - node := node // capture range variable for i := 0; i < count; i++ { wg.Add(1) go func() { diff --git a/network/p2p/tracer/internal/rpc_sent_tracker_test.go b/network/p2p/tracer/internal/rpc_sent_tracker_test.go index 938a998cf46..e5bc85c1d37 100644 --- a/network/p2p/tracer/internal/rpc_sent_tracker_test.go +++ b/network/p2p/tracer/internal/rpc_sent_tracker_test.go @@ -54,7 +54,6 @@ func TestRPCSentTracker_IHave(t *testing.T) { } iHaves := make([]*pb.ControlIHave, len(testCases)) for i, testCase := range testCases { - testCase := testCase iHaves[i] = &pb.ControlIHave{ MessageIDs: testCase.messageIDS, }