You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
help = "Whether or not to read models from the block number they were registered in."
190
191
)]
191
192
pubstrict_model_reader:bool,
193
+
194
+
/// The chunk size to use for batch requests.
195
+
#[arg(
196
+
long = "indexing.batch_chunk_size",
197
+
default_value_t = DEFAULT_BATCH_CHUNK_SIZE,
198
+
help = "The chunk size to use for batch requests. This is used to split the requests into smaller chunks to avoid overwhelming the provider and potentially running into issues."
let results_of_chunks:Vec<Vec<ProviderResponseData>> = try_join_all(futures)
920
+
.await
921
+
.with_context(|| {
922
+
format!(
923
+
"One or more batch requests failed during chunked execution. This could be due to the provider being overloaded. You can try reducing the batch chunk size. Total requests: {}. Batch chunk size: {}",
924
+
requests.len(),
925
+
self.config.batch_chunk_size
926
+
)
927
+
})?;
928
+
929
+
let flattened_results = results_of_chunks.into_iter().flatten().collect();
0 commit comments