You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -224,16 +224,19 @@ Crawler::create()
224
224
->setConcurrency(1) // now all urls will be crawled one by one
225
225
```
226
226
227
-
## Defining Crawl Limits
227
+
## Defining Crawl and Time Limits
228
228
229
229
By default, the crawler continues until it has crawled every page it can find. This behavior might cause issues if you are working in an environment with limitations such as a serverless environment.
230
230
231
231
The crawl behavior can be controlled with the following two options:
232
232
233
233
-**Total Crawl Limit** (`setTotalCrawlLimit`): This limit defines the maximal count of URLs to crawl.
234
234
-**Current Crawl Limit** (`setCurrentCrawlLimit`): This defines how many URLs are processed during the current crawl.
235
+
-**Total Execution Time Limit** (`setTotalExecutionTimeLimit`): This limit defines the maximal execution time of the crawl.
236
+
-**Current Execution Time Limit** (`setCurrentExecutionTimeLimit`): This limits the execution time of the current crawl.
235
237
236
-
Let's take a look at some examples to clarify the difference between these two methods.
238
+
Let's take a look at some examples to clarify the difference between `setTotalCrawlLimit` and `setCurrentCrawlLimit`.
239
+
The difference between `setTotalExecutionTimeLimit` and `setCurrentExecutionTimeLimit` will be the same.
0 commit comments