|
| 1 | +{ |
| 2 | + "title": "Web Checker", |
| 3 | + "description": "The web checker actor loads <b>URLs to check</b> and checks for common captchas, status codes returned from crawling, as well as calculates the price a user may pay. <b>TODO: Needs to be more descriptive!!</b>", |
| 4 | + "type": "object", |
| 5 | + "schemaVersion": 1, |
| 6 | + "properties": { |
| 7 | + "urlsToCheck": { |
| 8 | + "title": "URLs to check", |
| 9 | + "type": "array", |
| 10 | + "description": "A static list of URLs to check for captchas. To be able to add new URLs on the fly, enable the <b>Use request queue</b> option.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#start-urls' target='_blank' rel='noopener'>Start URLs</a> in README.", |
| 11 | + "sectionCaption": "Checker Options", |
| 12 | + "sectionDescription": "Options that will be passed to the checkers", |
| 13 | + "editor": "requestListSources", |
| 14 | + "prefill": [ |
| 15 | + { |
| 16 | + "url": "https://www.amazon.com/b?ie=UTF8&node=11392907011" |
| 17 | + } |
| 18 | + ] |
| 19 | + }, |
| 20 | + "proxyConfiguration": { |
| 21 | + "title": "Proxy Configuration", |
| 22 | + "type": "object", |
| 23 | + "description": "Specifies proxy servers that will be used by the scraper in order to hide its origin.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#proxy-configuration' target='_blank' rel='noopener'>Proxy configuration</a> in README.", |
| 24 | + "default": {}, |
| 25 | + "editor": "proxy", |
| 26 | + "prefill": { |
| 27 | + "useApifyProxy": false |
| 28 | + } |
| 29 | + }, |
| 30 | + "saveSnapshot": { |
| 31 | + "title": "Enabled", |
| 32 | + "type": "boolean", |
| 33 | + "description": "Will save HTML for Cheerio and HTML + screenshot for Puppeteer/Playwright", |
| 34 | + "editor": "checkbox", |
| 35 | + "groupCaption": "Save Snapshots" |
| 36 | + }, |
| 37 | + "linkSelector": { |
| 38 | + "title": "Link Selector", |
| 39 | + "type": "string", |
| 40 | + "description": "A CSS selector saying which links on the page (<code><a></code> elements with <code>href</code> attribute) shall be followed and added to the request queue. This setting only applies if <b>Use request queue</b> is enabled. To filter the links added to the queue, use the <b>Pseudo-URLs</b> setting.<br><br>If <b>Link selector</b> is empty, the page links are ignored.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#link-selector' target='_blank' rel='noopener'>Link selector</a> in README.", |
| 41 | + "sectionCaption": "Crawler Options", |
| 42 | + "sectionDescription": "Specific options that are relevant for crawlers", |
| 43 | + "editor": "textfield", |
| 44 | + "prefill": "a[href]", |
| 45 | + "minLength": 1 |
| 46 | + }, |
| 47 | + "pseudoUrls": { |
| 48 | + "title": "Pseudo-URLs", |
| 49 | + "type": "array", |
| 50 | + "description": "Specifies what kind of URLs found by <b>Link selector</b> should be added to the request queue. A pseudo-URL is a URL with regular expressions enclosed in <code>[]</code> brackets, e.g. <code>http://www.example.com/[.*]</code>. This setting only applies if the <b>Use request queue</b> option is enabled.<br><br>If <b>Pseudo-URLs</b> are omitted, the actor enqueues all links matched by the <b>Link selector</b>.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#pseudo-urls' target='_blank' rel='noopener'>Pseudo-URLs</a> in README.", |
| 51 | + "default": [], |
| 52 | + "editor": "pseudoUrls", |
| 53 | + "prefill": [ |
| 54 | + { |
| 55 | + "purl": "https://www.amazon.com[.*]/dp/[.*]" |
| 56 | + } |
| 57 | + ] |
| 58 | + }, |
| 59 | + "repeatChecksOnProvidedUrls": { |
| 60 | + "title": "Repeat checks on provided URLs", |
| 61 | + "type": "integer", |
| 62 | + "description": "Will access each URL multiple times. Useful to test the same URL or bypass blocking of the first page.", |
| 63 | + "editor": "number" |
| 64 | + }, |
| 65 | + "maxNumberOfPagesCheckedPerDomain": { |
| 66 | + "title": "Max number of pages checked per domain", |
| 67 | + "type": "integer", |
| 68 | + "description": "The maximum number of pages that the checker will load. The checker will stop when this limit is reached. It's always a good idea to set this limit in order to prevent excess platform usage for misconfigured scrapers. Note that the actual number of pages loaded might be slightly higher than this value.<br><br>If set to <code>0</code>, there is no limit.", |
| 69 | + "default": 100, |
| 70 | + "editor": "number" |
| 71 | + }, |
| 72 | + "maxConcurrentPagesCheckedPerDomain": { |
| 73 | + "title": "Maximum concurrent pages checked per domain", |
| 74 | + "type": "integer", |
| 75 | + "description": "Specifies the maximum number of pages that can be processed by the checker in parallel for one domain. The checker automatically increases and decreases concurrency based on available system resources. This option enables you to set an upper limit, for example to reduce the load on a target website.", |
| 76 | + "default": 50, |
| 77 | + "editor": "number", |
| 78 | + "minimum": 1 |
| 79 | + }, |
| 80 | + "maxConcurrentDomainsChecked": { |
| 81 | + "title": "Maximum number of concurrent domains checked", |
| 82 | + "type": "integer", |
| 83 | + "description": "Specifies the maximum number of domains that should be checked at a time. This setting is relevant when passing in more than one URL to check.", |
| 84 | + "default": 5, |
| 85 | + "editor": "number", |
| 86 | + "minimum": 1, |
| 87 | + "maximum": 10 |
| 88 | + }, |
| 89 | + "retireBrowserInstanceAfterRequestCount": { |
| 90 | + "title": "Retire browser instance after request count", |
| 91 | + "type": "integer", |
| 92 | + "description": "How often will the browser itself rotate. Pick a higher number for smaller consumption, pick a lower number to rotate (test) more proxies.", |
| 93 | + "default": 10, |
| 94 | + "editor": "number", |
| 95 | + "minimum": 1 |
| 96 | + } |
| 97 | + }, |
| 98 | + "required": ["urlsToCheck"] |
| 99 | +} |
0 commit comments