Skip to content

Commit 2e58992

Browse files
webrdanielbaldassevakatzinoB4nanbarjin
authored
docs: website design uplift (#2872)
Deployed here: https://webrdaniel.github.io/: It is not possible to navigate to the updated `/python` page as it lives in different repo. Changes: - navigation header (different on index page and javascript pages) - footer - search modal - new index page - new [/js](https://webrdaniel.github.io/js) page Figma design files are [here](https://www.figma.com/design/EK51g98Cjx8UzmKJOUMcVe/Crawlee-brand-exploration?node-id=768-26162&t=tARb0KC2maZWdIaZ-0) Content file: https://www.notion.so/apify/Crawlee-homepage-value-props-12af39950a2280d29d20edb080b2ece9?pvs=4 --------- Co-authored-by: Eva Baldasseroni <37815707+baldasseva@users.noreply.github.com> Co-authored-by: katzino <katzino@gmail.com> Co-authored-by: Martin Adámek <banan23@gmail.com> Co-authored-by: Jindřich Bär <jindrichbar@gmail.com>
1 parent 3c2a38a commit 2e58992

File tree

336 files changed

+5600
-1835
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

336 files changed

+5600
-1835
lines changed

CHANGELOG.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1256,7 +1256,7 @@ const crawler = new BasicCrawler({
12561256

12571257
#### How to use `sendRequest()`?
12581258

1259-
See [the Got Scraping guide](https://crawlee.dev/docs/guides/got-scraping).
1259+
See [the Got Scraping guide](https://crawlee.dev/js/docs/guides/got-scraping).
12601260

12611261
#### Removed options
12621262

@@ -1381,7 +1381,7 @@ Previously, you were able to have a browser pool that would mix Puppeteer and Pl
13811381

13821382
One small feature worth mentioning is the ability to handle requests with browser crawlers outside the browser. To do that, we can use a combination of `Request.skipNavigation` and `context.sendRequest()`.
13831383

1384-
Take a look at how to achieve this by checking out the [Skipping navigation for certain requests](https://crawlee.dev/docs/examples/skip-navigation) example!
1384+
Take a look at how to achieve this by checking out the [Skipping navigation for certain requests](https://crawlee.dev/js/docs/examples/skip-navigation) example!
13851385

13861386
### Logging
13871387

@@ -1441,14 +1441,14 @@ await Actor.main(async () => {
14411441

14421442
#### Events
14431443

1444-
Apify SDK (v2) exports `Apify.events`, which is an `EventEmitter` instance. With Crawlee, the events are managed by [`EventManager`](https://crawlee.dev/api/core/class/EventManager) class instead. We can either access it via `Actor.eventManager` getter, or use `Actor.on` and `Actor.off` shortcuts instead.
1444+
Apify SDK (v2) exports `Apify.events`, which is an `EventEmitter` instance. With Crawlee, the events are managed by [`EventManager`](https://crawlee.dev/js/api/core/class/EventManager) class instead. We can either access it via `Actor.eventManager` getter, or use `Actor.on` and `Actor.off` shortcuts instead.
14451445

14461446
```diff
14471447
-Apify.events.on(...);
14481448
+Actor.on(...);
14491449
```
14501450

1451-
> We can also get the [`EventManager`](https://crawlee.dev/api/core/class/EventManager) instance via `Configuration.getEventManager()`.
1451+
> We can also get the [`EventManager`](https://crawlee.dev/js/api/core/class/EventManager) instance via `Configuration.getEventManager()`.
14521452
14531453
In addition to the existing events, we now have an `exit` event fired when calling `Actor.exit()` (which is called at the end of `Actor.main()`). This event allows you to gracefully shut down any resources when `Actor.exit` is called.
14541454

MIGRATIONS.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Migration from 2.x.x to 3.0.0
2-
Check the v3 [upgrading guide](https://crawlee.dev/docs/upgrading/upgrading-to-v3).
2+
Check the v3 [upgrading guide](https://crawlee.dev/js/docs/upgrading/upgrading-to-v3).
33

44
# Migration from 1.x.x to 2.0.0
55
There should be no changes needed apart from upgrading your Node.js version to >= 15.10. If you encounter issues with `cheerio`, [read their CHANGELOG](https://github.com/cheeriojs/cheerio/releases). We bumped it from `rc.3` to `rc.10`.

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ Crawlee is available as the [`crawlee`](https://www.npmjs.com/package/crawlee) N
2828
2929
## Installation
3030

31-
We recommend visiting the [Introduction tutorial](https://crawlee.dev/docs/introduction) in Crawlee documentation for more information.
31+
We recommend visiting the [Introduction tutorial](https://crawlee.dev/js/docs/introduction) in Crawlee documentation for more information.
3232

3333
> Crawlee requires **Node.js 16 or higher**.
3434
@@ -78,7 +78,7 @@ const crawler = new PlaywrightCrawler({
7878
await crawler.run(['https://crawlee.dev']);
7979
```
8080

81-
By default, Crawlee stores data to `./storage` in the current working directory. You can override this directory via Crawlee configuration. For details, see [Configuration guide](https://crawlee.dev/docs/guides/configuration), [Request storage](https://crawlee.dev/docs/guides/request-storage) and [Result storage](https://crawlee.dev/docs/guides/result-storage).
81+
By default, Crawlee stores data to `./storage` in the current working directory. You can override this directory via Crawlee configuration. For details, see [Configuration guide](https://crawlee.dev/js/docs/guides/configuration), [Request storage](https://crawlee.dev/js/docs/guides/request-storage) and [Result storage](https://crawlee.dev/js/docs/guides/result-storage).
8282

8383
### Installing pre-release versions
8484

docs/examples/cheerio_crawler.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ const crawler = new CheerioCrawler({
2525

2626
// This function will be called for each URL to crawl.
2727
// It accepts a single parameter, which is an object with options as:
28-
// https://crawlee.dev/api/cheerio-crawler/interface/CheerioCrawlerOptions#requestHandler
28+
// https://crawlee.dev/js/api/cheerio-crawler/interface/CheerioCrawlerOptions#requestHandler
2929
// We use for demonstration only 2 of them:
3030
// - request: an instance of the Request class with information such as the URL that is being crawled and HTTP method
3131
// - $: the cheerio object containing parsed HTML

docs/examples/http_crawler.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ const crawler = new HttpCrawler({
2525

2626
// This function will be called for each URL to crawl.
2727
// It accepts a single parameter, which is an object with options as:
28-
// https://crawlee.dev/api/http-crawler/interface/HttpCrawlerOptions#requestHandler
28+
// https://crawlee.dev/js/api/http-crawler/interface/HttpCrawlerOptions#requestHandler
2929
// We use for demonstration only 2 of them:
3030
// - request: an instance of the Request class with information such as the URL that is being crawled and HTTP method
3131
// - body: the HTML code of the current page

docs/examples/jsdom_crawler.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ const crawler = new JSDOMCrawler({
2525

2626
// This function will be called for each URL to crawl.
2727
// It accepts a single parameter, which is an object with options as:
28-
// https://crawlee.dev/api/jsdom-crawler/interface/JSDOMCrawlerOptions#requestHandler
28+
// https://crawlee.dev/js/api/jsdom-crawler/interface/JSDOMCrawlerOptions#requestHandler
2929
// We use for demonstration only 2 of them:
3030
// - request: an instance of the Request class with information such as the URL that is being crawled and HTTP method
3131
// - window: the JSDOM window object

docs/experiments/request_locking.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,8 +82,8 @@ import { RequestQueueV2 } from 'crawlee';
8282
const queue = await RequestQueueV2.open('my-locking-queue');
8383
await queue.addRequests([
8484
{ url: 'https://crawlee.dev' },
85-
{ url: 'https://crawlee.dev/docs' },
86-
{ url: 'https://crawlee.dev/api' },
85+
{ url: 'https://crawlee.dev/js/docs' },
86+
{ url: 'https://crawlee.dev/js/api' },
8787
]);
8888
```
8989

docs/guides/docker_browser_js.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Specify the base Docker image. You can read more about
2-
# the available images at https://crawlee.dev/docs/guides/docker-images
2+
# the available images at https://crawlee.dev/js/docs/guides/docker-images
33
# You can also use any other image from Docker Hub.
44
FROM apify/actor-node-playwright-chrome:20
55

docs/guides/docker_browser_ts.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Specify the base Docker image. You can read more about
2-
# the available images at https://crawlee.dev/docs/guides/docker-images
2+
# the available images at https://crawlee.dev/js/docs/guides/docker-images
33
# You can also use any other image from Docker Hub.
44
FROM apify/actor-node-playwright-chrome:20 AS builder
55

docs/guides/docker_node_js.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Specify the base Docker image. You can read more about
2-
# the available images at https://crawlee.dev/docs/guides/docker-images
2+
# the available images at https://crawlee.dev/js/docs/guides/docker-images
33
# You can also use any other image from Docker Hub.
44
FROM apify/actor-node:20
55

0 commit comments

Comments
 (0)