-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Running the example project
Jeremy Chou edited this page Mar 7, 2022
·
2 revisions
This example illustrates how to share a spider’s requests queue across multiple spider instances, highly suitable for broad crawls.
Setup scrapy_redis package in your PYTHONPATH
$ cd example-project
$ scrapy crawl dmoz
... [dmoz] ...
^C
$ scrapy crawl dmoz
... [dmoz] DEBUG: Resuming crawl (9019 requests scheduled)
$ scrapy crawl dmoz
... [dmoz] DEBUG: Resuming crawl (8712 requests scheduled)
```bash
### Start one or more post-processing workers:
```bash
$ python process_items.py dmoz:items -v
...
Processing: Kilani Giftware (http://www.dmoz.org/Computers/Shopping/Gifts/)
Processing: NinjaGizmos.com (http://www.dmoz.org/Computers/Shopping/Gifts/)
...
```bash