|
8 | 8 | import pytest |
9 | 9 | from yarl import URL |
10 | 10 |
|
11 | | -from crawlee._request import UserData |
| 11 | +from crawlee._request import CrawleeRequestData, UserData |
12 | 12 | from crawlee._types import HttpMethod |
13 | 13 |
|
14 | 14 | from apify.request_loaders import ApifyRequestList |
@@ -65,6 +65,8 @@ async def test_request_list_open_request_types( |
65 | 65 | assert request.url == request_dict_input['url'] |
66 | 66 | assert request.payload == request_dict_input.get('payload', '').encode('utf-8') |
67 | 67 | expected_user_data = UserData() |
| 68 | + # `crawlee_data` must be present in user_data |
| 69 | + expected_user_data.crawlee_data = CrawleeRequestData() |
68 | 70 | if 'userData' in optional_input: |
69 | 71 | for key, value in optional_input['userData'].items(): |
70 | 72 | expected_user_data[key] = value |
@@ -181,6 +183,8 @@ async def test_request_list_open_from_url_additional_inputs(httpserver: HTTPServ |
181 | 183 | assert request.headers.root == example_start_url_input['headers'] |
182 | 184 | assert request.payload == str(example_start_url_input['payload']).encode('utf-8') |
183 | 185 | expected_user_data = UserData() |
| 186 | + # `crawlee_data` must be present in user_data |
| 187 | + expected_user_data.crawlee_data = CrawleeRequestData() |
184 | 188 | for key, value in example_start_url_input['userData'].items(): |
185 | 189 | expected_user_data[key] = value |
186 | 190 | assert request.user_data == expected_user_data |
|
0 commit comments