|
18 | 18 | import os |
19 | 19 | import unittest |
20 | 20 | from datetime import datetime |
21 | | -from unittest.mock import patch |
| 21 | +from types import SimpleNamespace |
| 22 | +from unittest.mock import patch, MagicMock |
22 | 23 |
|
23 | 24 | from sqlalchemy.orm import Session |
24 | 25 |
|
@@ -46,7 +47,7 @@ def test_handler_calls_main_function(self, mock_rebuild_func): |
46 | 47 |
|
47 | 48 | self.assertEqual(response["message"], "test") |
48 | 49 | mock_rebuild_func.assert_called_once_with( |
49 | | - dry_run=True, after_date="2024-01-01", latest_only=False |
| 50 | + dry_run=True, after_date="2024-01-01", latest_only=False, dataset_id=None |
50 | 51 | ) |
51 | 52 |
|
52 | 53 | @with_db_session(db_url=default_db_url) |
@@ -93,3 +94,101 @@ def test_rebuild_missing_dataset_files_processing( |
93 | 94 | self.assertIn("completed", response["message"]) |
94 | 95 | self.assertGreaterEqual(response["total_processed"], 0) |
95 | 96 | self.assertTrue(publish_mock.called or response["total_processed"] == 0) |
| 97 | + |
| 98 | + |
| 99 | +class TestRebuildSpecificDatasetFiles(unittest.TestCase): |
| 100 | + @patch( |
| 101 | + "tasks.dataset_files.rebuild_missing_dataset_files.rebuild_missing_dataset_files" |
| 102 | + ) |
| 103 | + def test_handler_calls_main_function(self, mock_rebuild_func): |
| 104 | + mock_rebuild_func.return_value = {"message": "test", "total_processed": 0} |
| 105 | + payload = {"dry_run": True, "after_date": "2024-01-01", "latest_only": False} |
| 106 | + |
| 107 | + response = rebuild_missing_dataset_files_handler(payload) |
| 108 | + |
| 109 | + self.assertEqual(response["message"], "test") |
| 110 | + mock_rebuild_func.assert_called_once_with( |
| 111 | + dry_run=True, after_date="2024-01-01", latest_only=False, dataset_id=None |
| 112 | + ) |
| 113 | + |
| 114 | + @patch( |
| 115 | + "tasks.dataset_files.rebuild_missing_dataset_files.rebuild_missing_dataset_files" |
| 116 | + ) |
| 117 | + def test_handler_forwards_dataset_id(self, mock_rebuild_func): |
| 118 | + payload = { |
| 119 | + "dry_run": False, |
| 120 | + "after_date": None, |
| 121 | + "latest_only": True, |
| 122 | + "dataset_id": "ds-123", |
| 123 | + } |
| 124 | + |
| 125 | + rebuild_missing_dataset_files_handler(payload) |
| 126 | + |
| 127 | + mock_rebuild_func.assert_called_once_with( |
| 128 | + dry_run=False, after_date=None, latest_only=True, dataset_id="ds-123" |
| 129 | + ) |
| 130 | + |
| 131 | + def test_rebuild_with_specific_dataset_id_publishes_one_message(self): |
| 132 | + dataset_stable_id = "ds-123" |
| 133 | + fake_feed = SimpleNamespace( |
| 134 | + producer_url="https://example.com", |
| 135 | + stable_id="feed-stable", |
| 136 | + id=42, |
| 137 | + authentication_type=None, |
| 138 | + authentication_info_url=None, |
| 139 | + api_key_parameter_name=None, |
| 140 | + ) |
| 141 | + fake_dataset = SimpleNamespace( |
| 142 | + stable_id=dataset_stable_id, hash="abc123", feed=fake_feed |
| 143 | + ) |
| 144 | + |
| 145 | + # Mock the chained SQLAlchemy calls: |
| 146 | + # db_session.query(Gtfsdataset).filter(...).options(...).count()/all() |
| 147 | + db_session = MagicMock() |
| 148 | + query_mock = MagicMock() |
| 149 | + filter_mock = MagicMock() |
| 150 | + options_mock = MagicMock() |
| 151 | + |
| 152 | + db_session.query.return_value = query_mock |
| 153 | + query_mock.filter.return_value = filter_mock |
| 154 | + filter_mock.options.return_value = options_mock |
| 155 | + |
| 156 | + options_mock.count.return_value = 1 |
| 157 | + options_mock.all.return_value = [fake_dataset] |
| 158 | + |
| 159 | + with patch.dict( |
| 160 | + os.environ, |
| 161 | + {"PROJECT_ID": "test-project", "DATASET_PROCESSING_TOPIC_NAME": "topic"}, |
| 162 | + clear=False, |
| 163 | + ), patch( |
| 164 | + "tasks.dataset_files.rebuild_missing_dataset_files.get_datasets_with_missing_files_query" |
| 165 | + ) as get_query_mock, patch( |
| 166 | + "tasks.dataset_files.rebuild_missing_dataset_files.publish_messages" |
| 167 | + ) as mock_publish: |
| 168 | + from tasks.dataset_files.rebuild_missing_dataset_files import ( |
| 169 | + rebuild_missing_dataset_files, |
| 170 | + Gtfsdataset, |
| 171 | + ) |
| 172 | + |
| 173 | + result = rebuild_missing_dataset_files( |
| 174 | + db_session=db_session, |
| 175 | + dry_run=False, |
| 176 | + after_date=None, |
| 177 | + latest_only=True, # ignored when dataset_id is provided |
| 178 | + dataset_id=dataset_stable_id, |
| 179 | + ) |
| 180 | + |
| 181 | + # Asserts |
| 182 | + get_query_mock.assert_not_called() # bypasses generic query when dataset_id is set |
| 183 | + db_session.query.assert_called_once_with(Gtfsdataset) |
| 184 | + query_mock.filter.assert_called_once() # filtered by stable_id |
| 185 | + options_mock.count.assert_called_once() |
| 186 | + options_mock.all.assert_called_once() |
| 187 | + |
| 188 | + self.assertEqual(result["total_processed"], 1) |
| 189 | + mock_publish.assert_called_once() |
| 190 | + |
| 191 | + messages_arg, project_id_arg, _topic_arg = mock_publish.call_args[0] |
| 192 | + self.assertEqual(project_id_arg, "test-project") |
| 193 | + self.assertEqual(len(messages_arg), 1) |
| 194 | + self.assertEqual(messages_arg[0]["dataset_stable_id"], dataset_stable_id) |
0 commit comments