|
1 | | -import unittest |
2 | 1 | import openml |
3 | 2 | import openml.evaluations |
4 | 3 | from openml.testing import TestBase |
@@ -29,6 +28,42 @@ def test_evaluation_list_filter_uploader(self): |
29 | 28 | # self.assertEquals(evaluations[run_id].uploader, uploader_id) |
30 | 29 |
|
31 | 30 |
|
| 31 | + def test_evaluation_list_filter_uploader(self): |
| 32 | + openml.config.server = self.production_server |
| 33 | + |
| 34 | + setup_id = 10 |
| 35 | + |
| 36 | + evaluations = openml.evaluations.list_evaluations("predictive_accuracy", setup=[setup_id]) |
| 37 | + |
| 38 | + self.assertGreater(len(evaluations), 100) |
| 39 | + for run_id in evaluations.keys(): |
| 40 | + self.assertEquals(evaluations[run_id].setup_id, setup_id) |
| 41 | + |
| 42 | + |
| 43 | + def test_evaluation_list_filter_flow(self): |
| 44 | + openml.config.server = self.production_server |
| 45 | + |
| 46 | + flow_id = 100 |
| 47 | + |
| 48 | + evaluations = openml.evaluations.list_evaluations("predictive_accuracy", flow=[flow_id]) |
| 49 | + |
| 50 | + self.assertGreater(len(evaluations), 2) |
| 51 | + for run_id in evaluations.keys(): |
| 52 | + self.assertEquals(evaluations[run_id].flow_id, flow_id) |
| 53 | + |
| 54 | + |
| 55 | + def test_evaluation_list_filter_run(self): |
| 56 | + openml.config.server = self.production_server |
| 57 | + |
| 58 | + run_id = 1 |
| 59 | + |
| 60 | + evaluations = openml.evaluations.list_evaluations("predictive_accuracy", id=[run_id]) |
| 61 | + |
| 62 | + self.assertEquals(len(evaluations), 1) |
| 63 | + for run_id in evaluations.keys(): |
| 64 | + self.assertEquals(evaluations[run_id].run_id, run_id) |
| 65 | + |
| 66 | + |
32 | 67 | def test_evaluation_list_limit(self): |
33 | 68 | openml.config.server = self.production_server |
34 | 69 |
|
|
0 commit comments