@@ -32,6 +32,7 @@ def tokenizer_mock():
3232 return tokenizer
3333
3434
35+ @pytest .mark .xfail (reason = "old and broken" , run = False )
3536@pytest .mark .smoke
3637@patch (f"{ process_dataset .__module__ } .guidellm_load_dataset" )
3738@patch (f"{ process_dataset .__module__ } .check_load_processor" )
@@ -119,6 +120,7 @@ def test_handle_error_strategy_too_short_prompt(tokenizer_mock):
119120 handle_error_strategy ("short" , 10 , tokenizer_mock )
120121
121122
123+ @pytest .mark .xfail (reason = "old and broken" , run = False )
122124@pytest .mark .smoke
123125@patch (f"{ process_dataset .__module__ } .save_dataset_to_file" )
124126@patch (f"{ process_dataset .__module__ } .Dataset" )
@@ -165,6 +167,7 @@ def test_process_dataset_non_empty(
165167 assert len (tokenizer_mock .encode (item ["prompt" ])) <= 3
166168
167169
170+ @pytest .mark .xfail (reason = "old and broken" , run = False )
168171@pytest .mark .sanity
169172@patch (f"{ process_dataset .__module__ } .Dataset" )
170173@patch (f"{ process_dataset .__module__ } .guidellm_load_dataset" )
@@ -195,6 +198,7 @@ def test_process_dataset_empty_after_processing(
195198 mock_dataset_class .from_list .assert_not_called ()
196199
197200
201+ @pytest .mark .xfail (reason = "old and broken" , run = False )
198202@pytest .mark .smoke
199203@patch (f"{ process_dataset .__module__ } .push_dataset_to_hub" )
200204@patch (f"{ process_dataset .__module__ } .Dataset" )
@@ -229,6 +233,7 @@ def test_process_dataset_push_to_hub_called(
229233 mock_push .assert_called_once_with ("id123" , mock_dataset_obj )
230234
231235
236+ @pytest .mark .xfail (reason = "old and broken" , run = False )
232237@pytest .mark .sanity
233238@patch (f"{ process_dataset .__module__ } .push_dataset_to_hub" )
234239@patch (f"{ process_dataset .__module__ } .Dataset" )
0 commit comments