@@ -287,8 +287,7 @@ async def create_chat_completion(
287287
288288 generators .append (generator )
289289 except ValueError as e :
290- # TODO: Use a vllm-specific Validation Error
291- return self .create_error_response (str (e ))
290+ return self .create_error_response (e )
292291
293292 assert len (generators ) == 1
294293 (result_generator ,) = generators
@@ -316,8 +315,7 @@ async def create_chat_completion(
316315 request_metadata ,
317316 )
318317 except ValueError as e :
319- # TODO: Use a vllm-specific Validation Error
320- return self .create_error_response (str (e ))
318+ return self .create_error_response (e )
321319
322320 async def _preprocess_chat (
323321 self ,
@@ -634,7 +632,7 @@ async def chat_completion_stream_generator(
634632 )
635633 except RuntimeError as e :
636634 logger .exception ("Error in reasoning parser creation." )
637- data = self .create_streaming_error_response (str ( e ) )
635+ data = self .create_streaming_error_response (e )
638636 yield f"data: { data } \n \n "
639637 yield "data: [DONE]\n \n "
640638 return
@@ -646,7 +644,7 @@ async def chat_completion_stream_generator(
646644 tool_parsers = [None ] * num_choices
647645 except Exception as e :
648646 logger .exception ("Error in tool parser creation." )
649- data = self .create_streaming_error_response (str ( e ) )
647+ data = self .create_streaming_error_response (e )
650648 yield f"data: { data } \n \n "
651649 yield "data: [DONE]\n \n "
652650 return
@@ -1286,9 +1284,8 @@ async def chat_completion_stream_generator(
12861284 )
12871285
12881286 except Exception as e :
1289- # TODO: Use a vllm-specific Validation Error
12901287 logger .exception ("Error in chat completion stream generator." )
1291- data = self .create_streaming_error_response (str ( e ) )
1288+ data = self .create_streaming_error_response (e )
12921289 yield f"data: { data } \n \n "
12931290 # Send the final done message after all response.n are finished
12941291 yield "data: [DONE]\n \n "
@@ -1313,8 +1310,7 @@ async def chat_completion_full_generator(
13131310 except asyncio .CancelledError :
13141311 return self .create_error_response ("Client disconnected" )
13151312 except ValueError as e :
1316- # TODO: Use a vllm-specific Validation Error
1317- return self .create_error_response (str (e ))
1313+ return self .create_error_response (e )
13181314
13191315 assert final_outputs is not None
13201316
@@ -1487,7 +1483,7 @@ def _create_text_choice(
14871483 )
14881484 except RuntimeError as e :
14891485 logger .exception ("Error in reasoning parser creation." )
1490- return self .create_error_response (str ( e ) )
1486+ return self .create_error_response (e )
14911487 # If the reasoning parser is enabled,
14921488 # tool calls are extracted exclusively from the content.
14931489 reasoning_content , content = reasoning_parser .extract_reasoning (output .text , request = request )
@@ -1572,7 +1568,7 @@ def _create_text_choice(
15721568 tool_parser = self .tool_parser (tokenizer )
15731569 except RuntimeError as e :
15741570 logger .exception ("Error in tool parser creation." )
1575- return self .create_error_response (str ( e ) )
1571+ return self .create_error_response (e )
15761572
15771573 tool_call_info = tool_parser .extract_tool_calls (content if content is not None else "" , request = request )
15781574 # In the OpenAI API the finish_reason is "tools_called"
0 commit comments