@@ -161,7 +161,11 @@ def test_client(mock_async_diffusion):
161161 app .state .engine_client = mock_async_diffusion
162162 app .state .diffusion_engine = mock_async_diffusion # Also set for health endpoint
163163 app .state .stage_configs = [{"stage_type" : "diffusion" }]
164- app .state .diffusion_model_name = "Qwen/Qwen-Image" # For models endpoint
164+
165+ from vllm .entrypoints .openai .serving_models import BaseModelPath
166+ from vllm_omni .entrypoints .openai .api_server import _DiffusionServingModels
167+ app .state .openai_serving_models = _DiffusionServingModels ([BaseModelPath (name = "Qwen/Qwen-Image" , model_path = "Qwen/Qwen-Image" )])
168+
165169 app .state .args = Namespace (
166170 default_sampling_params = '{"0": {"num_inference_steps":4, "guidance_scale":7.5}}' ,
167171 max_generated_image_size = 4096 , # 64*64
@@ -214,33 +218,6 @@ def test_health_endpoint_no_engine():
214218 assert data ["status" ] == "unhealthy"
215219
216220
217- def test_models_endpoint (test_client ):
218- """Test /v1/models endpoint for diffusion mode"""
219- response = test_client .get ("/v1/models" )
220- assert response .status_code == 200
221- data = response .json ()
222- assert data ["object" ] == "list"
223- assert len (data ["data" ]) == 1
224- assert data ["data" ][0 ]["id" ] == "Qwen/Qwen-Image"
225- assert data ["data" ][0 ]["object" ] == "model"
226-
227-
228- def test_models_endpoint_no_engine ():
229- """Test /v1/models endpoint when no engine is initialized"""
230- from fastapi import FastAPI
231-
232- from vllm_omni .entrypoints .openai .api_server import router
233-
234- app = FastAPI ()
235- app .include_router (router )
236- # Don't set any engine
237-
238- client = TestClient (app )
239- response = client .get ("/v1/models" )
240- assert response .status_code == 200
241- data = response .json ()
242- assert data ["object" ] == "list"
243- assert len (data ["data" ]) == 0
244221
245222
246223def test_generate_single_image (test_client ):
0 commit comments