You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
**Description:** Transformers v5 `AutoModelForCausalLM.from_config()` now resolves `auto_map` entries in config.json and tries to load the referenced dynamic module file. The `tiny_nemotron5_h_with_nemotron_tokenizer` test asset has `auto_map.AutoModelForCausalLM = "modeling_nemotron_h.NemotronHForCausalLM"` but the `modeling_nemotron_h.py` file is missing from the test asset directory.
423
+
424
+
**Stack trace:**
425
+
```
426
+
ray::DTensorPolicyWorker.__init__() (pid=3460983)
427
+
File "dtensor_policy_worker.py", line 275, in __init__
428
+
self.model = model_class.from_config(...)
429
+
File "transformers/models/auto/auto_factory.py", line 226, in from_config
**Status:** SKIPPED — needs `modeling_nemotron_h.py` added to the test asset, or the test asset config needs to reference a model class that ships with transformers.
448
+
449
+
## Err 9. FP8 + cpu_offload colocated test borderline timeout
450
+
451
+
**Description:**`test_vllm_generation_with_hf_training_colocated[False-True-fp8-False]` (async_engine=False, cpu_offload=True, fp8) takes 303s, exceeding the 300s `@pytest.mark.timeout`. The sibling variant `[True-False-fp8-False]` (async_engine=True, no cpu_offload) passes at 266s. This is a borderline timeout, likely not related to transformers v5.
452
+
453
+
**Reproduction:**
454
+
```bash
455
+
cd tests && uv run --no-sync pytest unit/models/generation/test_vllm_generation.py::test_vllm_generation_with_hf_training_colocated -k "False-True-fp8-False" --hf-gated -x -s
0 commit comments