Skip to content

Pull requests: vllm-project/compressed-tensors

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Switch test runners to use the vllm runners
#496 opened Oct 15, 2025 by dhuangnm Loading…
Update neuralmagic --> vllm-project for links
#495 opened Oct 15, 2025 by mgoin Loading…
[Attention] Support FP4 attention quantization
#491 opened Oct 14, 2025 by kylesayrs Loading…
Tensor Group Validation
#490 opened Oct 14, 2025 by kylesayrs Loading…
[Attention] R3 Attention Transform
#485 opened Oct 8, 2025 by kylesayrs Loading…
[MXFP4] Add calibration support
#440 opened Aug 28, 2025 by dsikka Draft
[MXFP4] Support MXFp4 Format
#439 opened Aug 28, 2025 by dsikka Draft
[Transform] Attention/Cache transforms
#436 opened Aug 26, 2025 by kylesayrs Loading…
[KV Cache] support kv cache int8 per channel quant
#398 opened Jul 19, 2025 by Eviannn Loading…
Optimize sparse 2:4 compression performance
#358 opened Jun 16, 2025 by rahul-tuli Draft
8 tasks done
relax setuptools_scm version requirement
#343 opened Jun 6, 2025 by envolution Loading…
ProTip! Filter pull requests by the default branch with base:main.