You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[AWQ] Support for Calibration Datasets of varying feature dimension (#1536)
SUMMARY:
AWQModifier currently expects all batches to have the same feature
dimension. Users are reporting this causes errors for vision-language
datasets. This PR adds support to AWQModifier for calibration dataset
batches with varying feature dimension. Rather than concatentating all
outputs to a single torch tensor, the lists are retained and passed
through one by one to compute loss. This removes the need for chunk
memory configuration, and the logic around that calculation.
* Resolves#1524
TEST PLAN:
- [x] Re-ran for `"meta-llama/Llama-3.2-3B-Instruct"`, wikitext PPL of
13.30 is better than the 14.08 previously achieved, because the dataset
is slightly different now.
- [x] Also confirmed [user-provided code in
#1524](#1524 (comment))
can be run with smaller `"Qwen/Qwen2.5-VL-7B-Instruct"` model up until
it tries to access jpg file on user's local machine.
---------
Signed-off-by: Brian Dellabetta <[email protected]>
0 commit comments