-
Notifications
You must be signed in to change notification settings - Fork 248
Update pull.yml to test snapshot saving and loading #1486
Conversation
test snapshot saving and loading
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1486
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5d098bd with merge base 083fdaf ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Fixed typos.
cuda-32.json because somebody would rather fail a job than accept a partil group
|
@jerryzh168 @Jack-Khuu can you please have a look what happens with reloading of the Int4 quantized Linear class from torchao? https://hud.pytorch.org/pr/pytorch/torchchat/1486#36825796920 shows this: pull / test-gpu-aoti-bfloat16 (cuda, stories15M) / linux-job (gh) |
Remove fp16 and fp32 int4 quantized models for now. @jerryzh168 Not sure why these dtypes are not compatible with int4 quantization?
|
Thanks for the find, it's using cuda so it should be using the new subclass APIs too hmmm torchchat/torchchat/utils/quantize.py Lines 114 to 117 in 53a1004
|
add DEVICE specification for snapshot and use device cpu
|
@Jack-Khuu updated to run at a minimum the CPU test correctly. Recommend we land that, then put a new PR to enable CUDA and you can assign to the right stakeholder working with torchao? |
|
I was just going to recommend that Sounds like a plan, thanks for following up |
|
thanks, will follow up later, created an issue to track here: pytorch/ao#1727 |
test snapshot saving and loading