- 
                Notifications
    You must be signed in to change notification settings 
- Fork 248
Update pull.yml to test snapshot saving and loading #1486
Conversation
test snapshot saving and loading
| 🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1486
 Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5d098bd with merge base 083fdaf ( This comment was automatically generated by Dr. CI and updates every 15 minutes. | 
Fixed typos.
cuda-32.json because somebody would rather fail a job than accept a partil group
| @jerryzh168 @Jack-Khuu can you please have a look what happens with reloading of the Int4 quantized Linear class from torchao? https://hud.pytorch.org/pr/pytorch/torchchat/1486#36825796920 shows this: pull / test-gpu-aoti-bfloat16 (cuda, stories15M) / linux-job (gh) | 
Remove fp16 and fp32 int4 quantized models for now. @jerryzh168 Not sure why these dtypes are not compatible with int4 quantization?
| Thanks for the find, it's using cuda so it should be using the new subclass APIs too hmmm torchchat/torchchat/utils/quantize.py Lines 114 to 117 in 53a1004 
 | 
add DEVICE specification for snapshot and use device cpu
| @Jack-Khuu updated to run at a minimum the CPU test correctly. Recommend we land that, then put a new PR to enable CUDA and you can assign to the right stakeholder working with torchao? | 
| I was just going to recommend that Sounds like a plan, thanks for following up | 
| thanks, will follow up later, created an issue to track here: pytorch/ao#1727 | 
test snapshot saving and loading