You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would like to run quantized gemma-3 for inference (4b,12b and 27b model variants). The project appear to support this (--quant option in run_multimodal script for instance), but the weights do not seem to be available in pytorch format on kaggle:
Could quantized weights be made available that work directly with run_multimodal --quant or I am missing some way to convert the bfloat16 weights?