ZeroGPU-LLM-Inference / quantize_to_awq_colab.ipynb
Alikestocode's picture
Lower Gemma AWQ group size to 16
f8c20fd
Open in Colab