Do you have plans to do a follow-up model release with quantization aware training as was done for Gemma 3?
https://developers.googleblog.com/en/gemma-3-quantized-aware...
Having 4 bit QAT versions of the larger models would be great for people who only have 16 or 24 GB of VRAM.