Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

https://huggingface.co/google/gemma-3-27b-it-qat-q4_0-gguf/t...

> 17 days ago

Anywaaay...

I'm literally asking, quite honestly, if this is just an 'after the fact' update literally weeks later, that they uploaded a bunch of models, or if there is something more significant about this I'm missing.



Hi! Omar from the Gemma team here.

Last time we only released the quantized GGUFs. Only llama.cpp users could use it (+ Ollama, but without vision).

Now, we released the unquantized checkpoints, so anyone can quantize themselves and use in their favorite tools, including Ollama with vision, MLX, LM Studio, etc. MLX folks also found that the model worked decently with 3 bits compared to naive 3-bit, so by releasing the unquantized checkpoints we allow further experimentation and research.

TL;DR. One was a release in a specific format/tool, we followed-up with a full release of artifacts that enable the community to do much more.


Hey Omar, is there any chance that Gemma 3 might get a speech (ASR/AST/TTS) release?


Probably the former... I see your confusion but it's really only a couple weeks at most. The news cycle is strong in you, grasshopper :)




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: