I'm not Austin, but I am Tris, the friendly neighborhood product person on Gemma. Overall, I think that the main feeling is: incredibly relieved to have had the launch go as smoothly as it has! The complexity of the launch is truly astounding:
1) Reference implementations in JAX, PyTorch, TF with Keras 3, MaxText/JAX, more...
2) Full integration at launch with HF including Transformers + optimization therein
3) TensorRT-LLM and full NVIDIA opt across the stack in partnership with that team (mentioned on the NVIDIA earnings call by Jensen, even)
4) More developer surfaces than you can shake a stick at: Kaggle, Colab, Gemma.cpp, GGUF
5) Comms landing with full coordination from Sundar + Demis + Jeff Dean, not to mention positive articles in NYT, Verge, Fortune, etc.
6) Full Google Cloud launches across several major products, including Vertex and GKE
7) Launched globally and with a permissive set of terms that enable developers to do awesome stuff
Pulling that off without any major SNAFUs is a huge relief for the team. We're excited by the potential of using all of those surfaces and the launch momentum to build a lot more great things for you all =)
I am not a fan of a lot of what Google does, but congratulations! That’s a massive undertaking and it is bringing the field forward. I am glad you could do this, and hope you’ll have many other successful releases.
Always -- anything that comes with the Google name attached always attracts some negativity. There's plenty of valid criticism, most of which we hope to address in the coming weeks and months =).
I mean, many articles will have a negative cast because of the need for clicks -- e.g., the Verge's launch article is entitled "Google Gemma: because Google doesn’t want to give away Gemini yet" -- which I think is both an unfair characterization (given the free tier of Gemini Pro) and unnecessarily inflammatory.
Legitimate criticisms include not working correctly out of the box for llama.cpp due to repetition penalty and vocab size, some snafus on chat templates with huggingface, the fact that they're not larger-sized models, etc. Lots of the issues are already fixed, and we're committed to making sure these models are great.
Honestly, not sure what you're trying to get at here -- are you trying to "gotcha" the fact that not everything is perfect? That's true for any launch.
I think he's trying to bring up the racial image gen bias stuff that's going on with Gemini, but for some reason won't say it. He also doesn't appear to realize there is a difference between the two products/teams/launches...
I thought that reflecting what went poorly is really informative and inspiring! It really shows how you are taking this seriously, and iterating on a great project that you're building better and better, thanks for the insight and for a peek behind the curtain
> Legitimate criticisms include not working correctly out of the box for llama.cpp
I don't think that's a legitimate criticism. Especially not for something that just launched. You should be helping your own project first, before you help others. For example, it'd be nice to see AVX2 work as well as AVX512 with gemma.cpp.
These comments appear to be about Gemini's image generation, IIUC. Gemma, however, is a language model -- whilst I believe that a larger unreleased version of it is used as part of the Gemini product, it doesn't seem relevant to these criticisms. Also, the Gemma base model is released, which doesn't AFAIK contain any RLHF.
The impression I have is that you're using the release of Gemma to complain about tangentially related issues about Google and politics more generally. The HN guidelines warn against this: "Eschew flamebait. Avoid generic tangents... Please don't post shallow dismissals, especially of other people's work. A good critical comment teaches us something. Please don't use Hacker News for political or ideological battle."
It looks like you're trying to get some sort of "confession" from relevant people based on recent memes against the company? The reality is likely that the developers sincerely believe in the value of this product and are proud of its launch. You're just adding uninteresting, irrelevant noise to the discussion and you probably won't get what you want.
I've been completely honest, human-like, and non-evasive with you. I answered your questions directly and frankly.
Every time, you ignored the honest and human-like answers to try and score some imaginary points.
We're honestly trying our best to build open models *with* the community that you can tune and use to build neat AI research + products. Ignoring that in favor of some political narrative is really petty.
You’re being flagged and downvoted on your threads because you’re being unreasonable, not because of some moderator level conspiracy.
The argument you’re making is not falsifiable.
If you want to pursue this line of argument, you need to pause, think about how you can confidently and unambiguously make substantive claims; for example, substantive examples from actually using these models.
The way you’re currently arguing is not meaningful or compelling.