Hacker News new | past | comments | ask | show | jobs | submit login

Gemma 3 is out! Multimodal (image + text), 128K context, supports 140+ languages, and comes in 1B, 4B, 12B, and 27B sizes with open weights & commercial use.

Gemma 3 model overview: https://ai.google.dev/gemma/docs/core

Huggingface collection: https://huggingface.co/collections/google/gemma-3-release-67...

ollama: https://ollama.com/library/gemma3




A kind of ancillary note, but it's amazing to me how fragmented this presentation and documentation is:

* the parent link is to storage.googleapis.com

* There's documentation on ai.google.dev

* The announcement blogpost is https://blog.google/technology/developers/gemma-3/

* you try it on https://aistudio.google.com/

It's helpful to have a top-level post like this, but can some PM please consolidate this into, IDK, ai.google.com/gemini?


Apparently ai.google.com currently redirects to ai.google, which is different from ai.google.dev where the Gemini stuff actually is.



I don't see how this actually matters - who cares if it it's different top level domains?


Two reasons it matters:

1) Discoverability

2) "System structure mirrors organization". I.E., it's an indicator of a fragmented and disorganized structure that's not likely to produce cohesive product results.


> System structure mirrors organization

You listed:

- one static pdf file stored on a CDN

- one company blog static website

- one developer documentation static website

- one interactive product URL

As much as I like to dunk on how messy things can be at Google I don't think this is a really good example. Apart from small startups I would be scared if you served all of them from the same base host.


The many domains is a problem because it suggests a many-teams approach to product development, and the more cooks in the kitchen, the more likely a repeat of Gemini 1’s rollout, which was a mess [0]. Basically I’m looking to see that Google cares about the meta-level user experience of finding, understanding, and using its products, and scattering key usage details around the internet is not a good sign. It suggests deeper process problems if a simple issue like this either didn’t get noticed or can’t get fixed.

[0] https://www.theguardian.com/technology/2024/mar/08/we-defini...


> "System structure mirrors organization"

Conway's Law is the general term for this concept https://en.wikipedia.org/wiki/Conway%27s_law


The ollama page shows Gemma 27B beating Deepseek v3 and o3-mini on lmarena. I'm very excited to try it out.


Same!


Doesn't yet work in LM Studio. Barfs an error when trying to load the model. (Error 6, whatever that means. Happy I missed the first 5.)


You need the newest llama.cpp and if you have an amd card and recently updated the drivers, roll them back. Most people complaining are using ROCm.

I assure you gemma 3 works fine in LM studio. Gguf and MLx are available.


> Barfs an error when trying to load the model

Since you're not using the official models (since they're not GGUFs), what exact model are you trying to use? The 3rd party you rely on might have screwed something up.


Please make sure to update to the latest llama.cpp version


> ollama: https://ollama.com/library/gemma3

Needs an ollama newer than 0.5.11. Probably the very-recently-released v0.6.0[1]:

> New Model:

> * Gemma 3: Google Gemma 3 model is now available in 1B, 4B, 12B, and 27B parameter sizes.

[1]: https://github.com/ollama/ollama/releases/tag/v0.6.0


Doesn't work on 0.5.13. Had to upgrade to 0.6.0.


> open weights

What exactly is this supposed to mean? That I can grab the weights by just downloading them, or something like that?

Because when I open up the HuggingFace repository, it asks me to "accept the conditions" (Google’s usage license). How is this different from any other proprietary binaries people distribute on the internet but let you run locally? Are other software (like 1Password for example) also "open software" because you can download it?


Replace "google" with "unsloth" in the browser address bar if you want to download them without signing up to hf


Regardless of where you get the weights, Google says you need to follow their terms and conditions for the model/weights:

> By using, reproducing, modifying, distributing, performing or displaying any portion or element of Gemma, Model Derivatives including via any Hosted Service, (each as defined below) (collectively, the "Gemma Services") or otherwise accepting the terms of this Agreement, you agree to be bound by this Agreement.

https://ai.google.dev/gemma/terms

Worth knowing if you're planning to use this model for production usage/with a business.

So once again, I don't understand what "open" is supposed to mean when they call models like these "open weights". What part exactly is "open"?


I don't disagree but even Linux has "Terms and conditions" of usage under it's license you really need to dig into what those are.

There's no doubt Gemma's license is less permissive than other models and that it has less community finetuners for that reason.


According to the OSI's open source definition, you can't put restrictions against persons or groups or fields of use. In the license, Linux is not restricted in what domain it will be used (good or bad).

Here's OSI's argument about this when Meta's llama put such limitations in their license: https://opensource.org/blog/metas-llama-2-license-is-not-ope...


can you link to Linux terms and conditions? search returned nothing.


I guess my comment was a bit wrong, Linux has "TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION" not usage.



i think generally these companies are too afraid of the obvious rejoinder to try actually enforcing these terms


Probably, up until they aren't. Are you willing to bet against Google's lawyers feeling daring in the future? As a private individual, I sure aren't, and I don't think I'd bet my (hypothetical) business on it either.


"Open weights" refers to a license that allows you to freely (or mostly freely) copy the model file (i.e. weights). An "open source" model would be possible to build from training data, but those hardly exist.


I'm still a huge fan of gemma-22b. Looking forward to this one!




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: