Hacker Newsnew | past | comments | ask | show | jobs | submit | GaggiX's commentslogin

I managed to run Balatro on a Trimui Brick on a very minimalist Linux distro, steamos should be easy.

It's now on Portmaster if anyone is curious -

https://portmaster.games/detail.html?name=balatro


That's rad!

Getting it to run in KDE was straightforward. I've gotta figure out why it instacrashes in SteamOS game mode.


From what I remember Glaze is using some small CLIP model and LPIPS (based on VGG) for their adversarial loss, that's why it's so ineffective to large, better trained model.

It use SD to do a style transfer on the image using image-to-image, then it use gradient descent on the image itself to lower the difference between CLIP embeddings of the original and style transfer image + trying to maintain LPIPS, then every step is normalized to not exceed a certain threshold from the original image.

So essentially it's an adversarial attack against a small CLIP model, even though today's models are much robust than that.


The people talking about semantics in the comment section seems to completely ignore the positive correlation of LLMs between accuracy and stated confidence, this is called calibration and this "old" blog post from a year ago already showed it, LLMs can know what they know: https://openai.com/index/introducing-simpleqa/

Opus: "an artistic work, especially one on a large scale."

The names Haiku, Sonnet, and Opus have not been chosen randomly.


And so much more intuitive than the OpenAI names for their models. I still don't get their naming scheme.

I never saw the cursor changing size to fit the button you are hovering on, it's pretty cool, I don't know if it's better but it's cool.

Looks cool. Feels horrible. I don't want my mouse cursor to morph into buttons or anything. I don't want it to be a lagging blob either.

iPadOS had it for about 5 years.

≥In general encoder+decoder models are much more efficient at infererence than decoder-only models because they run over the entire input all at once (which leverages parallel compute more effectively).

Decoder-only models also do this, the only difference is that they use a masked attention.


>or some sort of cracked way to pack a ton of parametric knowledge into a Flash Model.

More experts with a lower pertentage of active ones -> more sparsity.


They went too far, now the Flash model is competing with their Pro version. Better SWE-bench, better ARC-AGI 2 than 3.0 Pro. I imagine they are going to improve 3.0 Pro before it's no more in Preview.

Also I don't see it written in the blog post but Flash supports more granular settings for reasoning: minimal, low, medium, high (like openai models), while pro is only low and high.


"minimal" is a bit weird.

> Matches the “no thinking” setting for most queries. The model may think very minimally for complex coding tasks. Minimizes latency for chat or high throughput applications.

I'd prefer a hard "no thinking" rule than what this is.


It still supports the legacy mode of setting the budget, you can set it to 0 and it would be equivalent to none reasoning effort like gpt 5.1/5.2

I can confirm this is the case via the API, but annoyingly AI Studio doesn't let you do so.

> They went too far, now the Flash model is competing with their Pro version

Wasn't this the case with the 2.5 Flash models too? I remember being very confused at that time.


This is similar to how Anthropic has treated sonnet/opus as well. At least pre opus 4.5.

To me it seems like the big model has been "look what we can do", and the smaller model is "actually use this one though".


I'm not sure how I'm going to live with this!

Japan has an extremely high literacy rate.

Also Anki is not ugly at all in general, the interface of Hashcards looks much uglier to me.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: