Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> Llammas are showing the world that it doesn't take monopoly-level hardware to run those things.

LLaMA was not necessarily the model that did that. A fairer attribution might be BERT or GPT-Neo.



it was difficult to run all those models. now gamers follow youtube tutorials


Name 1 way GPT-Neo was harder to run than LLaMA.


In my experience GPT-Neo never produced any useful output.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: