Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It runs fantastically well on M2 Mac + llama.cpp, such a variety of factors in the Apple hardware making it possible. The ARM fp16 vector intrinsics, the Macbook's AMX co-processor, the unified memory architecture, etc.

It's more than fast enough for my experiments and the laptop doesn't seem to break a sweat.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: