Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

MacBook Pro M2 with 64GB of RAM. That's why I tend to be limited to Ollama and MLX - stuff that requires NVIDIA doesn't work for me locally.


MLX is slower than GGUFs on Macs.

On my M1 Max macbook pro, the GGUF version bartowski/google_gemma-3-27b-it-qat-GGUF is 15.6gb and runs at 17tok/sec, whereas mlx-community/gemma-3-27b-it-qat-4bit is 16.8gb and runs at 15tok/sec. Note that both of these are the new QAT 4bit quants.


No, in general mlx versions are always faster, ice tested most of them.


What TPS difference are you getting?


> MacBook Pro M2 with 64GB of RAM

Are there non-mac options with similar capabilities?


Yes, but I don't really know anything about those. https://www.reddit.com/r/LocalLLaMA/ is full of people running models on PCs with NVIDIA cards.

The unique benefit of an Apple Silicon Mac at the moment is that the 64GB of RAM is available to both the GPU and the CPU at once. With other hardware you usually need dedicated separate VRAM for the GPU.


It’s not out yet, but the upcoming Framework desktop [0] is supposed to have a similar unified memory setup.

[0] https://frame.work/desktop


Anything with the Radeon 8060S/Ryzen AI Max+ 395. One of the popular MiniPC Chinese brands has them for preorder[0] with shipping starting May 7th. Framework also has them, but shipping Q3.

0: https://www.gmktec.com/products/prepaid-deposit-amd-ryzen™-a...


I've never been able to get ROCm working reliably personally.


Nvidia Orin AGX if a desktop form factor works for you.


I remember seeing a post about someone running the full size DeepSeek model in a dual-Xeon server with a ton of RAM.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: