Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Qwen3.5 35B A3B is much much faster and fits if you get a 3 bit version. How fast are you getting 27B to run?

On my M3 Air w/ 24GB of memory 27B is 2 tok/s but 35B A3B is 14-22 tok/s which is actually usable.



Using ik_llama.cpp to run a 27B 4bpw quant on a RTX 3090, I get 1312 tok/s PP and 40.7 tok/s TG at zero context, dropping to 1009 tok/s PP and 36.2 tok/s TG at 40960 context.

35B A3B is faster but didn't do too well in my limited testing.


with regular llama.cpp on a 3070ti I get 60tok/s TG with the 9B model, it's quite impressive.


The 27B is rated slightly higher for SWE-bench.


Don't sleep on the 9B version either, I get much faster speeds and can't tell any difference in quality. On my 3070ti I get ~60tok/s with it, and half that with the 35B-A3B.


27B needs less memory and does better on benchmarks, but 35B-A3B seems to run roughly twice as fast.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: