It runs fantastically well on M2 Mac + llama.cpp, such a variety of factors in the Apple hardware making it possible. The ARM fp16 vector intrinsics, the Macbook's AMX co-processor, the unified memory architecture, etc.
It's more than fast enough for my experiments and the laptop doesn't seem to break a sweat.
It's more than fast enough for my experiments and the laptop doesn't seem to break a sweat.