Hacker News new | past | comments | ask | show | jobs | submit login

So assuming you utilized this with (4) x 4090s is there a theoretical comparative to performance vs the A6000 / other professional lines?



It depends on what you do with it and how much bandwidth it needs between the cards. For LLM inference with tensor parallelism (usually limited by VRAM read bandwidth, but little exchange needed) 2x 4090 will massively outperform a single A6000. For training, not so much.


I believe this is mostly for memory capacities. PCIe access between GPUs is slower than soldered RAM on a single GPU




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: