Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
dweekly
1 day ago
|
parent
|
context
|
favorite
| on:
Speed, scale and reliability: 25 years of Google d...
For TPU pods they use 3D torus topology with multi-terabit cross connects. For GPU, A3 Ultra instances offer "non-blocking 3.2 Tbps per server of GPU-to-GPU traffic over RoCE".
Is that the worst for training? Namely: do superior solutions exist?
Consider applying for YC's W25 batch! Applications are open till Nov 12.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search:
Is that the worst for training? Namely: do superior solutions exist?