Hacker News new | past | comments | ask | show | jobs | submit login

> The original CLIP was trained on 500x A100 Nvidia GPUs. The latest Open_CLIP trained on 1024x GPUs.

> We trained on the setup of 3x workstations, with 4x RTX 3090 consumer-grade GPUs in each, connected over 200 GBit InfiniBand HDR.

ok so 85x improvement on the GPU count (i suspect even better once you take into account the differences in consumer grade GPU) but i must still be missing something - where does it say it uses 100x less data?




Look at the “dataset” column: CLIP was trained on 400m images, UForm on 4m.


There are also dataset sizes for Albef and ViCHA.




Consider applying for YC's W25 batch! Applications are open till Nov 12.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: