Hacker News new | past | comments | ask | show | jobs | submit login

>You cannot train any GPT-2 models with an AMD GPU.

It seems like you can. I know of at least one person who has finetunned 1.5b on a 16GB AMD. I think u/sillysaurusx had some part in it, but apparently translating the code from CUDA was fairly easy.




There are also several people on Twitter who have mentioned training it on AMD GPUs.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: