Hacker News new | past | comments | ask | show | jobs | submit login

But it took the deepseek team a few weeks to replicate something at least close to o1.

If people can replicate 90% of your product in 6 weeks you have competition.




Not only a few weeks, but more importantly, it was cheap.

The moat for these big models were always expected to be capital expenditure for training costing billions. It's why these companies like openAI etc, are spending massively on compute - it's building a bigger moat (or trying to at least).

If it can be shown, which seems to have been, that you could use smarts and make use of compute more efficiently and cheaply, but achieve similar (or even better) results, the hardware moat bouyed by capital is no longer.

i'm actually glad tho. An opensourced version of these weights should ideally spur the type of innovation that stable diffusion did when theirs was released.


o1-preview was released Sep 12, 2024. So DeepSeek team probably had a couple of months.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: