ChatGPT lists clickable sources in a lot of nontrivial queries. Those sites don’t even need to pay OpenAI for the traffic (yet). If you ask „what’s happening in the world today“, you might get 20 links. How is this worse, exactly?
So either even higher costs and hope that a bug problem of LLMs get solved somehow.
Given how much data they need that will be pretty expensive, I mean really really expensive.
How many people can write good training data and how much per day?