It's not mindless brute-forcing, the details of the architecture, data, and training strategy still matter a lot (if you gave a modern datacenter to an AI researcher from the 60s they wouldn't get an LLM very quickly). The bitter lesson is that you should focus on adjusting your techniques so that they can take advantage of processing power to learn more about your problem themselves, instead of trying to hand-craft half the solution yourself to 'help' the part that's learning.
I found this article a little weak, but there is an interesting parallel.
The 10,000 hours thing is encouraging because the amount of effort you put in as far more important than your natural ability.
... Until you get to the point where everyone is already working as hard as humanly possible, at which point natural ability becomes the sorting function again.
They have researchers working for insane salaries just so they don't go to another frontier lab to share their ideas. If you think it is just "mindless bruteforce" you don't understand anything. The idea is that the most effective methods are ones that scale but those ideas are also then limited by the compute available.
https://en.wikipedia.org/wiki/Bitter_lesson