Anything with a linter means, at minimum, free verifiable rewards for RL (though whether something parses versus looks good is another story). That, plus, they have more data than anyone, and also it seems somewhat reasonable that stronger models could learn 'more' from a given instance or set of examples.
In this case, what the research says is that the approximations we have already been using for a long time are correct. "You're already right, keep doing what you're doing!" is not generally something people consider a "practical application".
I’m imagining a timeline where dolphins have accepted long ago that humans are intelligent because of our external manifestations of technology, and consequentially have developed much more sophisticated philosophy than humans based on having to understand paradigms well outside of their personal experiences.
Soon, we will find out through dolphin-Gemma that they have been talking to aliens for centuries, since the aliens tried to talk to us but failed due to our, and their, philosophical myopia… but the dolphins, with their non-circular philosophical understanding recognized alien communications and started the conversation before we finished the pyramids.
This seems like the type of work that, whether or not it itself is profound (which I believe it seems it is), will accelerate the forthbringing of derivative works that will definitely be profound.
Yeah, I had the same uncanny valley feeling when I read this phrase: "carrying centuries of Korean culture in every jar". What next: Recommended that we "delve" into the history of soy sauce production?
Nah, not a bot :) Maybe I leaned a little poetic there, but I do think there's something powerful about how much history and identity can be packed into something as humble as soy sauce...
I think this is really about the actively made decision to spend longer with each sensation than you 'normally' would - relative to your own individual baseline. Basically whether this is a tool in the person's metacognitive toolkit that they typically use or not. Like 'the feeling of the touch of a thought' on the mind itself, as opposed to a linear pre-simulation of a performance to play out loud. Or the difference between making what you say out loud a cheap, low entropy, lossy encoding of all that you're thinking, versus gathering all the aspects quietly and needing to find a way linearize all that incidental complexity.
If you want a hard problem that's far from predictable with tangible utility, why don't you work on something that helps humans with human problems, but whee you can still use your talents? Some research that could help understand a disease, or a new kind of medical technology, or something like that. I don't know how well it aligns with your salary goals but it might be meaningful.