Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It 100% is still intelligence. GPT-5 with Thinking still can't win at tic-tac-toe.


What if it's the desired outcome? Become more human-like (i.e. dumb) to make us feel better about ourselves? NI beats AI again!


> What if it's the desired outcome?

To be able to reason about the rules of a game so trivial that it has been solved for ages, so that it can figure out enough strategy to never not bring the game to a draw (if played against one who is playing to not lose), or a win (if played against someone who is leaving the bot an opening to win), as mentioned in [0] and probably a squillion other places?

Duh?

[0] <https://news.ycombinator.com/item?id=44919138>


Speaking of human-level capabilities, it looks like I totally failed to correctly read the section of your comment that I quoted. Shame on me.

However, I'd expect that "Appearing to fail to reason well enough to know how to always fail to lose, and -if the opportunity presents itself- win at one of the simplest games there is." is absolutely not a desired outcome for OpenAI, or any other company that's burning billions of dollars producing LLMs.

If their robot was currently reliably capable of adequate performance at Tic Tac Toe, it absolutely would be exhibiting that behavior.


Tic-tac-toe is solved and a draw can be forced 100% of the time...


That's exactly why it's so crazy that GPT-5 with Thinking still loses...


Ah, your first comment said "can't win". Which is different than "always loses".


Ah okay, well it will still lose some of the time, which is surprising. And it will lose in surprising way, e.g., thinking for 14 seconds and then making an extremely basic mistake like not seeing it already have two on a row and could just win.


.. and you can "program" a neural network — so simple it can be implemented by boxes full of marbles and simple rules about how to interact with the boxes — to learn by playing tictactoe until it always plays perfect games. This is frequently chosen as a lesson in how neural network training even works.

But I have a different challenge for you: train a human to play tictactoe, but never allow them to see the game visually, even in examples. You have to train them to play only by spoken words.

Point being that tictactoe is a visual game and when you're only teaching a model to learn from the vast sea of stream-of-tokens (similar to stream-of-phonemes) language, visual games like this aren't going to be well covered in the training set, nor is it going to be easy to generalize to playing them.


tic-tac toe is merely a visualization of a small arithmetic game "sum 3 digits to 15"

   618   
   753
   294

Well whatever your story is, I know with near certainty that no amount of scaffolding is going to get you from an LLM that can't figure out tic-tac-toe (but will confidently make bad moves) to something that can replace a human in an economically important job.


llm maximalists' apologies:

- but tokens are not letters - but humans fail too - just wait, we are on an S curve to AGI - but your prompt was incorrect - but I tried and here it works

Meanwhile, their claims:

- LLMs are performing at PhD levels. - AGI is around the corner - humanity will be wiped out - situational awareness report




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: