Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This is just silly. You’re saying that these models are completely incapable of what they’re doing and are only getting to answers from cheating. You can see this isn’t true very quickly when using them.

[Me] I want to make a bouquet to honor the home country of the first person to isolate Molybdenum. Be brief.

[ChatGPT-4] To honor Peter Jacob Hjelm, the Swedish chemist who first isolated Molybdenum in 1781, create a bouquet using blue and yellow flowers, representing the colors of the Swedish flag. Some suggestions are blue irises, yellow tulips, and blue delphiniums, arranged with green foliage for contrast.

I didn’t know the answer to that before asking. I also checked that it’s right. Soemtimes it’s right, sometimes it’s wrong. But if I walk up to a horse and ask it to stamp eight times, that’s going to happen approximately never. If I get it stamping at all, then maybe 1/20. This is much better than that.



Because you are using results of the training.


What human wasn't trained?

Even a child raised by wolves gets training.


A human isn't trained to program in Python by memorizing impressions of millions of examples correlated to a verbal description, from which he or she regurgitates mash-up that statistically correlates to the prompt, without having any idea what the content means.

Ask it to write the same program in some little known programming language. A human who knows how to code could do that by studying a small amount of reference material.


"Ask it to write the same program in some little known programming language. A human who knows how to code could do that by studying a small amount of reference material."

That depends on the language, the human, and their state of mind.


GPT-4 can do that.


LLM doesn't pick some results because LLM is smart, it picks results because it's trained. So analogy with a horse is valid.


You don’t get the horse analogy. The horse isn’t trained to count to eight. The horse is trained to stomp continuously. The performer yells “stop” at eight. It’s a carnival trick and a form of cherry-picking. The commenter was saying that ChatGPT doesn’t have the ability to solve these problems at all and people are just choosing the randomly correct answers. I think this is obviously not true.


The purpose of the horse analogy wasn't so much to equate ChatGPT to a horse that just stomps, but the behavior of the humans around it who believe that it can count. This can be the naive trainer too, who doesn't recognize that he's subconsciously cuing the horse. He has a confirmation bias whereby he or she rejects any evidence refuting the hypothesis that the horse cannot count.

There is a difference between solving and problem and presenting a solution, and that difference doesn't hinge on whether the solution is correct.


ChatGPT is making me realize how completely differently I see the concepts of knowledge and ability from so many people.


Writing programs requires an education. You can sort of fake it with training, if you're a machine that can train on a billion examples and reliably retain something from each one, without understanding any of them.


So your claim is that GPT-4 is a stochastic parrot?

Meanwhile, the creators at OpenAI claims that it is not.

How do we test your claim? What are the logical reasoning questions we can ask GPT-4 that it shouldn't be able to answer if it's just a stochastic parrot?

I keep asking people this, but so far no one has answered. Is the claim a personal belief that you don't want to discuss, similar to your religious beliefs?

If so, fine, but please make it clear that you're not making a scientific claim.


GPT-4 looks like a much more powerful system which can produce answers in ways that can't easily be explained/dismissed as "stochastic parroting".




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: