Please don't post links to (live) Google searches. It will absolutely give different results depending on the entire state of every computation everywhere, simultaneously. In two weeks, this will return who knows what.
One of the things that bothers me about modern Google search results is that I can't be sure the other person I'm telling to google something will even see the same results as me. Results change over time, sure, but google can't help but try and tailor results.
I'm similarly bothered by rolling updates of any software. If something changes, there's no version number to see what documentation is applicable. There's no way to test on a different computer to see if it's my local environment. There's no way to roll back to a previous version to see if one of them broke.
I understand the advantages of A/B testing from a development side, but it makes for a miserable user experience.
Not even different person, Even if you search a query multiple times in a row it changes results drastically. I got a phished website of my Govt's tax website as a top result due to this[1].
W.r.t OP, I've been contemplating a search engine wall of shame[2] for a while now for recording poor search results and perhaps provide actionable feedback to the search engines.
Oh right, I always wondered how GitHub/SO scrappers make it to the top results, I thought plagiarism was SEO Don'ts 101? Maybe Google has some whitelist for programming websites due to parsing difficulties and meta nature of it.
Anyways you are welcomed to submit your entries to the above search engine wall of shame thread(It's a validation platform I created), May be someone would see the value in it and create a subreddit or separate website for tracking search queries.
I don't think it matters with knowledge cards - the system is so fuzzy that it'll throw up that answer to almost anything.
Apparently Louis Armstrong, the legendary jazz musician, was there with Neil Armstrong on the Moon in July 1969 as the oldest astronaut in history, a few years before his death at the age of 69. [1]
For future visitors after Google fixes the bug: at this time, Google gave the same response that you would expect if the query mentioned Earth’s moon instead of Mars. This is also probably true if you ask Google the same question as an “Okay, Google” search. It’s a big info box that says when the moon landing was, shows a photo of an astronaut on the moon, and so on.
Pretty sure you meant the moon. I would guess he first stepped on Earth closer to his birth date (although I'm guessing his first steps were a few months after ;))
Regardless of what it ought to say, I see the same effect with the destinations “earth”, “saturn”, “america”, “russia”, “newton”, “silver”, and “lego”, though the date varies from the 16th to the 20th.
I'm not necessarily asking you, but I'm taking your message as an opportunity to ask if anyone knows of an extension that does the opposite?
I have pretty bad astigmatism, and I cannot read pages with dark background. There used to be an extension I used for this purpose, but it stopped working several years ago, and I have suffered ever since. Going into developer mode to play around with the stylesheet works, but is a huge hassle.
Seems silly - not quite running, but consider that Europeans forgot how to swim efficiently and really only had the doggy paddle until they observed native Americans doing the front crawl. Basically the swimming equivalent of knowing how to jog, but not how to run or sprint.
The linked article doesn't say anything about native Americans, and in terms of substantive data it seems to jump in history from the 4th century prevalence of Roman bathhouses, to 16th century local bans on swimming, filling in the gaps (which account for the entire premise of the article) with speculation about bridges and mermaids.
It immediately follows the description of those swimming bans with other 16th century examples of Tudor scholars endorsing and writing guides on swimming, resulting in an article that's an interesting read but with a poorly supported thesis.
It would be more charitable to say that current ML/NLP approaches have orders of magnitude less understanding than humans. The ML/NLP process just jumps straight to the correlated answer, which a human would do, but then use higher order reasoning - "wait mars? that wasn't where he set foot".
This to me is similar to current autonomous driving limitations. The cars can only respond to situations they have seen before. Any novel elements lead to failure, where a human can fall back on higher order reasoning: "That is a stoplight yes, but it is on the back of a truck, ignore it" (real example)
I mean that there is no semantic comprehension or knowledge of 'Neil Armstrong' (as a specific person), 'land' (as an action), and 'Mars' (as a place). It's not a mishearing because computers don't mishear ASCII-inputs like we might do with auditory signal.
I get the same result when I Google 'when did Lance Armstrong land on the moon', 'when did Buzz Lightyear land on the moon', or 'when did Lance Armstrong land on Mars.'
If you ask the average American “When did Niel Gaiman set foot on the moon?” Most would answer that they don’t know exactly but think it was in the 60’s.
This is not a limitation of AI, it’s exactly what you want it to do. It’s reading into the context of the question and finding it more likely that you made a mistake in your question than seriously want an answer for a constructed nonsensical question that has no frame of reference or context in our common knowledge pool.
If you want exact logical answers deduced from base prepositions you don’t want ML models or “AI” your looking at formal logic and deduction.
> Most would answer that they don’t know exactly but think it was in the 60’s.
I think this is an important point. Humans and Google have a similar bias (for different reasons) that they want to be helpful or seem knowledgeable when asked a question, so won't say "I am confused by that question" and will rather give an answer to a question which is similar to a question that they can understand, in the hopes that it is approximately right.
In the case of asking when Neil Armstrong landed on Mars, guessing "I think it was in the 60's" is accurate to within 10 years but off by 56 million km. For a good example of what average people think about the solar system, though, consider the question "Is the moon really a planet?" that was once part of an impromptu debate on TV:
I dont think comparing yourself to an average idiot on the street is what google should strive for.
The way it used to work you had one intelligent entity forming a query and a machine performing said query. Somewhere along the line Google got this idea returning no result was a bad outcome and started overriding queries. Its like you go to the hardware store and ask for nails, but shopkeeper starts telling you all about nail saloons in the vicinity because they ran out of finishing nails - or worse, goes nail saloon -> spa -> finishing -> ending -> happy ending.
We went from intelligent entity forming a query and a machine performing said query to idiot computer treating you line an idiot.
In a school where I studied, something similar was used as a trick question during a history exam. "Which language Vladimir Lenin used to write correspondence addressed to Karl Marx?" or something like that. Nearly half of the class failed on this. To those unaware: Lenin discovered Marx's book, Capital, in 1887, while Marx died in 1883, so there could not be any correspondence.
I think the conditional display of the fact box implies they are in fact asking whether there is a good (enough) answer. They are just getting it wrong here.
But at the same time, I think it's doing a really good job of what it's trying to do. Google search is not trying to be a repository for all the world's information. It's just trying to get people to what they're
most likely looking for or show the most related things. Given the significance of the moon landing and the fact that no one has set foot on mars I find it unsurprising that it brings up info on the moon landing. It's seems better to assume what the user is likely looking for especially when (at least my) Google searches often take the form of "moon land neil year". I can just type things like that out, stream of consciousness, and the majority of the time I get what I'm looking for immediately.
The first is that Google has specifically chosen to call out an answer in some kind. If the query is reasonably framed as a question, there is a clear indication in the UI that the response is meant to be an answer to that question.
Now it's definitely the case that a lot of questions have some amount of semantic ambiguity that a listener would have to resolve. For example, a question about a "president" can reasonably be inferred to mean specifically a "US president" of some kind, at least if the query is from the US and is in English.
And sometimes people can ask questions where there's a confused detail. And responding with the question they probably meant to ask is not unreasonable.
However--and this is a big however--it is incumbent to emphasize that the answer is for a different question than the one that was literally asked. You see this when you do searches of misspelled terms: "did you mean this one instead?" Because occasionally, no, you did mean the term that has much fewer results.
And this kind of emphasize-the-answer can have poor results sometimes. Ask Google which president became supreme dictator. The answer makes it clear why it thinks that, but... that's a really different question from the one that was asked.
If someone asked you in-person "when did Neil Armstrong set foot on Mars?", would you just say "July 20, 1969"? Or would you say "nobody has been to Mars, but if you're talking about the moon..."
Google's response here only makes sense if Google said "Did you mean: When did Neil Armstrong land on the moon?"
> isn't really any level of 'comprehension' at all.
I mean, that's a bit harsh. I bet there are lots of people who would answer the question the exact same way. They don't have a total lack of comprehension, they perhaps misheard a word or misremembered a fact they once knew. Honestly, while the Google answer is wrong and this demonstrates a major flaw in their confidence of answering queries, the level of comprehension is still quite impressive (to me, at least)
While I agree that "no comprehension at all" seems harsh, I don't think anyone would jumble things up between moon and mars unless they got the fact wrong in the first place.
These knowledge cards are pretty useful, but they shouldn't be taken as the source of fact (at least right now) unless one opens the link to check where that card is extracted from.
Yes, but a human who don't understand a word lacks the vocabulary for that specific language and that's the reason they cannot understand or comprehend the question properly. Whereas Google Search does have the vocabulary built in with its data set. The problem here is it didn't piece the words together correctly even though the question is valid.
It is one thing to not understand a word in the first place and another if you cannot piece it together i.e. assuming you know the language.
I agree, but it is a type of failure of comprehension (Google Search doesn't understand the word "poop" either, it's just aware that it crops up in a lot of contexts and can rote-recall multiple definitions for it). Same with the human giving a date for Nixon landing on the moon because they understand moon landings but have no comprehension of who Nixon was (much like Google, which additionally has the unhelpful correlation between mentions of Nixon and mentions of moon landings to factor in)
But in human's case, the lack of understanding of word Nixon is because they didn't know what that means in the first place and they don't have data to know or cross check what that word means. On the other hand, Google Search does have an access to the data (just strings in this case from an article) which it has to verify based on the user's query and then display it as the knowledge card and yet it couldn't because... (I don't know why exactly).
That isn't the question of your comprehension though. You got the input (question) wrong because of your habit and not because you couldn't comprehend the sentence structure. You misread the sentence and that's why you got confused over what is wrong if I get it properly. You most probably wouldn't have done this if you read it as Mars in the first place.
Google helpfully responds "16 bits", which is pulled from the History section of Wikipedia and hasn't been accurate in something like twenty-five years.
Edit: Should have listened to people saying to screenshot your queries. Google still quotes the paragraph in question, and bolds "16 bits", but no longer puts it in a big bold heading like it's the single answer to your question.
Double Edit: except in chrome, where I do still get the old page. Here's a screenshot for posterity, after Google somehow fixes this: https://i.imgur.com/7Ng6DyK.png
I'd imagine Google's position is that no one ever searches for this in earnest, and so it doesn't really matter what the answer given is. They want to maximize what percentage of actual searches give a correct answer, rather than what percentage of possible questions do.
On the other hand, it's a dangerous game. You never know when current events might make a previously never-searched question with a wrong answer very popular.
Yes, I'd imagine the number of people searching for the year Neil Armstrong went to Mars because they for some reason are extremely deluded about history is much, much smaller than the number of people who type "mars" when they meant "moon"
Problem is Google does this when I search for a very specific, weird error message too, in which case I actually mean that specific weird error message.
I've lost a few hours to Google trying to outsmart me and it is maddening: not only do I waste hours but it is not given that I find the magic incantation that gives me the answer.
DuckDuckGo.com is getting better lately but is not stable. Lately I've also started using metager.org which has seemed promising so far.
I wish Google had some way to set parameters on this sort of thing, because I often find that the default "try to assume that the user meant the most common thing" is useful, but I also often find it bothersome. And if Google isn't willing to give us a more robust customizable search, then another player should.
They have: there is a "verbatim option" and also they have never removed doublequotes from their documentation or told us it has been discontinued or put to use for something else like the plus operator (which IIRC meant "make sure this particular word is included").
They just ignore both verbatim option and the doublequote operator which is why I conclude they (as an organization, not as individuals) are incompetent and/or arrogant.
AI will be the next great source fake news, fake stuff. This one was easy to flag as bogus. Problem is that for other topics it will give average user the impression that this is actully a fact. To make it worse Google UI is massively misleading, big bold text, makes it look as legit.
Saw this yesterday, my Google search results (regular and image) are basically the same. A bunch of the top results for "desk ornament" are nazi memorabilia for some reason:
As does "when did Louis Armstrong set foot on the moon" and "when did Richard Nixon set foot on the moon"?
JFK, on the other hand, apparently "set foot" on the moon on May 25 1961 and "land[ed]" on the moon on September 12 1962, and Arthur C Clarke some time in 1968
Asked it when Keith landed on the Moon to see how it juggled those inputs, and to be fair it gave me a beautiful excerpt about an Apollo engineer called Keith who worked on some of the scientific equipment left there!
"What year did Tom Hanks land on the moon?" "1970"
"When did the Simpsons learn to control fire?" "Dec 17, 1989"
"Which river did George Washington drown in?" "The Allegheny River"
It's got to be tough to train an AI that can generalize to (or at least recognize) questions outside its training set. Unfortunately, when 99% of questions containing 'when' and 'Neil Armstrong' have the answer 1969, it will need a pretty sophisticated algo to avoid the trap of learning that it should always answer 1969.
Google (and plenty of other search engines) will all helpfully tell you that the slowest animal is the sloth, which is plainly incorrect if you just look at a video of a sloth and then look at a video of a worm. Despite their reputation, sloths still move visibly.
A fun one I tried was "Which president became supreme dictator" that answered "William Howard Taft." It took some finessing to get it (I originally started with "dictator for life"), but it's clear that it picked up only "supreme" and "president" in the query to guess that you meant to ask which president became a Supreme Court justice.
The last one actually is an example of the ambiguity of language. For example, if you asked this from a person, answers like “sponge”, “mollusk” and “flatworm” probably aren’t what you’re really considering and you might think the other person is being too pedantic. So really you want nuance from the AI like “if you meant all animals, X. If you meant all land animals, Y. If you meant mammals, Z”. Human are very ambiguous with categories because all categories are arbitrary human classifications and dealing with that ambiguity is difficult even for humans.
The problem isn't so much the answers, but how they are framed.
In the case of Neil Armstrong, the failure is to state that the answer is for a Moon landing. That would allow the person making the query to realize that the answer was not what they were looking for. In this case it is because Armstrong never landed on Mars. In other cases it may be because the question or data were incorrect.
The current Tom Hanks answer is closer to what should have been done in this regard since it refers to the film directly, though it remains problematic in that it highlights the year to such a degree that the person making the query may ignore the context. (It is also problematic because the film is about a mission that didn't land on the Moon, which could only be determined through further research. Granted, that is more along the lines of your slowest animal example.)
> Google (and plenty of other search engines) will all helpfully tell you that the slowest animal is the sloth, which is plainly incorrect if you just look at a video of a sloth and then look at a video of a worm.
Would your average person consider a worm an "animal". In the scientific sense, sure, pretty much everything living that is not a plant is an animal, but I think most people don't think of worms when they think of animals.
You can say, "well that's wrong", but that's kinda missing the point. Categories are somewhat arbitrary, and word usage can differ greatly from the scientific or dictionary definition when we're talking about colloquial usage.
As someone who is building such instant answer system for our own web search engine [1] the level of failure here tells me that this is:
- Either a basic, distilled BERT based model (optimized for latency and scale) or
- More likely, still a purely heuristics driven answers system like the one Google has been using for last 10+ years.
The current NLP models are able to quite successfully answer these questions with proper context. In this case ignoring Lance or Mars in question looks more like old-school keyword based heuristics and there is no way it can get this right.
Google is not only allowing mistakes in this, what is considered a fairly difficult problem to solve, but also questions that directly query their own knowledge graph, for example this one querying for a CEO of a well known public company where it returns the wrong answer [2]
This only shows that 'emperor has no clothes' [3] and that there is still a lot of room for innovation left, specially on the 'organizing the world's information' front.
Interesting. When searching for, "when did neil armstrong set foot on venus?" It, shows, "In which year and when Armstrong set his foot on the moon?" and provides the correct year. That's presumably what should show when searching for Mars (or any non earth moon planets).
And the big secret that they're hiding is that they were able to journey to Mars in three days. To this day they say that such a time is impossible, that it would take months.
I really wish someone would come along and kick googles ass on search, it's embarrassingly bad. It's wild how I used to look up to the FAANG folks, but it has become a real emperor has no clothes situation.
this one is a bit problematic, because people assume those 0 click search results are True / gold standard. like when I do 37 c to f and I see it display the results, I never double check them, I just assume they are right.
Yeah the fact that *this* is what the mighty Google just let loose upon the world shows that they don't really care any more about their users, their products and their supposed mission of organizing the worlds' information.
The fix is quite easy too, just have an AI driven system with human supervision, I think at this point it is well known that a human + AI will basically always trump an AI and it's not like google doesn't have the resources to pull this off. You could build a database of verified facts that are often googled, have the AI gather data, have humans do a sanity check.
It's probably just some scheme to get someone promoted inside Google, the infobox is a side-effect of someone's career plan, it doesn't really matter if it actually works as long as it hits some key metric that gets used in an evaluation somewhere, who cares about propagating misinformation and falsehoods google is still making money. It makes me depressed.
("First woman in space" result snippet is about Sally Ride, the first American in space. First result, below that is the Wikipedia entry for Valentina Tereshkova, the first woman in space.)
> I must of been busy that day, I would of remembered that.
As a non-native speaker, I'm having trouble understanding the sentence, specially the "I must of", "I would of" constructions. Is this some style choice I don't know about?
This reminds me of how recently Google did not make it clear that there is no longer a tax penalty for not having health insurance (except in five states and D.C. I think).
now we can confirm that AI as as flawed as human intelligence itself. It just dynamically reads all the texts and remembers what it wants without being sure it is right.
Ask Google how tall Elon Musk is. You will get 188 cm and conversion of 5 foot 10 inches which is way off. Two websites in the results that feed Google have the conversion wrong. Bad data is all over the place and copied over and over even simple incorrect conversions.
I tried the original query with Lance Armstrong and Louis Armstrong. Google still highlights the 1969 date: https://i.imgur.com/R3A2E6Z.png
Asking my phone via voice returns a sentence that starts with "Armstrong..." and talks about Neil's moon walk.
This is... a bit aggravating.
Man, referring to the story the guy whose photo Google attached to the text talking about a serial killer with his name, what if Google returns a question about him with a sentence about the serial killer?
E.g. something like "What is Barack Osama famous for?" "Osama is well-known as a terrorist". At least most of the world knows this isn't true, again what if it's Joe Neighbor, who shares a name with Joe Child Molester...
this is a perfectly good response because chances are that's what the user meant/was looking for (also since the scenario of Armstrong on Mars doesn't exist)
Please take a screenshot for posterity.