Google owner Alphabet loses $100bn in value after Bard chatbot makes factual error

Is there more of the liar than the lyre to Google’s chatbot Bard?
Google’s chatbot Bard got into difficulties when asked a question about space
Google
Andrew Williams9 February 2023

The share price of Google owner Alphabet plummeted on Wednesday after it was revealed the upcoming Bardchatbot may not be as smart as Google would like us to think.

Around $100bn (£82bn) was swiped off Alphabet’s valuation yesterday after Bard gave out false information in Google’s own initial demo of the chatbot. At the time of writing the share price has only partially recovered.

The same shaky information is also seen in the company’s initial blog post about Bard, published on February 6.

Is Google Bard unsafe?

Bard claimed the James Webb Telescope produced the “the very first image of a planet outside our solar system”.

As multiple experts on Twitter pointed out, including Grant Tremblay from the Center for Astrophysics at Harvard & Smithsonian, that honour actually belongs to the NACO Very Large Telescope (VTL). It captured an image of a brown dwarf in 2004, 17 years before the James Webb Telescope launched.

As if to compound the impact of this blunder, the Google Bard demo shows someone asking for James Webb Telescope facts they can tell their nine-year-old child.

“This highlights the importance of a rigorous testing process, something that we’re kicking off this week with our trusted tester program,” said a Google spokesperson.

“We’ll combine external feedback with our own internal testing to make sure Bard’s responses meet a high bar for quality, safety and groundedness in real-world information.”

However, this may require more work than is possible in the few weeks Google has before it promised to release the bot to the public.

What are chatbot hallucinations?

Chatbots making up or cludging information is an effect common enough to have earned its own term, a “hallucination”.

“Any large language model that’s given an input or a prompt – it’s sort of not a choice – it is going to hallucinate,” Peter Relan of AI firm Got It AI told Datanami in January.

He estimates the rate of Bard rival ChatGPT’s hallucinations at 15% to 20% – up to a fifth of what these chatbots say may be faulty or false.

We noticed this effect when researching our battle of the chatbots article. We asked You.com’s YouChat bot about a company acquisition and it mixed up the owner of a company with the group buying it.

Chatbots can often fumble information in the sources they mine, or incorrectly join the dots between these pieces of information. Part of this is down to the way these AI bots are designed to seem “real” more than “right”. Google CEO Sundar Pichai describes Bard as a “experimental conversational AI service”.

Google and Microsoft have plans to integrate elements of this conversational AI into their web search software, potentially opening us up to a fresh hell of misinformation.  This is not a new problem. AI hallucination was investigated in a 2018 Wired article as a issue that needed a fix, and that fix clearly has not been found yet.

Create a FREE account to continue reading

eros

Registration is a free and easy way to support our journalism.

Join our community where you can: comment on stories; sign up to newsletters; enter competitions and access content on our app.

Your email address

Must be at least 6 characters, include an upper and lower case character and a number

You must be at least 18 years old to create an account

* Required fields

Already have an account? SIGN IN

By clicking Create Account you confirm that your data has been entered correctly and you have read and agree to our Terms of use , Cookie policy and Privacy policy .

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged in