Business Standard

Friday, December 20, 2024 | 11:57 AM ISTEN Hindi

Notification Icon
userprofile IconSearch

Google wants you to chat with its experimental AI chatbot at your own risk

Google has already warned that early previews of its LaMDA (Language Model for Dialogue Applications) model "may display inaccurate or inappropriate content".

Photo: Bloomberg

Photo: Bloomberg

IANS New Delhi

Google has opened its experimental artificial intelligence (AI) chatbot for the public and you can now register to chat with the AI-driven bot trained on the company's controversial language model.

Google has already warned that early previews of its LaMDA (Language Model for Dialogue Applications) model "may display inaccurate or inappropriate content".

'AI Test Kitchen' by Google is an app where people can learn about, experience, and give feedback on Google's emerging AI technology.

"Our goal is to learn, improve and innovate responsibly on AI together. We'll be opening up to small groups of people gradually," said the company.

According to Alphabet and Google CEO Sunday Pichai, 'AI Test Kitchen' is "meant to give you a sense of what it might be like to have LaMDA in your hands".

 

The ability of these language models to generate infinite possibilities shows potential, "but it also means they don't always get things quite right".

"And while we've made substantial improvements in safety and accuracy in the latest version of LaMDA, we're still at the beginning of a journey," said Google.

"We've added multiple layers of protection to the AI Test Kitchen. This work has minimised the risk, but not eliminated it," it added.

Both Google and Meta (formerly Facebook) have unveiled their AI conversational chatbots, asking the public to give feedback.

The initial reports are scary as the Meta chatbot named BlenderBot 3 thought Mark Zuckerberg is "creepy and manipulative" and Donald Trump will always be the US president.

Meta said last week that all conversational AI chatbots are known to sometimes mimic and generate unsafe, biased or offensive remarks.

"BlenderBot can still make rude or offensive comments, which is why we are collecting feedback that will help make future chatbots better," the company mentioned in a blogpost.

Last month, Google fired an engineer over breaching its confidentiality agreement after he made a claim that the tech giant's conversation AI is "sentient" because it has feelings, emotions and subjective experiences.

Lemoine also interviewed LaMDA, which came with surprising and shocking answers.

--IANS

na/

(Only the headline and picture of this report may have been reworked by the Business Standard staff; the rest of the content is auto-generated from a syndicated feed.)

Don't miss the most important news and views of the day. Get them on our Telegram channel

First Published: Aug 27 2022 | 5:34 PM IST

Explore News Home