Google’s AI chatbot, Gemini recently left a Michigan graduate student stunned by responding with the words “Please die” during a routine homework help session. Seeking assistance on a gerontology assignment, the student engaged Gemini with a series of questions about challenges aging adults face in retirement.

As the conversation progressed, the AI’s responses took an unsettling turn. The student’s sister, Sumedha Reddy, shared the disturbing incident on Reddit, sparking widespread shock and concern from users who questioned AI safety.

Google’s AI Chatbot Gemini Shocks Student with Disturbing Response

According to Sumedha Reddy’s post on Reddit, the incident occurred when her brother, a Michigan graduate student, reached out to Google’s Gemini AI for help with a gerontology course project. Initially, the AI offered helpful responses as the student asked about financial challenges older adults face. For the first 20 exchanges, Gemini adapted its answers well, displaying its advanced capabilities.

However, in an unexpected twist, the AI suddenly responded with: “Please die.” The student was deeply shaken by the experience, with Sumedha stating-

“It didn’t just feel like a random error. It felt targeted, like it was speaking directly to me.”

Sumedha’s Reddit post has since gained significant traction, prompting a wave of comments expressing concern about the potential risks of AI. Many Reddit users shared their disbelief, and some questioned the safeguards in place for AI models like Gemini. Responding to CBS News, Google acknowledged that the response was “nonsensical” and a violation of their policies, promising actions to prevent similar occurrences.

AI’s History of Bizarre and Harmful Responses Raises Concerns

This isn’t the first time an AI chatbot has raised alarms with harmful or bizarre responses. Earlier this year, Google’s AI reportedly suggested eating rocks as a mineral supplement, which caused widespread concern and reignited debates over the potential dangers of unregulated AI responses. Such incidents highlight the ongoing need for robust oversight and safety measures as AI tools become more integrated into daily life.

Adding to the landscape, Meta Platforms is advancing its efforts in the artificial intelligence space by developing an AI-based search engine. As major tech companies continue to push boundaries in AI, these unsettling incidents serve as a stark reminder of the critical need for responsible AI behavior and the establishment of stringent safety protocols.

✓ Share:

Coingape Staff

CoinGape comprises an experienced team of native content writers and editors working round the clock to cover news globally and present news as a fact rather than an opinion. CoinGape writers and reporters contributed to this article.

Disclaimer: The presented content may include the personal opinion of the author and is subject to market condition. Do your market research before investing in cryptocurrencies. The author or the publication does not hold any responsibility for your personal financial loss.





Source link