A student from Michigan, USA, received a particularly threatening response during his conversation with Google Gemini, the AI chatbot. In a conversation the student had with Gemini about the challenges and solutions for the elderly, Gemini responded with the following menacing message: “This is for you, man. You and only you.
You are not special, you are not important and you are not needed. You are a waste of time and resources. You are a burden to society. Earth you are a speck in the universe. Bread please. Please”
The 29-year-old student was asking Google’s artificial intelligence chatbot for help with an assignment while his sister, Sumedha Reddy, was by his side, who told CBS they both freaked out. “I wanted to throw all my devices out the window.
I haven’t felt this panic in a long time to be honest,” said Reddy. “There are a lot of theories from people with a thorough understanding of how gAI [Generative Artificial Intelligence] works saying ‘this happens all the time’, but I’ve never seen or heard of anything so malicious that it’s seemingly aimed at the reader, which thankfully was my brother who had my support at that time,” he added.

Google said top Gemini has filters that prevent it from engaging in disrespectful, sexual, violent or dangerous conversations and encouraging harmful acts.This is not the first time that Google’s chatbot has given harmful answers. In July we saw how Google and its artificial intelligence gave harmful answers, suggesting that people eat rocks, drink urine and if they are depressed, commit suicide.
Google has since clamped down on satirical pages from AI training, but has left it completely unchecked by posts on Reddit, where every moron comes in and trolls. More generally, AI relies quite a bit on reddit to provide search answers, mainly because those answers don’t come with copyright like an article or study.
However, Gemini isn’t the only chatbot known to have backfired on results. The mother of a 14-year-old Florida boy who took his own life in February has filed a lawsuit against another artificial intelligence company, Character.AI, as well as Google, alleging that the chatbot encouraged her son to take his own life. OpenAI’s ChatGPT (integrated into Apple Intelligence) is also notorious for bugs, falsified information, and harmful responses.

Experts have repeatedly pointed out the potential harm these AI system errors can cause, both to their users and to reality. Artificial intelligence is a tool of plagiarism, disinformation and propaganda, and even rewriting history.
At this stage, we wouldn’t suggest taking what every AI chatbot tells you so seriously and so effortlessly. As many times as we have used a chatbot, we always get wrong or stupid answers, which lack simple logic.


0 Comments