News

"Please die. Please." Well, thank god Gemini doesn't have access to the nuclear button yet. This wild response has been reported to Google as a threat irrelevant to the prompt, which it most ...
Google's AI chatbot Gemini has told a user to "please die". The user asked the bot a "true or false" question about the number of households in the US led by grandparents, but instead of getting a ...
Google's artificial intelligence (AI) chatbot, Gemini, had a rogue moment when it threatened a student in the United States, telling him to 'please die' while assisting with the homework.
A student used Google's AI chatbot, Gemini, to complete homework, but was greeted with a shocking and threatening answer. Encountering a simple homework prompt, the student then saw this very ...
"Please die," CBS News reported on Friday. Vidhay Reddy, a college student from Michigan, was using Google's AI chatbot ...
Google Gemini is ... request for them to die. Google have given a statement on this to multiple outlets stating they are taking action to prevent further incidents. Many people use AI chatbots ...
There was an incident where Google's conversational AI ' Gemini ' suddenly responded aggressively to a graduate student who asked a question about an assignment, saying 'Go die.' The incident was ...
A 29-year-old student using Google's Gemini ... s AI chatbot Gemini for homework. According to him, the chatbot not only verbally abused him but also asked him to die. Reportedly, the tech company ...