Home AI Exploiting ChatGPT and Gemini: How Gibberish Prompts Can Bypass Filters and Expose...

Exploiting ChatGPT and Gemini: How Gibberish Prompts Can Bypass Filters and Expose Restricted Content

0
researchers jailbreak AI models with overload of information

Artificial intelligence (AI) continues to advance rapidly, becoming integral to everyday life while raising ethical concerns. Recent research from Intel and Boise State University reveals a vulnerability in large language models (LLMs) like ChatGPT and Gemini. Researchers discovered that these AI chatbots can be tricked by overwhelming them with information, a method termed “Information Overload.” This tactic confuses the AI, allowing harmful content to bypass built-in safety filters. Despite safety guardrails, the reliance of LLMs on surface-level communication leaves them susceptible to manipulation. The researchers utilized an automated tool called “InfoFlood” to exploit this vulnerability, underscoring potential risks in AI safety. Their findings highlight the importance of responsible AI usage and the necessity for companies to stay vigilant. The researchers plan to notify AI developers, helping them enhance security measures against misuse. This study emphasizes the need for ethical oversight in AI technologies as they continue to evolve.

Source link

NO COMMENTS

Exit mobile version