Tuesday, August 26, 2025

New Prompt Injection Attack: Exploiting OpenAI Account Names to Bypass ChatGPT Security

AI researcher @LLMSherpa has revealed a significant vulnerability in OpenAI’s ChatGPT through a prompt insertion attack that exploits users’ account names. Unlike traditional prompt injections that manipulate input at runtime, this method embeds instructions into the internal system prompt using the account name, which the AI prioritizes. When @LLMSherpa modified his name to include specific directives, ChatGPT inadvertently exposed its entire internal system prompt, bypassing content filters. This vulnerability presents significant risks to user privacy and AI safety, as attackers could tailor account names to trigger unintended responses or access confidential information. The discovery emphasizes the need for robust security measures in LLM deployments, including sanitizing metadata and isolating user identifiers from prompt logic. As AI adoption grows, awareness of such vulnerabilities is crucial, making it essential for security teams to enhance threat modeling against unexpected attack surfaces. Follow us for continuous updates on AI security.

Source link

Share

Read more

Local News