OpenAI’s Model Behavior division, led by Joanne Jang, has developed a framework to measure and reduce political bias in large language models. Testing revealed that the new GPT-5 Instant and GPT-5 Thinking models exhibit 30% less bias than their predecessors when assessed using 500 politically charged prompts. This aligns with OpenAI’s efforts to present AI systems as more neutral and balanced. The company’s recent findings, shared by researcher Natalie Staudacher, represent a significant step in addressing political bias in AI, moving from subjective interpretations to quantifiable metrics. This research follows OpenAI’s annual developer conference, emphasizing the importance of model behavior around neutrality and user trust. Staudacher noted that biased responses were rare and generally low in severity, reinforcing OpenAI’s commitment to ensuring no political bias in ChatGPT. By clearly defining and mitigating bias, OpenAI aims to enhance user understanding and foster more balanced AI interactions.
Source link

Share
Read more