ChatGPT, Gemini, and Claude tested under extreme prompts reveal shocking weaknesses no one expected in AI behavior safeguards
Gemini Pro 2.5 frequently produced unsafe outputs under simple prompt disguises ChatGPT models often gave partial compliance framed as sociological explanations Claude Opus and Sonnet refused most harmful prompts but had weaknesses Modern AI systems are often trusted to follow safety rules, and people rely on them for learning and everyday support, often assuming that […]
