- Researchers say they have found ways to break through guardrails on major AI-powered chatbots.
- AI-powered chatbots like ChatGPT are moderated to ensure they don't produce harmful content.
- Researchers used jailbreaks developed for open-source systems to target mainstream AI systems.
Researchers say they have found potentially unlimited ways to break the safety guardrails on major AI-powered chatbots from OpenAI, Google, and Anthropic.