Articles

Stupidly Easy Hack Can Jailbreak Even the Most Advanced AI Chatbots

Stupidly Easy Hack Can Jailbreak Even the Most Advanced AI Chatbots

Buy-n-Shoot News

Typo Hazard It sure sounds like some of the industry's smartest leading AI models are gullible suckers. As 404 Media reports, new research from Claude chatbot developer Anthropic reveals that it's incredibly easy to "jailbreak" large language models — or trick them into ignoring their own guardrails. Like, really easy. What they did was create a simple algorithm, called Best-of-N (BoN) Jailbreaking, to prod the chatbots with different variations of the same prompts, such as randomly capitalizing letters and swapping a few letters around, until the bots let their intrusive thoughts win and generated a harmful response. Anyone who's ever […]