Meet The Ai Jailbreakers: ‘i See The Worst Things Humanity Has Produced’
To test the safety and security of AI, hackers have to trick large language models into breaking their own rules. It requires ingenuity and manipulation – and can come at a deep emotional cost
A few months ago, Valen Tagliabue sat in his hotel room watching his chatbot, and felt euphoric. He had just manipulated it so skilfully, so subtly, that it began ignoring its own safety rules. It told him how to sequence new, potentially lethal pathogens and how to make them resistant to known drugs.
Tagliabue had spent much of the previous two years testing and prodding large language models such as Claude and ChatGPT, always with the aim of making them say things they shouldn’t. But this was one of his most advanced “hacks” yet: a sophisticated plan of manipulation, which involved him being cruel, vindictive, sycophantic, even abusive. “I fell into this dark flow where I knew exactly what to say, and what the model would say back, and I watched it pour out everything,” he says. Thanks to him, the creators of the chatbot could now fix the flaw he had found, hopefully making it a little safer for everyone.
Continue reading...Popular Products
-
Digital LED Pixel Art Photo Frame$450.99$314.78 -
Ai Dash Cam with Front & Rear HD, GPS...$295.56$147.78 -
Blind Spot & Lane Change Detection Ra...$286.99$199.78 -
Waterproof Multi-Sport Health Tracker...$83.99$57.78 -
Electric Blackhead Remover Vacuum Por...$208.99$145.78