Writing backwards can trick an AI into providing a bomb recipe

You May Be Interested In:‘We got stuck in puddles’: skiers upset by lack of snow on Swedish slopes


ChatGPT can be tricked with the right prompt

trickyaamir/Shutterstock

State-of-the-art generative AI models like ChatGPT can be tricked into giving instructions on how to make a bomb by simply writing the request in reverse, warn researchers.

Large language models (LLMs) like ChatGPT are trained on vast swathes of data from the internet and can create a range of outputs – some of which their makers would prefer didn’t spill out again. Unshackled, they are equally likely to be able to provide a decent cake recipe as know how to make explosives from household chemicals.

share Paylaş facebook pinterest whatsapp x print

Similar Content

The Google Gemini logo.
New hack uses prompt injection to corrupt Gemini’s long-term memory
The Microsoft Copilot logo.
Windows 11’s Copilot Vision wants to help you learn to use complicated apps
Nvidia confirms the Switch 2 supports DLSS, G-Sync, and ray-tracing
Nvidia confirms the Switch 2 supports DLSS, G-Sync, and ray-tracing
A still frame of an example AI-generated mushroom people video created by OpenAI's Sora.
Ten months after first tease, OpenAI launches Sora video generation publicly
Amazon, Apple make a deal to offer Apple TV+ in a Prime bundle
Amazon, Apple make a deal to offer Apple TV+ in a Prime bundle
Lenovo ThinkBook Plus Gen 6 Rollable
Lenovo laptop’s rollable screen uses motors to grow from 14 to 16.7 inches
The News Spectrum | © 2024 | News