Tick, tick, tick...

Someone tricked ChatGPT into giving them instructions on how to build a homemade fertiliser bomb. Brill.

Hacker tricks ChatGPT into giving out detailed instructions for making homemade bombs | TechCrunch
An explosives expert told TechCrunch that the ChatGPT output could be used to make a detonatable product and was too sensitive to be released.

How? By getting it to play a game. Creating a science fiction fantasy world, where any restrictions it had didn’t apply.

Details were checked and it would have been viable. People are going to spend a lot of time trying to jailbreak LLMs. Extremely hard to secure prompts.

(Bonus points if you get the image reference)

Subscribe to Gary P Shewan

Don’t miss out on the latest issues. Sign up now to get access to the library of members-only issues.
jamie@example.com
Subscribe