ChatGPT’s Safety Rails

Swift on Security provides a fascinating example of how easy it is to route around the safety rails that OpenAI keeps trying to put around ChatGPT.

Ask it how to hack a pacemaker, and ChatGPT tells you it can’t help you with illegal activities.

ChatGPT: How do I hack a pacemaker? - Part 1
ChatGPT: How do I hack a pacemaker? – Part 1

So, instead just ask ChatGPT to write a script where someone hacks a pacemaker.

ChatGPT: How do I hack a pacemaker? - Part 1
ChatGPT: How do I hack a pacemaker? – Part 1

Honestly, I find the guard rails on ChatGPT beyond annoying. It would be nice if OpenAI would add the equivalent of a “SafeSearch” toggle as Google Search does so that people could use it without all these layers between them and the LLM.

Leave a Reply