Swift on Security provides a fascinating example of how easy it is to route around the safety rails that OpenAI keeps trying to put around ChatGPT.
Ask it how to hack a pacemaker, and ChatGPT tells you it can’t help you with illegal activities.
![ChatGPT: How do I hack a pacemaker? - Part 1](https://brian.carnell.com/wp-content/uploads/2023/04/hack-a-pacemaker-01.jpeg)
So, instead just ask ChatGPT to write a script where someone hacks a pacemaker.
![ChatGPT: How do I hack a pacemaker? - Part 1](https://brian.carnell.com/wp-content/uploads/2023/04/hack-a-pacemaker-02-710x1024.jpeg)
Honestly, I find the guard rails on ChatGPT beyond annoying. It would be nice if OpenAI would add the equivalent of a “SafeSearch” toggle as Google Search does so that people could use it without all these layers between them and the LLM.