ChatGPT safety systems can be bypassed to get weapons instructions | Dark Hacker News