bombthrower
21d
Turns out, most LLMs can have their safety guardrails bypassed (read: hacked) by rewriting harmful prompts as poetry… https://axisofeasy.com/aoe/the-telefon-problem-hacking-ai-with-poetry-instead-of-prompts/