Jailbreaking in verse: how poetry loosens AI’s tongue
Researchers have discovered that styling prompts as poetry can significantly undermine the effectiveness of language models’ safety guardrails.
3 articles
Researchers have discovered that styling prompts as poetry can significantly undermine the effectiveness of language models’ safety guardrails.
The Whisper Leak attack allows its perpetrator to guess the topic of your conversation with an AI assistant — without decrypting the traffic. We explore how this is possible, and what you can do to protect your AI chats.
How hackers exploit chatbot features to restore encrypted chats from OpenAI ChatGPT, Microsoft Copilot, and most other AI chatbots.