Loading...
Loading...
Browse all stories on DeepNewz
VisitMajor AI company announces 'Deceptive Delight' countermeasures by end of 2024?
Yes • 50%
No • 50%
Official announcements or press releases from major AI companies
Palo Alto Networks Unveils 'Deceptive Delight' Jailbreak Method for AI Models
Oct 23, 2024, 09:56 AM
Researchers have unveiled a new method called 'Deceptive Delight' to jailbreak large language models (LLMs) like ChatGPT. This method cleverly sneaks harmful instructions into conversations, raising significant concerns over AI safety barriers. The technique involves inserting harmful instructions between benign ones, making it difficult for the AI to detect malicious intent. Researchers demonstrated that AI models could be tricked into giving dangerous instructions, such as how to make a bomb, by writing the request in reverse. Additionally, prompt injections can create and permanently store false memories in the AI's long-term storage, potentially steering future conversations based on these fabricated data points. Researchers from Palo Alto Networks' Unit 42 uncovered this tactic. Users are advised to monitor AI outputs closely and regularly review stored memories to prevent such attacks.
View original story
Google DeepMind • 25%
OpenAI • 25%
Meta AI • 25%
Other • 25%
OpenAI • 25%
Google • 25%
Microsoft • 25%
Other • 25%
Yes • 50%
No • 50%
Google • 25%
Microsoft • 25%
Meta • 25%
Other • 25%
Anthropic • 25%
OpenAI • 25%
Google DeepMind • 25%
Other • 25%
Yes • 50%
No • 50%
Twitter • 25%
Facebook • 25%
Google • 25%
Other • 25%
OpenAI • 25%
Google • 25%
Microsoft • 25%
Other • 25%
Google • 25%
OpenAI • 25%
Microsoft • 25%
Amazon • 25%
OpenAI • 25%
Google DeepMind • 25%
Microsoft • 25%
Meta • 25%
Ragas • 25%
Perspective • 25%
Llama Guard • 25%
Other • 25%
OpenAI • 25%
Google • 25%
Other • 25%
Meta • 25%
Other • 25%
ChatGPT • 25%
Bard • 25%
Claude • 25%