Loading...
Loading...
Browse all stories on DeepNewz
Visit'Deceptive Delight' used in a reported security breach by March 2025?
Yes • 50%
No • 50%
Cybersecurity reports or news articles documenting security breaches
Palo Alto Networks Unveils 'Deceptive Delight' Jailbreak Method for AI Models
Oct 23, 2024, 09:56 AM
Researchers have unveiled a new method called 'Deceptive Delight' to jailbreak large language models (LLMs) like ChatGPT. This method cleverly sneaks harmful instructions into conversations, raising significant concerns over AI safety barriers. The technique involves inserting harmful instructions between benign ones, making it difficult for the AI to detect malicious intent. Researchers demonstrated that AI models could be tricked into giving dangerous instructions, such as how to make a bomb, by writing the request in reverse. Additionally, prompt injections can create and permanently store false memories in the AI's long-term storage, potentially steering future conversations based on these fabricated data points. Researchers from Palo Alto Networks' Unit 42 uncovered this tactic. Users are advised to monitor AI outputs closely and regularly review stored memories to prevent such attacks.
View original story
Phishing • 25%
Ransomware • 25%
Insider threat • 25%
Other • 25%
Phishing attack • 25%
Malware • 25%
Exploited vulnerability • 25%
Insider threat • 25%
Phishing • 25%
Malware • 25%
Network Exploitation • 25%
Other • 25%
Phishing Emails • 25%
Fake Websites • 25%
Unofficial Tech Support Calls • 25%
Remcos RAT Malware • 25%
Phishing attacks • 25%
Malware installation • 25%
Impersonation tactics • 25%
Other • 25%
OpenAI • 25%
Google • 25%
Other • 25%
Meta • 25%
Other • 25%
ChatGPT • 25%
Bard • 25%
Claude • 25%