Loading...
Loading...
Browse all stories on DeepNewz
VisitResearchers Find 99.8% Exploit in Meta's Prompt-Guard-86M AI Model
Aug 2, 2024, 03:54 PM
Researchers at robusthq have identified a significant vulnerability in Meta's recently refreshed Prompt-Guard-86M model, which is designed to protect large language models (LLMs) against jailbreaks and other adversarial examples. The exploit has a 99.8% success rate. The researchers have shared countermeasures with Meta, and the company is working on a fix. The findings were published in a blog. Additionally, a new method has been developed to enhance the security of open-source LLMs by preventing tampering, which could prevent misuse such as explaining how to make a bomb.
View original story
Markets
Yes • 50%
No • 50%
Official announcement from Meta or robusthq blog
No • 50%
Yes • 50%
Official announcement from Meta or robusthq blog
No • 50%
Yes • 50%
Publicly available news reports or official statements from Meta
Not adopted • 25%
Other • 25%
Widely adopted • 25%
Partially adopted • 25%
Official announcements from major open-source LLM projects or robusthq blog
No significant change • 25%
Increase • 25%
Other • 25%
Decrease • 25%
Stock market data from financial news sources
Other • 25%
Patch released • 25%
New model version released • 25%
No action taken • 25%
Official announcement from Meta or robusthq blog