20% of Generative AI ‘Jailbreak’ Attacks are Successful
Generative AI jailbreak attacks, where models are instructed to ignore their safeguards, succeed 20% of the time, research has found. On average, adversaries need just 42 seconds and five interactions to break through. In some cases, attacks occur in as little as four seconds. These findings both highlight the significant vulnerabilities in current GenAI algorithms and the difficulty in preventing exploitations in real time. Of the successful attacks, 90% lead to sensitive data leaks, according…
Read More