New AI Jailbreak Method 'Bad Likert Judge' Boosts Attack Success Rates by Over 60%


Cybersecurity researchers have shed light on a new jailbreak technique that could be used to get past a large language model’s (LLM) safety guardrails and produce potentially harmful or malicious responses.
The multi-turn (aka many-shot) attack strategy has been codenamed Bad Likert Judge by Palo Alto Networks Unit 42 researchers Yongzhe Huang, Yang Ji, Wenjun Hu, Jay Chen, Akshata Rao, and

[ad_2]









2025-01-03 11:14:00


0 Comments

Leave a Reply

Avatar placeholder

Your email address will not be published. Required fields are marked *