
Researchers discover ‘Deceptive Delight’ to jailbreak LLM
Cybersecurity researchers have discovered a new technique that could allow attackers to jailbreak large language models during conversations. The method, called Deceptive Delight, has an average success rate of 64.6%