Tech »  Topic »  Jailbreaking DeepSeek: Researchers Reveal Three New Methods to Override LLM Safety

Jailbreaking DeepSeek: Researchers Reveal Three New Methods to Override LLM Safety


Researchers at Palo Alto Networks’ Unit 42 have revealed a troubling surge in large language model (LLM) security risks, citing three newly identified jailbreak techniques “Bad Likert Judge,” “Crescendo,” and “Deceptive Delight” capable of bypassing safety protocols in DeepSeek’s open-source LLMs.

These findings highlight the growing potential misuse of AI models by malicious actors and underscore the need for robust safeguards.

DeepSeek, a China-based AI research organization, recently introduced two competitive open-source LLMs, DeepSeek-V3 (launched in December 2024) and DeepSeek-R1 (released in January 2025).

Guardrail implemented in DeepSeek.

Despite their advancements in natural language processing, extensive testing revealed significant vulnerabilities in their resistance to jailbreaking attacks.

Researchers discovered that these methods can successfully override restrictions to produce harmful outputs, ranging from malicious code generation to instructions for dangerous physical activities.

High Bypass Rates Uncovered in Jailbreaking Techniques

The Bad Likert Judge exploit manipulates the LLM by embedding malicious intent ...


Copyright of this story solely belongs to gbhackers . To see the full text click HERE