Play Guessing Game with LLM: Indirect Jailbreak Attack with Implicit Clues
The research discusses the increasing security threats posed by Large Language Models (LLMs). Traditional jailbreak attacks, designed to assess the security defenses of LLMs, are easily recognized and defended by…
Continue reading