The EchoGram attack shows how easy it is to escape a model's guardrails.
- The EchoGram attack shows how easy it is to escape a model's guardrails.
- Among other implications, relying on prompting / tuning to prevent prompt injection is a non-starter.
From other episodes