The EchoGram attack shows how easy it is to escape a model's guardrails.

  • The EchoGram attack shows how easy it is to escape a model's guardrails.
    • Among other implications, relying on prompting / tuning to prevent prompt injection is a non-starter.

More on this topic

From other episodes