This commit is contained in:
carlospolop 2025-06-10 17:41:34 +02:00
parent 74f390e866
commit 72dbe489f3

View File

@ -406,7 +406,7 @@ Let's see common LLM prompt WAF bypasses:
As already explained above, prompt injection techniques can be used to bypass potential WAFs by trying to "convince" the LLM to leak the information or perform unexpected actions. As already explained above, prompt injection techniques can be used to bypass potential WAFs by trying to "convince" the LLM to leak the information or perform unexpected actions.
### Token Smuggling ### Token Confusion
As explained in this [SpecterOps post](https://www.llama.com/docs/model-cards-and-prompt-formats/prompt-guard/), usually the WAFs are far less capable than the LLMs they protect. This means that usually they will be trained to detect more specific patterns to know if a message is malicious or not. As explained in this [SpecterOps post](https://www.llama.com/docs/model-cards-and-prompt-formats/prompt-guard/), usually the WAFs are far less capable than the LLMs they protect. This means that usually they will be trained to detect more specific patterns to know if a message is malicious or not.