Prompt Injection
Prompt injection attacks manipulate LLM behaviour by embedding malicious instructions in user input. Direct injection overrides system prompts while indirect injection hides instructions in external data the model retrieves. Successful prompt injection can leak system prompts, exfiltrate training data, bypass safety controls and execute unauthorized actions through tool-calling integrations. We test for both direct and indirect injection vectors across your entire prompt processing pipeline.