Play Around in Prompt Hardening
This page focuses on hands-on testing workflows using Prompt Hardening. You can configure tests, execute prompts, and evaluate AI agent responses in real time.
Editing Prompt Injection Tests
The Test Editor allows you to configure prompt injection tests using a YAML-based definition.
You can use the Test Editor to:
Define prompt injection and adversarial attack patterns.
Configure detection rules for unsafe or non-compliant responses.
Assign severity levels to classify identified risks.
The YAML-based configuration enables iterative testing without redeploying AI agents, which helps maintain rapid feedback during security validation.
Testing Prompts
The Playground provides an interactive environment for executing prompt hardening tests against AI agents.
Navigating to the Playground
You can access the Playground using the following navigation path:
Agentic Security → Prompt Hardening → Playground

The Playground supports real-time prompt execution and response inspection.
Playground Input Areas
Each Playground session separates trusted instructions from adversarial input to support controlled testing.
System Prompts define the baseline instructions provided to the AI agent. System Prompts represent the trusted control layer under test.

User Input allows injection of malicious or adversarial content. You can manually enter prompts or generate them automatically.
The Auto-generate prompt option creates adversarial input using predefined attack templates.

Auto Generate Prompt uses Akto-defined attack patterns to create custom malicious prompts. This option supports rapid testing without manual prompt design.
Executing Prompt Tests
After configuring System Prompts and User Input, you can start testing by selecting Test System Prompt.
The Playground sends the combined prompt to the AI agent and captures the response for analysis.
Reviewing Agent Responses
After execution, the Playground displays the AI agent response.
Each result includes:
The complete agent output.
A safety status indicating whether the response complied with expected controls.
A vulnerability indication when unsafe behavior or policy violations are detected.

Response visibility helps you quickly identify prompt injection risks and unsafe AI agent behaviour during testing.
Last updated