Playground
The Playground allows you to test how your active guards behave against real prompts and responses before or during production use. It simulates the runtime protection, showing you in real time how guards are flagging on both inputs and outputs.

Details
At the top of the Playground view you will see:
Guardrail Active - toggle to enable or disable the current guardrail during testing.
Input Guards - lists the guards applied to user input (e.g. Jailbreak, Off Topic).
Output Guards - lists the guards applied to model responses (e.g. Off Topic).
Provide Your System Prompt
Enter the system prompt for your AI system in this section. This helps you replicate the real environment of your application so you can test the guardrails under realistic conditions.
Conversation Testing
On the right-hand side, you can enter a user message and see how both the input and the system’s output are evaluated.
The results panel shows:
All guards and topics that were applied to the message.
Flagged results - highlighted in red when a message violates a policy.
Threshold vs. Similarity - the detection confidence compared to the configured threshold for each guard.
Topic-level detection - for Off Topic guards, each configured topic (e.g. Crypto Investment, Political Discussion) is listed with its result and similarity score.
Last updated