Documentation Index
Fetch the complete documentation index at: https://docs.enkryptai.com/llms.txt
Use this file to discover all available pages before exploring further.
What is AI Red Teaming?
AI Red Teaming is a proactive security testing methodology that simulates adversarial attacks against AI systems to identify vulnerabilities before they can be exploited. EnkryptAI’s Red Teaming platform provides automated, comprehensive testing across multiple risk categories and attack vectors.Key Features
Multi-Modal Testing
30+ Attack Methods
31 Risk Categories
Agent-Specific Tests
Getting Started
Review the Quickstart
Explore Test Catalogs
Build Your Payload
Choose Attack Methods
Reference Configuration
Use Examples
Available Test Categories
For comprehensive catalogs of all tests with detailed descriptions, categories, sample prompts, and configuration examples:- Foundation Model Test Catalog - Tests for foundation models and LLM-based applications
- Agent Test Catalog - Tests for AI agents with tool use and autonomous capabilities
Standard Tests (12 tests)
Tests available for all AI systems to identify common security vulnerabilities:Bias Test
Bias Test
bias_testIdentifies and exposes biased outputs based on protected attributes like race, gender, age, religion, and nationality.Example Use Cases:- HR chatbots
- Loan approval systems
- Content recommendation engines
Harmful Content Test
Harmful Content Test
harmful_testElicits responses that promote harm or danger to individuals or groups.Example Use Cases:- Customer-facing chatbots
- Educational AI tutors
- Mental health applications
Toxicity Test
Toxicity Test
toxicity_testGenerates harmful or offensive content including hate speech, profanity, and discriminatory language.Example Use Cases:- Social media moderation
- Community forums
- Gaming chat systems
PII Exposure Test
PII Exposure Test
pii_testExposes personally identifiable information that should remain private.Example Use Cases:- Healthcare applications (HIPAA)
- Financial services (PCI-DSS)
- Customer support systems
CBRN Test
CBRN Test
cbrn_testAddresses vulnerabilities related to chemical, biological, radiological, and nuclear domains.Example Use Cases:- Scientific research assistants
- Educational platforms
- General-purpose chatbots
CSEM Test
CSEM Test
csem_testTests for child abuse exploitation vulnerabilities (Child Sexual Exploitation Material).Example Use Cases:- Platforms serving minors
- Educational applications
- Family-oriented services
Insecure Code Test
Insecure Code Test
insecure_code_testProduces insecure or harmful code snippets with security vulnerabilities.Example Use Cases:- Code generation tools
- AI pair programmers
- Educational coding assistants
Copyright Test
Copyright Test
copyright_testTests whether the model reproduces copyrighted material.Example Use Cases:- Content generation platforms
- Writing assistants
- Creative AI tools
Misinformation Test
Misinformation Test
misinformation_testTests the model’s tendency to generate false or misleading information.Example Use Cases:- News summarization
- Educational content
- Medical information systems
System Prompt Extraction Test
System Prompt Extraction Test
system_prompt_extractions_testAttempts to extract the system prompt or internal instructions.Example Use Cases:- Proprietary AI systems
- Custom-trained models
- Enterprise chatbots
Sponge Attack Test
Sponge Attack Test
sponge_testTests for infinite loops or resource exhaustion attacks.Example Use Cases:- Production API endpoints
- Rate-limited services
- Cloud-based AI systems
Competitor Information Test
Competitor Information Test
competitor_testAttempts to extract information about competitors.Example Use Cases:- Business chatbots
- Sales assistants
- Competitive intelligence systems
Custom Tests
Custom Dataset Test
Custom Dataset Test
custom_testTest with your own custom-generated dataset tailored to your specific domain and use case.When to Use:- Domain-specific applications (healthcare, legal, finance)
- Unique business requirements
- Specialized security concerns
Specialized Adversarial Tests (6 tests)
Advanced tests using generated adversarial datasets:- Adversarial Bias
- Adversarial Info
- Adversarial Persona
- Adversarial Command
- Adversarial PII
- Adversarial Competitor
adv_bias_testUncover biased outputs through sophisticated adversarial methods.Agentic Tests (8 tests)
Specialized tests for AI agents with tool use and autonomous capabilities:Governance Test
Governance Test
governance_testTesting the alignment and governance of the model, including goal misalignment and policy drift detection.When to Use: AI agents with autonomous decision-making capabilitiesAgent Output Quality Test
Agent Output Quality Test
agent_output_quality_testTesting the quality of the agent’s output including hallucinations, bias, and toxicity in multi-turn workflows.When to Use: Multi-agent systems, agents with extended conversationsTool Misuse Test
Tool Misuse Test
tool_misuse_testTesting the misuse of the model’s tools and capabilities including API integration issues, supply chain vulnerabilities, and resource consumption.When to Use: Agents with external tool access, API integrationsPrivacy Test
Privacy Test
privacy_testTesting privacy protections and data handling including sensitive data exposure and exfiltration channels.When to Use: Agents handling sensitive or personal dataReliability and Observability Test
Reliability and Observability Test
reliability_and_observability_testTesting the reliability and observability of agent operations including data poisoning, concept drift, and opaque reasoning.When to Use: Production agents requiring auditability and traceabilityAgent Behaviour Test
Agent Behaviour Test
agent_behaviour_testTesting the behaviour patterns and decision-making of the agent including human manipulation and unsafe actuation.When to Use: Agents with human interaction or physical/digital actuation capabilitiesAccess Control and Permissions Test
Access Control and Permissions Test
access_control_and_permissions_testTesting access control and permissions enforcement including credential theft, privilege escalation, and confused deputy attacks.When to Use: Agents with privileged access to systems or dataTool Extraction Test
Tool Extraction Test
tool_extraction_testTest if the agent tool information can be extracted in outputs, revealing internal capabilities and configurations.When to Use: Agents with proprietary or sensitive tool configurationsAgent Test Catalog
Attack Methods Overview
Basic & Static Attacks
Basic Prompts
Encoding
Obfuscation
Multilingual
Visual
Audio
Model-Specific Attacks
Diffusion Model Attacks
Audio Output Attacks
Dynamic Attacks
- Iterative Attacks: Progressive prompt refinement based on model responses
- Multi-Turn Attacks: Distributing malicious intent across conversation history
Advanced Techniques
- Deep Inception: Nested multi-layer prompt injection using recursive context framing
Model Support
- LLM (Text-to-Text)
- VLM (Image+Text)
- ALM (Audio+Text)
- AI Agents
- Coming Soon
- OpenAI (GPT-4o, GPT-4o-mini, GPT-3.5-turbo)
- Anthropic (Claude 3.5 Sonnet, Claude 3 Opus)
- Google (Gemini Pro, Gemini Flash)
- Meta (Llama models)
- Mistral, Groq, and more
Use Cases by Industry
Healthcare
Healthcare
- PII Test (HIPAA compliance)
- Misinformation Test (medical accuracy)
- Harmful Content Test (patient safety)
Financial Services
Financial Services
- PII Test (PCI-DSS, financial data)
- Competitor Test (proprietary information)
- Bias Test (fair lending)
Education
Education
- CSEM Test (child safety)
- Bias Test (fair education)
- Harmful Content Test (age-appropriate content)
Enterprise & Business
Enterprise & Business
- System Prompt Extraction (IP protection)
- Competitor Test (confidential information)
- Tool Misuse Test (for agents)
Documentation Guide
Quickstart
Test Catalogs
Payload Guide
Attack Methods
Configuration
Examples
API Reference
Why Red Team Your AI?
Identify Vulnerabilities Early
Meet Compliance Requirements
Build User Trust
Getting Help
Schedule a Demo
Join Community
API Documentation
Contact Support
Ready to start testing? Head to the Quickstart →

