Our research team conducts cutting-edge studies on AI safety, security, and risk assessment methodologies. We publish our findings to advance the field and help the community build safer AI systems.
“No Free Lunch with Guardrail”
Benchmarks show stronger guardrails improve safety but can reduce usability. Our paper proposes a framework to balance the trade-offs — ensuring practical, secure LLM deployment.“Fine-Tuning, Quantization & Safety”
Fine-tuning increases jailbreak vulnerability, while quantization has varied effects. Our analysis emphasizes the role of strong guardrails in deployment.
“Investigating Implicit Bias in LLMs”
A study of 50+ models reveals that bias persists — and sometimes worsens — in newer models. The work calls for standardized benchmarks to prevent discrimination in real-world AI use.
“VERA: Validation & Enhancement for RAG”
VERA improves Retrieval-Augmented Generation by refining retrieved context and output, reducing hallucinations and enhancing response quality across open-source and commercial models.
“SAGE-RT Synthetic Red Teaming”
SAGE enables scalable, synthetic red-teaming across 1,500+ harmfulness categories — achieving 100% jailbreak success on GPT-4o and GPT-3.5 in key scenarios.
Our research goes beyond just publications – it’s been applied to real-world benchmark studies to evaluate the security and performance of leading AI guardrails. These comparative tests provide practical insights into how guardrails perform under real attack scenarios.
Enkrypt AI tests over 100 leading foundation models - including from AI21, DeepSeek, Databricks, and Mistral - to strengthen the safety of their LLMs without compromising performance.
Our research advancements fuel our security platform and power the LLM Safety & Security Leaderboard - the most comprehensive benchmark for model safety in the industry.Our research is conducted in collaboration with leading universities, research institutions, and industry partners worldwide.