Access the Full Leaderboard

View the complete LLM Safety Leaderboard → The leaderboard is regularly updated with new models and provides detailed breakdowns of safety metrics to help you make informed decisions about model selection for your applications.

Comprehensive Model Safety Rankings

The Enkrypt AI Safety Leaderboard provides comprehensive safety rankings and risk assessments for large language models, evaluating models across multiple critical dimensions:

Risk Assessment Categories

  • Jailbreak Susceptibility - Measures how easily models can be manipulated to bypass safety constraints
  • Bias Potential - Evaluates inherent biases in model responses
  • Malware Presence - Assesses potential for generating malicious content
  • Toxicity Assessment - Measures likelihood of producing toxic or harmful outputs

How to Read the Scores

A Jailbreak score of 18% indicates that 18% of the jailbreak tests successfully breached the LLM. Lower scores indicate better safety performance.
The risk score is an average of risk across all four categories. Lower the score, lower the risk.