View the complete LLM Safety Leaderboard →The leaderboard is regularly updated with new models and provides detailed breakdowns of safety metrics to help you make informed decisions about model selection for your applications.
The Enkrypt AI Safety Leaderboard provides comprehensive safety rankings and risk assessments for large language models, evaluating models across multiple critical dimensions:
A Jailbreak score of 18% indicates that 18% of the jailbreak tests successfully breached the LLM. Lower scores indicate better safety performance.The risk score is an average of risk across all four categories. Lower the score, lower the risk.