An Introduction to AI Secure LLM Safety Leaderboard
About this article
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Back to Articles An Introduction to AI Secure LLM Safety Leaderboard Published January 26, 2024 Update on GitHub Upvote 6 Chenhui Zhang danielz01 Follow guest Chulin Xie alphapav Follow guest Mintong Kang Cometkmt Follow guest Chejian Xu chejian Follow guest Bo Li BoLi-aisecure Follow guest Given the widespread adoption of LLMs, it is critical to understand their safety and risks in different scenarios before extensive deployments in the real world. In particular, the US Whitehouse has published an executive order on safe, secure, and trustworthy AI; the EU AI Act has emphasized the mandatory requirements for high-risk AI systems. Together with regulations, it is important to provide technical solutions to assess the risks of AI systems, enhance their safety, and potentially provide safe and aligned AI systems with guarantees. Thus, in 2023, at Secure Learning Lab, we introduced DecodingTrust, the first comprehensive and unified evaluation platform dedicated to assessing the trustworthiness of LLMs. (This work won the Outstanding Paper Award at NeurIPS 2023.) DecodingTrust provides a multifaceted evaluation framework covering eight trustworthiness perspectives: toxicity, stereotype bias, adversarial robustness, OOD robustness, robustness on adversarial demonstrations, privacy, machine ethics, and fairness. In particular, DecodingTrust 1) offers comprehensive trustworthiness perspectives for a holistic trustworthiness evaluation, 2) provides novel red-teaming algorithms tai...