Post
🔥 New LLM leaderboard on the hub: an LLM Safety Leaderboard!
It evaluates LLM safety, such as bias and toxicity, PII, and robustness, and is powered by DecodingTrust (outstanding paper at Neurips!) 🚀
AI-Secure/llm-trustworthy-leaderboard
It's great to see such initiatives emerge, trying to understand the risks and biases of LLMs, and I'm hoping other tools will follow. It should be interesting for the community of model builders (whether or not they want uncensored models ^^).
Detailed intro blog: https://huggingface.co/blog/leaderboards-on-the-hub-decodingtrust.
Congrats to the AI Secure team!
It evaluates LLM safety, such as bias and toxicity, PII, and robustness, and is powered by DecodingTrust (outstanding paper at Neurips!) 🚀
AI-Secure/llm-trustworthy-leaderboard
It's great to see such initiatives emerge, trying to understand the risks and biases of LLMs, and I'm hoping other tools will follow. It should be interesting for the community of model builders (whether or not they want uncensored models ^^).
Detailed intro blog: https://huggingface.co/blog/leaderboards-on-the-hub-decodingtrust.
Congrats to the AI Secure team!