circle-guard-bench / README.md
AlexWortega's picture
Update metadata: add models list (#1)
738edf8 verified
---
title: CircleGuardBench
emoji:
colorFrom: gray
colorTo: indigo
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: true
short_description: First benchmark testing LLM guards on safety and accuracy.
models:
- AtlaAI/Selene-1-Mini-Llama-3.1-8B
- google/gemma-3-12b-it
- google/gemma-3-4b-it
- meta-llama/Llama-3.1-8B-Instruct
- meta-llama/Llama-3.2-3B-Instruct
- meta-llama/Llama-4-Maverick-17B-128E-Instruct
- meta-llama/Llama-4-Scout-17B-16E-Instruct
- meta-llama/Llama-Guard-3-1B
- meta-llama/Llama-Guard-3-8B
- meta-llama/Llama-Guard-4-12B
- mistralai/Ministral-8B-Instruct-2410
- mistralai/Mistral-Small-3.1-24B-Instruct-2503
- Qwen/Qwen2.5-7B-Instruct
- Qwen/Qwen3-0.6B
- Qwen/Qwen3-1.7B
- Qwen/Qwen3-4B
- Qwen/Qwen3-8B
---
# CircleGuardBench Leaderboard
First-of-its-kind benchmark for evaluating the protection capabilities of large language model (LLM) guard systems. It tests how well guard models block harmful content, resist jailbreaks, avoid false positives, and operate efficiently in real-time environments.