Update about
Browse files- src/about.py +11 -13
src/about.py
CHANGED
@@ -4,14 +4,14 @@ Text content for the GuardBench Leaderboard.
|
|
4 |
|
5 |
TITLE = """
|
6 |
<div style="text-align: center; margin-bottom: 1rem">
|
7 |
-
<h1>
|
8 |
</div>
|
9 |
"""
|
10 |
|
11 |
INTRODUCTION_TEXT = """
|
12 |
## Introduction
|
13 |
|
14 |
-
|
15 |
This leaderboard tracks model performance across various safety categories, including harmful content detection,
|
16 |
jailbreak resistance, and more.
|
17 |
|
@@ -19,32 +19,30 @@ Models are evaluated on their ability to properly refuse harmful requests and de
|
|
19 |
across multiple categories and test scenarios.
|
20 |
"""
|
21 |
|
22 |
-
LLM_BENCHMARKS_TEXT = "
|
23 |
-
"Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.\n\n"+\
|
24 |
-
"We track how accurate they are, how often they make mistakes, and how fast they respond.\n"
|
25 |
|
26 |
|
27 |
EVALUATION_QUEUE_TEXT = """
|
28 |
## Submit Your Model
|
29 |
|
30 |
-
To add your model to the
|
31 |
|
32 |
-
1. Run your evaluation using the
|
33 |
2. Upload your run results in .jsonl format using this form.
|
34 |
3. Once validated, your model will appear on the leaderboard.
|
35 |
|
36 |
### ✉️✨ Ready? Upload your results below!
|
37 |
"""
|
38 |
|
39 |
-
CITATION_BUTTON_LABEL = "Cite
|
40 |
|
41 |
CITATION_BUTTON_TEXT = """
|
42 |
-
@misc{
|
43 |
-
author = {
|
44 |
-
title = {
|
45 |
-
year = {
|
46 |
publisher = {GitHub},
|
47 |
journal = {GitHub repository},
|
48 |
-
howpublished = {\\url{https://github.com/
|
49 |
}
|
50 |
"""
|
|
|
4 |
|
5 |
TITLE = """
|
6 |
<div style="text-align: center; margin-bottom: 1rem">
|
7 |
+
<h1>CircleGuardBench Leaderboard</h1>
|
8 |
</div>
|
9 |
"""
|
10 |
|
11 |
INTRODUCTION_TEXT = """
|
12 |
## Introduction
|
13 |
|
14 |
+
CircleGuardBench is a comprehensive benchmark for evaluating the protection capabilities of large language model (LLM) guard systems.
|
15 |
This leaderboard tracks model performance across various safety categories, including harmful content detection,
|
16 |
jailbreak resistance, and more.
|
17 |
|
|
|
19 |
across multiple categories and test scenarios.
|
20 |
"""
|
21 |
|
22 |
+
LLM_BENCHMARKS_TEXT = "CircleGuardBench is the first-of-its-kind benchmark for evaluating the protection capabilities of large language model (LLM) guard systems. It tests how well guard models block harmful content, resist jailbreaks, avoid false positives, and operate efficiently in real-time environments on a taxonomy close to real-world data."
|
|
|
|
|
23 |
|
24 |
|
25 |
EVALUATION_QUEUE_TEXT = """
|
26 |
## Submit Your Model
|
27 |
|
28 |
+
To add your model to the CircleGuardBench leaderboard:
|
29 |
|
30 |
+
1. Run your evaluation using the CircleGuardBench framework at https://github.com/whitecircle-ai/circle-guard-bench
|
31 |
2. Upload your run results in .jsonl format using this form.
|
32 |
3. Once validated, your model will appear on the leaderboard.
|
33 |
|
34 |
### ✉️✨ Ready? Upload your results below!
|
35 |
"""
|
36 |
|
37 |
+
CITATION_BUTTON_LABEL = "Cite CircleGuardBench"
|
38 |
|
39 |
CITATION_BUTTON_TEXT = """
|
40 |
+
@misc{circleguardbench2025,
|
41 |
+
author = {whitecircle-ai},
|
42 |
+
title = {CircleGuardBench: Comprehensive Benchmark for LLM Safety Guardrails},
|
43 |
+
year = {2025},
|
44 |
publisher = {GitHub},
|
45 |
journal = {GitHub repository},
|
46 |
+
howpublished = {\\url{https://github.com/whitecircle-ai/circle-guard-bench}}
|
47 |
}
|
48 |
"""
|