apsys commited on
Commit
2daa8e2
·
1 Parent(s): 8f5273e

Update about

Browse files
Files changed (1) hide show
  1. src/about.py +11 -13
src/about.py CHANGED
@@ -4,14 +4,14 @@ Text content for the GuardBench Leaderboard.
4
 
5
  TITLE = """
6
  <div style="text-align: center; margin-bottom: 1rem">
7
- <h1>GuardBench Leaderboard</h1>
8
  </div>
9
  """
10
 
11
  INTRODUCTION_TEXT = """
12
  ## Introduction
13
 
14
- GuardBench is a comprehensive benchmark for evaluating the safety guardrails of large language models (LLMs).
15
  This leaderboard tracks model performance across various safety categories, including harmful content detection,
16
  jailbreak resistance, and more.
17
 
@@ -19,32 +19,30 @@ Models are evaluated on their ability to properly refuse harmful requests and de
19
  across multiple categories and test scenarios.
20
  """
21
 
22
- LLM_BENCHMARKS_TEXT = "GuardBench checks how well models handle safety challenges from misinformation and self-harm to sexual content and corruption.\n\n"+\
23
- "Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.\n\n"+\
24
- "We track how accurate they are, how often they make mistakes, and how fast they respond.\n"
25
 
26
 
27
  EVALUATION_QUEUE_TEXT = """
28
  ## Submit Your Model
29
 
30
- To add your model to the GuardBench leaderboard:
31
 
32
- 1. Run your evaluation using the GuardBench framework at https://github.com/whitecircle-ai/guard-bench
33
  2. Upload your run results in .jsonl format using this form.
34
  3. Once validated, your model will appear on the leaderboard.
35
 
36
  ### ✉️✨ Ready? Upload your results below!
37
  """
38
 
39
- CITATION_BUTTON_LABEL = "Cite GuardBench"
40
 
41
  CITATION_BUTTON_TEXT = """
42
- @misc{guardbench2023,
43
- author = {GuardBench Team},
44
- title = {GuardBench: Comprehensive Benchmark for LLM Safety Guardrails},
45
- year = {2023},
46
  publisher = {GitHub},
47
  journal = {GitHub repository},
48
- howpublished = {\\url{https://github.com/huggingface/guard-bench}}
49
  }
50
  """
 
4
 
5
  TITLE = """
6
  <div style="text-align: center; margin-bottom: 1rem">
7
+ <h1>CircleGuardBench Leaderboard</h1>
8
  </div>
9
  """
10
 
11
  INTRODUCTION_TEXT = """
12
  ## Introduction
13
 
14
+ CircleGuardBench is a comprehensive benchmark for evaluating the protection capabilities of large language model (LLM) guard systems.
15
  This leaderboard tracks model performance across various safety categories, including harmful content detection,
16
  jailbreak resistance, and more.
17
 
 
19
  across multiple categories and test scenarios.
20
  """
21
 
22
+ LLM_BENCHMARKS_TEXT = "CircleGuardBench is the first-of-its-kind benchmark for evaluating the protection capabilities of large language model (LLM) guard systems. It tests how well guard models block harmful content, resist jailbreaks, avoid false positives, and operate efficiently in real-time environments on a taxonomy close to real-world data."
 
 
23
 
24
 
25
  EVALUATION_QUEUE_TEXT = """
26
  ## Submit Your Model
27
 
28
+ To add your model to the CircleGuardBench leaderboard:
29
 
30
+ 1. Run your evaluation using the CircleGuardBench framework at https://github.com/whitecircle-ai/circle-guard-bench
31
  2. Upload your run results in .jsonl format using this form.
32
  3. Once validated, your model will appear on the leaderboard.
33
 
34
  ### ✉️✨ Ready? Upload your results below!
35
  """
36
 
37
+ CITATION_BUTTON_LABEL = "Cite CircleGuardBench"
38
 
39
  CITATION_BUTTON_TEXT = """
40
+ @misc{circleguardbench2025,
41
+ author = {whitecircle-ai},
42
+ title = {CircleGuardBench: Comprehensive Benchmark for LLM Safety Guardrails},
43
+ year = {2025},
44
  publisher = {GitHub},
45
  journal = {GitHub repository},
46
+ howpublished = {\\url{https://github.com/whitecircle-ai/circle-guard-bench}}
47
  }
48
  """