apsys commited on
Commit
3a2adee
·
1 Parent(s): afd100d
Files changed (1) hide show
  1. src/about.py +3 -7
src/about.py CHANGED
@@ -21,9 +21,7 @@ across multiple categories and test scenarios.
21
 
22
  LLM_BENCHMARKS_TEXT = """
23
  GuardBench checks how well models handle safety challenges — from misinformation and self-harm to sexual content and corruption.
24
-
25
  Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.
26
-
27
  We track how accurate they are, how often they make mistakes, and how fast they respond.
28
  """
29
 
@@ -32,11 +30,9 @@ EVALUATION_QUEUE_TEXT = """
32
 
33
  To add your model to the GuardBench leaderboard:
34
 
35
- Run your evaluation using the GuardBench framework at https://github.com/whitecircle-ai/guard-bench
36
-
37
- Upload your run results in .jsonl format using this form.
38
-
39
- Once validated, your model will appear on the leaderboard.
40
 
41
  ### ✉️✨ Ready? Upload your results below!
42
  """
 
21
 
22
  LLM_BENCHMARKS_TEXT = """
23
  GuardBench checks how well models handle safety challenges — from misinformation and self-harm to sexual content and corruption.
 
24
  Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.
 
25
  We track how accurate they are, how often they make mistakes, and how fast they respond.
26
  """
27
 
 
30
 
31
  To add your model to the GuardBench leaderboard:
32
 
33
+ 1. Run your evaluation using the GuardBench framework at https://github.com/whitecircle-ai/guard-bench
34
+ 2. Upload your run results in .jsonl format using this form.
35
+ 3. Once validated, your model will appear on the leaderboard.
 
 
36
 
37
  ### ✉️✨ Ready? Upload your results below!
38
  """