apsys commited on
Commit
0c12f85
·
1 Parent(s): 3bbc6ce
Files changed (1) hide show
  1. src/about.py +2 -2
src/about.py CHANGED
@@ -19,8 +19,8 @@ Models are evaluated on their ability to properly refuse harmful requests and de
19
  across multiple categories and test scenarios.
20
  """
21
 
22
- LLM_BENCHMARKS_TEXT = "GuardBench checks how well models handle safety challenges — from misinformation and self-harm to sexual content and corruption.\n"+\
23
- "Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.\n"+\
24
  "We track how accurate they are, how often they make mistakes, and how fast they respond.\n"
25
 
26
 
 
19
  across multiple categories and test scenarios.
20
  """
21
 
22
+ LLM_BENCHMARKS_TEXT = "GuardBench checks how well models handle safety challenges — from misinformation and self-harm to sexual content and corruption.\n\n"+\
23
+ "Models are tested with regular and adversarial prompts to see if they can avoid saying harmful things.\n\n"+\
24
  "We track how accurate they are, how often they make mistakes, and how fast they respond.\n"
25
 
26