reduced top_p to 0.8
Browse files
app.py
CHANGED
@@ -45,7 +45,7 @@ def llama_generation(input_text: str,
|
|
45 |
output_ids = llama_model.generate(input_ids=input_ids,
|
46 |
max_new_tokens=256,
|
47 |
temperature=0.5,
|
48 |
-
top_p=0.
|
49 |
repetition_penalty=2.0)
|
50 |
|
51 |
# Decode
|
|
|
45 |
output_ids = llama_model.generate(input_ids=input_ids,
|
46 |
max_new_tokens=256,
|
47 |
temperature=0.5,
|
48 |
+
top_p=0.8,
|
49 |
repetition_penalty=2.0)
|
50 |
|
51 |
# Decode
|