Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -23,7 +23,7 @@ def generate_text(prompt, temperature, top_p):
|
|
23 |
prompt_length = len(generated_text)
|
24 |
|
25 |
if input_tokens.size(1) > 512:
|
26 |
-
|
27 |
|
28 |
for _ in range(80): # Adjust the range to control the number of tokens generated
|
29 |
with torch.no_grad():
|
|
|
23 |
prompt_length = len(generated_text)
|
24 |
|
25 |
if input_tokens.size(1) > 512:
|
26 |
+
generated_text = "ERROR, CONTEXT SIZE EXCEEDED"
|
27 |
|
28 |
for _ in range(80): # Adjust the range to control the number of tokens generated
|
29 |
with torch.no_grad():
|