Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -303,7 +303,7 @@ agents =[
|
|
303 |
"PYTHON_CODE_DEV"
|
304 |
]
|
305 |
def generate(
|
306 |
-
|
307 |
):
|
308 |
seed = random.randint(1,1111111111111111)
|
309 |
|
@@ -320,14 +320,27 @@ def generate(
|
|
320 |
temperature = 1e-2
|
321 |
top_p = float(top_p)
|
322 |
|
323 |
-
|
324 |
-
|
325 |
-
|
326 |
-
|
327 |
-
|
328 |
-
|
329 |
-
|
330 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
331 |
|
332 |
formatted_prompt = format_prompt(prompt, history, max_history_turns=5) # Truncated history
|
333 |
logging.info(f"Formatted Prompt: {formatted_prompt}")
|
|
|
303 |
"PYTHON_CODE_DEV"
|
304 |
]
|
305 |
def generate(
|
306 |
+
prompt, history, agent_name=agents[0], sys_prompt="", temperature=0.9, max_new_tokens=2048, top_p=0.95, repetition_penalty=1.0,
|
307 |
):
|
308 |
seed = random.randint(1,1111111111111111)
|
309 |
|
|
|
320 |
temperature = 1e-2
|
321 |
top_p = float(top_p)
|
322 |
|
323 |
+
|
324 |
+
|
325 |
+
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
326 |
+
"""Generates text in chunks to avoid token limit errors."""
|
327 |
+
sentences = nltk.sent_tokenize(input_text)
|
328 |
+
generated_text = []
|
329 |
+
generator = pipeline('text-generation', model=model)
|
330 |
+
|
331 |
+
for sentence in sentences:
|
332 |
+
# Tokenize the sentence and check if it's within the limit
|
333 |
+
tokens = generator.tokenizer(sentence).input_ids
|
334 |
+
if len(tokens) + max_tokens_to_generate <= 32768:
|
335 |
+
# Generate text for this chunk
|
336 |
+
response = generator(sentence, max_length=max_tokens_to_generate, **generation_parameters)
|
337 |
+
generated_text.append(response[0]['generated_text'])
|
338 |
+
else:
|
339 |
+
# Handle cases where the sentence is too long
|
340 |
+
# You could split the sentence further or skip it
|
341 |
+
print(f"Sentence too long: {sentence}")
|
342 |
+
|
343 |
+
return ''.join(generated_text)
|
344 |
|
345 |
formatted_prompt = format_prompt(prompt, history, max_history_turns=5) # Truncated history
|
346 |
logging.info(f"Formatted Prompt: {formatted_prompt}")
|