acecalisto3 commited on
Commit
703bf32
·
verified ·
1 Parent(s): 724611a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +22 -9
app.py CHANGED
@@ -303,7 +303,7 @@ agents =[
303
  "PYTHON_CODE_DEV"
304
  ]
305
  def generate(
306
- prompt, history, agent_name=agents[0], sys_prompt="", temperature=0.9, max_new_tokens=2048, top_p=0.95, repetition_penalty=1.0,
307
  ):
308
  seed = random.randint(1,1111111111111111)
309
 
@@ -320,14 +320,27 @@ def generate(
320
  temperature = 1e-2
321
  top_p = float(top_p)
322
 
323
- generate_kwargs = dict(
324
- temperature=temperature,
325
- max_new_tokens=max_new_tokens, # Reduced max_new_tokens
326
- top_p=top_p,
327
- repetition_penalty=repetition_penalty,
328
- do_sample=True,
329
- seed=seed,
330
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
331
 
332
  formatted_prompt = format_prompt(prompt, history, max_history_turns=5) # Truncated history
333
  logging.info(f"Formatted Prompt: {formatted_prompt}")
 
303
  "PYTHON_CODE_DEV"
304
  ]
305
  def generate(
306
+ prompt, history, agent_name=agents[0], sys_prompt="", temperature=0.9, max_new_tokens=2048, top_p=0.95, repetition_penalty=1.0,
307
  ):
308
  seed = random.randint(1,1111111111111111)
309
 
 
320
  temperature = 1e-2
321
  top_p = float(top_p)
322
 
323
+
324
+
325
+ def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
326
+ """Generates text in chunks to avoid token limit errors."""
327
+ sentences = nltk.sent_tokenize(input_text)
328
+ generated_text = []
329
+ generator = pipeline('text-generation', model=model)
330
+
331
+ for sentence in sentences:
332
+ # Tokenize the sentence and check if it's within the limit
333
+ tokens = generator.tokenizer(sentence).input_ids
334
+ if len(tokens) + max_tokens_to_generate <= 32768:
335
+ # Generate text for this chunk
336
+ response = generator(sentence, max_length=max_tokens_to_generate, **generation_parameters)
337
+ generated_text.append(response[0]['generated_text'])
338
+ else:
339
+ # Handle cases where the sentence is too long
340
+ # You could split the sentence further or skip it
341
+ print(f"Sentence too long: {sentence}")
342
+
343
+ return ''.join(generated_text)
344
 
345
  formatted_prompt = format_prompt(prompt, history, max_history_turns=5) # Truncated history
346
  logging.info(f"Formatted Prompt: {formatted_prompt}")