acecalisto3 commited on
Commit
894a973
·
verified ·
1 Parent(s): a829eeb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -4
app.py CHANGED
@@ -126,8 +126,7 @@ def generate(prompt, history, agent_name=agents[0], sys_prompt="", temperature=0
126
 
127
  if VERBOSE:
128
  logging.info(LOG_RESPONSE.format(resp)) # Log the response
129
- return resp
130
-
131
 
132
  def compress_history(purpose, task, history, directory):
133
  resp = run_gpt(
@@ -332,8 +331,7 @@ def generate(
332
 
333
  if VERBOSE:
334
  logging.info(LOG_RESPONSE.format(resp)) # Log the response
335
- return resp
336
-
337
 
338
  def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
339
  """Generates text in chunks to avoid token limit errors."""
 
126
 
127
  if VERBOSE:
128
  logging.info(LOG_RESPONSE.format(resp)) # Log the response
129
+ return resp # Return resp here
 
130
 
131
  def compress_history(purpose, task, history, directory):
132
  resp = run_gpt(
 
331
 
332
  if VERBOSE:
333
  logging.info(LOG_RESPONSE.format(resp)) # Log the response
334
+ return resp # Return resp here
 
335
 
336
  def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
337
  """Generates text in chunks to avoid token limit errors."""