Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -126,8 +126,7 @@ def generate(prompt, history, agent_name=agents[0], sys_prompt="", temperature=0
|
|
126 |
|
127 |
if VERBOSE:
|
128 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
129 |
-
return resp
|
130 |
-
|
131 |
|
132 |
def compress_history(purpose, task, history, directory):
|
133 |
resp = run_gpt(
|
@@ -332,8 +331,7 @@ def generate(
|
|
332 |
|
333 |
if VERBOSE:
|
334 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
335 |
-
return resp
|
336 |
-
|
337 |
|
338 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
339 |
"""Generates text in chunks to avoid token limit errors."""
|
|
|
126 |
|
127 |
if VERBOSE:
|
128 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
129 |
+
return resp # Return resp here
|
|
|
130 |
|
131 |
def compress_history(purpose, task, history, directory):
|
132 |
resp = run_gpt(
|
|
|
331 |
|
332 |
if VERBOSE:
|
333 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
334 |
+
return resp # Return resp here
|
|
|
335 |
|
336 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
337 |
"""Generates text in chunks to avoid token limit errors."""
|