Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -84,7 +84,7 @@ else:
|
|
84 |
import pickle
|
85 |
with open('nn_model.pkl', 'rb') as f:
|
86 |
nn = pickle.load(f)
|
87 |
-
|
88 |
def get_relevant_documents(query, k=5):
|
89 |
"""
|
90 |
Retrieves the k most relevant documents to the query.
|
@@ -101,7 +101,7 @@ def get_relevant_documents(query, k=5):
|
|
101 |
elapsed_time = time.time() - start_time
|
102 |
print(f"[PERF] get_relevant_documents took {elapsed_time:.2f} seconds")
|
103 |
return relevant_docs
|
104 |
-
|
105 |
def generate_response(question, history):
|
106 |
import time
|
107 |
start_time = time.time()
|
@@ -115,7 +115,7 @@ def generate_response(question, history):
|
|
115 |
elapsed_time = time.time() - start_time
|
116 |
print(f"[PERF] generate_response took {elapsed_time:.2f} seconds")
|
117 |
return response
|
118 |
-
|
119 |
def _generate_response_gpu(question, history):
|
120 |
print(f"\n[LOG] Received question: {question}")
|
121 |
|
@@ -201,7 +201,7 @@ Provide the AutoGen v0.4 agent code that fulfills the user's request. Utilize fe
|
|
201 |
# Update chat history with new message pair
|
202 |
history.append((question, response))
|
203 |
return history
|
204 |
-
|
205 |
def _generate_response_cpu(question, history):
|
206 |
print(f"[LOG] Running on CPU")
|
207 |
try:
|
|
|
84 |
import pickle
|
85 |
with open('nn_model.pkl', 'rb') as f:
|
86 |
nn = pickle.load(f)
|
87 |
+
@spaces.GPU
|
88 |
def get_relevant_documents(query, k=5):
|
89 |
"""
|
90 |
Retrieves the k most relevant documents to the query.
|
|
|
101 |
elapsed_time = time.time() - start_time
|
102 |
print(f"[PERF] get_relevant_documents took {elapsed_time:.2f} seconds")
|
103 |
return relevant_docs
|
104 |
+
@spaces.GPU
|
105 |
def generate_response(question, history):
|
106 |
import time
|
107 |
start_time = time.time()
|
|
|
115 |
elapsed_time = time.time() - start_time
|
116 |
print(f"[PERF] generate_response took {elapsed_time:.2f} seconds")
|
117 |
return response
|
118 |
+
@spaces.GPU
|
119 |
def _generate_response_gpu(question, history):
|
120 |
print(f"\n[LOG] Received question: {question}")
|
121 |
|
|
|
201 |
# Update chat history with new message pair
|
202 |
history.append((question, response))
|
203 |
return history
|
204 |
+
@spaces.GPU
|
205 |
def _generate_response_cpu(question, history):
|
206 |
print(f"[LOG] Running on CPU")
|
207 |
try:
|