Sambhavnoobcoder commited on
Commit
568558e
·
verified ·
1 Parent(s): f52965e

removed temp caching

Browse files
Files changed (1) hide show
  1. app.py +3 -5
app.py CHANGED
@@ -84,7 +84,7 @@ def handle_query(query, faiss_index, embeddings_texts, model):
84
  prompt=f"Based on the following context:\n\n{combined_text}\n\nAnswer the following question: {query}",
85
  max_output_tokens=200
86
  )
87
- generated_text = response.result
88
  except Exception as e:
89
  print(f"Error generating text: {e}")
90
  generated_text = "An error occurred while generating the response."
@@ -105,7 +105,7 @@ def generate_concise_response(prompt, context):
105
  prompt=f"{prompt}\n\nContext: {context}\n\nAnswer:",
106
  max_output_tokens=200
107
  )
108
- return response.result
109
  except Exception as e:
110
  print(f"Error generating concise response: {e}")
111
  return "An error occurred while generating the concise response."
@@ -128,14 +128,13 @@ def chatbot(message, history):
128
  response, sources = handle_query(message, faiss_index, all_texts, embedding_model)
129
  print("Query:", message)
130
  print("Response:", response)
131
- total_text = response
132
  if sources:
133
  print("Sources:", sources)
134
  relevant_source = ""
135
  for source in sources:
136
  relevant_source += source + "\n"
137
  total_text += "\n\nSources:\n" + relevant_source
138
-
139
  else:
140
  print("Sources: None of the provided sources were used.")
141
  print("----")
@@ -149,7 +148,6 @@ def chatbot(message, history):
149
 
150
  return total_text
151
 
152
-
153
  iface = gr.ChatInterface(
154
  chatbot,
155
  title="LLM Research Assistant",
 
84
  prompt=f"Based on the following context:\n\n{combined_text}\n\nAnswer the following question: {query}",
85
  max_output_tokens=200
86
  )
87
+ generated_text = response.result if response else "No response generated."
88
  except Exception as e:
89
  print(f"Error generating text: {e}")
90
  generated_text = "An error occurred while generating the response."
 
105
  prompt=f"{prompt}\n\nContext: {context}\n\nAnswer:",
106
  max_output_tokens=200
107
  )
108
+ return response.result if response else "No response generated."
109
  except Exception as e:
110
  print(f"Error generating concise response: {e}")
111
  return "An error occurred while generating the concise response."
 
128
  response, sources = handle_query(message, faiss_index, all_texts, embedding_model)
129
  print("Query:", message)
130
  print("Response:", response)
131
+ total_text = response if response else "No response generated."
132
  if sources:
133
  print("Sources:", sources)
134
  relevant_source = ""
135
  for source in sources:
136
  relevant_source += source + "\n"
137
  total_text += "\n\nSources:\n" + relevant_source
 
138
  else:
139
  print("Sources: None of the provided sources were used.")
140
  print("----")
 
148
 
149
  return total_text
150
 
 
151
  iface = gr.ChatInterface(
152
  chatbot,
153
  title="LLM Research Assistant",