awacke1 commited on
Commit
0ed0f57
·
1 Parent(s): fe1b62b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -4
app.py CHANGED
@@ -187,7 +187,7 @@ def chat_with_model(prompt, document_section, model_choice='Llama-2-7b-chat-hf')
187
  res_box.markdown(f'*{result}*')
188
 
189
  except:
190
- st.write('Stream llm issue')
191
 
192
  full_reply_content = result
193
  st.write("Elapsed time:")
@@ -198,9 +198,6 @@ def chat_with_model(prompt, document_section, model_choice='Llama-2-7b-chat-hf')
198
  readitaloud(full_reply_content)
199
  return result
200
 
201
- except:
202
- st.write('Llama model is asleep. Starting up now on A10 - please give 5 minutes then retry as KEDA scales up from zero to activate running container(s).')
203
-
204
  # Chat and Chat with files
205
  def chat_with_model2(prompt, document_section, model_choice='gpt-3.5-turbo'):
206
  model = model_choice
@@ -236,6 +233,7 @@ def chat_with_model2(prompt, document_section, model_choice='gpt-3.5-turbo'):
236
  res_box.markdown(f'*{result}*')
237
  except:
238
  st.write(' ')
 
239
  full_reply_content = all_content
240
  st.write("Elapsed time:")
241
  st.write(time.time() - start_time)
 
187
  res_box.markdown(f'*{result}*')
188
 
189
  except:
190
+ st.write('.')
191
 
192
  full_reply_content = result
193
  st.write("Elapsed time:")
 
198
  readitaloud(full_reply_content)
199
  return result
200
 
 
 
 
201
  # Chat and Chat with files
202
  def chat_with_model2(prompt, document_section, model_choice='gpt-3.5-turbo'):
203
  model = model_choice
 
233
  res_box.markdown(f'*{result}*')
234
  except:
235
  st.write(' ')
236
+
237
  full_reply_content = all_content
238
  st.write("Elapsed time:")
239
  st.write(time.time() - start_time)