ai: Ship internal training data.
Browse files
jarvis.py
CHANGED
@@ -39,6 +39,8 @@ from pptx import Presentation
|
|
39 |
|
40 |
os.system("apt-get update -q -y && apt-get install -q -y tesseract-ocr tesseract-ocr-eng tesseract-ocr-ind libleptonica-dev libtesseract-dev")
|
41 |
|
|
|
|
|
42 |
LINUX_SERVER_HOSTS = [host for host in json.loads(os.getenv("LINUX_SERVER_HOST", "[]")) if host]
|
43 |
LINUX_SERVER_HOSTS_MARKED = set()
|
44 |
LINUX_SERVER_HOSTS_ATTEMPTS = {}
|
@@ -158,7 +160,6 @@ async def fetch_response_async(host, provider_key, selected_model, messages, mod
|
|
158 |
raise
|
159 |
|
160 |
async def chat_with_model_async(history, user_input, selected_model_display, sess):
|
161 |
-
global ACTIVE_CANDIDATE
|
162 |
if not get_available_items(LINUX_SERVER_PROVIDER_KEYS, LINUX_SERVER_PROVIDER_KEYS_MARKED) or not get_available_items(LINUX_SERVER_HOSTS, LINUX_SERVER_HOSTS_MARKED):
|
163 |
return RESPONSES["RESPONSE_3"]
|
164 |
if not hasattr(sess, "session_id"):
|
@@ -167,7 +168,10 @@ async def chat_with_model_async(history, user_input, selected_model_display, ses
|
|
167 |
model_config = MODEL_CONFIG.get(selected_model, DEFAULT_CONFIG)
|
168 |
messages = [{"role": "user", "content": user} for user, _ in history]
|
169 |
messages += [{"role": "assistant", "content": assistant} for _, assistant in history if assistant]
|
|
|
|
|
170 |
messages.append({"role": "user", "content": user_input})
|
|
|
171 |
if ACTIVE_CANDIDATE is not None:
|
172 |
try:
|
173 |
return await fetch_response_async(ACTIVE_CANDIDATE[0], ACTIVE_CANDIDATE[1], selected_model, messages, model_config, sess.session_id)
|
|
|
39 |
|
40 |
os.system("apt-get update -q -y && apt-get install -q -y tesseract-ocr tesseract-ocr-eng tesseract-ocr-ind libleptonica-dev libtesseract-dev")
|
41 |
|
42 |
+
INTERNAL_TRAINING_DATA = os.getenv("INTERNAL_TRAINING_DATA", "")
|
43 |
+
|
44 |
LINUX_SERVER_HOSTS = [host for host in json.loads(os.getenv("LINUX_SERVER_HOST", "[]")) if host]
|
45 |
LINUX_SERVER_HOSTS_MARKED = set()
|
46 |
LINUX_SERVER_HOSTS_ATTEMPTS = {}
|
|
|
160 |
raise
|
161 |
|
162 |
async def chat_with_model_async(history, user_input, selected_model_display, sess):
|
|
|
163 |
if not get_available_items(LINUX_SERVER_PROVIDER_KEYS, LINUX_SERVER_PROVIDER_KEYS_MARKED) or not get_available_items(LINUX_SERVER_HOSTS, LINUX_SERVER_HOSTS_MARKED):
|
164 |
return RESPONSES["RESPONSE_3"]
|
165 |
if not hasattr(sess, "session_id"):
|
|
|
168 |
model_config = MODEL_CONFIG.get(selected_model, DEFAULT_CONFIG)
|
169 |
messages = [{"role": "user", "content": user} for user, _ in history]
|
170 |
messages += [{"role": "assistant", "content": assistant} for _, assistant in history if assistant]
|
171 |
+
if INTERNAL_TRAINING_DATA:
|
172 |
+
messages.insert(0, {"role": "system", "content": INTERNAL_TRAINING_DATA})
|
173 |
messages.append({"role": "user", "content": user_input})
|
174 |
+
global ACTIVE_CANDIDATE
|
175 |
if ACTIVE_CANDIDATE is not None:
|
176 |
try:
|
177 |
return await fetch_response_async(ACTIVE_CANDIDATE[0], ACTIVE_CANDIDATE[1], selected_model, messages, model_config, sess.session_id)
|