Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -24,7 +24,7 @@ class OllamaClient:
|
|
24 |
print(f"Error listing models: {e}")
|
25 |
return []
|
26 |
|
27 |
-
def chat_completion(self, messages, max_tokens=4000, stream=True, temperature=0.3, top_p=0.
|
28 |
# Convert messages to Ollama format
|
29 |
ollama_messages = []
|
30 |
for msg in messages:
|
@@ -343,9 +343,9 @@ with gr.Blocks(theme="gstaff/xkcd",
|
|
343 |
title="Offline Sensitive Survey Data Analysis") as demo:
|
344 |
gr.HTML(
|
345 |
"""
|
346 |
-
<div style="text-align: center; max-width:
|
347 |
<h1 style="font-size: 3em; font-weight: 600; margin: 0.5em;">Offline Sensitive Survey Data Analysis</h1>
|
348 |
-
<h3 style="font-size: 1.2em; margin: 1em;">Leveraging Ollama Inference Server</h3>
|
349 |
</div>
|
350 |
"""
|
351 |
)
|
@@ -366,9 +366,9 @@ with gr.Blocks(theme="gstaff/xkcd",
|
|
366 |
placeholder="Ask me anything about the uploaded data file... ",
|
367 |
container=False
|
368 |
)
|
369 |
-
with gr.Row():
|
370 |
-
clear = gr.ClearButton([msg, chatbot])
|
371 |
send = gr.Button("Send")
|
|
|
372 |
|
373 |
with gr.Column(scale=1):
|
374 |
gr.Markdown("### Upload File \nSupport: CSV, Parquet files, Text")
|
@@ -384,13 +384,13 @@ with gr.Blocks(theme="gstaff/xkcd",
|
|
384 |
choices=[],
|
385 |
interactive=True
|
386 |
)
|
387 |
-
refresh_models = gr.Button("
|
388 |
|
389 |
with gr.Accordion("Advanced Settings ⚙️", open=False):
|
390 |
system_message = gr.Textbox(label="Override System Message 📝", value="")
|
391 |
-
max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens
|
392 |
-
temperature = gr.Slider(minimum=0, maximum=1, value=0.3, label="Temperature
|
393 |
-
top_p = gr.Slider(minimum=0, maximum=1, value=0.
|
394 |
|
395 |
# Function to load available models
|
396 |
def load_models():
|
|
|
24 |
print(f"Error listing models: {e}")
|
25 |
return []
|
26 |
|
27 |
+
def chat_completion(self, messages, max_tokens=4000, stream=True, temperature=0.3, top_p=0.7):
|
28 |
# Convert messages to Ollama format
|
29 |
ollama_messages = []
|
30 |
for msg in messages:
|
|
|
343 |
title="Offline Sensitive Survey Data Analysis") as demo:
|
344 |
gr.HTML(
|
345 |
"""
|
346 |
+
<div style="text-align: center; max-width: 1000px; margin: 0 auto;">
|
347 |
<h1 style="font-size: 3em; font-weight: 600; margin: 0.5em;">Offline Sensitive Survey Data Analysis</h1>
|
348 |
+
<h3 style="font-size: 1.2em; margin: 1em;">Leveraging your Local Ollama Inference Server</h3>
|
349 |
</div>
|
350 |
"""
|
351 |
)
|
|
|
366 |
placeholder="Ask me anything about the uploaded data file... ",
|
367 |
container=False
|
368 |
)
|
369 |
+
with gr.Row():
|
|
|
370 |
send = gr.Button("Send")
|
371 |
+
clear = gr.ClearButton([msg, chatbot])
|
372 |
|
373 |
with gr.Column(scale=1):
|
374 |
gr.Markdown("### Upload File \nSupport: CSV, Parquet files, Text")
|
|
|
384 |
choices=[],
|
385 |
interactive=True
|
386 |
)
|
387 |
+
refresh_models = gr.Button("Select Model")
|
388 |
|
389 |
with gr.Accordion("Advanced Settings ⚙️", open=False):
|
390 |
system_message = gr.Textbox(label="Override System Message 📝", value="")
|
391 |
+
max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens (maximum number of words for generated response)")
|
392 |
+
temperature = gr.Slider(minimum=0, maximum=1, value=0.3, label="Temperature (higher = more creative)")
|
393 |
+
top_p = gr.Slider(minimum=0, maximum=1, value=0.7, label="Top P (word choices by probability threshold)")
|
394 |
|
395 |
# Function to load available models
|
396 |
def load_models():
|