Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -68,9 +68,9 @@ def run_huggingface_model(model, messages, max_tokens, temperature, top_p):
|
|
68 |
payload = {
|
69 |
"messages": messages,
|
70 |
"max_tokens": max_tokens,
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
# "inputs": json.dumps(messages),
|
75 |
# "seed": random.randint(1, 1000000000),
|
76 |
# "parameters": {
|
@@ -84,13 +84,15 @@ def run_huggingface_model(model, messages, max_tokens, temperature, top_p):
|
|
84 |
response = requests.post(model, headers=headers, json=payload, timeout=30)
|
85 |
|
86 |
if response.status_code != 200:
|
87 |
-
print(response.content)
|
88 |
response = json.loads(response.content)
|
89 |
print("ERROR: " + response["error"])
|
90 |
else:
|
91 |
print(response.content)
|
92 |
-
|
93 |
-
|
|
|
|
|
|
|
94 |
return result
|
95 |
|
96 |
def run_together_model(model, messages, max_tokens, temperature, top_p):
|
|
|
68 |
payload = {
|
69 |
"messages": messages,
|
70 |
"max_tokens": max_tokens,
|
71 |
+
"temperature": temperature,
|
72 |
+
"top_p": top_p,
|
73 |
+
"stream": False
|
74 |
# "inputs": json.dumps(messages),
|
75 |
# "seed": random.randint(1, 1000000000),
|
76 |
# "parameters": {
|
|
|
84 |
response = requests.post(model, headers=headers, json=payload, timeout=30)
|
85 |
|
86 |
if response.status_code != 200:
|
|
|
87 |
response = json.loads(response.content)
|
88 |
print("ERROR: " + response["error"])
|
89 |
else:
|
90 |
print(response.content)
|
91 |
+
try:
|
92 |
+
response = json.loads(response.content)
|
93 |
+
result = response["choices"][0]["message"]["content"]
|
94 |
+
catch:
|
95 |
+
result = response.content
|
96 |
return result
|
97 |
|
98 |
def run_together_model(model, messages, max_tokens, temperature, top_p):
|