Spaces:
Sleeping
Sleeping
Update app.js
Browse files
app.js
CHANGED
@@ -1,459 +1,499 @@
|
|
1 |
// app.js
|
2 |
-
|
3 |
-
const http = require('http');
|
4 |
-
const { v4: uuidv4 } = require('uuid');
|
5 |
const fetch = require('node-fetch');
|
|
|
|
|
6 |
|
7 |
-
|
8 |
-
const
|
9 |
-
|
10 |
-
//
|
11 |
-
const
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
];
|
25 |
|
26 |
-
//
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
object: "list",
|
37 |
-
data: models,
|
38 |
-
};
|
39 |
-
|
40 |
-
res.writeHead(200, { 'Content-Type': 'application/json' });
|
41 |
-
res.end(JSON.stringify(responseBody));
|
42 |
-
}
|
43 |
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
openaiPayload.messages.forEach(msg => {
|
53 |
-
if (msg.role === 'system') {
|
54 |
-
// 保存system消息,Julep可能有特殊处理方式
|
55 |
-
systemMessage = msg;
|
56 |
-
} else {
|
57 |
-
userMessages.push(msg);
|
58 |
-
}
|
59 |
-
});
|
60 |
-
}
|
61 |
-
|
62 |
-
// 创建Julep格式的payload
|
63 |
-
const julepPayload = {
|
64 |
-
messages: userMessages.map(msg => ({
|
65 |
-
role: msg.role,
|
66 |
-
content: msg.content,
|
67 |
-
name: msg.name,
|
68 |
-
tool_call_id: msg.tool_call_id,
|
69 |
-
tool_calls: msg.tool_calls ? msg.tool_calls.map(tc => ({
|
70 |
-
type: tc.type,
|
71 |
-
function: tc.function ? {
|
72 |
-
name: tc.function.name,
|
73 |
-
arguments: tc.function.arguments
|
74 |
-
} : undefined,
|
75 |
-
integration: tc.integration,
|
76 |
-
system: tc.system,
|
77 |
-
api_call: tc.api_call,
|
78 |
-
computer_20241022: tc.computer_20241022,
|
79 |
-
text_editor_20241022: tc.text_editor_20241022,
|
80 |
-
bash_20241022: tc.bash_20241022,
|
81 |
-
id: tc.id,
|
82 |
-
})) : undefined,
|
83 |
-
})),
|
84 |
-
tools: openaiPayload.tools,
|
85 |
-
tool_choice: openaiPayload.tool_choice,
|
86 |
-
recall: openaiPayload.recall,
|
87 |
-
save: openaiPayload.save,
|
88 |
-
model: openaiPayload.model,
|
89 |
-
stream: openaiPayload.stream,
|
90 |
-
stop: openaiPayload.stop,
|
91 |
-
seed: openaiPayload.seed,
|
92 |
-
max_tokens: openaiPayload.max_tokens,
|
93 |
-
logit_bias: openaiPayload.logit_bias,
|
94 |
-
response_format: openaiPayload.response_format,
|
95 |
-
agent: openaiPayload.agent,
|
96 |
-
repetition_penalty: openaiPayload.repetition_penalty,
|
97 |
-
length_penalty: openaiPayload.length_penalty,
|
98 |
-
min_p: openaiPayload.min_p,
|
99 |
-
frequency_penalty: openaiPayload.frequency_penalty,
|
100 |
-
presence_penalty: openaiPayload.presence_penalty,
|
101 |
-
temperature: openaiPayload.temperature,
|
102 |
-
top_p: openaiPayload.top_p,
|
103 |
-
};
|
104 |
-
|
105 |
-
// 如果有system消息,添加到Julep payload中
|
106 |
-
// Julep API可能使用system_prompt字段或其他方式处理system消息
|
107 |
-
if (systemMessage) {
|
108 |
-
julepPayload.system_prompt = systemMessage.content;
|
109 |
-
}
|
110 |
-
|
111 |
-
// 删除可能错误包含的session_id
|
112 |
-
delete julepPayload.session_id;
|
113 |
-
|
114 |
-
return julepPayload;
|
115 |
}
|
116 |
|
117 |
-
|
118 |
-
|
119 |
-
|
120 |
-
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
-
|
125 |
-
index: choice.index,
|
126 |
-
message: {
|
127 |
-
role: choice.message?.role || "assistant",
|
128 |
-
content: choice.message?.content || "",
|
129 |
-
tool_calls: choice.message?.tool_calls ? choice.message.tool_calls.map(tc => ({
|
130 |
-
id: tc.id,
|
131 |
-
type: tc.type,
|
132 |
-
function: tc.function ? {
|
133 |
-
name: tc.function.name,
|
134 |
-
arguments: tc.function.arguments
|
135 |
-
} : undefined,
|
136 |
-
})) : undefined,
|
137 |
-
},
|
138 |
-
finish_reason: choice.finish_reason,
|
139 |
-
})),
|
140 |
-
usage: julepData.usage ? {
|
141 |
-
prompt_tokens: julepData.usage.prompt_tokens,
|
142 |
-
completion_tokens: julepData.usage.completion_tokens,
|
143 |
-
total_tokens: julepData.usage.total_tokens,
|
144 |
-
} : undefined,
|
145 |
-
};
|
146 |
-
|
147 |
-
return openaiResponse;
|
148 |
}
|
149 |
|
150 |
-
//
|
151 |
-
function
|
152 |
-
|
153 |
-
id: sessionId,
|
154 |
-
object: "chat.completion.chunk",
|
155 |
-
created: Math.floor(Date.now() / 1000),
|
156 |
-
model: model,
|
157 |
-
choices: julepChunk.choices.map(choice => {
|
158 |
-
const openaiChoice = {
|
159 |
-
index: choice.index,
|
160 |
-
delta: {
|
161 |
-
role: choice.delta?.role,
|
162 |
-
content: choice.delta?.content,
|
163 |
-
tool_calls: choice.delta?.tool_calls ? choice.delta.tool_calls.map(tc => ({
|
164 |
-
id: tc.id,
|
165 |
-
type: tc.type,
|
166 |
-
function: tc.function ? {
|
167 |
-
name: tc.function.name,
|
168 |
-
arguments: tc.function.arguments
|
169 |
-
} : undefined,
|
170 |
-
})) : undefined,
|
171 |
-
},
|
172 |
-
finish_reason: choice.finish_reason,
|
173 |
-
};
|
174 |
-
|
175 |
-
// 清理空的delta字段
|
176 |
-
if (openaiChoice.delta.role === undefined) delete openaiChoice.delta.role;
|
177 |
-
if (openaiChoice.delta.content === undefined) delete openaiChoice.delta.content;
|
178 |
-
if (openaiChoice.delta.tool_calls === undefined) delete openaiChoice.delta.tool_calls;
|
179 |
-
if (Object.keys(openaiChoice.delta).length === 0 && openaiChoice.finish_reason === undefined) {
|
180 |
-
delete openaiChoice.delta;
|
181 |
-
}
|
182 |
-
|
183 |
-
return openaiChoice;
|
184 |
-
}),
|
185 |
-
};
|
186 |
-
|
187 |
-
return openaiChunk;
|
188 |
}
|
189 |
|
190 |
-
//
|
191 |
-
async function
|
192 |
-
|
193 |
-
|
194 |
-
body += chunk.toString();
|
195 |
-
});
|
196 |
|
197 |
-
|
198 |
-
|
199 |
-
|
200 |
-
|
201 |
-
|
202 |
-
|
203 |
-
|
204 |
-
|
205 |
-
|
206 |
-
|
207 |
-
|
208 |
-
|
209 |
-
|
210 |
-
|
211 |
-
// 检查是否有system消息,如果有,将其添加到agent的描述中
|
212 |
-
let systemPrompt = "";
|
213 |
-
if (openaiPayload.messages && Array.isArray(openaiPayload.messages)) {
|
214 |
-
const systemMessage = openaiPayload.messages.find(msg => msg.role === 'system');
|
215 |
-
if (systemMessage) {
|
216 |
-
systemPrompt = systemMessage.content;
|
217 |
}
|
218 |
-
|
219 |
-
|
220 |
-
|
221 |
-
|
222 |
-
|
223 |
-
|
224 |
-
|
225 |
-
|
226 |
-
|
227 |
-
|
228 |
-
|
229 |
-
|
230 |
-
|
231 |
-
|
232 |
-
|
233 |
-
|
234 |
-
|
235 |
-
|
236 |
-
|
237 |
-
|
238 |
-
|
239 |
-
|
240 |
-
|
241 |
-
|
242 |
-
|
243 |
-
|
244 |
-
|
245 |
-
|
246 |
-
|
247 |
-
|
248 |
-
|
249 |
-
|
250 |
-
|
251 |
-
|
252 |
-
|
|
|
|
|
|
|
253 |
},
|
254 |
-
|
255 |
-
|
256 |
-
|
257 |
-
|
258 |
-
|
259 |
-
|
260 |
-
|
261 |
-
|
262 |
-
|
263 |
-
|
264 |
-
|
265 |
-
|
266 |
-
const julepUrl = `${JULEP_API_BASE_URL}/sessions/${sessionId}/chat`;
|
267 |
-
|
268 |
-
const julepResponse = await fetch(julepUrl, {
|
269 |
-
method: "POST",
|
270 |
-
headers: {
|
271 |
-
'Content-Type': 'application/json',
|
272 |
-
'Authorization': req.headers.authorization
|
273 |
},
|
274 |
-
|
275 |
-
|
276 |
-
|
277 |
-
// 处理流式响应
|
278 |
-
if (openaiPayload.stream && julepResponse.headers.get('content-type')?.includes('text/event-stream')) {
|
279 |
-
// 设置正确的响应头
|
280 |
-
res.writeHead(200, {
|
281 |
-
'Content-Type': 'text/event-stream',
|
282 |
-
'Cache-Control': 'no-cache',
|
283 |
-
'Connection': 'keep-alive',
|
284 |
-
'Transfer-Encoding': 'chunked'
|
285 |
-
});
|
286 |
|
287 |
-
|
288 |
-
|
289 |
-
|
290 |
-
|
291 |
-
const
|
292 |
-
|
293 |
-
|
294 |
-
|
295 |
-
|
296 |
-
|
297 |
-
|
298 |
-
|
299 |
-
|
300 |
-
|
301 |
-
|
302 |
-
|
303 |
-
|
304 |
-
|
305 |
-
|
306 |
-
|
307 |
-
|
308 |
-
|
309 |
-
|
310 |
-
|
311 |
-
res.write(
|
312 |
-
|
313 |
-
}
|
314 |
-
|
315 |
-
try {
|
316 |
-
// 解析Julep的JSON响应
|
317 |
-
const julepChunk = JSON.parse(data);
|
318 |
-
// 转换为OpenAI格式
|
319 |
-
const openaiChunk = convertJulepChunkToOpenai(julepChunk, openaiPayload.model || 'julep-model', sessionId);
|
320 |
-
// 发送给客户端
|
321 |
-
res.write(`data: ${JSON.stringify(openaiChunk)}\n\n`);
|
322 |
-
} catch (parseError) {
|
323 |
-
console.error("Error parsing Julep stream chunk:", parseError, "Raw data:", data);
|
324 |
-
// 可选:发送错误信息给客户端
|
325 |
-
// res.write(`data: ${JSON.stringify({ error: "Failed to parse response" })}\n\n`);
|
326 |
-
}
|
327 |
-
} else if (line.trim() && !line.startsWith(':')) {
|
328 |
-
// 记录非注释的未知行类型
|
329 |
-
console.warn("Received unexpected stream line:", line);
|
330 |
}
|
331 |
-
|
332 |
-
|
333 |
-
|
334 |
-
|
335 |
-
|
336 |
-
|
337 |
-
|
338 |
-
|
339 |
-
|
340 |
-
|
341 |
-
|
342 |
-
|
343 |
-
|
344 |
-
|
345 |
-
|
346 |
-
|
347 |
-
const openaiChunk = convertJulepChunkToOpenai(julepChunk, openaiPayload.model || 'julep-model', sessionId);
|
348 |
-
res.write(`data: ${JSON.stringify(openaiChunk)}\n\n`);
|
349 |
-
} catch (e) {
|
350 |
-
console.error("Error parsing final chunk:", e);
|
351 |
}
|
352 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
353 |
}
|
354 |
-
|
355 |
-
|
356 |
-
|
357 |
-
|
358 |
-
|
359 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
360 |
});
|
361 |
-
|
362 |
-
|
363 |
-
|
364 |
-
|
365 |
-
|
366 |
-
|
367 |
-
|
368 |
-
|
369 |
-
|
370 |
-
|
371 |
-
|
372 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
373 |
});
|
374 |
-
|
375 |
-
|
376 |
-
|
377 |
-
|
378 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
379 |
});
|
380 |
-
|
381 |
-
|
382 |
-
|
383 |
-
|
384 |
-
|
385 |
-
|
386 |
-
|
387 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
388 |
} catch (error) {
|
389 |
-
|
390 |
-
|
391 |
-
|
|
|
|
|
|
|
392 |
}
|
393 |
-
|
394 |
-
}
|
395 |
|
396 |
-
//
|
397 |
-
|
398 |
-
|
399 |
-
|
400 |
-
// 处理CORS预检请求
|
401 |
-
if (req.method === 'OPTIONS') {
|
402 |
-
res.writeHead(204, {
|
403 |
-
'Access-Control-Allow-Origin': '*',
|
404 |
-
'Access-Control-Allow-Methods': 'GET, POST, OPTIONS',
|
405 |
-
'Access-Control-Allow-Headers': 'Content-Type, Authorization',
|
406 |
-
'Access-Control-Max-Age': '86400'
|
407 |
-
});
|
408 |
-
res.end();
|
409 |
-
return;
|
410 |
-
}
|
411 |
-
|
412 |
-
// 为所有响应添加CORS头
|
413 |
-
res.setHeader('Access-Control-Allow-Origin', '*');
|
414 |
-
|
415 |
-
// 处理/v1/models端点
|
416 |
-
if (url.pathname === '/v1/models' && req.method === 'GET') {
|
417 |
-
listModelsHandler(res);
|
418 |
-
}
|
419 |
-
// 处理/v1/chat/completions端点
|
420 |
-
else if (url.pathname === '/v1/chat/completions' && req.method === 'POST') {
|
421 |
-
chatCompletionsHandler(req, res);
|
422 |
-
}
|
423 |
-
// 处理根路径,返回简单的HTML页面
|
424 |
-
else if (url.pathname === '/' && req.method === 'GET') {
|
425 |
-
res.writeHead(200, { 'Content-Type': 'text/html' });
|
426 |
-
res.end(`
|
427 |
-
<!DOCTYPE html>
|
428 |
-
<html>
|
429 |
-
<head>
|
430 |
-
<title>Julep API Proxy</title>
|
431 |
-
<style>
|
432 |
-
body { font-family: Arial, sans-serif; max-width: 800px; margin: 0 auto; padding: 20px; }
|
433 |
-
h1 { color: #333; }
|
434 |
-
pre { background: #f4f4f4; padding: 10px; border-radius: 5px; }
|
435 |
-
</style>
|
436 |
-
</head>
|
437 |
-
<body>
|
438 |
-
<h1>Julep API Proxy</h1>
|
439 |
-
<p>This server proxies requests to the Julep API using OpenAI-compatible endpoints:</p>
|
440 |
-
<ul>
|
441 |
-
<li><code>GET /v1/models</code> - List available models</li>
|
442 |
-
<li><code>POST /v1/chat/completions</code> - Create a chat completion</li>
|
443 |
-
</ul>
|
444 |
-
<p>Server is running on port ${PORT}</p>
|
445 |
-
</body>
|
446 |
-
</html>
|
447 |
-
`);
|
448 |
-
}
|
449 |
-
// 返回404
|
450 |
-
else {
|
451 |
-
res.writeHead(404, { 'Content-Type': 'application/json' });
|
452 |
-
res.end(JSON.stringify({ error: "Not Found" }));
|
453 |
-
}
|
454 |
});
|
455 |
|
456 |
-
//
|
457 |
-
|
458 |
-
|
459 |
});
|
|
|
|
1 |
// app.js
|
2 |
+
const express = require('express');
|
|
|
|
|
3 |
const fetch = require('node-fetch');
|
4 |
+
const app = express();
|
5 |
+
const PORT = process.env.PORT || 7860;
|
6 |
|
7 |
+
// Julep API Base URL (fixed)
|
8 |
+
const JULEP_API_BASE = "https://api.julep.ai/api";
|
9 |
+
|
10 |
+
// Hardcoded list of models (Agent IDs in this context)
|
11 |
+
const HARDCODED_MODELS = [
|
12 |
+
'mistral-large-2411', 'o1', 'text-embedding-3-large', 'vertex_ai/text-embedding-004',
|
13 |
+
'claude-3.5-haiku', 'cerebras/llama-4-scout-17b-16e-instruct', 'llama-3.1-8b',
|
14 |
+
'magnum-v4-72b', 'voyage-multilingual-2', 'claude-3-haiku', 'gpt-4o',
|
15 |
+
'BAAI/bge-m3', 'openrouter/meta-llama/llama-4-maverick', 'openrouter/meta-llama/llama-4-scout',
|
16 |
+
'claude-3.5-sonnet', 'hermes-3-llama-3.1-70b', 'claude-3.5-sonnet-20240620',
|
17 |
+
'qwen-2.5-72b-instruct', 'l3.3-euryale-70b', 'gpt-4o-mini', 'cerebras/llama-3.3-70b',
|
18 |
+
'o1-preview', 'gemini-1.5-pro-latest', 'l3.1-euryale-70b', 'claude-3-sonnet',
|
19 |
+
'Alibaba-NLP/gte-large-en-v1.5', 'openrouter/meta-llama/llama-4-scout:free',
|
20 |
+
'llama-3.1-70b', 'eva-qwen-2.5-72b', 'claude-3.5-sonnet-20241022', 'gemini-2.0-flash',
|
21 |
+
'deepseek-chat', 'o1-mini', 'eva-llama-3.33-70b', 'gemini-2.5-pro-preview-03-25',
|
22 |
+
'gemini-1.5-pro', 'gpt-4-turbo', 'openrouter/meta-llama/llama-4-maverick:free',
|
23 |
+
'o3-mini', 'claude-3.7-sonnet', 'voyage-3', 'cerebras/llama-3.1-8b', 'claude-3-opus'
|
24 |
];
|
25 |
|
26 |
+
// --- Helper Functions ---
|
27 |
+
|
28 |
+
// Define acceptable log levels
|
29 |
+
const LogLevels = {
|
30 |
+
DEBUG: 'debug',
|
31 |
+
INFO: 'info',
|
32 |
+
WARN: 'warn',
|
33 |
+
ERROR: 'error',
|
34 |
+
TRACE: 'trace'
|
35 |
+
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
|
37 |
+
function log(level, message, data = null) {
|
38 |
+
// Basic check if the console object has the method
|
39 |
+
if (typeof console[level] === 'function') {
|
40 |
+
console[level](`[${level.toUpperCase()}] ${new Date().toISOString()} - ${message}${data ? ':' : ''}`, data !== null ? JSON.stringify(data, null, 2) : '');
|
41 |
+
} else {
|
42 |
+
// Fallback for potentially missing methods like 'trace' in some environments
|
43 |
+
console.log(`[${level.toUpperCase()}] ${new Date().toISOString()} - ${message}${data ? ':' : ''}`, data !== null ? JSON.stringify(data, null, 2) : '');
|
44 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
}
|
46 |
|
47 |
+
function getJulepApiKey(req) {
|
48 |
+
const authHeader = req.headers.authorization;
|
49 |
+
if (authHeader && authHeader.startsWith("Bearer ")) {
|
50 |
+
log(LogLevels.DEBUG, 'Extracted Julep API Key successfully.');
|
51 |
+
return authHeader.substring(7);
|
52 |
+
}
|
53 |
+
log(LogLevels.WARN, 'Could not extract Julep API Key from Authorization header.');
|
54 |
+
return null;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
}
|
56 |
|
57 |
+
// Helper for small delays
|
58 |
+
function sleep(ms) {
|
59 |
+
return new Promise(resolve => setTimeout(resolve, ms));
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
60 |
}
|
61 |
|
62 |
+
// Note: Using fire-and-forget for background tasks
|
63 |
+
async function cleanupJulepResources(agentId, sessionId, headers) {
|
64 |
+
log(LogLevels.INFO, 'Attempting Julep resource cleanup.', { agentId, sessionId });
|
65 |
+
const cleanupPromises = [];
|
|
|
|
|
66 |
|
67 |
+
// Define cleanup logic as separate async functions for clarity
|
68 |
+
const deleteResource = async (url, type, id) => {
|
69 |
+
try {
|
70 |
+
log(LogLevels.DEBUG, `Sending DELETE request for ${type} ${id} to: ${url}`);
|
71 |
+
const response = await fetch(url, { method: "DELETE", headers });
|
72 |
+
const responseText = await response.text(); // Get text regardless of status
|
73 |
+
if (!response.ok) {
|
74 |
+
log(LogLevels.WARN, `Cleanup failed for ${type} ${id}: ${response.status} ${response.statusText}`, { body: responseText });
|
75 |
+
} else {
|
76 |
+
log(LogLevels.INFO, `Cleanup successful for ${type} ${id}.`, { status: response.status, body: responseText });
|
77 |
+
}
|
78 |
+
} catch (err) {
|
79 |
+
log(LogLevels.ERROR, `Cleanup error during fetch for ${type} ${id}: ${err instanceof Error ? err.message : String(err)}`, { error: err });
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
}
|
81 |
+
};
|
82 |
+
|
83 |
+
if (sessionId) {
|
84 |
+
const sessionDeleteUrl = `${JULEP_API_BASE}/sessions/${sessionId}`;
|
85 |
+
cleanupPromises.push(deleteResource(sessionDeleteUrl, 'session', sessionId));
|
86 |
+
}
|
87 |
+
if (agentId) {
|
88 |
+
const agentDeleteUrl = `${JULEP_API_BASE}/agents/${agentId}`;
|
89 |
+
// Add a small delay before deleting the agent, sometimes helps if session deletion is slow
|
90 |
+
await sleep(100);
|
91 |
+
cleanupPromises.push(deleteResource(agentDeleteUrl, 'agent', agentId));
|
92 |
+
}
|
93 |
+
|
94 |
+
if (cleanupPromises.length > 0) {
|
95 |
+
log(LogLevels.DEBUG, `Waiting for ${cleanupPromises.length} cleanup promises.`);
|
96 |
+
// Run cleanup in background
|
97 |
+
Promise.allSettled(cleanupPromises)
|
98 |
+
.then(results => {
|
99 |
+
log(LogLevels.INFO, 'Cleanup promises settled.', { results });
|
100 |
+
})
|
101 |
+
.catch(error => {
|
102 |
+
log(LogLevels.ERROR, 'Unexpected error during Promise.allSettled for cleanup.', { error });
|
103 |
+
});
|
104 |
+
} else {
|
105 |
+
log(LogLevels.INFO, 'No Julep resources to clean up.');
|
106 |
+
}
|
107 |
+
}
|
108 |
+
|
109 |
+
// Helper to format Julep ToolCall delta to OpenAI format
|
110 |
+
function toolCallDeltaToOpenAI(julepToolCalls) {
|
111 |
+
if (!julepToolCalls) return undefined;
|
112 |
+
return julepToolCalls.map((toolCall, index) => ({
|
113 |
+
index: toolCall.index ?? index,
|
114 |
+
id: toolCall.id,
|
115 |
+
type: "function",
|
116 |
+
function: {
|
117 |
+
name: toolCall.function?.name,
|
118 |
+
arguments: toolCall.function?.arguments,
|
119 |
},
|
120 |
+
}));
|
121 |
+
}
|
122 |
+
|
123 |
+
// Helper to format Julep ToolCall message to OpenAI format
|
124 |
+
function toolCallMessageToOpenAI(julepToolCalls) {
|
125 |
+
if (!julepToolCalls) return undefined;
|
126 |
+
return julepToolCalls.map(toolCall => ({
|
127 |
+
id: toolCall.id,
|
128 |
+
type: "function",
|
129 |
+
function: {
|
130 |
+
name: toolCall.function?.name,
|
131 |
+
arguments: toolCall.function?.arguments,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
132 |
},
|
133 |
+
}));
|
134 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
135 |
|
136 |
+
// Helper function to simulate streaming from a complete response
|
137 |
+
async function simulateStream(julepChatData, requestedModel, res) {
|
138 |
+
log(LogLevels.INFO, 'Starting stream simulation.');
|
139 |
+
try {
|
140 |
+
const baseChunk = {
|
141 |
+
id: julepChatData.id || `chatcmpl-sim-${Date.now()}`,
|
142 |
+
object: "chat.completion.chunk",
|
143 |
+
created: Math.floor(new Date(julepChatData.created_at || Date.now()).getTime() / 1000),
|
144 |
+
model: requestedModel,
|
145 |
+
system_fingerprint: julepChatData.system_fingerprint || null,
|
146 |
+
};
|
147 |
+
|
148 |
+
for (const [index, choice] of julepChatData.choices.entries()) {
|
149 |
+
log(LogLevels.DEBUG, `Simulating stream for choice index ${index}.`);
|
150 |
+
const role = choice.message?.role;
|
151 |
+
const content = choice.message?.content;
|
152 |
+
const toolCallsInput = choice.message?.tool_calls;
|
153 |
+
const toolCallsDelta = toolCallsInput ? toolCallDeltaToOpenAI(toolCallsInput) : undefined; // Format as delta
|
154 |
+
const finishReason = choice.finish_reason;
|
155 |
+
|
156 |
+
// 1. Send role chunk
|
157 |
+
if (role) {
|
158 |
+
const roleChunk = { ...baseChunk, choices: [{ index: index, delta: { role: role }, finish_reason: null }] };
|
159 |
+
log(LogLevels.DEBUG, 'Sending role chunk:', roleChunk);
|
160 |
+
res.write(`data: ${JSON.stringify(roleChunk)}\n\n`);
|
161 |
+
await sleep(5);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
162 |
}
|
163 |
+
|
164 |
+
// 2. Send tool calls chunk(s) if they exist
|
165 |
+
if (toolCallsDelta && toolCallsDelta.length > 0) {
|
166 |
+
const toolCallDeltaChunk = { ...baseChunk, choices: [{ index: index, delta: { tool_calls: toolCallsDelta }, finish_reason: null }] };
|
167 |
+
log(LogLevels.DEBUG, 'Sending tool_calls chunk:', toolCallDeltaChunk);
|
168 |
+
res.write(`data: ${JSON.stringify(toolCallDeltaChunk)}\n\n`);
|
169 |
+
await sleep(5);
|
170 |
+
}
|
171 |
+
|
172 |
+
// 3. Stream content
|
173 |
+
if (content && typeof content === 'string') {
|
174 |
+
log(LogLevels.DEBUG, `Streaming content for choice ${index} (length: ${content.length})`);
|
175 |
+
for (const char of content) {
|
176 |
+
const contentChunk = { ...baseChunk, choices: [{ index: index, delta: { content: char }, finish_reason: null }] };
|
177 |
+
res.write(`data: ${JSON.stringify(contentChunk)}\n\n`);
|
178 |
+
await sleep(2); // Simulate typing delay
|
|
|
|
|
|
|
|
|
179 |
}
|
180 |
+
log(LogLevels.DEBUG, `Finished streaming content for choice ${index}`);
|
181 |
+
} else if (content) {
|
182 |
+
// Send non-string content as a single chunk (might be structured JSON etc.)
|
183 |
+
const contentChunk = { ...baseChunk, choices: [{ index: index, delta: { content: JSON.stringify(content) }, finish_reason: null }] };
|
184 |
+
log(LogLevels.DEBUG, 'Sending non-string content chunk:', contentChunk);
|
185 |
+
res.write(`data: ${JSON.stringify(contentChunk)}\n\n`);
|
186 |
+
await sleep(5);
|
187 |
+
}
|
188 |
+
|
189 |
+
// 4. Send finish reason chunk
|
190 |
+
if (finishReason) {
|
191 |
+
const finishChunk = { ...baseChunk, choices: [{ index: index, delta: {}, finish_reason: finishReason }] };
|
192 |
+
log(LogLevels.DEBUG, 'Sending finish reason chunk:', finishChunk);
|
193 |
+
res.write(`data: ${JSON.stringify(finishChunk)}\n\n`);
|
194 |
+
await sleep(5);
|
195 |
+
}
|
196 |
+
}
|
197 |
+
|
198 |
+
// 5. Send DONE marker
|
199 |
+
log(LogLevels.INFO, 'Sending [DONE] marker.');
|
200 |
+
res.write('data: [DONE]\n\n');
|
201 |
+
res.end();
|
202 |
+
log(LogLevels.INFO, 'Stream simulation completed successfully.');
|
203 |
+
|
204 |
+
} catch (error) {
|
205 |
+
log(LogLevels.ERROR, `Error during stream simulation: ${error instanceof Error ? error.message : String(error)}`, { error: error });
|
206 |
+
// Try to end the response if possible
|
207 |
+
try {
|
208 |
+
if (!res.headersSent) {
|
209 |
+
res.status(500).json({ error: 'Stream simulation error' });
|
210 |
+
} else {
|
211 |
+
res.end();
|
212 |
}
|
213 |
+
} catch (endError) {
|
214 |
+
log(LogLevels.ERROR, 'Error ending response after stream error', { error: endError });
|
215 |
+
}
|
216 |
+
}
|
217 |
+
}
|
218 |
+
|
219 |
+
// --- Middleware ---
|
220 |
+
app.use(express.json());
|
221 |
+
|
222 |
+
// CORS middleware
|
223 |
+
app.use((req, res, next) => {
|
224 |
+
res.header('Access-Control-Allow-Origin', '*');
|
225 |
+
res.header('Access-Control-Allow-Methods', 'GET, POST, OPTIONS');
|
226 |
+
res.header('Access-Control-Allow-Headers', 'Authorization, Content-Type');
|
227 |
+
|
228 |
+
if (req.method === 'OPTIONS') {
|
229 |
+
return res.status(204).end();
|
230 |
+
}
|
231 |
+
next();
|
232 |
+
});
|
233 |
+
|
234 |
+
// --- Routes ---
|
235 |
+
app.get('/v1/models', async (req, res) => {
|
236 |
+
log(LogLevels.INFO, 'Handling /v1/models request.');
|
237 |
+
const julepApiKey = getJulepApiKey(req);
|
238 |
+
if (!julepApiKey) {
|
239 |
+
log(LogLevels.WARN, 'Unauthorized /v1/models request (missing API key).');
|
240 |
+
// Optionally allow models request without key, or enforce it
|
241 |
+
}
|
242 |
+
|
243 |
+
const now = Math.floor(Date.now() / 1000);
|
244 |
+
const openaiModels = HARDCODED_MODELS.map((modelId) => ({
|
245 |
+
id: modelId, object: "model", created: now, owned_by: "julep",
|
246 |
+
permission: [{ id: `modelperm-${modelId}-${now}`, object: "model_permission", created: now, allow_create_engine: false, allow_sampling: true, allow_logprobs: true, allow_search_indices: false, allow_view: true, allow_fine_tuning: false, organization: "*", group: null, is_blocking: false, }],
|
247 |
+
root: modelId, parent: null,
|
248 |
+
}));
|
249 |
+
log(LogLevels.DEBUG, 'Returning hardcoded models list.');
|
250 |
+
res.json({ data: openaiModels, object: "list" });
|
251 |
+
});
|
252 |
+
|
253 |
+
app.post('/v1/chat/completions', async (req, res) => {
|
254 |
+
log(LogLevels.INFO, 'Handling /v1/chat/completions request.');
|
255 |
+
const julepApiKey = getJulepApiKey(req);
|
256 |
+
if (!julepApiKey) {
|
257 |
+
log(LogLevels.ERROR, 'Unauthorized chat completions request: Missing Julep API Key.');
|
258 |
+
return res.status(401).send("Unauthorized: Missing or invalid Authorization header");
|
259 |
+
}
|
260 |
+
|
261 |
+
// Define headers early, use this single object throughout
|
262 |
+
const headers = {
|
263 |
+
"Authorization": `Bearer ${julepApiKey}`,
|
264 |
+
"Content-Type": "application/json",
|
265 |
+
};
|
266 |
+
log(LogLevels.DEBUG, 'Julep API request headers prepared (key omitted).', { "Content-Type": headers["Content-Type"] });
|
267 |
+
|
268 |
+
let agentId = null;
|
269 |
+
let sessionId = null;
|
270 |
+
let requestBody = req.body;
|
271 |
+
|
272 |
+
try {
|
273 |
+
const { model, messages, stream, ...rest } = requestBody;
|
274 |
+
const clientRequestedStream = stream === true;
|
275 |
+
log(LogLevels.INFO, `Request details: model=${model}, clientRequestedStream=${clientRequestedStream}`);
|
276 |
+
|
277 |
+
// Validate essential parameters
|
278 |
+
if (!model || !messages || !Array.isArray(messages) || messages.length === 0) {
|
279 |
+
log(LogLevels.ERROR, 'Invalid request body: "model" and "messages" are required.', { model, messages });
|
280 |
+
return res.status(400).send("Invalid request body. 'model' and 'messages' are required.");
|
281 |
+
}
|
282 |
+
if (!HARDCODED_MODELS.includes(model)) {
|
283 |
+
log(LogLevels.ERROR, `Invalid model requested: ${model}`);
|
284 |
+
return res.status(400).send(`Invalid model: ${model}. Please use one of the available models.`);
|
285 |
+
}
|
286 |
+
log(LogLevels.DEBUG, 'Request parameters validated.');
|
287 |
+
|
288 |
+
// --- Agent and Session Creation ---
|
289 |
+
// 2. Create Agent
|
290 |
+
const createAgentUrl = `${JULEP_API_BASE}/agents`;
|
291 |
+
const createAgentBody = {
|
292 |
+
name: `temp-openai-${model}-${Date.now()}`,
|
293 |
+
model: model,
|
294 |
+
about: `Temporary agent for OpenAI model ${model}`,
|
295 |
+
};
|
296 |
+
log(LogLevels.INFO, 'Attempting to create Julep Agent.', { url: createAgentUrl, body: createAgentBody });
|
297 |
+
const createAgentResponse = await fetch(createAgentUrl, {
|
298 |
+
method: "POST",
|
299 |
+
headers,
|
300 |
+
body: JSON.stringify(createAgentBody)
|
301 |
});
|
302 |
+
log(LogLevels.DEBUG, `Create Agent response status: ${createAgentResponse.status}`);
|
303 |
+
|
304 |
+
if (!createAgentResponse.ok) {
|
305 |
+
const errorStatus = createAgentResponse.status;
|
306 |
+
const errorStatusText = createAgentResponse.statusText;
|
307 |
+
let errorText = "[Could not read error body]";
|
308 |
+
try {
|
309 |
+
errorText = await createAgentResponse.text();
|
310 |
+
} catch (e) {
|
311 |
+
log(LogLevels.WARN, `Could not read error text from createAgentResponse: ${e instanceof Error ? e.message : String(e)}`);
|
312 |
+
}
|
313 |
+
log(LogLevels.ERROR, `Error creating Julep Agent: ${errorStatus} - ${errorText}`);
|
314 |
+
return res.status(errorStatus).send(`Error creating Julep Agent: ${errorStatusText} - ${errorText}`);
|
315 |
+
}
|
316 |
+
|
317 |
+
let agentData;
|
318 |
+
try {
|
319 |
+
const agentResponseText = await createAgentResponse.text();
|
320 |
+
log(LogLevels.DEBUG, 'Create Agent raw response text:', agentResponseText);
|
321 |
+
agentData = JSON.parse(agentResponseText);
|
322 |
+
log(LogLevels.INFO, 'Julep Agent created successfully.', { agentData });
|
323 |
+
agentId = agentData.id;
|
324 |
+
} catch (e) {
|
325 |
+
log(LogLevels.ERROR, `Failed to parse Julep Agent creation response JSON: ${e instanceof Error ? e.message : String(e)}`, { error: e });
|
326 |
+
// Attempt cleanup (fire-and-forget)
|
327 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after agent parse error', err));
|
328 |
+
return res.status(500).send(`Internal Server Error: Failed to parse Julep Agent response. ${e instanceof Error ? e.message : String(e)}`);
|
329 |
+
}
|
330 |
+
|
331 |
+
// 3. Create Session
|
332 |
+
const createSessionUrl = `${JULEP_API_BASE}/sessions`;
|
333 |
+
const createSessionBody = { agent: agentId }; // Julep API uses agent
|
334 |
+
log(LogLevels.INFO, 'Attempting to create Julep Session.', { url: createSessionUrl, body: createSessionBody });
|
335 |
+
const createSessionResponse = await fetch(createSessionUrl, {
|
336 |
+
method: "POST",
|
337 |
+
headers,
|
338 |
+
body: JSON.stringify(createSessionBody)
|
339 |
});
|
340 |
+
log(LogLevels.DEBUG, `Create Session response status: ${createSessionResponse.status}`);
|
341 |
+
|
342 |
+
if (!createSessionResponse.ok) {
|
343 |
+
const errorStatus = createSessionResponse.status;
|
344 |
+
const errorStatusText = createSessionResponse.statusText;
|
345 |
+
let errorText = "[Could not read error body]";
|
346 |
+
try {
|
347 |
+
errorText = await createSessionResponse.text();
|
348 |
+
} catch (e) {
|
349 |
+
log(LogLevels.WARN, `Could not read error text from createSessionResponse: ${e instanceof Error ? e.message : String(e)}`);
|
350 |
+
}
|
351 |
+
log(LogLevels.ERROR, `Error creating Julep Session: ${errorStatus} - ${errorText}`);
|
352 |
+
// Cleanup the agent we just created (fire-and-forget)
|
353 |
+
cleanupJulepResources(agentId, null, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after session creation error', err));
|
354 |
+
return res.status(errorStatus).send(`Error creating Julep Session: ${errorStatusText} - ${errorText}`);
|
355 |
+
}
|
356 |
+
|
357 |
+
let sessionData;
|
358 |
+
try {
|
359 |
+
const sessionResponseText = await createSessionResponse.text();
|
360 |
+
log(LogLevels.DEBUG, 'Create Session raw response text:', sessionResponseText);
|
361 |
+
sessionData = JSON.parse(sessionResponseText);
|
362 |
+
log(LogLevels.INFO, 'Julep Session created successfully.', { sessionData });
|
363 |
+
sessionId = sessionData.id;
|
364 |
+
} catch (e) {
|
365 |
+
log(LogLevels.ERROR, `Failed to parse Julep Session creation response JSON: ${e instanceof Error ? e.message : String(e)}`, { error: e });
|
366 |
+
// Cleanup agent and session (fire-and-forget)
|
367 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after session parse error', err));
|
368 |
+
return res.status(500).send(`Internal Server Error: Failed to parse Julep Session response. ${e instanceof Error ? e.message : String(e)}`);
|
369 |
+
}
|
370 |
+
|
371 |
+
// --- Perform Chat Completion (ALWAYS non-streaming to Julep) ---
|
372 |
+
// 4. Send Chat Request to Julep
|
373 |
+
const chatUrl = `${JULEP_API_BASE}/sessions/${sessionId}/chat`;
|
374 |
+
const chatBodyToJulep = {
|
375 |
+
messages: messages.map((msg) => ({
|
376 |
+
role: msg.role,
|
377 |
+
content: typeof msg.content === 'string' ? msg.content : JSON.stringify(msg.content),
|
378 |
+
// Include tool_calls if present in the input message (OpenAI format)
|
379 |
+
tool_calls: msg.tool_calls, // Assuming Julep accepts OpenAI tool call format here
|
380 |
+
tool_call_id: msg.tool_call_id // If it's a tool response message
|
381 |
+
})),
|
382 |
+
stream: false, // Force non-streaming
|
383 |
+
...rest, // Pass through other OpenAI parameters like temperature, top_p, etc.
|
384 |
+
};
|
385 |
+
log(LogLevels.INFO, 'Sending Chat request to Julep (forced non-stream).', { url: chatUrl });
|
386 |
+
log(LogLevels.DEBUG, 'Julep Chat Request Body:', chatBodyToJulep);
|
387 |
+
const chatResponse = await fetch(chatUrl, {
|
388 |
+
method: "POST",
|
389 |
+
headers,
|
390 |
+
body: JSON.stringify(chatBodyToJulep)
|
391 |
});
|
392 |
+
log(LogLevels.DEBUG, `Julep Chat response status: ${chatResponse.status}`);
|
393 |
+
|
394 |
+
// --- Handle Julep Response ---
|
395 |
+
if (!chatResponse.ok) {
|
396 |
+
const errorStatus = chatResponse.status;
|
397 |
+
const errorStatusText = chatResponse.statusText;
|
398 |
+
let errorText = "[Could not read error body]";
|
399 |
+
try {
|
400 |
+
errorText = await chatResponse.text();
|
401 |
+
} catch (e) {
|
402 |
+
log(LogLevels.WARN, `Could not read error text from chatResponse: ${e instanceof Error ? e.message : String(e)}`);
|
403 |
+
}
|
404 |
+
log(LogLevels.ERROR, `Error during Julep Chat Completion: ${errorStatus} - ${errorText}`);
|
405 |
+
// Cleanup agent and session (fire-and-forget)
|
406 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after chat error', err));
|
407 |
+
return res.status(errorStatus).send(`Error during Julep Chat Completion: ${errorStatusText} - ${errorText}`);
|
408 |
+
}
|
409 |
+
|
410 |
+
// Julep request was successful, get the full JSON body
|
411 |
+
let julepChatData;
|
412 |
+
try {
|
413 |
+
const chatResponseText = await chatResponse.text();
|
414 |
+
log(LogLevels.DEBUG, 'Julep Chat raw response text:', chatResponseText);
|
415 |
+
julepChatData = JSON.parse(chatResponseText);
|
416 |
+
log(LogLevels.INFO, 'Julep chat completion successful.', { responseId: julepChatData.id })
|
417 |
+
log(LogLevels.DEBUG, 'Julep Chat response data:', julepChatData);
|
418 |
+
} catch (e) {
|
419 |
+
log(LogLevels.ERROR, `Failed to parse Julep Chat response JSON (status was OK): ${e instanceof Error ? e.message : String(e)}`, { error: e });
|
420 |
+
// Cleanup agent and session (fire-and-forget)
|
421 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after chat parse error', err));
|
422 |
+
return res.status(500).send(`Internal Server Error: Failed to parse Julep Chat response. ${e instanceof Error ? e.message : String(e)}`);
|
423 |
+
}
|
424 |
+
|
425 |
+
// *** Trigger cleanup NOW (fire-and-forget), before returning the response/stream ***
|
426 |
+
log(LogLevels.INFO, 'Julep chat successful, queueing cleanup.');
|
427 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed after successful chat', err));
|
428 |
+
|
429 |
+
// --- Format and Return Response to Client ---
|
430 |
+
// Access the actual chat response data
|
431 |
+
const julepResponseData = julepChatData;
|
432 |
+
if (!julepResponseData || !julepResponseData.choices) {
|
433 |
+
log(LogLevels.ERROR, 'Julep response format unexpected. Missing "response" or "response.choices".', { julepChatData });
|
434 |
+
return res.status(500).send('Internal Server Error: Unexpected format from Julep API.');
|
435 |
+
}
|
436 |
+
|
437 |
+
if (clientRequestedStream) {
|
438 |
+
log(LogLevels.INFO, 'Client requested stream, starting simulation.');
|
439 |
+
// Set headers for SSE
|
440 |
+
res.setHeader('Content-Type', 'text/event-stream');
|
441 |
+
res.setHeader('Cache-Control', 'no-cache');
|
442 |
+
res.setHeader('Connection', 'keep-alive');
|
443 |
+
res.flushHeaders(); // Flush the headers to establish SSE with client
|
444 |
+
|
445 |
+
// Start simulation
|
446 |
+
simulateStream(julepResponseData, model, res)
|
447 |
+
.catch(streamErr => {
|
448 |
+
log(LogLevels.ERROR, 'Stream simulation task failed.', { error: streamErr });
|
449 |
+
});
|
450 |
+
} else {
|
451 |
+
log(LogLevels.INFO, 'Client requested non-streaming response.');
|
452 |
+
// Format julepResponseData to OpenAI format
|
453 |
+
const openaiCompletion = {
|
454 |
+
id: julepResponseData.id || `chatcmpl-${Date.now()}`,
|
455 |
+
object: "chat.completion",
|
456 |
+
created: Math.floor(new Date(julepResponseData.created_at || Date.now()).getTime() / 1000),
|
457 |
+
model: model, // Use the originally requested model
|
458 |
+
choices: julepResponseData.choices.map((choice) => ({
|
459 |
+
index: choice.index,
|
460 |
+
message: {
|
461 |
+
role: choice.message.role,
|
462 |
+
content: choice.message.content,
|
463 |
+
// Use toolCallMessageToOpenAI here for the completed message format
|
464 |
+
tool_calls: choice.message.tool_calls ? toolCallMessageToOpenAI(choice.message.tool_calls) : undefined
|
465 |
+
},
|
466 |
+
finish_reason: choice.finish_reason
|
467 |
+
})),
|
468 |
+
usage: julepResponseData.usage ? {
|
469 |
+
prompt_tokens: julepResponseData.usage.prompt_tokens,
|
470 |
+
completion_tokens: julepResponseData.usage.completion_tokens,
|
471 |
+
total_tokens: julepResponseData.usage.total_tokens
|
472 |
+
} : undefined,
|
473 |
+
system_fingerprint: julepResponseData.system_fingerprint || null,
|
474 |
+
};
|
475 |
+
log(LogLevels.DEBUG, 'Formatted non-streaming OpenAI response:', openaiCompletion);
|
476 |
+
log(LogLevels.INFO, 'Returning non-streaming JSON response to client.');
|
477 |
+
res.json(openaiCompletion);
|
478 |
+
}
|
479 |
+
|
480 |
} catch (error) {
|
481 |
+
// Catch errors from initial parsing, validation, or unexpected issues within the try block
|
482 |
+
log(LogLevels.ERROR, `Error in handleChatCompletions (outer catch): ${error instanceof Error ? error.message : String(error)}`, { error: error, agentId, sessionId });
|
483 |
+
// Use the headers defined at the start if available
|
484 |
+
// Attempt cleanup (fire-and-forget)
|
485 |
+
cleanupJulepResources(agentId, sessionId, headers).catch(err => log(LogLevels.ERROR, 'Background cleanup failed in outer catch block', err));
|
486 |
+
res.status(500).send(`Internal Server Error: ${error instanceof Error ? error.message : String(error)}`);
|
487 |
}
|
488 |
+
});
|
|
|
489 |
|
490 |
+
// Health check endpoint for Hugging Face
|
491 |
+
app.get('/', (req, res) => {
|
492 |
+
res.send('Julep API Proxy is running');
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
493 |
});
|
494 |
|
495 |
+
// Start the server
|
496 |
+
app.listen(PORT, () => {
|
497 |
+
log(LogLevels.INFO, `Server running on port ${PORT}`);
|
498 |
});
|
499 |
+
|