Spaces:
Sleeping
Sleeping
feat(models): add `nvidia/Llama-3.1-Nemotron-70B-Instruct-HF` (#1527)
Browse files- chart/env/prod.yaml +17 -0
- src/lib/server/models.ts +7 -1
chart/env/prod.yaml
CHANGED
@@ -137,6 +137,23 @@ envVars:
|
|
137 |
}
|
138 |
]
|
139 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
140 |
{
|
141 |
"name": "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
142 |
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
|
|
|
137 |
}
|
138 |
]
|
139 |
},
|
140 |
+
{
|
141 |
+
"name": "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
|
142 |
+
"modelUrl": "https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
|
143 |
+
"websiteUrl": "https://www.nvidia.com/",
|
144 |
+
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nvidia-logo.png",
|
145 |
+
"description": "Nvidia's latest Llama fine-tune, topping alignment benchmarks and optimized for instruction following.",
|
146 |
+
"parameters": {
|
147 |
+
"stop": ["<|eot_id|>"],
|
148 |
+
"temperature": 0.6,
|
149 |
+
"truncate": 28672,
|
150 |
+
"max_new_tokens": 2048
|
151 |
+
},
|
152 |
+
"endpoint": {
|
153 |
+
"type": "openai",
|
154 |
+
"baseURL": "https://api-inference.huggingface.co/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF/v1"
|
155 |
+
}
|
156 |
+
},
|
157 |
{
|
158 |
"name": "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
159 |
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
|
src/lib/server/models.ts
CHANGED
@@ -319,7 +319,13 @@ const hasInferenceAPI = async (m: Awaited<ReturnType<typeof processModel>>) => {
|
|
319 |
return false;
|
320 |
}
|
321 |
|
322 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
323 |
|
324 |
if (!r.ok) {
|
325 |
logger.warn(`Failed to check if ${m.id} has inference API: ${r.statusText}`);
|
|
|
319 |
return false;
|
320 |
}
|
321 |
|
322 |
+
let r: Response;
|
323 |
+
try {
|
324 |
+
r = await fetch(`https://huggingface.co/api/models/${m.id}`);
|
325 |
+
} catch (e) {
|
326 |
+
console.log(e);
|
327 |
+
return false;
|
328 |
+
}
|
329 |
|
330 |
if (!r.ok) {
|
331 |
logger.warn(`Failed to check if ${m.id} has inference API: ${r.statusText}`);
|