nsarrazin HF Staff commited on
Commit
c5ee468
·
unverified ·
1 Parent(s): dc6961c

feat(models): add `nvidia/Llama-3.1-Nemotron-70B-Instruct-HF` (#1527)

Browse files
Files changed (2) hide show
  1. chart/env/prod.yaml +17 -0
  2. src/lib/server/models.ts +7 -1
chart/env/prod.yaml CHANGED
@@ -137,6 +137,23 @@ envVars:
137
  }
138
  ]
139
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
140
  {
141
  "name": "meta-llama/Llama-3.2-11B-Vision-Instruct",
142
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
 
137
  }
138
  ]
139
  },
140
+ {
141
+ "name": "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
142
+ "modelUrl": "https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
143
+ "websiteUrl": "https://www.nvidia.com/",
144
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nvidia-logo.png",
145
+ "description": "Nvidia's latest Llama fine-tune, topping alignment benchmarks and optimized for instruction following.",
146
+ "parameters": {
147
+ "stop": ["<|eot_id|>"],
148
+ "temperature": 0.6,
149
+ "truncate": 28672,
150
+ "max_new_tokens": 2048
151
+ },
152
+ "endpoint": {
153
+ "type": "openai",
154
+ "baseURL": "https://api-inference.huggingface.co/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF/v1"
155
+ }
156
+ },
157
  {
158
  "name": "meta-llama/Llama-3.2-11B-Vision-Instruct",
159
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
src/lib/server/models.ts CHANGED
@@ -319,7 +319,13 @@ const hasInferenceAPI = async (m: Awaited<ReturnType<typeof processModel>>) => {
319
  return false;
320
  }
321
 
322
- const r = await fetch(`https://huggingface.co/api/models/${m.id}`);
 
 
 
 
 
 
323
 
324
  if (!r.ok) {
325
  logger.warn(`Failed to check if ${m.id} has inference API: ${r.statusText}`);
 
319
  return false;
320
  }
321
 
322
+ let r: Response;
323
+ try {
324
+ r = await fetch(`https://huggingface.co/api/models/${m.id}`);
325
+ } catch (e) {
326
+ console.log(e);
327
+ return false;
328
+ }
329
 
330
  if (!r.ok) {
331
  logger.warn(`Failed to check if ${m.id} has inference API: ${r.statusText}`);