Spaces:
Running
on
Zero
Running
on
Zero
Update README.md
Browse files
README.md
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
---
|
2 |
-
title:
|
3 |
emoji: 🧠
|
4 |
colorFrom: pink
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
7 |
-
sdk_version: 5.
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
license: apache-2.0
|
@@ -77,4 +77,4 @@ Use the dropdown to select any of these:
|
|
77 |
3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
|
78 |
4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
|
79 |
5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
|
80 |
-
6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.
|
|
|
1 |
---
|
2 |
+
title: ZeroGPU-LLM-Inference
|
3 |
emoji: 🧠
|
4 |
colorFrom: pink
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
7 |
+
sdk_version: 5.29.0
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
license: apache-2.0
|
|
|
77 |
3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
|
78 |
4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
|
79 |
5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
|
80 |
+
6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.
|