Luigi commited on
Commit
371669a
·
verified ·
1 Parent(s): 076c1f2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
- title: 'ZeroGPU-LLM-Inference'
3
  emoji: 🧠
4
  colorFrom: pink
5
  colorTo: purple
6
  sdk: gradio
7
- sdk_version: 5.25.2
8
  app_file: app.py
9
  pinned: false
10
  license: apache-2.0
@@ -77,4 +77,4 @@ Use the dropdown to select any of these:
77
  3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
78
  4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
79
  5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
80
- 6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.
 
1
  ---
2
+ title: ZeroGPU-LLM-Inference
3
  emoji: 🧠
4
  colorFrom: pink
5
  colorTo: purple
6
  sdk: gradio
7
+ sdk_version: 5.29.0
8
  app_file: app.py
9
  pinned: false
10
  license: apache-2.0
 
77
  3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
78
  4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
79
  5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
80
+ 6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.