Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
---
|
| 2 |
-
title:
|
| 3 |
emoji: 🧠
|
| 4 |
colorFrom: pink
|
| 5 |
colorTo: purple
|
| 6 |
sdk: gradio
|
| 7 |
-
sdk_version: 5.
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
license: apache-2.0
|
|
@@ -77,4 +77,4 @@ Use the dropdown to select any of these:
|
|
| 77 |
3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
|
| 78 |
4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
|
| 79 |
5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
|
| 80 |
-
6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.
|
|
|
|
| 1 |
---
|
| 2 |
+
title: ZeroGPU-LLM-Inference
|
| 3 |
emoji: 🧠
|
| 4 |
colorFrom: pink
|
| 5 |
colorTo: purple
|
| 6 |
sdk: gradio
|
| 7 |
+
sdk_version: 5.29.0
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
license: apache-2.0
|
|
|
|
| 77 |
3. After up to *Search Timeout* seconds, snippets merge into the system prompt.
|
| 78 |
4. The selected model pipeline is loaded (bf16→f16→f32 fallback) on ZeroGPU.
|
| 79 |
5. Prompt is formatted—any `<think>…</think>` blocks will be streamed as separate “💭 Thought.”
|
| 80 |
+
6. Tokens stream to the Chatbot UI. Press **Cancel** to stop mid-generation.
|