Spaces:
Running
Running
Set better defaults for free-tier users: Qwen3-1.7B model, 1024 max tokens, search disabled
Browse files
app.py
CHANGED
|
@@ -593,11 +593,11 @@ with gr.Blocks(title="LLM Inference with ZeroGPU") as demo:
|
|
| 593 |
gr.Markdown("Interact with the model. Select parameters and chat below.")
|
| 594 |
with gr.Row():
|
| 595 |
with gr.Column(scale=3):
|
| 596 |
-
model_dd = gr.Dropdown(label="Select Model", choices=list(MODELS.keys()), value=
|
| 597 |
-
search_chk = gr.Checkbox(label="Enable Web Search", value=
|
| 598 |
sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
|
| 599 |
gr.Markdown("### Generation Parameters")
|
| 600 |
-
max_tok = gr.Slider(64, 16384, value=
|
| 601 |
temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
|
| 602 |
k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
|
| 603 |
p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
|
|
|
|
| 593 |
gr.Markdown("Interact with the model. Select parameters and chat below.")
|
| 594 |
with gr.Row():
|
| 595 |
with gr.Column(scale=3):
|
| 596 |
+
model_dd = gr.Dropdown(label="Select Model", choices=list(MODELS.keys()), value="Qwen3-1.7B")
|
| 597 |
+
search_chk = gr.Checkbox(label="Enable Web Search", value=False)
|
| 598 |
sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
|
| 599 |
gr.Markdown("### Generation Parameters")
|
| 600 |
+
max_tok = gr.Slider(64, 16384, value=1024, step=32, label="Max Tokens")
|
| 601 |
temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
|
| 602 |
k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
|
| 603 |
p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
|