Luigi commited on
Commit
2cae073
·
1 Parent(s): de766da

Set better defaults for free-tier users: Qwen3-1.7B model, 1024 max tokens, search disabled

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -593,11 +593,11 @@ with gr.Blocks(title="LLM Inference with ZeroGPU") as demo:
593
  gr.Markdown("Interact with the model. Select parameters and chat below.")
594
  with gr.Row():
595
  with gr.Column(scale=3):
596
- model_dd = gr.Dropdown(label="Select Model", choices=list(MODELS.keys()), value=list(MODELS.keys())[0])
597
- search_chk = gr.Checkbox(label="Enable Web Search", value=True)
598
  sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
599
  gr.Markdown("### Generation Parameters")
600
- max_tok = gr.Slider(64, 16384, value=2048, step=32, label="Max Tokens")
601
  temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
602
  k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
603
  p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
 
593
  gr.Markdown("Interact with the model. Select parameters and chat below.")
594
  with gr.Row():
595
  with gr.Column(scale=3):
596
+ model_dd = gr.Dropdown(label="Select Model", choices=list(MODELS.keys()), value="Qwen3-1.7B")
597
+ search_chk = gr.Checkbox(label="Enable Web Search", value=False)
598
  sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
599
  gr.Markdown("### Generation Parameters")
600
+ max_tok = gr.Slider(64, 16384, value=1024, step=32, label="Max Tokens")
601
  temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
602
  k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
603
  p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")