Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
-
from huggingface_hub import InferenceClient
|
| 3 |
import os
|
| 4 |
import json
|
| 5 |
|
|
@@ -50,33 +50,22 @@ def get_prompt(name):
|
|
| 50 |
"""Retrieve a system prompt by name."""
|
| 51 |
return system_prompts.get(name, "")
|
| 52 |
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
except Exception as e:
|
| 61 |
-
return [f"Error fetching models: {str(e)}"]
|
| 62 |
|
| 63 |
# Gradio Interface
|
| 64 |
with gr.Blocks() as demo:
|
| 65 |
-
gr.Markdown("## Hugging Face Chatbot with
|
| 66 |
|
| 67 |
with gr.Row():
|
| 68 |
with gr.Column():
|
| 69 |
-
|
| 70 |
-
task_selector = gr.Dropdown(
|
| 71 |
-
choices=["text-generation", "image-classification", "text-classification", "translation"],
|
| 72 |
-
label="Select Task",
|
| 73 |
-
value="text-generation"
|
| 74 |
-
)
|
| 75 |
-
|
| 76 |
-
# Model selector
|
| 77 |
-
model_selector = gr.Dropdown(choices=[], label="Select Model")
|
| 78 |
|
| 79 |
-
# System prompt and input
|
| 80 |
system_prompt_name = gr.Dropdown(choices=list(system_prompts.keys()), label="Select System Prompt")
|
| 81 |
system_prompt_content = gr.TextArea(label="System Prompt", value=get_prompt("default"), lines=4)
|
| 82 |
save_prompt_button = gr.Button("Save System Prompt")
|
|
@@ -87,14 +76,14 @@ with gr.Blocks() as demo:
|
|
| 87 |
with gr.Column():
|
| 88 |
output = gr.TextArea(label="Model Response", interactive=False, lines=10)
|
| 89 |
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
models = fetch_models(task)
|
| 93 |
-
print(f"Models:{models}")
|
| 94 |
-
return gr.Dropdown.update(choices=models, value=models[0] if models else None)
|
| 95 |
|
| 96 |
-
|
| 97 |
-
|
|
|
|
|
|
|
|
|
|
| 98 |
save_prompt_button.click(update_prompt, inputs=[system_prompt_name, system_prompt_content], outputs=[])
|
| 99 |
submit_button.click(chat_with_model, inputs=[user_input, system_prompt_content, model_selector], outputs=[output])
|
| 100 |
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
+
from huggingface_hub import InferenceClient
|
| 3 |
import os
|
| 4 |
import json
|
| 5 |
|
|
|
|
| 50 |
"""Retrieve a system prompt by name."""
|
| 51 |
return system_prompts.get(name, "")
|
| 52 |
|
| 53 |
+
# List of available models
|
| 54 |
+
available_models = [
|
| 55 |
+
"gpt-3.5-turbo",
|
| 56 |
+
"gpt-4",
|
| 57 |
+
"HuggingFaceH4/zephyr-7b-beta",
|
| 58 |
+
"HuggingFaceH4/zephyr-7b-alpha"
|
| 59 |
+
]
|
|
|
|
|
|
|
| 60 |
|
| 61 |
# Gradio Interface
|
| 62 |
with gr.Blocks() as demo:
|
| 63 |
+
gr.Markdown("## Hugging Face Chatbot with Gradio")
|
| 64 |
|
| 65 |
with gr.Row():
|
| 66 |
with gr.Column():
|
| 67 |
+
model_selector = gr.Dropdown(choices=available_models, label="Select Model", value=available_models[0])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
|
|
|
|
| 69 |
system_prompt_name = gr.Dropdown(choices=list(system_prompts.keys()), label="Select System Prompt")
|
| 70 |
system_prompt_content = gr.TextArea(label="System Prompt", value=get_prompt("default"), lines=4)
|
| 71 |
save_prompt_button = gr.Button("Save System Prompt")
|
|
|
|
| 76 |
with gr.Column():
|
| 77 |
output = gr.TextArea(label="Model Response", interactive=False, lines=10)
|
| 78 |
|
| 79 |
+
def load_prompt(name):
|
| 80 |
+
return get_prompt(name)
|
|
|
|
|
|
|
|
|
|
| 81 |
|
| 82 |
+
system_prompt_name.change(
|
| 83 |
+
lambda name: (name, get_prompt(name)),
|
| 84 |
+
inputs=[system_prompt_name],
|
| 85 |
+
outputs=[system_prompt_name, system_prompt_content]
|
| 86 |
+
)
|
| 87 |
save_prompt_button.click(update_prompt, inputs=[system_prompt_name, system_prompt_content], outputs=[])
|
| 88 |
submit_button.click(chat_with_model, inputs=[user_input, system_prompt_content, model_selector], outputs=[output])
|
| 89 |
|