Spaces:
Runtime error
Runtime error
| from dotenv import load_dotenv, find_dotenv | |
| from fasthtml.common import * | |
| from groq import Groq | |
| # Set up the app, including daisyui and tailwind for the chat component | |
| tlink = Script(src="https://cdn.tailwindcss.com"), | |
| dlink = Link(rel="stylesheet", href="https://cdn.jsdelivr.net/npm/daisyui@4.11.1/dist/full.min.css") | |
| app = FastHTML(hdrs=(tlink, dlink, picolink)) | |
| load_dotenv(find_dotenv()) #os.environ["GROQ_API_KEY"] | |
| client = Groq() | |
| sp = """You are a helpful and concise assistant.""" | |
| messages = [] | |
| # Chat message component, polling if message is still being generated | |
| def ChatMessage(msg_idx): | |
| msg = messages[msg_idx] | |
| text = "..." if msg['content'] == "" else msg['content'] | |
| bubble_class = f"chat-bubble-{'primary' if msg['role'] == 'user' else 'secondary'}" | |
| chat_class = f"chat-{'end' if msg['role'] == 'user' else 'start'}" | |
| generating = 'generating' in messages[msg_idx] and messages[msg_idx]['generating'] | |
| print(generating) | |
| stream_args = {"hx_trigger":"every 0.1s", "hx_swap":"outerHTML", "hx_get":f"/chat_message/{msg_idx}"} | |
| return Div(Div(msg['role'], cls="chat-header"), | |
| Div(text, cls=f"chat-bubble {bubble_class}"), | |
| cls=f"chat {chat_class}", id=f"chat-message-{msg_idx}", | |
| **stream_args if generating else {}) | |
| # Route that gets polled while streaming | |
| def get_chat_message(msg_idx:int): | |
| if msg_idx >= len(messages): return "" | |
| return ChatMessage(msg_idx) | |
| # The input field for the user message. Also used to clear the | |
| # input field after sending a message via an OOB swap | |
| def ChatInput(): | |
| return Input(type="text", name='msg', id='msg-input', | |
| placeholder="Type a message", | |
| cls="input input-bordered w-full", hx_swap_oob='true') | |
| # The main screen | |
| def get(): | |
| page = Body(H1('Agentic Bot 1002'), | |
| Div(*[ChatMessage(msg) for msg in messages], | |
| id="chatlist", cls="chat-box h-[73vh] overflow-y-auto"), | |
| Form(Group(ChatInput(), Button("Send", cls="btn btn-primary")), | |
| hx_post="/", hx_target="#chatlist", hx_swap="beforeend", | |
| cls="flex space-x-2 mt-2", | |
| ), cls="p-4 max-w-lg mx-auto") | |
| return Title('Agentic Bot 1002'), page | |
| # Run the chat model in a separate thread | |
| def get_response(r, idx): #.choices[0].delta.content | |
| for chunk in r: | |
| if chunk.choices[0].delta.content is not None: | |
| messages[idx]["content"] += chunk.choices[0].delta.content | |
| messages[idx]["generating"] = False | |
| # Handle the form submission | |
| def post(msg:str): | |
| messages.append({"role":"system", "content":sp}) | |
| idx = len(messages) | |
| messages.append({"role":"user", "content":msg}) | |
| # r = cli(messages, sp=sp, stream=True) # Send message to chat model (with streaming) | |
| # Remove 'generating' key-value pair if it exists | |
| for entry in messages: | |
| if 'generating' in entry: | |
| del entry['generating'] | |
| stream = client.chat.completions.create( | |
| messages=messages, | |
| model="llama3-8b-8192", | |
| temperature=0.5, | |
| max_tokens=1024, | |
| top_p=1, | |
| stop=None, | |
| stream=True, | |
| ) | |
| messages.append({"role":"assistant", "generating":True, "content":""}) # Response initially blank | |
| get_response(stream, idx+1) # Start a new thread to fill in content | |
| return (ChatMessage(idx), # The user's message | |
| ChatMessage(idx+1), # The chatbot's response | |
| ChatInput()) # And clear the input field via an OOB swap | |
| if __name__ == '__main__': uvicorn.run("polling:app", host='0.0.0.0', port=8000, reload=True) |