File size: 3,675 Bytes
2a88707
 
34d0739
a05fede
34d0739
a05fede
34d0739
 
2a88707
34d0739
 
 
2a88707
34d0739
2a88707
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
34d0739
 
 
2a88707
 
 
 
34d0739
2a88707
34d0739
2a88707
34d0739
 
 
 
 
2763883
34d0739
2a88707
 
 
a05fede
34d0739
 
 
a05fede
34d0739
2a88707
34d0739
2a88707
2763883
2a88707
 
 
34d0739
2a88707
 
34d0739
 
 
 
 
2a88707
 
34d0739
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
import streamlit as st
import cohere
import os

st.set_page_config(page_title="Cohere Chat", layout="wide")

AI_PFP = "media/pfps/cohere-pfp.png"
USER_PFP = "media/pfps/user-pfp.jpg"

if not os.path.exists(AI_PFP) or not os.path.exists(USER_PFP):
    st.error("Missing profile pictures in media/pfps directory")
    st.stop()

model_info = {
    "command-a-03-2025": {
        "description": "Command A is our most performant model to date, excelling at tool use, agents, retrieval augmented generation (RAG), and multilingual use cases.",
        "context": "256K",
        "output": "8K"
    },
    "command-r7b-12-2024": {
        "description": "Small, fast update excelling at RAG, tool use, and complex reasoning tasks.",
        "context": "128K",
        "output": "4K"
    },
    "command-r-plus-04-2024": {
        "description": "Instruction-following model for complex RAG workflows and multi-step tool use.",
        "context": "128K",
        "output": "4K"
    },
    "command-r-plus": {
        "description": "Alias for command-r-plus-04-2024.",
        "context": "128K",
        "output": "4K"
    },
    "command-r-08-2024": {
        "description": "Updated Command R model from August 2024.",
        "context": "128K",
        "output": "4K"
    },
    "command-r-03-2024": {
        "description": "Instruction-following model for code generation, RAG, and agents.",
        "context": "128K",
        "output": "4K"
    },
    "command-r": {
        "description": "Alias for command-r-03-2024.",
        "context": "128K",
        "output": "4K"
    },
    "command": {
        "description": "Conversational model with long context capabilities.",
        "context": "4K",
        "output": "4K"
    },
    "command-nightly": {
        "description": "Experimental nightly build (not for production).",
        "context": "128K",
        "output": "4K"
    },
    "command-light": {
        "description": "Faster lightweight version of command.",
        "context": "4K",
        "output": "4K"
    },
    "command-light-nightly": {
        "description": "Experimental nightly build of command-light.",
        "context": "128K",
        "output": "4K"
    }
}

with st.sidebar:
    st.title("Settings")
    api_key = st.text_input("Cohere API Key", type="password")
    selected_model = st.selectbox("Model", options=list(model_info.keys()))
    st.divider()
    st.image(AI_PFP, width=60)
    st.subheader(selected_model)
    st.markdown(model_info[selected_model]["description"])
    st.caption(f"Context: {model_info[selected_model]['context']}")
    st.caption(f"Output: {model_info[selected_model]['output']}")

st.title(f"Chat - {selected_model}")

if "messages" not in st.session_state:
    st.session_state.messages = []

for msg in st.session_state.messages:
    with st.chat_message(msg["role"], avatar=USER_PFP if msg["role"] == "user" else AI_PFP):
        st.markdown(msg["content"])

if prompt := st.chat_input("Message..."):
    if not api_key:
        st.error("API key required")
        st.stop()
    
    st.session_state.messages.append({"role": "user", "content": prompt})
    with st.chat_message("user", avatar=USER_PFP):
        st.markdown(prompt)

    try:
        co = cohere.ClientV2(api_key)
        with st.chat_message("assistant", avatar=AI_PFP):
            response = co.chat(model=selected_model, messages=st.session_state.messages)
            reply = response.text if hasattr(response, 'text') else "Error processing response"
            st.markdown(reply)
        st.session_state.messages.append({"role": "assistant", "content": reply})
    
    except Exception as e:
        st.error(f"Error: {str(e)}")