Spaces:
Running
Running
File size: 5,152 Bytes
ea97ae9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 |
# model_options.py
"""
Centralized model catalog for Glossarion UIs.
Returned list should mirror the main GUI model dropdown.
"""
from typing import List
def get_model_options() -> List[str]:
return [
# OpenAI Models
"gpt-4o", "gpt-4o-mini", "gpt-4-turbo", "gpt-4.1-nano", "gpt-4.1-mini", "gpt-4.1",
"gpt-3.5-turbo", "gpt-3.5-turbo-16k", "gpt-4", "gpt-4-32k",
"gpt-5-mini","gpt-5","gpt-5-nano",
"o1-preview", "o1-mini", "o3", "o4-mini",
# Google Gemini Models
"gemini-2.0-flash","gemini-2.0-flash-lite",
"gemini-2.5-flash","gemini-2.5-flash-lite", "gemini-2.5-pro", "gemini-pro", "gemini-pro-vision",
# Anthropic Claude Models
"claude-opus-4-20250514", "claude-sonnet-4-20250514",
"claude-3-5-sonnet-20241022", "claude-3-7-sonnet-20250219",
"claude-3-opus-20240229", "claude-3-sonnet-20240229", "claude-3-haiku-20240307",
"claude-2.1", "claude-2", "claude-instant-1.2",
# Grok Models
"grok-4-0709", "grok-4-fast", "grok-4-fast-reasoning", "grok-4-fast-reasoning-latest", "grok-3", "grok-3-mini",
# Vertex AI Model Garden - Claude models (confirmed)
"claude-4-opus@20250514",
"claude-4-sonnet@20250514",
"claude-opus-4@20250514",
"claude-sonnet-4@20250514",
"claude-3-7-sonnet@20250219",
"claude-3-5-sonnet@20240620",
"claude-3-5-sonnet-v2@20241022",
"claude-3-opus@20240229",
"claude-3-sonnet@20240229",
"claude-3-haiku@20240307",
# Alternative format with vertex_ai prefix
"vertex/claude-3-7-sonnet@20250219",
"vertex/claude-3-5-sonnet@20240620",
"vertex/claude-3-opus@20240229",
"vertex/claude-4-opus@20250514",
"vertex/claude-4-sonnet@20250514",
"vertex/gemini-1.5-pro",
"vertex/gemini-1.5-flash",
"vertex/gemini-2.0-flash",
"vertex/gemini-2.5-pro",
"vertex/gemini-2.5-flash",
"vertex/gemini-2.5-flash-lite",
# Chute AI
"chutes/openai/gpt-oss-120b",
"chutes/deepseek-ai/DeepSeek-V3.1",
# DeepSeek Models
"deepseek-chat", "deepseek-coder", "deepseek-coder-33b-instruct",
# Mistral Models
"mistral-large", "mistral-medium", "mistral-small", "mistral-tiny",
"mixtral-8x7b-instruct", "mixtral-8x22b", "codestral-latest",
# Meta Llama Models (via Together/other providers)
"llama-2-7b-chat", "llama-2-13b-chat", "llama-2-70b-chat",
"llama-3-8b-instruct", "llama-3-70b-instruct", "codellama-34b-instruct",
# Yi Models
"yi-34b-chat", "yi-34b-chat-200k", "yi-6b-chat",
# Qwen Models
"qwen-72b-chat", "qwen-14b-chat", "qwen-7b-chat", "qwen-plus", "qwen-turbo",
# Cohere Models
"command", "command-light", "command-nightly", "command-r", "command-r-plus",
# AI21 Models
"j2-ultra", "j2-mid", "j2-light", "jamba-instruct",
# Perplexity Models
"perplexity-70b-online", "perplexity-7b-online", "pplx-70b-online", "pplx-7b-online",
# Groq Models (usually with suffix)
"llama-3-70b-groq", "llama-3-8b-groq", "mixtral-8x7b-groq",
# Chinese Models
"glm-4", "glm-3-turbo", "chatglm-6b", "chatglm2-6b", "chatglm3-6b",
"baichuan-13b-chat", "baichuan2-13b-chat",
"moonshot-v1-8k", "moonshot-v1-32k", "moonshot-v1-128k",
# Other Models
"falcon-40b-instruct", "falcon-7b-instruct",
"phi-2", "phi-3-mini", "phi-3-small", "phi-3-medium",
"orca-2-13b", "orca-2-7b",
"vicuna-13b", "vicuna-7b",
"alpaca-7b",
"wizardlm-70b", "wizardlm-13b",
"openchat-3.5",
# For POE, prefix with 'poe/'
"poe/gpt-4", "poe/gpt-4o", "poe/gpt-4.5", "poe/gpt-4.1",
"poe/claude-3-opus", "poe/claude-4-opus", "poe/claude-3-sonnet", "poe/claude-4-sonnet",
"poe/claude", "poe/Assistant",
"poe/gemini-2.5-flash", "poe/gemini-2.5-pro",
# For OR, prevfix with 'or/'
"or/google/gemini-2.5-pro",
"or/google/gemini-2.5-flash",
"or/google/gemini-2.5-flash-lite",
"or/openai/gpt-5",
"or/openai/gpt-5-mini",
"or/openai/gpt-5-nano",
"or/openai/chatgpt-4o-latest",
"or/deepseek/deepseek-r1-0528:free",
"or/google/gemma-3-27b-it:free",
# For ElectronHub, prefix with 'eh/'
"eh/gpt-4", "eh/gpt-3.5-turbo", "eh/claude-3-opus", "eh/claude-3-sonnet",
"eh/llama-2-70b-chat", "eh/yi-34b-chat-200k", "eh/mistral-large",
"eh/gemini-pro", "eh/deepseek-coder-33b",
# Last Resort
"deepl", # Will use DeepL API
"google-translate-free", # Uses free web endpoint (no key)
"google-translate", # Will use Google Cloud Translate
]
|