Spaces:
Running
Running
add gemma 3n, smallthinker 4b0.6b and falcon h1-1.5b
Browse files
app.py
CHANGED
|
@@ -24,6 +24,28 @@ cancel_event = threading.Event()
|
|
| 24 |
# Torch-Compatible Model Definitions with Adjusted Descriptions
|
| 25 |
# ------------------------------
|
| 26 |
MODELS = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
"Qwen/Qwen3-14B-FP8": {"repo_id": "Qwen/Qwen3-14B-FP8", "description": "Qwen/Qwen3-14B-FP8"},
|
| 28 |
#"Qwen/Qwen3-32B-FP8": {"repo_id": "Qwen/Qwen3-32B-FP8", "description": "Qwen/Qwen3-32B-FP8"},
|
| 29 |
"DeepSeek-R1-0528-Qwen3-8B": {"repo_id": "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B", "description": "DeepSeek-R1-0528-Qwen3-8B"},
|
|
|
|
| 24 |
# Torch-Compatible Model Definitions with Adjusted Descriptions
|
| 25 |
# ------------------------------
|
| 26 |
MODELS = {
|
| 27 |
+
# … your existing entries …
|
| 28 |
+
|
| 29 |
+
# Gemma 3n “effective” variants (official Google repos)
|
| 30 |
+
"Gemma-3n-E2B": {
|
| 31 |
+
"repo_id": "google/gemma-3n-E2B",
|
| 32 |
+
"description":"Gemma 3n base model with effective 2 B parameters (≈2 GB VRAM)"
|
| 33 |
+
},
|
| 34 |
+
"Gemma-3n-E4B": {
|
| 35 |
+
"repo_id": "google/gemma-3n-E4B",
|
| 36 |
+
"description":"Gemma 3n base model with effective 4 B parameters (≈3 GB VRAM)"
|
| 37 |
+
},
|
| 38 |
+
|
| 39 |
+
# PowerInfer SmallThinker (instruction‑tuned)
|
| 40 |
+
"SmallThinker-4BA0.6B-Instruct": {
|
| 41 |
+
"repo_id": "PowerInfer/SmallThinker-4BA0.6B-Instruct",
|
| 42 |
+
"description":"SmallThinker 4 B backbone with 0.6 B activated parameters, instruction‑tuned"
|
| 43 |
+
},
|
| 44 |
+
# TIIUAE Falcon‑H1 (instruction‑tuned)
|
| 45 |
+
"Falcon-H1-1.5B-Instruct": {
|
| 46 |
+
"repo_id": "tiiuae/Falcon-H1-1.5B-Instruct",
|
| 47 |
+
"description":"Falcon‑H1 model with 1.5 B parameters, instruction‑tuned"
|
| 48 |
+
},
|
| 49 |
"Qwen/Qwen3-14B-FP8": {"repo_id": "Qwen/Qwen3-14B-FP8", "description": "Qwen/Qwen3-14B-FP8"},
|
| 50 |
#"Qwen/Qwen3-32B-FP8": {"repo_id": "Qwen/Qwen3-32B-FP8", "description": "Qwen/Qwen3-32B-FP8"},
|
| 51 |
"DeepSeek-R1-0528-Qwen3-8B": {"repo_id": "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B", "description": "DeepSeek-R1-0528-Qwen3-8B"},
|