Luigi commited on
Commit
048cfc4
·
verified ·
1 Parent(s): b9efb74

add 4 models from qwen3 family

Browse files
Files changed (1) hide show
  1. app.py +18 -8
app.py CHANGED
@@ -27,14 +27,24 @@ cancel_event = threading.Event()
27
  # ------------------------------
28
  MODELS = {
29
  # ~30.5B total parameters (MoE: 3.3B activated)
30
- # "Qwen3-30B-A3B-Thinking-2507-FP8": {
31
- # "repo_id": "Qwen/Qwen3-30B-A3B-Thinking-2507-FP8",
32
- # "description": "FP8-quantized MoE model with 30.5B total parameters (3.3B activated), 128 experts (8 activated), 48 layers, and native 262,144-token context. Optimized for complex reasoning tasks with enhanced thinking capabilities in mathematics, coding, science, and agent benchmarks. Supports only thinking mode; includes automatic reasoning delimiters."
33
- # },
34
- # "Qwen3-30B-A3B-Instruct-2507-FP8": {
35
- # "repo_id": "Qwen/Qwen3-30B-A3B-Instruct-2507-FP8",
36
- # "description": "FP8-quantized instruct-tuned variant of Qwen3-30B-A3B (30.5B total params, 3.3B activated), featuring strong general capabilities in instruction following, tool usage, text generation, and 256K long-context understanding. Ideal for agentic and multi-turn dialogue applications."
37
- # },
 
 
 
 
 
 
 
 
 
 
38
 
39
  # ~235B total parameters (MoE: 22B activated) — included for reference if added later
40
  # "Qwen3-235B-A22B-Thinking": { ... },
 
27
  # ------------------------------
28
  MODELS = {
29
  # ~30.5B total parameters (MoE: 3.3B activated)
30
+ "Qwen3-30B-A3B-Instruct-2507-AWQ-4bit": {
31
+ "repo_id": "cpatonn/Qwen3-30B-A3B-Instruct-2507-AWQ-4bit",
32
+ "description": "4-bit AWQ quantized instruct-tuned MoE model based on Qwen3-30B-A3B-Instruct-2507. Features 30.5B total parameters (3.3B activated), 128 experts (8 activated), 48 layers, and native 262,144-token context. Excels in instruction following, logical reasoning, multilingualism, coding, and long-context understanding. Supports only non-thinking mode (no <think> blocks). Quantized using AWQ (W4A16) with lm_head and gating layers preserved in higher precision."
33
+ },
34
+ "Qwen3-30B-A3B-Thinking-2507-AWQ-4bit": {
35
+ "repo_id": "cpatonn/Qwen3-30B-A3B-Thinking-2507-AWQ-4bit",
36
+ "description": "4-bit AWQ quantized thinking-mode MoE model based on Qwen3-30B-A3B-Thinking-2507. Contains 30.5B total parameters (3.3B activated), 128 experts (8 activated), 48 layers, and 262,144-token native context. Optimized for deep reasoning in mathematics, science, coding, and agent tasks. Outputs include automatic reasoning delimiters (<think>...</think>). Quantized with AWQ (W4A16), preserving lm_head and expert gating layers."
37
+ },
38
+
39
+ # ~80B total parameters (MoE: 3B activated)
40
+ "Qwen3-Next-80B-A3B-Instruct-AWQ-4bit": {
41
+ "repo_id": "cpatonn/Qwen3-Next-80B-A3B-Instruct-AWQ-4bit",
42
+ "description": "4-bit AWQ quantized instruct model from the Qwen3-Next series. Features 80B total parameters (3B activated), hybrid Gated DeltaNet + Gated Attention architecture, 512 experts (10 activated + 1 shared), and native 262,144-token context (extendable to 1M tokens with YaRN). Delivers performance comparable to Qwen3-235B on many benchmarks while offering superior ultra-long-context efficiency. Supports only non-thinking mode. Note: May require re-quantization for stable inference (as of Sept 2025)."
43
+ },
44
+ "Qwen3-Next-80B-A3B-Thinking-AWQ-4bit": {
45
+ "repo_id": "cpatonn/Qwen3-Next-80B-A3B-Thinking-AWQ-4bit",
46
+ "description": "4-bit AWQ quantized thinking-mode variant of Qwen3-Next-80B-A3B. Combines 80B total parameters (3B activated), hybrid attention (Gated DeltaNet + Gated Attention), and 512-expert MoE (10 activated + 1 shared) for advanced reasoning over ultra-long contexts (natively 262K, extendable to 1M tokens). Designed for complex problem-solving with automatic reasoning trace generation. Quantized using AWQ; intended for high-end agentic and analytical workloads."
47
+ },
48
 
49
  # ~235B total parameters (MoE: 22B activated) — included for reference if added later
50
  # "Qwen3-235B-A22B-Thinking": { ... },