4-bit LLaDA model
#9
by
chentianqi
- opened
Thank you for your great works. We have released the 4-bit GPTQ quantized LLaDA model on Hugging Face:
Based on the published evaluation code, we have evaluated the quantized base model. The results are as follows:
| Dataset | GPTQ-4bit | FP16 |
|---|---|---|
| MMLU | 65.20 | 65.90 |
| CMMLU | 69.23 | 69.90 |
| ARC-Challenge | 45.48 | 47.90 |