Tamil-Qwen3-4B-Inst
Tamil-Qwen3-4B-Inst is a lightweight Tamil-English bilingual instruction-tuned model designed for efficient deployment and strong performance on instruction-following tasks.
Model Overview
We developed a specialized model by adapting state-of-the-art open-source base models through:
- Continual Pretraining on the Tamil Wikipedia dataset
- Instruction Finetuning using high-quality, human-annotated Tamil instruction datasets from the Aya Dataset
Model Summary
| Feature | Description |
|---|---|
| Base Model | Qwen3-4B(trained in tamil wikipedia dataset) |
| Parameters | 4B |
| Datasets | https://huggingface.co/datasets/CohereLabs/aya_dataset,https://huggingface.co/datasets/wikimedia/wikipedia/viewer/20231101.ta |
| Training Precision | bfloat16 |
| Epochs (Total) | 6 |
| Languages | Tamil, English |
| Use Case | Instruction following, conversational AI, and Tamil language tasks |
Prompting Format
Prompt Template Without Input
{system_prompt}
### Instruction:
{instruction or query}
### Response:
{response}
Prompt Template With Input
{system_prompt}
### Instruction:
{instruction or query}
### Input:
{input}
### Response:
{response}
Citation
If you use this model in your research, please cite:
@misc{tamilqwen3_4b_inst,
title={Tamil-Qwen3-4B-Inst: Efficient Bilingual Instruction-Tuned Model},
author={AITamilNadu},
year={2025},
url={https://huggingface.co/aitamilnadu/Tamil-Qwen3-4B-Inst}
}
Evaluation and Benchmarks
| Benchmark | Score |
|---|---|
| Average | 52.08% |
| ARC Challenge | 45.48% |
| HellaSwag | 61.64% |
| MMLU | 56.05% |
| TruthfulQA | 39.58% |
| Winogrande | 59.43% |
| GSM8K | 40.64% |
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support