Update README.md
Browse files
README.md
CHANGED
|
@@ -5,3 +5,36 @@ base_model:
|
|
| 5 |
- Qwen/Qwen2.5-7B-Instruct
|
| 6 |
---
|
| 7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
- Qwen/Qwen2.5-7B-Instruct
|
| 6 |
---
|
| 7 |
|
| 8 |
+
# MachineLearningLM
|
| 9 |
+
|
| 10 |
+
## model summary
|
| 11 |
+
|
| 12 |
+
Can LLMs learn from 1,000 in-context examples?
|
| 13 |
+
|
| 14 |
+
Introducing **MachineLearningLM** 🧪📊 — a model continuously pretrained on millions of synthetic tabular ML tasks, enabling robust many-shot in-context learning.
|
| 15 |
+
|
| 16 |
+
📈 **Scales from 8 to 1,024 examples**
|
| 17 |
+
📈 **~15% improvement** on unseen tabular tasks compared to o3-mini / GPT-5-mini / Qwen-2.5-7B
|
| 18 |
+
🌲 **Random-Forest–level robustness**
|
| 19 |
+
🧠 **MMLU score: 75.4%**
|
| 20 |
+
|
| 21 |
+
📄 Read the paper: https://arxiv.org/abs/2509.06806
|
| 22 |
+
|
| 23 |
+
<img src="https://github.githubassets.com/favicons/favicon.svg" width="16" height="16"> GitHub: https://github.com/HaoAreYuDong/MachineLearningLM
|
| 24 |
+
|
| 25 |
+
## evaluation and validation
|
| 26 |
+
|
| 27 |
+
We have developed an automated evaluation framework — simply configure the parameters to easily perform validation and evaluation. The code is now open-sourced at our GitHub.
|
| 28 |
+
|
| 29 |
+
### **Quick Start**
|
| 30 |
+
|
| 31 |
+
```bash
|
| 32 |
+
pip install -r requirement.txt
|
| 33 |
+
python ./src/evaluation/model_pred/dl_model_pred.py \
|
| 34 |
+
--input_dir ./demo_input.jsonl \
|
| 35 |
+
--output_dir ./demo_output.jsonl \
|
| 36 |
+
--model_name hf_repo/model_name
|
| 37 |
+
```
|
| 38 |
+
|
| 39 |
+
For more usage details, please visit our GitHub.
|
| 40 |
+
|