eval_name
stringlengths 12
111
| Precision
stringclasses 3
values | Type
stringclasses 6
values | T
stringclasses 6
values | Weight type
stringclasses 2
values | Architecture
stringclasses 52
values | Model
stringlengths 355
689
| fullname
stringlengths 4
102
| Model sha
stringlengths 0
40
| Average ⬆️
float64 1.03
52
| Hub License
stringclasses 26
values | Hub ❤️
int64 0
5.9k
| #Params (B)
int64 -1
140
| Available on the hub
bool 2
classes | MoE
bool 2
classes | Flagged
bool 2
classes | Chat Template
bool 2
classes | CO₂ cost (kg)
float64 0.03
107
| IFEval Raw
float64 0
0.9
| IFEval
float64 0
90
| BBH Raw
float64 0.27
0.75
| BBH
float64 0.81
63.5
| MATH Lvl 5 Raw
float64 0
0.51
| MATH Lvl 5
float64 0
50.7
| GPQA Raw
float64 0.22
0.44
| GPQA
float64 0
24.9
| MUSR Raw
float64 0.29
0.6
| MUSR
float64 0
38.5
| MMLU-PRO Raw
float64 0.1
0.73
| MMLU-PRO
float64 0
70
| Merged
bool 2
classes | Official Providers
bool 2
classes | Upload To Hub Date
stringclasses 424
values | Submission Date
stringclasses 169
values | Generation
int64 0
10
| Base Model
stringlengths 4
102
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0-hero_Matter-0.2-7B-DPO_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/0-hero/Matter-0.2-7B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">0-hero/Matter-0.2-7B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/0-hero__Matter-0.2-7B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
0-hero/Matter-0.2-7B-DPO
|
26a66f0d862e2024ce4ad0a09c37052ac36e8af6
| 8.805656
|
apache-2.0
| 3
| 7
| true
| false
| false
| true
| 0.609587
| 0.330279
| 33.027921
| 0.359625
| 10.055525
| 0.008308
| 0.830816
| 0.259228
| 1.230425
| 0.381375
| 5.871875
| 0.116356
| 1.817376
| false
| false
|
2024-04-13
|
2024-08-05
| 0
|
0-hero/Matter-0.2-7B-DPO
|
01-ai_Yi-1.5-34B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B
|
4b486f81c935a2dadde84c6baa1e1370d40a098f
| 25.621318
|
apache-2.0
| 47
| 34
| true
| false
| false
| false
| 11.351699
| 0.284117
| 28.411725
| 0.597639
| 42.749363
| 0.151813
| 15.181269
| 0.365772
| 15.436242
| 0.423604
| 11.217188
| 0.466589
| 40.732122
| false
| true
|
2024-05-11
|
2024-06-12
| 0
|
01-ai/Yi-1.5-34B
|
01-ai_Yi-1.5-34B-32K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-32K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-32K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-32K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-32K
|
2c03a29761e4174f20347a60fbe229be4383d48b
| 26.67756
|
apache-2.0
| 36
| 34
| true
| false
| false
| false
| 11.577314
| 0.311869
| 31.186917
| 0.601569
| 43.381847
| 0.151057
| 15.10574
| 0.363255
| 15.100671
| 0.439823
| 14.077865
| 0.470911
| 41.212323
| false
| true
|
2024-05-15
|
2024-06-12
| 0
|
01-ai/Yi-1.5-34B-32K
|
01-ai_Yi-1.5-34B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-Chat
|
f3128b2d02d82989daae566c0a7eadc621ca3254
| 32.892233
|
apache-2.0
| 257
| 34
| true
| false
| false
| true
| 11.211922
| 0.606676
| 60.667584
| 0.608375
| 44.262826
| 0.249245
| 24.924471
| 0.364933
| 15.324385
| 0.428198
| 13.058073
| 0.452045
| 39.116061
| false
| true
|
2024-05-10
|
2024-06-12
| 0
|
01-ai/Yi-1.5-34B-Chat
|
01-ai_Yi-1.5-34B-Chat-16K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-Chat-16K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-Chat-16K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-Chat-16K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-Chat-16K
|
ff74452e11f0f749ab872dc19b1dd3813c25c4d8
| 29.239909
|
apache-2.0
| 27
| 34
| true
| false
| false
| true
| 3.387011
| 0.45645
| 45.645
| 0.610022
| 44.536157
| 0.203927
| 20.392749
| 0.338087
| 11.744966
| 0.43976
| 13.736719
| 0.454455
| 39.383865
| false
| true
|
2024-05-15
|
2024-07-15
| 0
|
01-ai/Yi-1.5-34B-Chat-16K
|
01-ai_Yi-1.5-6B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-6B
|
cab51fce425b4c1fb19fccfdd96bd5d0908c1657
| 16.695346
|
apache-2.0
| 30
| 6
| true
| false
| false
| false
| 1.184757
| 0.26166
| 26.166017
| 0.449258
| 22.027905
| 0.063444
| 6.344411
| 0.313758
| 8.501119
| 0.437406
| 13.309115
| 0.314412
| 23.823508
| false
| true
|
2024-05-11
|
2024-08-10
| 0
|
01-ai/Yi-1.5-6B
|
01-ai_Yi-1.5-6B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-6B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-6B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-6B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-6B-Chat
|
3f64d3f159c6ad8494227bb77e2a7baef8cd808b
| 20.983906
|
apache-2.0
| 41
| 6
| true
| false
| false
| true
| 0.96884
| 0.514527
| 51.452701
| 0.457131
| 23.678723
| 0.054381
| 5.438066
| 0.302013
| 6.935123
| 0.439177
| 14.030469
| 0.319315
| 24.368351
| false
| true
|
2024-05-11
|
2024-10-22
| 0
|
01-ai/Yi-1.5-6B-Chat
|
01-ai_Yi-1.5-9B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B
|
8cfde9604384c50137bee480b8cef8a08e5ae81d
| 22.141313
|
apache-2.0
| 47
| 8
| true
| false
| false
| false
| 0.734446
| 0.293584
| 29.358436
| 0.514294
| 30.500717
| 0.113293
| 11.329305
| 0.379195
| 17.225951
| 0.432781
| 12.03099
| 0.391622
| 32.402482
| false
| true
|
2024-05-11
|
2024-06-12
| 0
|
01-ai/Yi-1.5-9B
|
01-ai_Yi-1.5-9B-32K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-32K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-32K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-32K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-32K
|
116561dfae63af90f9d163b43077629e0e916bb1
| 19.78461
|
apache-2.0
| 18
| 8
| true
| false
| false
| false
| 0.784037
| 0.230311
| 23.031113
| 0.496332
| 28.937012
| 0.106495
| 10.649547
| 0.35906
| 14.541387
| 0.418615
| 10.826823
| 0.376496
| 30.721779
| false
| true
|
2024-05-15
|
2024-06-12
| 0
|
01-ai/Yi-1.5-9B-32K
|
01-ai_Yi-1.5-9B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-Chat
|
bc87d8557c98dc1e5fdef6ec23ed31088c4d3f35
| 27.894417
|
apache-2.0
| 135
| 8
| true
| false
| false
| true
| 0.726772
| 0.604553
| 60.455259
| 0.555906
| 36.952931
| 0.127644
| 12.76435
| 0.334732
| 11.297539
| 0.425906
| 12.838281
| 0.397523
| 33.058141
| false
| true
|
2024-05-10
|
2024-06-12
| 0
|
01-ai/Yi-1.5-9B-Chat
|
01-ai_Yi-1.5-9B-Chat-16K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-Chat-16K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-Chat-16K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-Chat-16K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-Chat-16K
|
2b397e5f0fab87984efa66856c5c4ed4bbe68b50
| 23.035282
|
apache-2.0
| 34
| 8
| true
| false
| false
| true
| 0.792373
| 0.421404
| 42.14041
| 0.515338
| 31.497609
| 0.134441
| 13.444109
| 0.308725
| 7.829978
| 0.409906
| 10.038281
| 0.399352
| 33.261303
| false
| true
|
2024-05-15
|
2024-06-12
| 0
|
01-ai/Yi-1.5-9B-Chat-16K
|
01-ai_Yi-34B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B
|
e1e7da8c75cfd5c44522228599fd4d2990cedd1c
| 22.385715
|
apache-2.0
| 1,287
| 34
| true
| false
| false
| false
| 12.828742
| 0.304575
| 30.457519
| 0.54571
| 35.542431
| 0.052115
| 5.21148
| 0.366611
| 15.548098
| 0.411854
| 9.648438
| 0.441157
| 37.906324
| false
| true
|
2023-11-01
|
2024-06-12
| 0
|
01-ai/Yi-34B
|
01-ai_Yi-34B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B-200K
|
8ac1a1ebe011df28b78ccd08012aeb2222443c77
| 19.887594
|
apache-2.0
| 316
| 34
| true
| false
| false
| false
| 12.751928
| 0.154249
| 15.424851
| 0.544182
| 36.02211
| 0.049849
| 4.984894
| 0.356544
| 14.205817
| 0.381719
| 9.414844
| 0.453457
| 39.27305
| false
| true
|
2023-11-06
|
2024-06-12
| 0
|
01-ai/Yi-34B-200K
|
01-ai_Yi-34B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B-Chat
|
2e528b6a80fb064a0a746c5ca43114b135e30464
| 23.962312
|
apache-2.0
| 345
| 34
| true
| false
| false
| true
| 12.562848
| 0.469889
| 46.988878
| 0.556087
| 37.623988
| 0.046828
| 4.682779
| 0.338087
| 11.744966
| 0.397844
| 8.363802
| 0.409325
| 34.369459
| false
| true
|
2023-11-22
|
2024-06-12
| 0
|
01-ai/Yi-34B-Chat
|
01-ai_Yi-6B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B
|
7f7fb7662fd8ec09029364f408053c954986c8e5
| 13.611617
|
apache-2.0
| 371
| 6
| true
| false
| false
| false
| 0.549275
| 0.289338
| 28.933785
| 0.430923
| 19.408505
| 0.015861
| 1.586103
| 0.269295
| 2.572707
| 0.393687
| 7.044271
| 0.299119
| 22.124335
| false
| true
|
2023-11-01
|
2024-06-12
| 0
|
01-ai/Yi-6B
|
01-ai_Yi-6B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B-200K
|
4a74338e778a599f313e9fa8f5bc08c717604420
| 11.933158
|
apache-2.0
| 173
| 6
| true
| false
| false
| false
| 0.563212
| 0.084331
| 8.433069
| 0.428929
| 20.14802
| 0.01435
| 1.435045
| 0.281879
| 4.250559
| 0.45874
| 16.842448
| 0.284408
| 20.489805
| false
| true
|
2023-11-06
|
2024-06-12
| 0
|
01-ai/Yi-6B-200K
|
01-ai_Yi-6B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B-Chat
|
01f7fabb6cfb26efeb764da4a0a19cad2c754232
| 14.004357
|
apache-2.0
| 63
| 6
| true
| false
| false
| true
| 0.555333
| 0.339521
| 33.952136
| 0.41326
| 17.000167
| 0.006798
| 0.679758
| 0.294463
| 5.928412
| 0.368792
| 3.565625
| 0.3061
| 22.900044
| false
| true
|
2023-11-22
|
2024-06-12
| 0
|
01-ai/Yi-6B-Chat
|
01-ai_Yi-9B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-9B
|
b4a466d95091696285409f1dcca3028543cb39da
| 17.774103
|
apache-2.0
| 186
| 8
| true
| false
| false
| false
| 0.765332
| 0.270878
| 27.087794
| 0.493961
| 27.626956
| 0.053625
| 5.362538
| 0.317953
| 9.060403
| 0.405406
| 8.909115
| 0.35738
| 28.597813
| false
| true
|
2024-03-01
|
2024-06-12
| 0
|
01-ai/Yi-9B
|
01-ai_Yi-9B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-9B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-9B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-9B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-9B-200K
|
8c93accd5589dbb74ee938e103613508c4a9b88d
| 17.74214
|
apache-2.0
| 75
| 8
| true
| false
| false
| false
| 0.774491
| 0.232709
| 23.270921
| 0.47933
| 26.492495
| 0.067221
| 6.722054
| 0.315436
| 8.724832
| 0.429406
| 12.109115
| 0.362201
| 29.133422
| false
| true
|
2024-03-15
|
2024-06-12
| 0
|
01-ai/Yi-9B-200K
|
01-ai_Yi-Coder-9B-Chat_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-Coder-9B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-Coder-9B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-Coder-9B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-Coder-9B-Chat
|
356a1f8d4e4a606d0b879e54191ca809918576b8
| 16.872696
|
apache-2.0
| 191
| 8
| true
| false
| false
| true
| 0.909766
| 0.481704
| 48.17041
| 0.48142
| 25.943153
| 0.033233
| 3.323263
| 0.247483
| 0
| 0.399177
| 7.963802
| 0.24252
| 15.83555
| false
| true
|
2024-08-21
|
2024-09-12
| 1
|
01-ai/Yi-Coder-9B
|
152334H_miqu-1-70b-sf_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/152334H/miqu-1-70b-sf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">152334H/miqu-1-70b-sf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/152334H__miqu-1-70b-sf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
152334H/miqu-1-70b-sf
|
1dca4cce36f01f2104ee2e6b97bac6ff7bb300c1
| 29.059643
| 219
| 68
| false
| false
| false
| false
| 6.098986
| 0.518174
| 51.8174
| 0.610236
| 43.807147
| 0.122356
| 12.23565
| 0.350671
| 13.422819
| 0.458208
| 17.209375
| 0.422789
| 35.86547
| false
| false
|
2024-01-30
|
2024-06-26
| 0
|
152334H/miqu-1-70b-sf
|
|
1TuanPham_T-VisStar-7B-v0.1_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/1TuanPham/T-VisStar-7B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">1TuanPham/T-VisStar-7B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/1TuanPham__T-VisStar-7B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
1TuanPham/T-VisStar-7B-v0.1
|
b111b59971c14b46c888b96723ff7f3c7b6fd92f
| 19.044104
|
apache-2.0
| 2
| 7
| true
| false
| false
| true
| 1.269513
| 0.360704
| 36.070404
| 0.50522
| 30.243834
| 0.05136
| 5.135952
| 0.285235
| 4.697987
| 0.4375
| 13.554167
| 0.321061
| 24.562278
| true
| false
|
2024-09-19
|
2024-09-22
| 0
|
1TuanPham/T-VisStar-7B-v0.1
|
1TuanPham_T-VisStar-v0.1_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/1TuanPham/T-VisStar-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">1TuanPham/T-VisStar-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/1TuanPham__T-VisStar-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
1TuanPham/T-VisStar-v0.1
|
c9779bd9630a533f7e42fd8effcca69623d48c9c
| 19.044104
|
apache-2.0
| 2
| 7
| true
| false
| false
| true
| 0.624384
| 0.360704
| 36.070404
| 0.50522
| 30.243834
| 0.05136
| 5.135952
| 0.285235
| 4.697987
| 0.4375
| 13.554167
| 0.321061
| 24.562278
| true
| false
|
2024-09-19
|
2024-09-20
| 0
|
1TuanPham/T-VisStar-v0.1
|
3rd-Degree-Burn_L-3.1-Science-Writer-8B_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/L-3.1-Science-Writer-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/L-3.1-Science-Writer-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__L-3.1-Science-Writer-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/L-3.1-Science-Writer-8B
|
d9bb11fb02f8eca3aec408912278e513377115da
| 21.07862
| 0
| 8
| false
| false
| false
| false
| 0.709678
| 0.42625
| 42.625013
| 0.504131
| 29.199301
| 0.102719
| 10.271903
| 0.274329
| 3.243848
| 0.395948
| 11.69349
| 0.364943
| 29.438165
| false
| false
|
2024-11-19
| 0
|
Removed
|
||
3rd-Degree-Burn_Llama-3.1-8B-Squareroot_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-3.1-8B-Squareroot" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-3.1-8B-Squareroot</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-3.1-8B-Squareroot-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-3.1-8B-Squareroot
|
2bec01c2c5d53276eac2222c80190eb44ab2e6af
| 10.581747
|
apache-2.0
| 1
| 8
| true
| false
| false
| true
| 0.98705
| 0.221344
| 22.134381
| 0.346094
| 8.618064
| 0.227341
| 22.734139
| 0.256711
| 0.894855
| 0.308917
| 0.78125
| 0.17495
| 8.327793
| true
| false
|
2024-10-10
|
2024-10-10
| 1
|
3rd-Degree-Burn/Llama-3.1-8B-Squareroot (Merge)
|
3rd-Degree-Burn_Llama-3.1-8B-Squareroot-v1_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-3.1-8B-Squareroot-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1
|
09339d9c3b118ae3c6e7beab8b84347471990988
| 7.597362
| 0
| 8
| false
| false
| false
| true
| 0.772749
| 0.289238
| 28.923811
| 0.334277
| 6.515145
| 0.061934
| 6.193353
| 0.255872
| 0.782998
| 0.334063
| 1.757812
| 0.112699
| 1.411052
| false
| false
|
2024-11-10
| 0
|
Removed
|
||
3rd-Degree-Burn_Llama-Squared-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-Squared-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-Squared-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-Squared-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-Squared-8B
|
f30737e92b3a3fa0ef2a3f3ade487cc94ad34400
| 12.233544
| 0
| 8
| false
| false
| false
| true
| 1.011112
| 0.275524
| 27.55245
| 0.443103
| 21.277103
| 0.045317
| 4.531722
| 0.271812
| 2.908277
| 0.308948
| 1.951823
| 0.236619
| 15.179891
| false
| false
|
2024-10-08
| 0
|
Removed
|
||
4season_final_model_test_v2_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/4season/final_model_test_v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">4season/final_model_test_v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/4season__final_model_test_v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
4season/final_model_test_v2
|
cf690c35d9cf0b0b6bf034fa16dbf88c56fe861c
| 21.91554
|
apache-2.0
| 0
| 21
| true
| false
| false
| false
| 1.081038
| 0.319113
| 31.911329
| 0.634205
| 47.41067
| 0.013595
| 1.359517
| 0.327181
| 10.290828
| 0.431448
| 12.43099
| 0.352809
| 28.089908
| false
| false
|
2024-05-20
|
2024-06-27
| 0
|
4season/final_model_test_v2
|
AALF_FuseChat-Llama-3.1-8B-Instruct-preview_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/FuseChat-Llama-3.1-8B-Instruct-preview" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/FuseChat-Llama-3.1-8B-Instruct-preview</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__FuseChat-Llama-3.1-8B-Instruct-preview-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/FuseChat-Llama-3.1-8B-Instruct-preview
|
f740497979293c90fa1cfaa7c446016e107cc2c1
| 25.610368
| 0
| 8
| false
| false
| false
| true
| 0.688619
| 0.718958
| 71.895792
| 0.511989
| 30.848065
| 0.070242
| 7.024169
| 0.305369
| 7.38255
| 0.382
| 6.15
| 0.373255
| 30.361628
| false
| false
|
2024-11-20
| 0
|
Removed
|
||
AALF_FuseChat-Llama-3.1-8B-SFT-preview_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/FuseChat-Llama-3.1-8B-SFT-preview" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/FuseChat-Llama-3.1-8B-SFT-preview</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__FuseChat-Llama-3.1-8B-SFT-preview-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/FuseChat-Llama-3.1-8B-SFT-preview
|
601f2b8c448acc5686656d3979ed732ce050b827
| 27.374839
| 0
| 8
| false
| false
| false
| true
| 0.684308
| 0.72805
| 72.805046
| 0.52403
| 32.536782
| 0.114048
| 11.404834
| 0.30453
| 7.270694
| 0.402
| 9.75
| 0.374335
| 30.481678
| false
| false
|
2024-11-21
| 0
|
Removed
|
||
AALF_gemma-2-27b-it-SimPO-37K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/gemma-2-27b-it-SimPO-37K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/gemma-2-27b-it-SimPO-37K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__gemma-2-27b-it-SimPO-37K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/gemma-2-27b-it-SimPO-37K
|
27f15219df2000a16955c9403c3f38b5f3413b3d
| 9.298079
|
gemma
| 18
| 27
| true
| false
| false
| true
| 9.997722
| 0.240653
| 24.065258
| 0.391134
| 15.307881
| 0
| 0
| 0.280201
| 4.026846
| 0.34876
| 1.595052
| 0.197141
| 10.79344
| false
| false
|
2024-08-13
|
2024-09-05
| 2
|
google/gemma-2-27b
|
AALF_gemma-2-27b-it-SimPO-37K-100steps_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/gemma-2-27b-it-SimPO-37K-100steps" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/gemma-2-27b-it-SimPO-37K-100steps</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__gemma-2-27b-it-SimPO-37K-100steps-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/gemma-2-27b-it-SimPO-37K-100steps
|
d5cbf18b2eb90b77f5ddbb74cfcaeedfa692c90c
| 9.894336
|
gemma
| 11
| 27
| true
| false
| false
| true
| 9.856735
| 0.256764
| 25.676427
| 0.393082
| 15.261078
| 0
| 0
| 0.288591
| 5.145414
| 0.332917
| 0.78125
| 0.212517
| 12.501847
| false
| false
|
2024-08-13
|
2024-09-21
| 2
|
google/gemma-2-27b
|
AELLM_gemma-2-aeria-infinity-9b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AELLM/gemma-2-aeria-infinity-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AELLM/gemma-2-aeria-infinity-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AELLM__gemma-2-aeria-infinity-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AELLM/gemma-2-aeria-infinity-9b
|
24e1de07258925d5ddb52134b66e2eb0d698dc11
| 28.344029
| 1
| 9
| false
| false
| false
| true
| 3.003789
| 0.7594
| 75.93995
| 0.598334
| 42.090214
| 0
| 0
| 0.333893
| 11.185682
| 0.401969
| 9.046094
| 0.38622
| 31.802231
| false
| false
|
2024-10-09
|
2024-10-09
| 1
|
AELLM/gemma-2-aeria-infinity-9b (Merge)
|
|
AELLM_gemma-2-lyco-infinity-9b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AELLM/gemma-2-lyco-infinity-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AELLM/gemma-2-lyco-infinity-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AELLM__gemma-2-lyco-infinity-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AELLM/gemma-2-lyco-infinity-9b
|
2941a682fcbcfea3f1485c9e0691cc1d9edc742e
| 27.204937
| 0
| 10
| false
| false
| false
| true
| 2.97852
| 0.731648
| 73.164758
| 0.583953
| 39.787539
| 0
| 0
| 0.32802
| 10.402685
| 0.400635
| 8.91276
| 0.378657
| 30.961879
| false
| false
|
2024-10-09
|
2024-10-09
| 1
|
AELLM/gemma-2-lyco-infinity-9b (Merge)
|
|
AGI-0_Artificium-llama3.1-8B-001_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AGI-0/Artificium-llama3.1-8B-001" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AGI-0/Artificium-llama3.1-8B-001</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AGI-0__Artificium-llama3.1-8B-001-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AGI-0/Artificium-llama3.1-8B-001
|
6bf3dcca3b75a06a4e04e5f944e709cccf4673fd
| 19.063822
|
unknown
| 33
| 8
| true
| false
| false
| true
| 1.860327
| 0.524769
| 52.476872
| 0.425622
| 19.348898
| 0.110272
| 11.02719
| 0.26594
| 2.12528
| 0.379458
| 5.165625
| 0.318152
| 24.239066
| false
| false
|
2024-08-16
|
2024-09-08
| 0
|
AGI-0/Artificium-llama3.1-8B-001
|
AGI-0_smartllama3.1-8B-001_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AGI-0/smartllama3.1-8B-001" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AGI-0/smartllama3.1-8B-001</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AGI-0__smartllama3.1-8B-001-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AGI-0/smartllama3.1-8B-001
|
974d5ee685f1be003a1d8d08e907fe672d225035
| 20.23573
|
unknown
| 33
| 8
| true
| false
| false
| false
| 0.718834
| 0.351787
| 35.178659
| 0.467018
| 24.857737
| 0.11858
| 11.858006
| 0.306208
| 7.494407
| 0.438646
| 14.397396
| 0.348654
| 27.628177
| false
| false
|
2024-08-16
|
2024-11-25
| 0
|
AGI-0/smartllama3.1-8B-001
|
AI-MO_NuminaMath-7B-CoT_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-MO/NuminaMath-7B-CoT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-MO/NuminaMath-7B-CoT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-MO__NuminaMath-7B-CoT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-MO/NuminaMath-7B-CoT
|
ff7e3044218efe64128bd9c21f9ec66c3de04324
| 13.097309
|
apache-2.0
| 18
| 6
| true
| false
| false
| true
| 0.745989
| 0.268854
| 26.885442
| 0.431419
| 19.152364
| 0.088369
| 8.836858
| 0.26594
| 2.12528
| 0.330344
| 0.826302
| 0.286818
| 20.757609
| false
| false
|
2024-07-15
|
2024-09-10
| 1
|
deepseek-ai/deepseek-math-7b-base
|
AI-MO_NuminaMath-7B-TIR_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-MO/NuminaMath-7B-TIR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-MO/NuminaMath-7B-TIR</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-MO__NuminaMath-7B-TIR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-MO/NuminaMath-7B-TIR
|
c6e394cc0579423c9cde6df6cc192c07dae73388
| 11.815723
|
apache-2.0
| 321
| 6
| true
| false
| false
| false
| 1.07411
| 0.275624
| 27.562423
| 0.414369
| 16.873547
| 0.018882
| 1.888218
| 0.258389
| 1.118568
| 0.350927
| 4.199219
| 0.273271
| 19.252364
| false
| false
|
2024-07-04
|
2024-07-11
| 1
|
deepseek-ai/deepseek-math-7b-base
|
AI-Sweden-Models_Llama-3-8B-instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-Sweden-Models/Llama-3-8B-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-Sweden-Models/Llama-3-8B-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-Sweden-Models__Llama-3-8B-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-Sweden-Models/Llama-3-8B-instruct
|
4e1c955228bdb4d69c1c4560e8d5872312a8f033
| 13.777204
|
llama3
| 10
| 8
| true
| false
| false
| true
| 1.166111
| 0.240128
| 24.012841
| 0.417346
| 18.388096
| 0.004532
| 0.453172
| 0.26594
| 2.12528
| 0.477094
| 19.936719
| 0.259724
| 17.747119
| false
| false
|
2024-06-01
|
2024-06-27
| 2
|
meta-llama/Meta-Llama-3-8B
|
AI-Sweden-Models_gpt-sw3-40b_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
GPT2LMHeadModel
|
<a target="_blank" href="https://huggingface.co/AI-Sweden-Models/gpt-sw3-40b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-Sweden-Models/gpt-sw3-40b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-Sweden-Models__gpt-sw3-40b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-Sweden-Models/gpt-sw3-40b
|
1af27994df1287a7fac1b10d60e40ca43a22a385
| 4.734433
|
other
| 10
| 39
| true
| false
| false
| false
| 2.959819
| 0.14703
| 14.702988
| 0.326774
| 6.894934
| 0.009063
| 0.906344
| 0.234899
| 0
| 0.36324
| 2.838281
| 0.127576
| 3.064051
| false
| false
|
2023-02-22
|
2024-06-26
| 0
|
AI-Sweden-Models/gpt-sw3-40b
|
AbacusResearch_Jallabi-34B_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AbacusResearch/Jallabi-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AbacusResearch/Jallabi-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AbacusResearch__Jallabi-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AbacusResearch/Jallabi-34B
|
f65696da4ed82c9a20e94b200d9dccffa07af682
| 25.972084
|
apache-2.0
| 2
| 34
| true
| false
| false
| false
| 3.286492
| 0.35286
| 35.286041
| 0.602338
| 43.615765
| 0.039275
| 3.927492
| 0.338926
| 11.856823
| 0.482177
| 20.238802
| 0.468168
| 40.90758
| false
| false
|
2024-03-01
|
2024-06-27
| 0
|
AbacusResearch/Jallabi-34B
|
Alibaba-NLP_gte-Qwen2-7B-instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Alibaba-NLP/gte-Qwen2-7B-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Alibaba-NLP__gte-Qwen2-7B-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Alibaba-NLP/gte-Qwen2-7B-instruct
|
e26182b2122f4435e8b3ebecbf363990f409b45b
| 13.40618
|
apache-2.0
| 226
| 7
| true
| false
| false
| true
| 2.172113
| 0.22554
| 22.554045
| 0.449514
| 21.925482
| 0.03852
| 3.851964
| 0.244966
| 0
| 0.355854
| 6.315104
| 0.332114
| 25.790485
| false
| false
|
2024-06-15
|
2024-08-05
| 0
|
Alibaba-NLP/gte-Qwen2-7B-instruct
|
ArliAI_ArliAI-RPMax-12B-v1.1_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/ArliAI/ArliAI-RPMax-12B-v1.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ArliAI/ArliAI-RPMax-12B-v1.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ArliAI__ArliAI-RPMax-12B-v1.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
ArliAI/ArliAI-RPMax-12B-v1.1
|
645db1cf8ad952eb57854a133e8e15303b898b04
| 20.812694
|
apache-2.0
| 41
| 12
| true
| false
| false
| true
| 1.833402
| 0.534885
| 53.488522
| 0.475182
| 24.809063
| 0.102719
| 10.271903
| 0.281879
| 4.250559
| 0.361844
| 5.563802
| 0.338431
| 26.492317
| false
| false
|
2024-08-31
|
2024-09-05
| 0
|
ArliAI/ArliAI-RPMax-12B-v1.1
|
ArliAI_Llama-3.1-8B-ArliAI-RPMax-v1.1_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ArliAI__Llama-3.1-8B-ArliAI-RPMax-v1.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
|
540bd352e59c63900af91b95a932b33aaee70c76
| 23.916967
|
llama3
| 29
| 8
| true
| false
| false
| true
| 0.892745
| 0.635902
| 63.590163
| 0.501561
| 28.787014
| 0.129909
| 12.990937
| 0.283557
| 4.474273
| 0.357688
| 5.310938
| 0.355136
| 28.348478
| false
| false
|
2024-08-23
|
2024-09-19
| 0
|
ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
|
Artples_L-MChat-7b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Artples/L-MChat-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Artples/L-MChat-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Artples__L-MChat-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Artples/L-MChat-7b
|
e10137f5cbfc1b73068d6473e4a87241cca0b3f4
| 21.225905
|
apache-2.0
| 2
| 7
| true
| false
| false
| true
| 0.592226
| 0.529665
| 52.966462
| 0.460033
| 24.201557
| 0.09139
| 9.138973
| 0.305369
| 7.38255
| 0.402865
| 8.12474
| 0.32987
| 25.54115
| true
| false
|
2024-04-02
|
2024-07-07
| 1
|
Artples/L-MChat-7b (Merge)
|
Artples_L-MChat-Small_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/Artples/L-MChat-Small" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Artples/L-MChat-Small</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Artples__L-MChat-Small-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Artples/L-MChat-Small
|
52484c277f6062c12dc6d6b6397ee0d0c21b0126
| 14.891449
|
mit
| 1
| 2
| true
| false
| false
| true
| 0.465511
| 0.328706
| 32.870561
| 0.482256
| 26.856516
| 0.017372
| 1.73716
| 0.267617
| 2.348993
| 0.369594
| 9.265885
| 0.246426
| 16.269577
| true
| false
|
2024-04-11
|
2024-07-07
| 1
|
Artples/L-MChat-Small (Merge)
|
Aryanne_SuperHeart_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Aryanne/SuperHeart" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Aryanne/SuperHeart</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Aryanne__SuperHeart-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Aryanne/SuperHeart
|
02b5050d7e600ce3db81a19638f6043c895d60cf
| 25.267673
|
llama3.1
| 1
| 8
| true
| false
| false
| false
| 0.903959
| 0.519223
| 51.922344
| 0.521538
| 31.893554
| 0.138973
| 13.897281
| 0.301174
| 6.823266
| 0.443573
| 14.713281
| 0.391207
| 32.356309
| true
| false
|
2024-09-23
|
2024-09-23
| 1
|
Aryanne/SuperHeart (Merge)
|
AtAndDev_Qwen2.5-1.5B-continuous-learnt_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AtAndDev/Qwen2.5-1.5B-continuous-learnt" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AtAndDev/Qwen2.5-1.5B-continuous-learnt</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AtAndDev__Qwen2.5-1.5B-continuous-learnt-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AtAndDev/Qwen2.5-1.5B-continuous-learnt
|
01c0981db9cf0f146fe050065f17343af75a8aa6
| 16.518524
| 0
| 1
| false
| false
| false
| true
| 0.673035
| 0.460521
| 46.052142
| 0.425775
| 19.537666
| 0.074773
| 7.477341
| 0.26594
| 2.12528
| 0.363646
| 3.789063
| 0.281167
| 20.129654
| false
| false
|
2024-10-13
| 0
|
Removed
|
||
AtAndDev_Qwen2.5-1.5B-continuous-learnt_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AtAndDev/Qwen2.5-1.5B-continuous-learnt" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AtAndDev/Qwen2.5-1.5B-continuous-learnt</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AtAndDev__Qwen2.5-1.5B-continuous-learnt-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AtAndDev/Qwen2.5-1.5B-continuous-learnt
|
01c0981db9cf0f146fe050065f17343af75a8aa6
| 16.45133
| 0
| 1
| false
| false
| false
| true
| 0.688585
| 0.451054
| 45.105431
| 0.42747
| 19.766409
| 0.085347
| 8.534743
| 0.270134
| 2.684564
| 0.362281
| 2.551823
| 0.280585
| 20.065012
| false
| false
|
2024-10-18
| 0
|
Removed
|
||
Aurel9_testmerge-7b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Aurel9/testmerge-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Aurel9/testmerge-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Aurel9__testmerge-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Aurel9/testmerge-7b
|
b5f0a72d981b5b2c6bd6294093c6956d88477a3e
| 20.994478
| 0
| 7
| false
| false
| false
| false
| 0.476464
| 0.397998
| 39.799842
| 0.518959
| 32.792793
| 0.067221
| 6.722054
| 0.300336
| 6.711409
| 0.465865
| 17.133073
| 0.305269
| 22.807698
| false
| false
|
2024-11-16
|
2024-11-16
| 1
|
Aurel9/testmerge-7b (Merge)
|
|
Azure99_blossom-v5-32b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5-32b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5-32b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5-32b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5-32b
|
ccd4d86e3de01187043683dea1e28df904f7408e
| 26.352555
|
apache-2.0
| 4
| 32
| true
| false
| false
| true
| 5.688
| 0.523544
| 52.35442
| 0.595455
| 42.883056
| 0.10423
| 10.422961
| 0.311242
| 8.165548
| 0.402
| 8.35
| 0.423454
| 35.939347
| false
| false
|
2024-04-29
|
2024-09-21
| 0
|
Azure99/blossom-v5-32b
|
Azure99_blossom-v5-llama3-8b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5-llama3-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5-llama3-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5-llama3-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5-llama3-8b
|
91ea35e2e65516988021e4bb3b908e3e497e05c2
| 14.473082
|
apache-2.0
| 4
| 8
| true
| false
| false
| true
| 0.872153
| 0.434293
| 43.429323
| 0.418491
| 18.306535
| 0.043807
| 4.380665
| 0.265101
| 2.013423
| 0.367021
| 5.310938
| 0.220578
| 13.397606
| false
| false
|
2024-04-20
|
2024-09-21
| 0
|
Azure99/blossom-v5-llama3-8b
|
Azure99_blossom-v5.1-34b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5.1-34b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5.1-34b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5.1-34b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5.1-34b
|
2c803204f5dbf4ce37e2df98eb0205cdc53de10d
| 28.599286
|
apache-2.0
| 5
| 34
| true
| false
| false
| true
| 9.591483
| 0.569656
| 56.965629
| 0.610911
| 44.147705
| 0.1571
| 15.70997
| 0.309564
| 7.941834
| 0.392792
| 7.298958
| 0.455785
| 39.531619
| false
| false
|
2024-05-19
|
2024-07-27
| 0
|
Azure99/blossom-v5.1-34b
|
Azure99_blossom-v5.1-9b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5.1-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5.1-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5.1-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5.1-9b
|
6044a3dc1e04529fe883aa513d37f266a320d793
| 24.871504
|
apache-2.0
| 2
| 8
| true
| false
| false
| true
| 2.21572
| 0.508582
| 50.858167
| 0.534329
| 34.201244
| 0.116314
| 11.63142
| 0.33557
| 11.409396
| 0.399396
| 8.024479
| 0.397939
| 33.104314
| false
| false
|
2024-05-15
|
2024-07-24
| 0
|
Azure99/blossom-v5.1-9b
|
BAAI_Gemma2-9B-IT-Simpo-Infinity-Preference_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Gemma2-9B-IT-Simpo-Infinity-Preference-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
|
028a91b1a4f14d365c6db08093b03348455c7bad
| 20.984069
| 14
| 9
| false
| false
| false
| true
| 5.86346
| 0.317638
| 31.763831
| 0.597946
| 42.190844
| 0
| 0
| 0.339765
| 11.96868
| 0.396573
| 8.104948
| 0.386885
| 31.876108
| false
| false
|
2024-08-28
|
2024-09-05
| 2
|
google/gemma-2-9b
|
|
BAAI_Infinity-Instruct-3M-0613-Llama3-70B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0613-Llama3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0613-Llama3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0613-Llama3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0613-Llama3-70B
|
9fc53668064bdda22975ca72c5a287f8241c95b3
| 34.697075
|
apache-2.0
| 5
| 70
| true
| false
| false
| true
| 10.526907
| 0.682113
| 68.211346
| 0.664161
| 51.327161
| 0.162387
| 16.238671
| 0.358221
| 14.42953
| 0.45226
| 16.532552
| 0.472989
| 41.443189
| false
| false
|
2024-06-27
|
2024-06-28
| 0
|
BAAI/Infinity-Instruct-3M-0613-Llama3-70B
|
BAAI_Infinity-Instruct-3M-0613-Mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0613-Mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0613-Mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0613-Mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0613-Mistral-7B
|
c7a742e539ec264b9eaeefe2aed29e92e8a7ebd6
| 22.180237
|
apache-2.0
| 11
| 7
| true
| false
| false
| true
| 0.949375
| 0.531987
| 53.198735
| 0.495823
| 28.992936
| 0.074773
| 7.477341
| 0.296141
| 6.152125
| 0.435083
| 13.252083
| 0.316074
| 24.0082
| false
| false
|
2024-06-21
|
2024-06-27
| 0
|
BAAI/Infinity-Instruct-3M-0613-Mistral-7B
|
BAAI_Infinity-Instruct-3M-0625-Llama3-70B_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Llama3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Llama3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Llama3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Llama3-70B
|
6d8ceada57e55cff3503191adc4d6379ff321fe2
| 36.142217
|
apache-2.0
| 3
| 70
| true
| false
| false
| true
| 10.430955
| 0.744212
| 74.421202
| 0.667034
| 52.028162
| 0.179003
| 17.900302
| 0.357383
| 14.317673
| 0.461656
| 18.340365
| 0.45861
| 39.845597
| false
| false
|
2024-07-09
|
2024-08-30
| 0
|
BAAI/Infinity-Instruct-3M-0625-Llama3-70B
|
BAAI_Infinity-Instruct-3M-0625-Llama3-8B_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Llama3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Llama3-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Llama3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Llama3-8B
|
7be7c0ff1e35c3bb781c47222da99a1724f5f1da
| 21.60936
|
apache-2.0
| 3
| 8
| true
| false
| false
| true
| 0.858004
| 0.605027
| 60.502688
| 0.495499
| 28.988222
| 0.061178
| 6.117825
| 0.275168
| 3.355705
| 0.371208
| 5.667708
| 0.325216
| 25.02401
| false
| false
|
2024-07-09
|
2024-07-13
| 0
|
BAAI/Infinity-Instruct-3M-0625-Llama3-8B
|
BAAI_Infinity-Instruct-3M-0625-Mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Mistral-7B
|
302e3ae0bcc50dae3fb69fc1b08b518398e8c407
| 22.843425
|
apache-2.0
| 3
| 7
| true
| false
| false
| true
| 0.785797
| 0.586742
| 58.674207
| 0.493967
| 28.823289
| 0.076284
| 7.628399
| 0.286913
| 4.9217
| 0.42724
| 12.238281
| 0.322972
| 24.774675
| false
| false
|
2024-07-09
|
2024-08-05
| 0
|
BAAI/Infinity-Instruct-3M-0625-Mistral-7B
|
BAAI_Infinity-Instruct-3M-0625-Qwen2-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Qwen2-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Qwen2-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Qwen2-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Qwen2-7B
|
503c24156d7682458686a7b5324f7f886e63470d
| 24.135357
|
apache-2.0
| 8
| 7
| true
| false
| false
| true
| 1.330078
| 0.555393
| 55.539302
| 0.534591
| 34.656829
| 0.068731
| 6.873112
| 0.312919
| 8.389262
| 0.38876
| 6.461719
| 0.396027
| 32.891918
| false
| false
|
2024-07-09
|
2024-08-05
| 0
|
BAAI/Infinity-Instruct-3M-0625-Qwen2-7B
|
BAAI_Infinity-Instruct-3M-0625-Yi-1.5-9B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B
|
a42c86c61b98ca4fdf238d688fe6ea11cf414d29
| 27.943551
|
apache-2.0
| 3
| 8
| true
| false
| false
| true
| 1.116801
| 0.518598
| 51.859843
| 0.550912
| 35.378707
| 0.151813
| 15.181269
| 0.354027
| 13.870246
| 0.457531
| 16.72474
| 0.411818
| 34.646498
| false
| false
|
2024-07-09
|
2024-08-05
| 0
|
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B
|
BAAI_Infinity-Instruct-7M-0729-Llama3_1-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-0729-Llama3_1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B
|
0aca33fd7500a781d041e8bf7e5e3789b03f54f4
| 23.094956
|
llama3.1
| 8
| 8
| true
| false
| false
| true
| 0.866805
| 0.613195
| 61.319521
| 0.507734
| 30.888805
| 0.106495
| 10.649547
| 0.292785
| 5.704698
| 0.357844
| 5.297135
| 0.32239
| 24.710033
| false
| false
|
2024-08-02
|
2024-08-05
| 0
|
BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B
|
BAAI_Infinity-Instruct-7M-0729-mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-0729-mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-0729-mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-0729-mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-0729-mistral-7B
|
36651591cb13346ecbde23832013e024029700fa
| 22.914335
|
apache-2.0
| 3
| 7
| true
| false
| false
| true
| 0.799261
| 0.616193
| 61.619281
| 0.496381
| 28.697915
| 0.064955
| 6.495468
| 0.290268
| 5.369128
| 0.406188
| 10.040104
| 0.327377
| 25.264111
| false
| false
|
2024-07-25
|
2024-08-05
| 0
|
BAAI/Infinity-Instruct-7M-0729-mistral-7B
|
BAAI_Infinity-Instruct-7M-Gen-Llama3_1-70B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-Llama3_1-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B
|
1ef63c4993a8c723c9695c827295c17080a64435
| 37.10681
|
llama3.1
| 17
| 70
| true
| false
| false
| true
| 11.069121
| 0.733546
| 73.354588
| 0.66952
| 52.498947
| 0.229607
| 22.960725
| 0.375839
| 16.778523
| 0.453906
| 16.971615
| 0.460688
| 40.076463
| false
| false
|
2024-07-25
|
2024-09-26
| 0
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B
|
BAAI_Infinity-Instruct-7M-Gen-Llama3_1-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-Llama3_1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B
|
56f9c2845ae024eb8b1dd9ea0d8891cbaf33c596
| 23.094956
|
llama3.1
| 8
| 8
| true
| false
| false
| true
| 0.91714
| 0.613195
| 61.319521
| 0.507734
| 30.888805
| 0.106495
| 10.649547
| 0.292785
| 5.704698
| 0.357844
| 5.297135
| 0.32239
| 24.710033
| false
| false
|
2024-08-02
|
2024-08-29
| 0
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B
|
BAAI_Infinity-Instruct-7M-Gen-mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-mistral-7B
|
82c83d670a8954f4250547b53a057dea1fbd460d
| 22.888939
|
apache-2.0
| 3
| 7
| true
| false
| false
| true
| 0.824635
| 0.614669
| 61.466908
| 0.496381
| 28.697915
| 0.064955
| 6.495468
| 0.290268
| 5.369128
| 0.406188
| 10.040104
| 0.327377
| 25.264111
| false
| false
|
2024-07-25
|
2024-08-29
| 0
|
BAAI/Infinity-Instruct-7M-Gen-mistral-7B
|
BAAI_OPI-Llama-3.1-8B-Instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/OPI-Llama-3.1-8B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/OPI-Llama-3.1-8B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__OPI-Llama-3.1-8B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/OPI-Llama-3.1-8B-Instruct
|
48504799d009b4e1b29e6d2948a7cde68acdc3b0
| 8.305018
|
llama3.1
| 1
| 8
| true
| false
| false
| true
| 0.671657
| 0.207455
| 20.745511
| 0.355122
| 9.768712
| 0
| 0
| 0.274329
| 3.243848
| 0.323302
| 3.579427
| 0.212434
| 12.492612
| false
| false
|
2024-09-06
|
2024-09-21
| 2
|
meta-llama/Meta-Llama-3.1-8B
|
BEE-spoke-data_Meta-Llama-3-8Bee_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/Meta-Llama-3-8Bee" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/Meta-Llama-3-8Bee</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__Meta-Llama-3-8Bee-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/Meta-Llama-3-8Bee
|
8143e34e77a49a30ec2617c5c9cc22cb3cda2287
| 14.544519
|
llama3
| 0
| 8
| true
| false
| false
| false
| 0.83038
| 0.195066
| 19.506576
| 0.462636
| 24.199033
| 0.041541
| 4.154079
| 0.313758
| 8.501119
| 0.365406
| 6.242448
| 0.321975
| 24.663859
| false
| false
|
2024-04-28
|
2024-07-04
| 1
|
meta-llama/Meta-Llama-3-8B
|
BEE-spoke-data_smol_llama-101M-GQA_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-101M-GQA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-101M-GQA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-101M-GQA
|
bb26643db413bada7e0c3c50752bf9da82403dba
| 3.918895
|
apache-2.0
| 28
| 0
| true
| false
| false
| false
| 0.119606
| 0.138437
| 13.843712
| 0.301756
| 3.198004
| 0
| 0
| 0.25755
| 1.006711
| 0.371271
| 4.275521
| 0.110705
| 1.189421
| false
| false
|
2023-10-26
|
2024-07-06
| 0
|
BEE-spoke-data/smol_llama-101M-GQA
|
BEE-spoke-data_smol_llama-220M-GQA_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-GQA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-GQA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-GQA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-GQA
|
8845b1d3c0bc73522ef2700aab467183cbdca9f7
| 6.401567
|
apache-2.0
| 12
| 0
| true
| false
| false
| false
| 0.163613
| 0.238605
| 23.860468
| 0.303167
| 3.037843
| 0
| 0
| 0.255872
| 0.782998
| 0.405875
| 9.067708
| 0.114943
| 1.660387
| false
| false
|
2023-12-22
|
2024-06-26
| 0
|
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_smol_llama-220M-GQA-fineweb_edu_bfloat16
|
bfloat16
|
🟩 continuously pretrained
|
🟩
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-GQA-fineweb_edu-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu
|
dec16b41d5e94070dbc1f8449a554373fd4cc1d1
| 6.516558
|
apache-2.0
| 1
| 0
| true
| false
| false
| false
| 0.161876
| 0.198812
| 19.881248
| 0.292905
| 2.314902
| 0
| 0
| 0.259228
| 1.230425
| 0.43676
| 14.261719
| 0.112699
| 1.411052
| false
| false
|
2024-06-08
|
2024-06-26
| 1
|
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_smol_llama-220M-openhermes_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-openhermes" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-openhermes</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-openhermes-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-openhermes
|
fb4bcd4b7eee363baacb4176a26cea2aaeb173f4
| 4.761772
|
apache-2.0
| 5
| 0
| true
| false
| false
| false
| 0.154426
| 0.155523
| 15.55229
| 0.302752
| 3.107692
| 0
| 0
| 0.267617
| 2.348993
| 0.384729
| 6.224479
| 0.112035
| 1.337175
| false
| false
|
2023-12-30
|
2024-09-21
| 1
|
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_tFINE-900m-e16-d32-flan_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-flan</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-flan-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-flan
|
d9ffec9798402d13d8f2c56ec3de3ad092445297
| 4.433887
|
apache-2.0
| 0
| 0
| true
| false
| false
| false
| 2.456006
| 0.150577
| 15.057714
| 0.302804
| 4.411894
| 0
| 0
| 0.233221
| 0
| 0.372417
| 3.71875
| 0.130735
| 3.414967
| false
| false
|
2024-09-06
|
2024-09-13
| 1
|
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024
|
b1e2f12f5224be9f7da0cb5ff30e1bbb3f10f6ca
| 5.823653
|
apache-2.0
| 0
| 0
| true
| false
| false
| false
| 2.600608
| 0.132067
| 13.206736
| 0.313779
| 4.737018
| 0
| 0
| 0.254195
| 0.559284
| 0.439271
| 13.808854
| 0.12367
| 2.630024
| false
| false
|
2024-09-10
|
2024-09-14
| 2
|
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-e16-d32-instruct_2e_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-instruct_2e-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e
|
4c626138c9f4e0c3eafe74b2755eb89334c7ca59
| 5.681552
|
apache-2.0
| 0
| 0
| true
| false
| false
| false
| 2.516619
| 0.140286
| 14.028555
| 0.313457
| 5.01307
| 0
| 0
| 0.259228
| 1.230425
| 0.420698
| 11.18724
| 0.12367
| 2.630024
| false
| false
|
2024-09-17
|
2024-09-22
| 3
|
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-instruct-orpo_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-instruct-orpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-instruct-orpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-instruct-orpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-instruct-orpo
|
e0a21c79bac74442252d36e2c01403afa3f0971b
| 3.431957
|
apache-2.0
| 0
| 0
| true
| false
| false
| true
| 2.574962
| 0.132992
| 13.299157
| 0.302209
| 3.267301
| 0
| 0
| 0.259228
| 1.230425
| 0.340854
| 1.106771
| 0.115193
| 1.688091
| false
| false
|
2024-09-22
|
2024-09-23
| 0
|
BEE-spoke-data/tFINE-900m-instruct-orpo
|
BSC-LT_salamandra-7b_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BSC-LT/salamandra-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BSC-LT/salamandra-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BSC-LT__salamandra-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BSC-LT/salamandra-7b
|
bf30739316ceac4b624583a27ec96dfc401179e8
| 5.641971
|
apache-2.0
| 13
| 7
| true
| false
| false
| false
| 0.189289
| 0.136738
| 13.67383
| 0.351661
| 10.157422
| 0
| 0
| 0.270134
| 2.684564
| 0.350094
| 1.861719
| 0.149269
| 5.474291
| false
| false
|
2024-09-30
|
2024-11-22
| 0
|
BSC-LT/salamandra-7b
|
BSC-LT_salamandra-7b-instruct_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BSC-LT/salamandra-7b-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BSC-LT/salamandra-7b-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BSC-LT__salamandra-7b-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BSC-LT/salamandra-7b-instruct
|
77ddccbc7d9f9ffd55a8535365e8eebc493ccb8e
| 10.080539
|
apache-2.0
| 32
| 7
| true
| false
| false
| true
| 1.147504
| 0.245074
| 24.507418
| 0.385132
| 14.688129
| 0.002266
| 0.226586
| 0.264262
| 1.901566
| 0.413437
| 10.213021
| 0.180519
| 8.946513
| false
| false
|
2024-09-30
|
2024-11-22
| 1
|
BSC-LT/salamandra-7b-instruct (Merge)
|
Ba2han_Llama-Phi-3_DoRA_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Ba2han/Llama-Phi-3_DoRA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Ba2han/Llama-Phi-3_DoRA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Ba2han__Llama-Phi-3_DoRA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Ba2han/Llama-Phi-3_DoRA
|
36f99064a7be8ba475c2ee5c5424e95c263ccb87
| 25.318838
|
mit
| 6
| 3
| true
| false
| false
| true
| 0.533136
| 0.513053
| 51.305314
| 0.551456
| 37.249164
| 0.112538
| 11.253776
| 0.326342
| 10.178971
| 0.406927
| 9.532552
| 0.391539
| 32.393248
| false
| false
|
2024-05-15
|
2024-06-26
| 0
|
Ba2han/Llama-Phi-3_DoRA
|
BenevolenceMessiah_Qwen2.5-72B-2x-Instruct-TIES-v1.0_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BenevolenceMessiah__Qwen2.5-72B-2x-Instruct-TIES-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0
|
459891ec78c9bbed2836a8bba706e1707db10231
| 34.185749
| 0
| 72
| false
| false
| false
| true
| 17.350892
| 0.54735
| 54.734992
| 0.727311
| 61.911495
| 0.093656
| 9.365559
| 0.36745
| 15.659955
| 0.420667
| 12.016667
| 0.562832
| 51.425827
| false
| false
|
2024-11-11
|
2024-11-24
| 1
|
BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0 (Merge)
|
|
BenevolenceMessiah_Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MixtralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BenevolenceMessiah__Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0
|
d90f6e36584dc9b367461701e83c833bdeb736f2
| 15.096268
|
apache-2.0
| 0
| 28
| true
| true
| false
| false
| 3.334797
| 0.301153
| 30.115316
| 0.490867
| 26.877991
| 0.043051
| 4.305136
| 0.262584
| 1.677852
| 0.407979
| 8.930729
| 0.268035
| 18.670582
| true
| false
|
2024-09-21
|
2024-09-22
| 1
|
BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0 (Merge)
|
BlackBeenie_Bloslain-8B-v0.2_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Bloslain-8B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Bloslain-8B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Bloslain-8B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Bloslain-8B-v0.2
|
ebcb7f9f30bc172523a827d1ddefeb52b1aba494
| 23.803914
| 1
| 8
| false
| false
| false
| false
| 0.691763
| 0.502337
| 50.233713
| 0.511088
| 30.662902
| 0.145015
| 14.501511
| 0.306208
| 7.494407
| 0.407573
| 10.446615
| 0.365359
| 29.484338
| false
| false
|
2024-11-19
|
2024-11-19
| 1
|
BlackBeenie/Bloslain-8B-v0.2 (Merge)
|
|
BlackBeenie_Llama-3.1-8B-pythonic-passthrough-merge_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Llama-3.1-8B-pythonic-passthrough-merge-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge
|
3ec46616f5b34821b3b928938931295f92e49213
| 7.311462
| 0
| 20
| false
| false
| false
| false
| 3.58329
| 0.231586
| 23.158553
| 0.345385
| 9.359905
| 0.006042
| 0.60423
| 0.268456
| 2.46085
| 0.377812
| 4.593229
| 0.133228
| 3.692007
| false
| false
|
2024-11-06
|
2024-11-06
| 1
|
BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge (Merge)
|
|
BlackBeenie_Neos-Gemma-2-9b_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Gemma-2-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Gemma-2-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Gemma-2-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Gemma-2-9b
|
56dbbb4f972be887e5b57311a8a32e148e98d154
| 25.211313
|
apache-2.0
| 1
| 9
| true
| false
| false
| true
| 2.679092
| 0.587567
| 58.756655
| 0.550298
| 35.638851
| 0.082326
| 8.232628
| 0.322987
| 9.731544
| 0.36175
| 5.785417
| 0.398105
| 33.122784
| false
| false
|
2024-11-11
|
2024-11-11
| 1
|
BlackBeenie/Neos-Gemma-2-9b (Merge)
|
BlackBeenie_Neos-Llama-3.1-8B_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Llama-3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Llama-3.1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Llama-3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Llama-3.1-8B
|
9b48520ec1a777be0f1fd88f95454d85ac568407
| 19.461825
|
apache-2.0
| 1
| 8
| true
| false
| false
| true
| 0.793867
| 0.494394
| 49.439376
| 0.4425
| 21.080123
| 0.129154
| 12.915408
| 0.268456
| 2.46085
| 0.37499
| 5.740365
| 0.326213
| 25.134826
| false
| false
|
2024-11-12
|
2024-11-12
| 1
|
BlackBeenie/Neos-Llama-3.1-8B (Merge)
|
BlackBeenie_Neos-Llama-3.1-base_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Llama-3.1-base" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Llama-3.1-base</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Llama-3.1-base-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Llama-3.1-base
|
d4af4d73ba5fea0275fd1e3ba5102a79ac8009db
| 3.968795
| 0
| 4
| false
| false
| false
| true
| 1.409285
| 0.175082
| 17.508212
| 0.293034
| 2.221447
| 0
| 0
| 0.237416
| 0
| 0.349906
| 2.838281
| 0.111203
| 1.244829
| false
| false
|
2024-11-11
|
2024-11-11
| 0
|
BlackBeenie/Neos-Llama-3.1-base
|
|
BlackBeenie_Neos-Phi-3-14B-v0.1_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Phi3ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Phi-3-14B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Phi-3-14B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Phi-3-14B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Phi-3-14B-v0.1
|
0afb7cc74a94f11f2695dc92788cdc6e28325f9c
| 26.843485
|
apache-2.0
| 0
| 13
| true
| false
| false
| true
| 0.909626
| 0.402245
| 40.224493
| 0.621193
| 46.631387
| 0.166918
| 16.691843
| 0.305369
| 7.38255
| 0.412542
| 10.534375
| 0.456366
| 39.596262
| false
| false
|
2024-11-27
|
2024-11-27
| 1
|
BlackBeenie/Neos-Phi-3-14B-v0.1 (Merge)
|
BlackBeenie_llama-3-luminous-merged_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/llama-3-luminous-merged" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/llama-3-luminous-merged</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__llama-3-luminous-merged-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/llama-3-luminous-merged
|
64288dd8e3305f2dc11d84fe0c653f351b2e8a9d
| 21.480108
| 0
| 8
| false
| false
| false
| false
| 0.763854
| 0.432345
| 43.234507
| 0.515392
| 30.643687
| 0.07855
| 7.854985
| 0.292785
| 5.704698
| 0.414896
| 10.628646
| 0.377327
| 30.814125
| false
| false
|
2024-09-15
|
2024-10-11
| 1
|
BlackBeenie/llama-3-luminous-merged (Merge)
|
|
BlackBeenie_llama-3.1-8B-Galore-openassistant-guanaco_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__llama-3.1-8B-Galore-openassistant-guanaco-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco
|
828fa03c10e9085700b7abbe26f95067fab010fd
| 18.072101
| 1
| 8
| false
| false
| false
| false
| 0.85682
| 0.263484
| 26.348422
| 0.521337
| 31.444705
| 0.048338
| 4.833837
| 0.300336
| 6.711409
| 0.440625
| 14.578125
| 0.320645
| 24.516105
| false
| false
|
2024-10-16
|
2024-10-19
| 0
|
BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco
|
|
BoltMonkey_DreadMix_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/DreadMix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/DreadMix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__DreadMix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/DreadMix
|
ab5dbaaff606538db73b6fd89aa169760104a566
| 28.661027
| 0
| 8
| false
| false
| false
| true
| 1.614205
| 0.709491
| 70.949082
| 0.54351
| 34.845015
| 0.149547
| 14.954683
| 0.299497
| 6.599553
| 0.421219
| 13.61901
| 0.378989
| 30.998818
| false
| false
|
2024-10-12
|
2024-10-13
| 1
|
BoltMonkey/DreadMix (Merge)
|
|
BoltMonkey_NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated
|
969e4c9b41e733a367f5ea18ed50a6171b5e2357
| 27.726282
|
llama3.1
| 2
| 8
| true
| false
| false
| true
| 1.640513
| 0.799891
| 79.989096
| 0.515199
| 30.7599
| 0.116314
| 11.63142
| 0.28104
| 4.138702
| 0.401875
| 9.467708
| 0.373338
| 30.370863
| true
| false
|
2024-10-01
|
2024-10-10
| 1
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated (Merge)
|
BoltMonkey_NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated
|
969e4c9b41e733a367f5ea18ed50a6171b5e2357
| 21.345511
|
llama3.1
| 2
| 8
| true
| false
| false
| false
| 0.774319
| 0.459023
| 45.902317
| 0.518544
| 30.793785
| 0.093656
| 9.365559
| 0.274329
| 3.243848
| 0.40826
| 9.532552
| 0.363115
| 29.235003
| true
| false
|
2024-10-01
|
2024-10-01
| 1
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated (Merge)
|
BoltMonkey_SuperNeuralDreadDevil-8b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/SuperNeuralDreadDevil-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/SuperNeuralDreadDevil-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__SuperNeuralDreadDevil-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/SuperNeuralDreadDevil-8b
|
804d5864127e603abec179a159b43f446246fafc
| 21.847726
| 1
| 8
| false
| false
| false
| true
| 2.405331
| 0.485801
| 48.580101
| 0.515108
| 30.606714
| 0.090634
| 9.063444
| 0.285235
| 4.697987
| 0.415948
| 10.426823
| 0.349402
| 27.711288
| false
| false
|
2024-10-13
|
2024-10-13
| 1
|
BoltMonkey/SuperNeuralDreadDevil-8b (Merge)
|
|
BrainWave-ML_llama3.2-3B-maths-orpo_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BrainWave-ML/llama3.2-3B-maths-orpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BrainWave-ML/llama3.2-3B-maths-orpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BrainWave-ML__llama3.2-3B-maths-orpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BrainWave-ML/llama3.2-3B-maths-orpo
|
d149d83d8e8f3883421d800848fec85766181923
| 5.076083
|
apache-2.0
| 2
| 3
| true
| false
| false
| false
| 0.707219
| 0.204907
| 20.490742
| 0.291178
| 2.347041
| 0
| 0
| 0.259228
| 1.230425
| 0.357531
| 4.52474
| 0.116772
| 1.863549
| false
| false
|
2024-10-24
|
2024-10-24
| 2
|
meta-llama/Llama-3.2-3B-Instruct
|
BramVanroy_GEITje-7B-ultra_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/GEITje-7B-ultra" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/GEITje-7B-ultra</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__GEITje-7B-ultra-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/GEITje-7B-ultra
|
d4552cdc6f015754646464d8411aa4f6bcdba8e8
| 10.909606
|
cc-by-nc-4.0
| 37
| 7
| true
| false
| false
| true
| 0.619523
| 0.372344
| 37.234427
| 0.377616
| 12.879913
| 0.009063
| 0.906344
| 0.262584
| 1.677852
| 0.328979
| 1.522396
| 0.20113
| 11.236702
| false
| false
|
2024-01-27
|
2024-10-28
| 3
|
mistralai/Mistral-7B-v0.1
|
BramVanroy_fietje-2_bfloat16
|
bfloat16
|
🟩 continuously pretrained
|
🟩
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2
|
3abe75d01094b713368e3d911ffb78a2d66ead22
| 9.027007
|
mit
| 6
| 2
| true
| false
| false
| false
| 0.312539
| 0.209803
| 20.980332
| 0.403567
| 15.603676
| 0.009063
| 0.906344
| 0.254195
| 0.559284
| 0.369563
| 5.161979
| 0.198554
| 10.950428
| false
| false
|
2024-04-09
|
2024-10-28
| 1
|
microsoft/phi-2
|
BramVanroy_fietje-2-chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2-chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2-chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2-chat
|
364e785d90438b787b94e33741a930c9932353c0
| 10.388869
|
mit
| 1
| 2
| true
| false
| false
| true
| 0.399033
| 0.291736
| 29.173593
| 0.414975
| 17.718966
| 0.005287
| 0.528701
| 0.239933
| 0
| 0.35276
| 3.195052
| 0.205452
| 11.716903
| false
| false
|
2024-04-29
|
2024-10-28
| 3
|
microsoft/phi-2
|
BramVanroy_fietje-2-instruct_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2-instruct
|
b7b44797cd52eda1182667217e8371dbdfee4976
| 10.196192
|
mit
| 2
| 2
| true
| false
| false
| true
| 0.324395
| 0.278996
| 27.89964
| 0.413607
| 17.57248
| 0.005287
| 0.528701
| 0.233221
| 0
| 0.336917
| 2.914583
| 0.210356
| 12.261746
| false
| false
|
2024-04-27
|
2024-10-28
| 2
|
microsoft/phi-2
|
Casual-Autopsy_L3-Umbral-Mind-RP-v2.0-8B_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Casual-Autopsy__L3-Umbral-Mind-RP-v2.0-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
|
b46c066ea8387264858dc3461f382e7b42fd9c48
| 25.911927
|
llama3
| 12
| 8
| true
| false
| false
| true
| 0.988385
| 0.712263
| 71.226346
| 0.526241
| 32.486278
| 0.110272
| 11.02719
| 0.286913
| 4.9217
| 0.368667
| 5.55
| 0.37234
| 30.260047
| true
| false
|
2024-06-26
|
2024-07-02
| 1
|
Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B (Merge)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.