lbourdois commited on
Commit
be216d9
·
verified ·
1 Parent(s): 6f3ae97

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +58 -44
README.md CHANGED
@@ -1,45 +1,59 @@
1
- ---
2
- base_model:
3
- - fblgit/cybertron-v4-qw7B-MGS
4
- - Qwen/Qwen2.5-7B-Instruct
5
- - Tsunami-th/Tsunami-0.5x-7B-Instruct
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- license: apache-2.0
11
- ---
12
- # merge
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- ## Merge Details
17
- ### Merge Method
18
-
19
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
20
-
21
- ### Models Merged
22
-
23
- The following models were included in the merge:
24
- * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
25
- * [Tsunami-th/Tsunami-0.5x-7B-Instruct](https://huggingface.co/Tsunami-th/Tsunami-0.5x-7B-Instruct)
26
-
27
- ### Configuration
28
-
29
- The following YAML configuration was used to produce this model:
30
-
31
- ```yaml
32
- models:
33
- - model: Tsunami-th/Tsunami-0.5x-7B-Instruct
34
- parameters:
35
- weight: 1.0
36
- - model: fblgit/cybertron-v4-qw7B-MGS
37
- parameters:
38
- weight: 1.0
39
- merge_method: task_arithmetic
40
- base_model: Qwen/Qwen2.5-7B-Instruct
41
- parameters:
42
- normalize: true
43
- dtype: bfloat16
44
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ```
 
1
+ ---
2
+ base_model:
3
+ - fblgit/cybertron-v4-qw7B-MGS
4
+ - Qwen/Qwen2.5-7B-Instruct
5
+ - Tsunami-th/Tsunami-0.5x-7B-Instruct
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ license: apache-2.0
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ ---
26
+ # merge
27
+
28
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
29
+
30
+ ## Merge Details
31
+ ### Merge Method
32
+
33
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
34
+
35
+ ### Models Merged
36
+
37
+ The following models were included in the merge:
38
+ * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
39
+ * [Tsunami-th/Tsunami-0.5x-7B-Instruct](https://huggingface.co/Tsunami-th/Tsunami-0.5x-7B-Instruct)
40
+
41
+ ### Configuration
42
+
43
+ The following YAML configuration was used to produce this model:
44
+
45
+ ```yaml
46
+ models:
47
+ - model: Tsunami-th/Tsunami-0.5x-7B-Instruct
48
+ parameters:
49
+ weight: 1.0
50
+ - model: fblgit/cybertron-v4-qw7B-MGS
51
+ parameters:
52
+ weight: 1.0
53
+ merge_method: task_arithmetic
54
+ base_model: Qwen/Qwen2.5-7B-Instruct
55
+ parameters:
56
+ normalize: true
57
+ dtype: bfloat16
58
+
59
  ```