lbourdois commited on
Commit
1d9bd51
·
verified ·
1 Parent(s): 51b48a6

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +167 -153
README.md CHANGED
@@ -1,153 +1,167 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - mergekit
5
- - merge
6
- base_model:
7
- - Qwen/Qwen2.5-7B
8
- - rombodawg/Rombos-LLM-V2.5-Qwen-7b
9
- - fblgit/cybertron-v4-qw7B-MGS
10
- - sethuiyer/Qwen2.5-7B-Anvita
11
- model-index:
12
- - name: Qwen2.5-7B-Instruct-Fusion
13
- results:
14
- - task:
15
- type: text-generation
16
- name: Text Generation
17
- dataset:
18
- name: IFEval (0-Shot)
19
- type: HuggingFaceH4/ifeval
20
- args:
21
- num_few_shot: 0
22
- metrics:
23
- - type: inst_level_strict_acc and prompt_level_strict_acc
24
- value: 69.62
25
- name: strict accuracy
26
- source:
27
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
28
- name: Open LLM Leaderboard
29
- - task:
30
- type: text-generation
31
- name: Text Generation
32
- dataset:
33
- name: BBH (3-Shot)
34
- type: BBH
35
- args:
36
- num_few_shot: 3
37
- metrics:
38
- - type: acc_norm
39
- value: 36.18
40
- name: normalized accuracy
41
- source:
42
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
43
- name: Open LLM Leaderboard
44
- - task:
45
- type: text-generation
46
- name: Text Generation
47
- dataset:
48
- name: MATH Lvl 5 (4-Shot)
49
- type: hendrycks/competition_math
50
- args:
51
- num_few_shot: 4
52
- metrics:
53
- - type: exact_match
54
- value: 19.94
55
- name: exact match
56
- source:
57
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
58
- name: Open LLM Leaderboard
59
- - task:
60
- type: text-generation
61
- name: Text Generation
62
- dataset:
63
- name: GPQA (0-shot)
64
- type: Idavidrein/gpqa
65
- args:
66
- num_few_shot: 0
67
- metrics:
68
- - type: acc_norm
69
- value: 7.27
70
- name: acc_norm
71
- source:
72
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
73
- name: Open LLM Leaderboard
74
- - task:
75
- type: text-generation
76
- name: Text Generation
77
- dataset:
78
- name: MuSR (0-shot)
79
- type: TAUR-Lab/MuSR
80
- args:
81
- num_few_shot: 0
82
- metrics:
83
- - type: acc_norm
84
- value: 12.95
85
- name: acc_norm
86
- source:
87
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
88
- name: Open LLM Leaderboard
89
- - task:
90
- type: text-generation
91
- name: Text Generation
92
- dataset:
93
- name: MMLU-PRO (5-shot)
94
- type: TIGER-Lab/MMLU-Pro
95
- config: main
96
- split: test
97
- args:
98
- num_few_shot: 5
99
- metrics:
100
- - type: acc
101
- value: 38.53
102
- name: accuracy
103
- source:
104
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
105
- name: Open LLM Leaderboard
106
- ---
107
- # merge
108
-
109
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
110
-
111
- ## Merge Details
112
- ### Merge Method
113
-
114
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
115
-
116
- ### Models Merged
117
-
118
- The following models were included in the merge:
119
- * [rombodawg/Rombos-LLM-V2.5-Qwen-7b](https://huggingface.co/rombodawg/Rombos-LLM-V2.5-Qwen-7b)
120
- * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
121
- * [sethuiyer/Qwen2.5-7B-Anvita](https://huggingface.co/sethuiyer/Qwen2.5-7B-Anvita)
122
-
123
- ### Configuration
124
-
125
- The following YAML configuration was used to produce this model:
126
-
127
- ```yaml
128
- models:
129
- - model: rombodawg/Rombos-LLM-V2.5-Qwen-7b
130
- - model: fblgit/cybertron-v4-qw7B-MGS
131
- - model: sethuiyer/Qwen2.5-7B-Anvita
132
- merge_method: model_stock
133
- base_model: Qwen/Qwen2.5-7B
134
- parameters:
135
- normalize: false
136
- int8_mask: true
137
- dtype: float16
138
-
139
- ```
140
-
141
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
142
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_bunnycore__Qwen2.5-7B-Instruct-Fusion)
143
-
144
- | Metric |Value|
145
- |-------------------|----:|
146
- |Avg. |30.75|
147
- |IFEval (0-Shot) |69.62|
148
- |BBH (3-Shot) |36.18|
149
- |MATH Lvl 5 (4-Shot)|19.94|
150
- |GPQA (0-shot) | 7.27|
151
- |MuSR (0-shot) |12.95|
152
- |MMLU-PRO (5-shot) |38.53|
153
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ base_model:
7
+ - Qwen/Qwen2.5-7B
8
+ - rombodawg/Rombos-LLM-V2.5-Qwen-7b
9
+ - fblgit/cybertron-v4-qw7B-MGS
10
+ - sethuiyer/Qwen2.5-7B-Anvita
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ model-index:
26
+ - name: Qwen2.5-7B-Instruct-Fusion
27
+ results:
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: IFEval (0-Shot)
33
+ type: HuggingFaceH4/ifeval
34
+ args:
35
+ num_few_shot: 0
36
+ metrics:
37
+ - type: inst_level_strict_acc and prompt_level_strict_acc
38
+ value: 69.62
39
+ name: strict accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: BBH (3-Shot)
48
+ type: BBH
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 36.18
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: hendrycks/competition_math
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 19.94
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ args:
80
+ num_few_shot: 0
81
+ metrics:
82
+ - type: acc_norm
83
+ value: 7.27
84
+ name: acc_norm
85
+ source:
86
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MuSR (0-shot)
93
+ type: TAUR-Lab/MuSR
94
+ args:
95
+ num_few_shot: 0
96
+ metrics:
97
+ - type: acc_norm
98
+ value: 12.95
99
+ name: acc_norm
100
+ source:
101
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: MMLU-PRO (5-shot)
108
+ type: TIGER-Lab/MMLU-Pro
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 38.53
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen2.5-7B-Instruct-Fusion
119
+ name: Open LLM Leaderboard
120
+ ---
121
+ # merge
122
+
123
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
124
+
125
+ ## Merge Details
126
+ ### Merge Method
127
+
128
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
129
+
130
+ ### Models Merged
131
+
132
+ The following models were included in the merge:
133
+ * [rombodawg/Rombos-LLM-V2.5-Qwen-7b](https://huggingface.co/rombodawg/Rombos-LLM-V2.5-Qwen-7b)
134
+ * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
135
+ * [sethuiyer/Qwen2.5-7B-Anvita](https://huggingface.co/sethuiyer/Qwen2.5-7B-Anvita)
136
+
137
+ ### Configuration
138
+
139
+ The following YAML configuration was used to produce this model:
140
+
141
+ ```yaml
142
+ models:
143
+ - model: rombodawg/Rombos-LLM-V2.5-Qwen-7b
144
+ - model: fblgit/cybertron-v4-qw7B-MGS
145
+ - model: sethuiyer/Qwen2.5-7B-Anvita
146
+ merge_method: model_stock
147
+ base_model: Qwen/Qwen2.5-7B
148
+ parameters:
149
+ normalize: false
150
+ int8_mask: true
151
+ dtype: float16
152
+
153
+ ```
154
+
155
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
156
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_bunnycore__Qwen2.5-7B-Instruct-Fusion)
157
+
158
+ | Metric |Value|
159
+ |-------------------|----:|
160
+ |Avg. |30.75|
161
+ |IFEval (0-Shot) |69.62|
162
+ |BBH (3-Shot) |36.18|
163
+ |MATH Lvl 5 (4-Shot)|19.94|
164
+ |GPQA (0-shot) | 7.27|
165
+ |MuSR (0-shot) |12.95|
166
+ |MMLU-PRO (5-shot) |38.53|
167
+