lbourdois commited on
Commit
955f49a
·
verified ·
1 Parent(s): 7f5fc98

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +64 -51
README.md CHANGED
@@ -1,51 +1,64 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-7B
4
- - kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
5
- - Qwen/Qwen2.5-7B-Instruct
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
-
11
- ---
12
- # output-model-directory
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- ## Merge Details
17
- ### Merge Method
18
-
19
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
20
-
21
- ### Models Merged
22
-
23
- The following models were included in the merge:
24
- * [kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000](https://huggingface.co/kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000)
25
- * [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct)
26
-
27
- ### Configuration
28
-
29
- The following YAML configuration was used to produce this model:
30
-
31
- ```yaml
32
- base_model: Qwen/Qwen2.5-7B
33
- dtype: bfloat16
34
- merge_method: ties
35
- models:
36
- - model: kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
37
- parameters:
38
- density: 1
39
- weight: 1
40
- - model: Qwen/Qwen2.5-7B-Instruct
41
- parameters:
42
- density: 1
43
- weight: 1
44
- parameters:
45
- density: 1
46
- int8_mask: true
47
- normalize: true
48
- weight: 1
49
- tokenizer_source: kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
50
-
51
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-7B
4
+ - kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
5
+ - Qwen/Qwen2.5-7B-Instruct
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ language:
11
+ - zho
12
+ - eng
13
+ - fra
14
+ - spa
15
+ - por
16
+ - deu
17
+ - ita
18
+ - rus
19
+ - jpn
20
+ - kor
21
+ - vie
22
+ - tha
23
+ - ara
24
+ ---
25
+ # output-model-directory
26
+
27
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
28
+
29
+ ## Merge Details
30
+ ### Merge Method
31
+
32
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
33
+
34
+ ### Models Merged
35
+
36
+ The following models were included in the merge:
37
+ * [kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000](https://huggingface.co/kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000)
38
+ * [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct)
39
+
40
+ ### Configuration
41
+
42
+ The following YAML configuration was used to produce this model:
43
+
44
+ ```yaml
45
+ base_model: Qwen/Qwen2.5-7B
46
+ dtype: bfloat16
47
+ merge_method: ties
48
+ models:
49
+ - model: kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
50
+ parameters:
51
+ density: 1
52
+ weight: 1
53
+ - model: Qwen/Qwen2.5-7B-Instruct
54
+ parameters:
55
+ density: 1
56
+ weight: 1
57
+ parameters:
58
+ density: 1
59
+ int8_mask: true
60
+ normalize: true
61
+ weight: 1
62
+ tokenizer_source: kamruzzaman-asif/qwen2.5_7B_instruct_base_lora_merged_0_15000
63
+
64
+ ```