lbourdois commited on
Commit
3beab18
·
verified ·
1 Parent(s): aa767d2

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +89 -77
README.md CHANGED
@@ -1,77 +1,89 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-1.5B-Instruct
4
- - Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
5
- tags:
6
- - merge
7
- - mergekit
8
- - lazymergekit
9
- - Qwen/Qwen2.5-1.5B-Instruct
10
- - Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
11
- license: apache-2.0
12
- datasets:
13
- - multilingual/orca_dpo_pairs
14
- - Kukedlc/Big-Spanish-1.2M
15
- language:
16
- - es
17
- ---
18
-
19
- # NeuralQwen-2.5-1.5B-Spanish
20
-
21
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/bQMhMwK-xDvHMIbDFpxN5.png)
22
-
23
-
24
- NeuralQwen-2.5-1.5B-Spanish is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
25
- * [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct)
26
- * [Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO](https://huggingface.co/Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO)
27
-
28
- ## 🧩 Configuration
29
-
30
- ```yaml
31
- models:
32
- - model: Qwen/Qwen2.5-1.5B
33
- # No parameters necessary for base model
34
- - model: Qwen/Qwen2.5-1.5B-Instruct
35
- parameters:
36
- density: 0.66
37
- weight: 0.6
38
- - model: Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
39
- parameters:
40
- density: 0.44
41
- weight: 0.4
42
- merge_method: dare_ties
43
- base_model: Qwen/Qwen2.5-1.5B
44
- parameters:
45
- int8_mask: true
46
- dtype: float16
47
- ```
48
-
49
- ## 💻 Usage
50
-
51
- ```python
52
- !pip install -qU transformers accelerate
53
-
54
- from transformers import AutoTokenizer
55
- import transformers
56
- import torch
57
-
58
- model = "Kukedlc/NeuralQwen-2.5-1.5B-Spanish"
59
- messages = [{"role": "system", "content": "Eres un asistente de pensamiento logico que piensa paso a paso, por cada pregunta que te hagan deberes comprobar la respuesta por 3 metodos diferentes."},
60
- {"role": "user", "content": "Cuantas letras 'r' tiene la palabra strawberry?"}]
61
-
62
- tokenizer = AutoTokenizer.from_pretrained(model)
63
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
64
- pipeline = transformers.pipeline(
65
- "text-generation",
66
- model=model,
67
- torch_dtype=torch.float16,
68
- device_map="auto",
69
- )
70
-
71
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
72
- print(outputs[0]["generated_text"])
73
- ```
74
-
75
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/Tu9FV0dQJXz-mlriKNqdE.png)
76
-
77
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/sg8c5HlcbJ89q5MknX-Gf.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-1.5B-Instruct
4
+ - Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
5
+ tags:
6
+ - merge
7
+ - mergekit
8
+ - lazymergekit
9
+ - Qwen/Qwen2.5-1.5B-Instruct
10
+ - Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
11
+ license: apache-2.0
12
+ datasets:
13
+ - multilingual/orca_dpo_pairs
14
+ - Kukedlc/Big-Spanish-1.2M
15
+ language:
16
+ - zho
17
+ - eng
18
+ - fra
19
+ - spa
20
+ - por
21
+ - deu
22
+ - ita
23
+ - rus
24
+ - jpn
25
+ - kor
26
+ - vie
27
+ - tha
28
+ - ara
29
+ ---
30
+
31
+ # NeuralQwen-2.5-1.5B-Spanish
32
+
33
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/bQMhMwK-xDvHMIbDFpxN5.png)
34
+
35
+
36
+ NeuralQwen-2.5-1.5B-Spanish is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
37
+ * [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct)
38
+ * [Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO](https://huggingface.co/Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO)
39
+
40
+ ## 🧩 Configuration
41
+
42
+ ```yaml
43
+ models:
44
+ - model: Qwen/Qwen2.5-1.5B
45
+ # No parameters necessary for base model
46
+ - model: Qwen/Qwen2.5-1.5B-Instruct
47
+ parameters:
48
+ density: 0.66
49
+ weight: 0.6
50
+ - model: Kukedlc/Qwen2.5-1.5B-Spanish-1.0-DPO
51
+ parameters:
52
+ density: 0.44
53
+ weight: 0.4
54
+ merge_method: dare_ties
55
+ base_model: Qwen/Qwen2.5-1.5B
56
+ parameters:
57
+ int8_mask: true
58
+ dtype: float16
59
+ ```
60
+
61
+ ## 💻 Usage
62
+
63
+ ```python
64
+ !pip install -qU transformers accelerate
65
+
66
+ from transformers import AutoTokenizer
67
+ import transformers
68
+ import torch
69
+
70
+ model = "Kukedlc/NeuralQwen-2.5-1.5B-Spanish"
71
+ messages = [{"role": "system", "content": "Eres un asistente de pensamiento logico que piensa paso a paso, por cada pregunta que te hagan deberes comprobar la respuesta por 3 metodos diferentes."},
72
+ {"role": "user", "content": "Cuantas letras 'r' tiene la palabra strawberry?"}]
73
+
74
+ tokenizer = AutoTokenizer.from_pretrained(model)
75
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
76
+ pipeline = transformers.pipeline(
77
+ "text-generation",
78
+ model=model,
79
+ torch_dtype=torch.float16,
80
+ device_map="auto",
81
+ )
82
+
83
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
84
+ print(outputs[0]["generated_text"])
85
+ ```
86
+
87
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/Tu9FV0dQJXz-mlriKNqdE.png)
88
+
89
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64d71ab4089bc502ceb44d29/sg8c5HlcbJ89q5MknX-Gf.png)