Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +91 -77
README.md CHANGED
@@ -1,78 +1,92 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-1.5B-Instruct
4
- base_model_relation: finetune
5
- library_name: peft
6
- pipeline_tag: text-generation
7
- tags:
8
- - mergekit
9
- - merge
10
- - llama-factory
11
- - lora
12
- datasets:
13
- - allura-org/fujin-cleaned-stage-1
14
- - Dampfinchen/Creative_Writing_Multiturn
15
- - ToastyPigeon/SpringDragon
16
- - allura-org/medquad_sharegpt
17
- - allura-org/scienceqa_sharegpt
18
- - Alignment-Lab-AI/orcamath-sharegpt
19
- ---
20
- # Q25-1.5-VeoLu-R2
21
- ![made with StableNoobAI-IterSPO in sd-webui-forge](veolu.png)
22
- [*A source of life and hope for the land.*](https://www.youtube.com/watch?v=TJRq1Ag2Wmw)
23
-
24
- Q25-1.5B-Veo Lu is a tiny General-Purpose Creative model, made up of a merge of bespoke finetunes on Qwen 2.5-1.5B-Instruct.
25
-
26
- Inspired by the success of [MN-12B-Mag Mell](https://huggingface.co/inflatebot/MN-12B-Mag-Mell-R1) and [MS-Meadowlark-22B](https://huggingface.co/allura-org/MS-Meadowlark-22B), Veo Lu was trained on a healthy, balanced diet of of Internet fiction, roleplaying, adventuring, and reasoning/general knowledge.
27
-
28
- The components of Veo Lu are:
29
-
30
- * Bard (pretrain, writing): [Fujin (Cleaned/extended Rosier)](https://huggingface.co/datasets/allura-org/fujin-cleaned-stage-1)
31
- * Scribe (pretrain, roleplay): [Creative Writing Multiturn](https://huggingface.co/datasets/Dampfinchen/Creative_Writing_Multiturn)
32
- * Cartographer (pretrain, adventuring): [SpringDragon](https://huggingface.co/datasets/ToastyPigeon/SpringDragon)
33
- * Alchemist (SFT, science/reasoning): [ScienceQA,](https://huggingface.co/datasets/allura-org/scienceqa_sharegpt) [MedquadQA,](https://huggingface.co/datasets/allura-org/medquad_sharegpt) [Orca Math Word Problems](https://huggingface.co/datasets/Alignment-Lab-AI/orcamath-sharegpt)
34
-
35
- This model is capable of carrying on a scene without going completely off the rails. That being said, it only has 1.5B parameters. So please, for the love of God, *manage your expectations.*
36
- Since it's Qwen, use ChatML formatting. Turn the temperature down to ~0.7-0.8 and try a dash of rep-pen.
37
-
38
- GGUFs coming soon, but honestly, the full-precision model is 3.5GB in size. You might wanna have a go at running this unquantized with vLLM.
39
- ```
40
- pip install vllm
41
- vllm serve Alfitaria/Q25-1.5B-VeoLu --max-model-len 16384 --max-num-seqs 1
42
- ```
43
-
44
- Made by inflatebot.
45
-
46
- Special thanks to our friends at [Allura](https://huggingface.co/allura-org), and especially to [Auri](https://huggingface.co/AuriAetherwiing), who basically held my hand through the whole process. Her effort and enthusiasm carried this project forward.
47
-
48
- ### Configuration
49
-
50
- The following YAML configuration was used to produce this model:
51
-
52
- ```yaml
53
- base_model: Qwen/Qwen2.5-1.5B-Instruct
54
- dtype: bfloat16
55
- merge_method: task_arithmetic
56
- parameters:
57
- normalize: 1.0
58
- slices:
59
- - sources:
60
- - layer_range: [0, 28]
61
- model: bard
62
- parameters:
63
- weight: 1.0
64
- - layer_range: [0, 28]
65
- model: scribe
66
- parameters:
67
- weight: 1.0
68
- - layer_range: [0, 28]
69
- model: cartographer
70
- parameters:
71
- weight: 1.0
72
- - layer_range: [0, 28]
73
- model: alchemist
74
- parameters:
75
- weight: 1.0
76
- - layer_range: [0, 28]
77
- model: Qwen/Qwen2.5-1.5B-Instruct
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
  ```
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-1.5B-Instruct
4
+ base_model_relation: finetune
5
+ library_name: peft
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ - llama-factory
11
+ - lora
12
+ datasets:
13
+ - allura-org/fujin-cleaned-stage-1
14
+ - Dampfinchen/Creative_Writing_Multiturn
15
+ - ToastyPigeon/SpringDragon
16
+ - allura-org/medquad_sharegpt
17
+ - allura-org/scienceqa_sharegpt
18
+ - Alignment-Lab-AI/orcamath-sharegpt
19
+ language:
20
+ - zho
21
+ - eng
22
+ - fra
23
+ - spa
24
+ - por
25
+ - deu
26
+ - ita
27
+ - rus
28
+ - jpn
29
+ - kor
30
+ - vie
31
+ - tha
32
+ - ara
33
+ ---
34
+ # Q25-1.5-VeoLu-R2
35
+ ![made with StableNoobAI-IterSPO in sd-webui-forge](veolu.png)
36
+ [*A source of life and hope for the land.*](https://www.youtube.com/watch?v=TJRq1Ag2Wmw)
37
+
38
+ Q25-1.5B-Veo Lu is a tiny General-Purpose Creative model, made up of a merge of bespoke finetunes on Qwen 2.5-1.5B-Instruct.
39
+
40
+ Inspired by the success of [MN-12B-Mag Mell](https://huggingface.co/inflatebot/MN-12B-Mag-Mell-R1) and [MS-Meadowlark-22B](https://huggingface.co/allura-org/MS-Meadowlark-22B), Veo Lu was trained on a healthy, balanced diet of of Internet fiction, roleplaying, adventuring, and reasoning/general knowledge.
41
+
42
+ The components of Veo Lu are:
43
+
44
+ * Bard (pretrain, writing): [Fujin (Cleaned/extended Rosier)](https://huggingface.co/datasets/allura-org/fujin-cleaned-stage-1)
45
+ * Scribe (pretrain, roleplay): [Creative Writing Multiturn](https://huggingface.co/datasets/Dampfinchen/Creative_Writing_Multiturn)
46
+ * Cartographer (pretrain, adventuring): [SpringDragon](https://huggingface.co/datasets/ToastyPigeon/SpringDragon)
47
+ * Alchemist (SFT, science/reasoning): [ScienceQA,](https://huggingface.co/datasets/allura-org/scienceqa_sharegpt) [MedquadQA,](https://huggingface.co/datasets/allura-org/medquad_sharegpt) [Orca Math Word Problems](https://huggingface.co/datasets/Alignment-Lab-AI/orcamath-sharegpt)
48
+
49
+ This model is capable of carrying on a scene without going completely off the rails. That being said, it only has 1.5B parameters. So please, for the love of God, *manage your expectations.*
50
+ Since it's Qwen, use ChatML formatting. Turn the temperature down to ~0.7-0.8 and try a dash of rep-pen.
51
+
52
+ GGUFs coming soon, but honestly, the full-precision model is 3.5GB in size. You might wanna have a go at running this unquantized with vLLM.
53
+ ```
54
+ pip install vllm
55
+ vllm serve Alfitaria/Q25-1.5B-VeoLu --max-model-len 16384 --max-num-seqs 1
56
+ ```
57
+
58
+ Made by inflatebot.
59
+
60
+ Special thanks to our friends at [Allura](https://huggingface.co/allura-org), and especially to [Auri](https://huggingface.co/AuriAetherwiing), who basically held my hand through the whole process. Her effort and enthusiasm carried this project forward.
61
+
62
+ ### Configuration
63
+
64
+ The following YAML configuration was used to produce this model:
65
+
66
+ ```yaml
67
+ base_model: Qwen/Qwen2.5-1.5B-Instruct
68
+ dtype: bfloat16
69
+ merge_method: task_arithmetic
70
+ parameters:
71
+ normalize: 1.0
72
+ slices:
73
+ - sources:
74
+ - layer_range: [0, 28]
75
+ model: bard
76
+ parameters:
77
+ weight: 1.0
78
+ - layer_range: [0, 28]
79
+ model: scribe
80
+ parameters:
81
+ weight: 1.0
82
+ - layer_range: [0, 28]
83
+ model: cartographer
84
+ parameters:
85
+ weight: 1.0
86
+ - layer_range: [0, 28]
87
+ model: alchemist
88
+ parameters:
89
+ weight: 1.0
90
+ - layer_range: [0, 28]
91
+ model: Qwen/Qwen2.5-1.5B-Instruct
92
  ```