yamatazen commited on
Commit
d2afd76
·
verified ·
1 Parent(s): c639f01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -42
README.md CHANGED
@@ -1,42 +1,45 @@
1
- ---
2
- base_model:
3
- - natong19/Mistral-Nemo-Instruct-2407-abliterated
4
- - yamatazen/Himeyuri-Magnum-12B
5
- - shisa-ai/shisa-v2-mistral-nemo-12b
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
-
11
- ---
12
- # Shisa-K-12B
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- ## Merge Details
17
- ### Merge Method
18
-
19
- This model was merged using the [Karcher Mean](https://en.wikipedia.org/wiki/Karcher_mean) merge method.
20
-
21
- ### Models Merged
22
-
23
- The following models were included in the merge:
24
- * [natong19/Mistral-Nemo-Instruct-2407-abliterated](https://huggingface.co/natong19/Mistral-Nemo-Instruct-2407-abliterated)
25
- * [yamatazen/Himeyuri-Magnum-12B](https://huggingface.co/yamatazen/Himeyuri-Magnum-12B)
26
- * [shisa-ai/shisa-v2-mistral-nemo-12b](https://huggingface.co/shisa-ai/shisa-v2-mistral-nemo-12b)
27
-
28
- ### Configuration
29
-
30
- The following YAML configuration was used to produce this model:
31
-
32
- ```yaml
33
- merge_method: karcher
34
- dtype: bfloat16
35
- out_dtype: bfloat16
36
- models:
37
- - model: natong19/Mistral-Nemo-Instruct-2407-abliterated
38
- - model: shisa-ai/shisa-v2-mistral-nemo-12b
39
- - model: yamatazen/Himeyuri-Magnum-12B
40
- tokenizer:
41
- source: natong19/Mistral-Nemo-Instruct-2407-abliterated
42
- ```
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - natong19/Mistral-Nemo-Instruct-2407-abliterated
4
+ - yamatazen/Himeyuri-Magnum-12B
5
+ - shisa-ai/shisa-v2-mistral-nemo-12b
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ language:
11
+ - en
12
+ - ja
13
+ ---
14
+ ![image/png](https://huggingface.co/yamatazen/Shisa-K-12B/resolve/main/Shisa-K-12B.png?download=true)
15
+ # Shisa-K-12B
16
+
17
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
+
19
+ ## Merge Details
20
+ ### Merge Method
21
+
22
+ This model was merged using the [Karcher Mean](https://en.wikipedia.org/wiki/Karcher_mean) merge method.
23
+
24
+ ### Models Merged
25
+
26
+ The following models were included in the merge:
27
+ * [natong19/Mistral-Nemo-Instruct-2407-abliterated](https://huggingface.co/natong19/Mistral-Nemo-Instruct-2407-abliterated)
28
+ * [yamatazen/Himeyuri-Magnum-12B](https://huggingface.co/yamatazen/Himeyuri-Magnum-12B)
29
+ * [shisa-ai/shisa-v2-mistral-nemo-12b](https://huggingface.co/shisa-ai/shisa-v2-mistral-nemo-12b)
30
+
31
+ ### Configuration
32
+
33
+ The following YAML configuration was used to produce this model:
34
+
35
+ ```yaml
36
+ merge_method: karcher
37
+ dtype: bfloat16
38
+ out_dtype: bfloat16
39
+ models:
40
+ - model: natong19/Mistral-Nemo-Instruct-2407-abliterated
41
+ - model: shisa-ai/shisa-v2-mistral-nemo-12b
42
+ - model: yamatazen/Himeyuri-Magnum-12B
43
+ tokenizer:
44
+ source: natong19/Mistral-Nemo-Instruct-2407-abliterated
45
+ ```