OddTheGreat commited on
Commit
a1f8fe6
·
verified ·
1 Parent(s): 5fa4363

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -29
README.md CHANGED
@@ -7,42 +7,32 @@ library_name: transformers
7
  tags:
8
  - mergekit
9
  - merge
10
-
 
 
 
 
11
  ---
12
- # merge
 
 
 
 
13
 
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
 
16
- ## Merge Details
17
- ### Merge Method
18
 
19
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [mrfakename/mistral-small-3.1-24b-instruct-2503-hf](https://huggingface.co/mrfakename/mistral-small-3.1-24b-instruct-2503-hf) as a base.
20
 
21
- ### Models Merged
22
 
23
- The following models were included in the merge:
24
- * [Gryphe/Pantheon-RP-1.8-24b-Small-3.1](https://huggingface.co/Gryphe/Pantheon-RP-1.8-24b-Small-3.1)
25
- * [OddTheGreat/Apparatus_24B](https://huggingface.co/OddTheGreat/Apparatus_24B)
26
 
27
- ### Configuration
28
 
29
- The following YAML configuration was used to produce this model:
30
 
31
- ```yaml
32
- models:
33
- - model: Gryphe/Pantheon-RP-1.8-24b-Small-3.1
34
- parameters:
35
- density: 0.1
36
- weight: 0.1
37
- - model: OddTheGreat/Apparatus_24B
38
- parameters:
39
- density: 1
40
- weight: 1
41
 
42
- merge_method: ties
43
- base_model: mrfakename/mistral-small-3.1-24b-instruct-2503-hf
44
- parameters:
45
- normalize: false
46
- int8_mask: true
47
- dtype: float16
48
- ```
 
7
  tags:
8
  - mergekit
9
  - merge
10
+ - roleplay
11
+ - creative
12
+ language:
13
+ - en
14
+ - ru
15
  ---
16
+ # Core
17
+
18
+ This is a merge of pretrained language models
19
+
20
+ With new Mistral recently released, and being slightly better than it predecessor, i wanted to update Apparatus.
21
 
22
+ Also, i tested new pantheon, and i like how it mimic human expressions and mannerisms, so idea of new merge was born.
23
 
24
+ Goal of this merge is transfer Apparatus to new Mistral 3.1, and enhance it dialogue capabilities while preserving it stability and ru performance.
 
25
 
26
+ It seems that i succeed. Model is smart enough, instruction-following, decently creative and stable.
27
 
28
+ tested on 250 answers, narration is really good, dialogues too, swipes make difference.
29
 
30
+ Russian performance still here, no problems with full ru cards and partly translatd ones.
 
 
31
 
32
+ Qvink Memory seems to break something. With it on, replies becomes much shorter.
33
 
34
+ Better use prebuilt V7 format in ST, but chatML also work fine. T1.01 XTC 0.1 0.1
35
 
36
+ Is it better than Apparatus? i cannot tell, so your feedback is apperciated.
 
 
 
 
 
 
 
 
 
37
 
38
+ P.S this model not include vision, but as soon i figure how to merge it in, i will update it.