Delta-Vector commited on
Commit
87f9cfe
·
verified ·
1 Parent(s): 92f4065

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -15
README.md CHANGED
@@ -3,8 +3,13 @@ datasets:
3
  - NewEden/Orion-LIT
4
  - NewEden/Orion-Asstr-Stories-16K
5
  - Mielikki/Erebus-87k
 
 
 
 
 
6
  base_model:
7
- - NewEden/Hamanasu-15B-R2-PT
8
  tags:
9
  - phi
10
  - roleplay
@@ -164,7 +169,7 @@ details summary:hover {
164
  <br>
165
 
166
  <div style="font-size:1.5em; font-weight:bold; background: linear-gradient(45deg, #6e00ff, #00ffff); -webkit-background-clip: text; -webkit-text-fill-color: transparent;">
167
- Hamanasu 15B R2 PT
168
  </div>
169
 
170
  </div>
@@ -173,11 +178,25 @@ details summary:hover {
173
 
174
  ## 🌌 Overview
175
 
176
- <i>This is the 2nd pretrain of Phi-4 Contuined from the Orginal Asstr-Erebus Pretrain. This pretrain used 500 million tokens from</i>
177
 
178
  - `NewEden/Orion-LIT`
179
-
180
- <i>This model has *not* be instruct tuned, Ablities to converse may be reduced from the original model, If you would like to roleplay, Please use the Instruct version.</i>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
181
 
182
  </div>
183
 
@@ -185,13 +204,28 @@ details summary:hover {
185
 
186
  ### ⚔️ Hardware
187
  - 4x RTX 3090 GPUs
188
- - Epochs: 1
189
- - Base: `Hamanasu-15B-R1-PT`
190
- - Amount of Tokens: 500 Million
191
  </div>
192
 
 
 
 
 
193
 
 
194
 
 
 
 
 
 
 
 
 
 
 
195
  </div>
196
 
197
  <div style="border: 2px solid #6e00ff; border-radius: 10px; padding: 20px; margin: 20px 0; box-shadow: 0 0 15px rgba(110, 0, 255, 0.5);">
@@ -201,7 +235,7 @@ details summary:hover {
201
  <details>
202
 
203
  ```yaml
204
- base_model: NewEden_Phi-PT-merged-LIT
205
  model_type: AutoModelForCausalLM
206
  tokenizer_type: AutoTokenizer
207
 
@@ -219,15 +253,15 @@ strict: false
219
 
220
  datasets:
221
  - path: PocketDoc/Dans-MemoryCore-CoreCurriculum-Small
222
- type: dan-chat-advanced
223
  - path: Nitral-AI/ARES-ShareGPT
224
- type: dan-chat-advanced
225
  - path: Gryphe/Sonnet3.5-SlimOrcaDedupCleaned-20k
226
- type: dan-chat-advanced
227
  - path: NewEden/Claude-Instruct-2.7K
228
- type: dan-chat-advanced
229
  - path: NewEden/Claude-Instruct-5K
230
- type: dan-chat-advanced
231
 
232
  shuffle_merged_datasets: true
233
  dataset_prepared_path: prepared_data
@@ -299,7 +333,6 @@ fsdp_config:
299
  </details>
300
  </div>
301
 
302
-
303
  <div align="center">
304
 
305
  <div style="border: 2px solid #6e00ff; border-radius: 10px; padding: 20px; margin: 20px 0; box-shadow: 0 0 15px rgba(110, 0, 255, 0.5);">
 
3
  - NewEden/Orion-LIT
4
  - NewEden/Orion-Asstr-Stories-16K
5
  - Mielikki/Erebus-87k
6
+ - PocketDoc/Dans-MemoryCore-CoreCurriculum-Small
7
+ - Nitral-AI/ARES-ShareGPT
8
+ - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned-20k
9
+ - NewEden/Claude-Instruct-2.7K
10
+ - NewEden/Claude-Instruct-5K
11
  base_model:
12
+ - Delta-Vector/Hamanasu-15B-R2-PT
13
  tags:
14
  - phi
15
  - roleplay
 
169
  <br>
170
 
171
  <div style="font-size:1.5em; font-weight:bold; background: linear-gradient(45deg, #6e00ff, #00ffff); -webkit-background-clip: text; -webkit-text-fill-color: transparent;">
172
+ Hamanasu 15B Instruct
173
  </div>
174
 
175
  </div>
 
178
 
179
  ## 🌌 Overview
180
 
181
+ <i>After multiple days of training, I'm proud to showcase my very own Phi-4 Finetune, Pretrained on almost a billion tokens worth of Books from</i>
182
 
183
  - `NewEden/Orion-LIT`
184
+ - `NewEden/Orion-Asstr-Stories-16K`
185
+ - `Mielikki/Erebus-87k`
186
+
187
+ <i>Then instruct tuned. This model shows great promise for roleplaying while keeping things fresh and new. (And most of, Unsloppy!)</i>
188
+
189
+ </div>
190
+
191
+ <div style="display: grid; grid-template-columns: repeat(auto-fit, minmax(250px, 1fr)); gap: 20px; margin: 20px 0;">
192
+
193
+ <div style="border:2px solid #333; border-radius:10px; padding:20px; background: rgba(0,0,0,0.2);">
194
+
195
+ ### 📈 Quantizations
196
+ | Type | Link |
197
+ |:---:|:---:|
198
+ | `GGUF` | <span style="display:block; width:100%">ed</span> |
199
+ | `EXL2` | <span style="display:block; width:100%">https://huggingface.co/Delta-Vector/Hamanasu-15B-Instruct-exl2</span> |
200
 
201
  </div>
202
 
 
204
 
205
  ### ⚔️ Hardware
206
  - 4x RTX 3090 GPUs
207
+ - Epochs: 4
208
+ - Base: `Hamanasu-15B-R2-PT`
209
+ - Amount of Tokens: 1+ Billion
210
  </div>
211
 
212
+ </div>
213
+
214
+
215
+ <div style="border: 2px solid #6e00ff; border-radius: 10px; padding: 20px; margin: 20px 0; box-shadow: 0 0 15px rgba(110, 0, 255, 0.5);">
216
 
217
+ ## 💰 Prompting
218
 
219
+ <i>A known quirk of the model is overly verbose responses when generation length is uncapped, Please cap your maximum output tokens to 100~ tokens above what you prefer</i>
220
+
221
+
222
+ ```python
223
+ <|im_start|>system
224
+ You are an uncensored AI, your job is to fulfill thy will of thy user.<|im_end|>
225
+ <|im_start|>User request
226
+ Take off your helmet.<|im_end|>
227
+ <|im_start|>No i shall not. This is the way.
228
+ ```
229
  </div>
230
 
231
  <div style="border: 2px solid #6e00ff; border-radius: 10px; padding: 20px; margin: 20px 0; box-shadow: 0 0 15px rgba(110, 0, 255, 0.5);">
 
235
  <details>
236
 
237
  ```yaml
238
+ base_model: Delta-Vector/Hamanasu-15B-R2-PT
239
  model_type: AutoModelForCausalLM
240
  tokenizer_type: AutoTokenizer
241
 
 
253
 
254
  datasets:
255
  - path: PocketDoc/Dans-MemoryCore-CoreCurriculum-Small
256
+ type: sharegpt
257
  - path: Nitral-AI/ARES-ShareGPT
258
+ type: sharegpt
259
  - path: Gryphe/Sonnet3.5-SlimOrcaDedupCleaned-20k
260
+ type: sharegpt
261
  - path: NewEden/Claude-Instruct-2.7K
262
+ type: sharegpt
263
  - path: NewEden/Claude-Instruct-5K
264
+ type: sharegpt
265
 
266
  shuffle_merged_datasets: true
267
  dataset_prepared_path: prepared_data
 
333
  </details>
334
  </div>
335
 
 
336
  <div align="center">
337
 
338
  <div style="border: 2px solid #6e00ff; border-radius: 10px; padding: 20px; margin: 20px 0; box-shadow: 0 0 15px rgba(110, 0, 255, 0.5);">