File size: 7,722 Bytes
23393ec
c3714d8
23393ec
0be55f4
305470f
df5c76e
 
 
0be55f4
f22dcc8
0be55f4
 
 
2e2e929
 
 
f22dcc8
2854d81
4403e1e
2e2e929
 
f22dcc8
2854d81
4403e1e
2e2e929
 
f22dcc8
2854d81
4403e1e
2e2e929
bccea64
f22dcc8
bccea64
 
 
 
 
f22dcc8
bccea64
 
 
 
 
f22dcc8
bccea64
 
 
 
 
f22dcc8
bccea64
 
4403e1e
bccea64
f22dcc8
 
 
 
31e7cd6
f22dcc8
 
 
 
 
4403e1e
f22dcc8
 
 
 
 
4403e1e
f22dcc8
ae874c4
 
 
 
4285d7e
 
 
 
 
 
 
 
 
 
 
 
 
 
d116a88
4285d7e
 
 
 
 
 
 
 
 
 
 
 
 
85919c1
 
 
 
 
3b4d8e6
85919c1
 
 
 
 
 
 
 
 
305470f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
---
license: other
---

For better/best results, use "Player" and "Monika" like so:

\nPlayer: (prompt)\nMonika:

# lm1_05042023b
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (1 small text file)
* "Raw" (pretty messy) dataset, currently recreating and reformatting + adding DDLC+ dialogue
* From base LLaMA-7b, trained on really low settings for 15 hours on just a CPU via ooba webui

Noting the last remark, while the lora works it was really just for getting more familiar with these things and seeing if we could train something on just a CPU...

# lm2_08152023
* Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
* From chat LLaMA-2-7b
* Lora of [Delphi v0.1](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.1)

# lm2_08152023a
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
* From chat LLaMA-2-7b (testing our new dataset)
* Lora of [Delphi v0.2](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2)

# lm2_08152023b
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
* Lora of [Delphi v0.2a](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2a)

# lm2_08162023c
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* 2 epochs
* From chat LLaMA-2-7b
* Lora of Delphi v0.2b

# lm2_08162023d
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* 200 steps
* From chat LLaMA-2-7b 
* Lora of Delphi v0.2c

# lm2_08162023e
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
* 2 epochs (overfitted)
* From chat LLaMA-2-7b 
* Lora of Delphi v0.2d

# lm2_08162023f
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
* 150 steps (cut-off before overfit)
* From chat LLaMA-2-7b 
* Lora of [Delphi v0.2e](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)

# llama-2-7b-chat-monika-v0.3 (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch
* From base LLaMA-2-7b 
* Lora of Delphi/Monika v0.3

# llama-2-7b-chat-monika-v0.3a (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch
* From chat LLaMA-2-7b 
* Lora of [Delphi/Monika v0.3a](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3a)

# llama-2-7b-chat-monika-v0.3b (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From chat LLaMA-2-7b 
* Lora of [Delphi/Monika v0.3b](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3b)

# llama-2-7b-chat-monika-v0.3c (~08/21/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs + changed some hyperparams (smaller lora, faster training)
* From chat LLaMA-2-7b 
* Lora of [Delphi/Monika v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1)

# LLilmonix3b-v0.1 loras (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1/1a for 1/3 epochs
* From red pajama 3b
* Loras of [LLilmonix3b-v0.1](https://huggingface.co/922-CA/LLilmonix3b-v0.1) and [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.1a)

# LLilmonix3b-v0.2 loras (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2/2a for 1/3 epochs
* From Open LLaMA 3b
* With Lora of [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.2)

# LLilmonix3b-v0.3 (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs
* From Orca Mini 3b
* With Lora of [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.2)

# llama-2-13b-chat-monika-v0.3d (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch with hyperparams for smaller lora
* From LLaMA-2-13b

# llama-2-13b-chat-monika-v0.3e (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs with hyperparams for smaller lora
* From LLaMA-2-13b

# ch1bika-v0.1 (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [TinyLLaMA-1.1b-step-50k-105b](https://huggingface.co/PY007/TinyLlama-1.,1B-step-50K-105b)

# LLilmonix3b-v0.4 (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [Open LLaMA 3b v2](https://huggingface.co/openlm-research/open_llama_3b_v2)

# llama-2-7b-monika-v0.3h-Air2.1-a (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [Airoboros-l2-7b-2.1](https://huggingface.co/jondurbin/airoboros-l2-7b-2.1)

# l2-7b-monika-v0.3m (09/07/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From chat LLaMA-2-7b 
* Lora of [l2-7b-monika-ddlc-v0.3m](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m)

# l2-7b-monika-v0.3m-Kv2-b (09/08/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From [Kimiko-LLaMA-2-7b](https://huggingface.co/johnwick123forevr/Llama2-chat-kimiko-Sharded-2gb)
* 2 epochs
* Lora of [l2-7b-monika-ddlc-v0.3m-Kv2-b](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m-Kv2-b)

# l2-7b-monika-v0.3m-Kv2-c (09/08/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From [Kimiko-LLaMA-2-7b](https://huggingface.co/johnwick123forevr/Llama2-chat-kimiko-Sharded-2gb)
* 3 epochs
* Lora of [l2-7b-monika-ddlc-v0.3m-Kv2-c](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m-Kv2-c)