Update README.md
Browse files
README.md
CHANGED
@@ -13,19 +13,19 @@ Noting the last remark, while the lora works it was really just for getting more
|
|
13 |
* Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
|
14 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
15 |
* From chat LLaMA-2-7b
|
16 |
-
* Lora of Delphi v0.1
|
17 |
|
18 |
# lm2_08152023a
|
19 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
|
20 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
21 |
* From chat LLaMA-2-7b (testing our new dataset)
|
22 |
-
* Lora of Delphi v0.2
|
23 |
|
24 |
# lm2_08152023b
|
25 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
|
26 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
27 |
* From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
|
28 |
-
* Lora of Delphi v0.2a
|
29 |
|
30 |
# lm2_08162023c
|
31 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
|
@@ -53,7 +53,7 @@ Noting the last remark, while the lora works it was really just for getting more
|
|
53 |
* 150 steps (cut-off before overfit)
|
54 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
55 |
* From chat LLaMA-2-7b
|
56 |
-
* Lora of Delphi v0.2e
|
57 |
|
58 |
# llama-2-7b-chat-monika-v0.3 (~08/20/2023)
|
59 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
@@ -67,18 +67,18 @@ Noting the last remark, while the lora works it was really just for getting more
|
|
67 |
* 1 epoch
|
68 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
69 |
* From chat LLaMA-2-7b
|
70 |
-
* Lora of Delphi/Monika v0.3a
|
71 |
|
72 |
# llama-2-7b-chat-monika-v0.3b (~08/20/2023)
|
73 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
74 |
* 2 epochs
|
75 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
76 |
* From chat LLaMA-2-7b
|
77 |
-
* Lora of Delphi/Monika v0.3b
|
78 |
|
79 |
# llama-2-7b-chat-monika-v0.3c (~08/21/2023)
|
80 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
81 |
* 3 epochs + changed some hyperparams (smaller lora, faster training)
|
82 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
83 |
* From chat LLaMA-2-7b
|
84 |
-
* Lora of Delphi/Monika v0.
|
|
|
13 |
* Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
|
14 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
15 |
* From chat LLaMA-2-7b
|
16 |
+
* Lora of [Delphi v0.1](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.1)
|
17 |
|
18 |
# lm2_08152023a
|
19 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
|
20 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
21 |
* From chat LLaMA-2-7b (testing our new dataset)
|
22 |
+
* Lora of [Delphi v0.2](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2)
|
23 |
|
24 |
# lm2_08152023b
|
25 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
|
26 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
27 |
* From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
|
28 |
+
* Lora of [Delphi v0.2a](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2a)
|
29 |
|
30 |
# lm2_08162023c
|
31 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
|
|
|
53 |
* 150 steps (cut-off before overfit)
|
54 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
55 |
* From chat LLaMA-2-7b
|
56 |
+
* Lora of [Delphi v0.2e](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)
|
57 |
|
58 |
# llama-2-7b-chat-monika-v0.3 (~08/20/2023)
|
59 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
|
|
67 |
* 1 epoch
|
68 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
69 |
* From chat LLaMA-2-7b
|
70 |
+
* Lora of [Delphi/Monika v0.3a](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3a)
|
71 |
|
72 |
# llama-2-7b-chat-monika-v0.3b (~08/20/2023)
|
73 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
74 |
* 2 epochs
|
75 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
76 |
* From chat LLaMA-2-7b
|
77 |
+
* Lora of [Delphi/Monika v0.3b](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3b)
|
78 |
|
79 |
# llama-2-7b-chat-monika-v0.3c (~08/21/2023)
|
80 |
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
|
81 |
* 3 epochs + changed some hyperparams (smaller lora, faster training)
|
82 |
* Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
|
83 |
* From chat LLaMA-2-7b
|
84 |
+
* Lora of [Delphi/Monika v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1)
|