922CA commited on
Commit
4403e1e
·
1 Parent(s): c3714d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -13,19 +13,19 @@ Noting the last remark, while the lora works it was really just for getting more
13
  * Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
14
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
15
  * From chat LLaMA-2-7b
16
- * Lora of Delphi v0.1
17
 
18
  # lm2_08152023a
19
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
20
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
21
  * From chat LLaMA-2-7b (testing our new dataset)
22
- * Lora of Delphi v0.2
23
 
24
  # lm2_08152023b
25
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
26
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
27
  * From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
28
- * Lora of Delphi v0.2a
29
 
30
  # lm2_08162023c
31
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
@@ -53,7 +53,7 @@ Noting the last remark, while the lora works it was really just for getting more
53
  * 150 steps (cut-off before overfit)
54
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
55
  * From chat LLaMA-2-7b
56
- * Lora of Delphi v0.2e
57
 
58
  # llama-2-7b-chat-monika-v0.3 (~08/20/2023)
59
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
@@ -67,18 +67,18 @@ Noting the last remark, while the lora works it was really just for getting more
67
  * 1 epoch
68
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
69
  * From chat LLaMA-2-7b
70
- * Lora of Delphi/Monika v0.3a
71
 
72
  # llama-2-7b-chat-monika-v0.3b (~08/20/2023)
73
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
74
  * 2 epochs
75
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
76
  * From chat LLaMA-2-7b
77
- * Lora of Delphi/Monika v0.3b
78
 
79
  # llama-2-7b-chat-monika-v0.3c (~08/21/2023)
80
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
81
  * 3 epochs + changed some hyperparams (smaller lora, faster training)
82
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
83
  * From chat LLaMA-2-7b
84
- * Lora of Delphi/Monika v0.3c (compute ran out before merge was completed)
 
13
  * Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
14
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
15
  * From chat LLaMA-2-7b
16
+ * Lora of [Delphi v0.1](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.1)
17
 
18
  # lm2_08152023a
19
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
20
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
21
  * From chat LLaMA-2-7b (testing our new dataset)
22
+ * Lora of [Delphi v0.2](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2)
23
 
24
  # lm2_08152023b
25
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
26
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
27
  * From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
28
+ * Lora of [Delphi v0.2a](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2a)
29
 
30
  # lm2_08162023c
31
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
 
53
  * 150 steps (cut-off before overfit)
54
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
55
  * From chat LLaMA-2-7b
56
+ * Lora of [Delphi v0.2e](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)
57
 
58
  # llama-2-7b-chat-monika-v0.3 (~08/20/2023)
59
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
 
67
  * 1 epoch
68
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
69
  * From chat LLaMA-2-7b
70
+ * Lora of [Delphi/Monika v0.3a](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3a)
71
 
72
  # llama-2-7b-chat-monika-v0.3b (~08/20/2023)
73
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
74
  * 2 epochs
75
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
76
  * From chat LLaMA-2-7b
77
+ * Lora of [Delphi/Monika v0.3b](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3b)
78
 
79
  # llama-2-7b-chat-monika-v0.3c (~08/21/2023)
80
  * Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
81
  * 3 epochs + changed some hyperparams (smaller lora, faster training)
82
  * Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
83
  * From chat LLaMA-2-7b
84
+ * Lora of [Delphi/Monika v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1)