File size: 7,722 Bytes
23393ec c3714d8 23393ec 0be55f4 305470f df5c76e 0be55f4 f22dcc8 0be55f4 2e2e929 f22dcc8 2854d81 4403e1e 2e2e929 f22dcc8 2854d81 4403e1e 2e2e929 f22dcc8 2854d81 4403e1e 2e2e929 bccea64 f22dcc8 bccea64 f22dcc8 bccea64 f22dcc8 bccea64 f22dcc8 bccea64 4403e1e bccea64 f22dcc8 31e7cd6 f22dcc8 4403e1e f22dcc8 4403e1e f22dcc8 ae874c4 4285d7e d116a88 4285d7e 85919c1 3b4d8e6 85919c1 305470f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 |
---
license: other
---
For better/best results, use "Player" and "Monika" like so:
\nPlayer: (prompt)\nMonika:
# lm1_05042023b
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (1 small text file)
* "Raw" (pretty messy) dataset, currently recreating and reformatting + adding DDLC+ dialogue
* From base LLaMA-7b, trained on really low settings for 15 hours on just a CPU via ooba webui
Noting the last remark, while the lora works it was really just for getting more familiar with these things and seeing if we could train something on just a CPU...
# lm2_08152023
* Fine-tuned on Monika dialogue (dataset of 12 items, manually edited)
* From chat LLaMA-2-7b
* Lora of [Delphi v0.1](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.1)
# lm2_08152023a
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
* From chat LLaMA-2-7b (testing our new dataset)
* Lora of [Delphi v0.2](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2)
# lm2_08152023b
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
* Lora of [Delphi v0.2a](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2a)
# lm2_08162023c
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* 2 epochs
* From chat LLaMA-2-7b
* Lora of Delphi v0.2b
# lm2_08162023d
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
* 200 steps
* From chat LLaMA-2-7b
* Lora of Delphi v0.2c
# lm2_08162023e
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
* 2 epochs (overfitted)
* From chat LLaMA-2-7b
* Lora of Delphi v0.2d
# lm2_08162023f
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
* 150 steps (cut-off before overfit)
* From chat LLaMA-2-7b
* Lora of [Delphi v0.2e](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)
# llama-2-7b-chat-monika-v0.3 (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch
* From base LLaMA-2-7b
* Lora of Delphi/Monika v0.3
# llama-2-7b-chat-monika-v0.3a (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch
* From chat LLaMA-2-7b
* Lora of [Delphi/Monika v0.3a](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3a)
# llama-2-7b-chat-monika-v0.3b (~08/20/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From chat LLaMA-2-7b
* Lora of [Delphi/Monika v0.3b](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3b)
# llama-2-7b-chat-monika-v0.3c (~08/21/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs + changed some hyperparams (smaller lora, faster training)
* From chat LLaMA-2-7b
* Lora of [Delphi/Monika v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1)
# LLilmonix3b-v0.1 loras (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1/1a for 1/3 epochs
* From red pajama 3b
* Loras of [LLilmonix3b-v0.1](https://huggingface.co/922-CA/LLilmonix3b-v0.1) and [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.1a)
# LLilmonix3b-v0.2 loras (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2/2a for 1/3 epochs
* From Open LLaMA 3b
* With Lora of [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.2)
# LLilmonix3b-v0.3 (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs
* From Orca Mini 3b
* With Lora of [LLilmonix3b-v0.1a](https://huggingface.co/922-CA/LLilmonix3b-v0.2)
# llama-2-13b-chat-monika-v0.3d (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 1 epoch with hyperparams for smaller lora
* From LLaMA-2-13b
# llama-2-13b-chat-monika-v0.3e (08/26/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by Nous Hermes 13b to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 3 epochs with hyperparams for smaller lora
* From LLaMA-2-13b
# ch1bika-v0.1 (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [TinyLLaMA-1.1b-step-50k-105b](https://huggingface.co/PY007/TinyLlama-1.,1B-step-50K-105b)
# LLilmonix3b-v0.4 (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [Open LLaMA 3b v2](https://huggingface.co/openlm-research/open_llama_3b_v2)
# llama-2-7b-monika-v0.3h-Air2.1-a (09/05/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* 2 epochs
* From [Airoboros-l2-7b-2.1](https://huggingface.co/jondurbin/airoboros-l2-7b-2.1)
# l2-7b-monika-v0.3m (09/07/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From chat LLaMA-2-7b
* Lora of [l2-7b-monika-ddlc-v0.3m](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m)
# l2-7b-monika-v0.3m-Kv2-b (09/08/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From [Kimiko-LLaMA-2-7b](https://huggingface.co/johnwick123forevr/Llama2-chat-kimiko-Sharded-2gb)
* 2 epochs
* Lora of [l2-7b-monika-ddlc-v0.3m-Kv2-b](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m-Kv2-b)
# l2-7b-monika-v0.3m-Kv2-c (09/08/2023)
* Fine-tuned on Monika dialogue from DDLC, reddit, and twitter (dataset of ~600 items augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn into multi-turn chat dialogue + 1st dataset of 12 items)
* From [Kimiko-LLaMA-2-7b](https://huggingface.co/johnwick123forevr/Llama2-chat-kimiko-Sharded-2gb)
* 3 epochs
* Lora of [l2-7b-monika-ddlc-v0.3m-Kv2-c](https://huggingface.co/922-CA/l2-7b-monika-ddlc-v0.3m-Kv2-c)
|