metadata
license: openrail
lm1_05042023b
- Trained on Monika dialogue from DDLC, reddit, and twitter (1 small text file)
- "Raw" (pretty messy) dataset, currently recreating and reformatting + adding DDLC+ dialogue
- From base LLaMA-7b, trained on really low settings for 15 hours on just a CPU via ooba webui
Noting the last remark, while the lora works it was really just for getting more familiar with these things and seeing if we could train something on just a CPU...
lm2_08152023
- Trained on Monika dialogue (dataset of 12 items, manually edited)
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b
- Lora of Delphi v0.1
lm2_08152023a
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of 520 items)
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b (testing our new dataset)
- Lora of Delphi v0.2
lm2_08152023b
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b (was going to try airoboros, and then nous hermes. But would always OOM or crash- will try in near future)
- Lora of Delphi v0.2a
lm2_08162023c
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
- 2 epochs
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b
- Lora of Delphi v0.2b
lm2_08162023d
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items, further cleaned)
- 200 steps
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b
- Lora of Delphi v0.2c
lm2_08162023e
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
- 2 epochs (overfitted)
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b
- Lora of Delphi v0.2d
lm2_08162023f
- Trained on Monika dialogue from DDLC, reddit, and twitter (dataset of ~426 items + 1st dataset of 12 items)
- 150 steps (cut-off before overfit)
- Formatted in chat/RP style, replacing "Human" and "Assistant" with "Player" and "Monika"
- From chat LLaMA-2-7b
- Lora of Delphi v0.2e