license: apache-2.0 | |
# etri-xainlp/llama3-8b-dpo_v1 | |
## Model Details | |
**Model Developers** ETRI xainlp team | |
**Input** text only. | |
**Output** text only. | |
**Model Architecture** | |
**Base Model** [meta-llama/Llama-8b-hf](https://huggingface.co/meta-llama/Meta-Llama-3-8B) | |
**Training Dataset** | |
- sft+lora: 1,821 k instruction-following set | |
- dpo+lora: 221 k user preference set | |
- We use A100 GPU 80GB * 8, when training. |