--- language: - en - ko license: cc-by-nc-4.0 tags: - dnotitia - nlp - llm - slm - conversation - chat - gguf base_model: - dnotitia/Llama-DNA-1.0-8B-Instruct library_name: transformers pipeline_tag: text-generation --- # DNA 1.0 8B Instruct

**DNA 1.0 8B Instruct** is a state-of-the-art (**SOTA**) bilingual language model based on Llama architecture, specifically optimized for Korean language understanding and generation, while also maintaining strong English capabilities. The model was developed through a sophisticated process involving model merging via spherical linear interpolation (**SLERP**) with Llama 3.1 8B Instruct, and underwent knowledge distillation (**KD**) using Llama 3.1 405B as the teacher model. It was extensively trained through continual pre-training (**CPT**) with a high-quality Korean dataset. The training pipeline was completed with supervised fine-tuning (**SFT**) and direct preference optimization (**DPO**) to align with human preferences and enhance instruction-following abilities. DNA 1.0 8B Instruct was fine-tuned on approximately 10B tokens of carefully curated data and has undergone extensive instruction tuning to enhance its ability to follow complex instructions and engage in natural conversations. - **Developed by:** Dnotitia Inc. - **Supported Languages:** Korean, English - **Model Release Date:** Dec 10, 2024 - **Vocab Size:** 128,256 - **Context Length:** 131,072 tokens (128k) - **License:** CC BY-NC 4.0

## Quickstart We offer weights in `F32`, `F16` formats and quantized weights in `Q8_0`, `Q6_K`, `Q5_K`, `Q4_K`, `Q3_K` and `Q2_K` formats. You can run GGUF weights with `llama.cpp` as follows: 1. Install `llama.cpp`. Please refer to the [llama.cpp repository](https://github.com/ggerganov/llama.cpp) for more details. 2. Download DNA 1.0 8B Instruct model in GGUF format. ```bash # Install huggingface_hub if not already installed $ pip install huggingface_hub[cli] # Download the GGUF weights $ huggingface-cli download dnotitia/Llama-DNA-1.0-8B-Instruct-GGUF \ --include "Llama-DNA-1.0-8B-Instruct-Q8_0.gguf" \ --local-dir . ``` 3. Run the model with `llama.cpp` in conversational mode. ```bash $ llama-cli -cnv -m ./Llama-DNA-1.0-8B-Instruct-Q8_0.gguf \ -p "You are a helpful assistant, Dnotitia DNA." ``` ## Run Locally For end users, we introduce two ways to run DNA 1.0 8B Instruct model locally. > **Note** > > We recommend using a repetition penalty not exceeding 1.0 for better generation quality. ### llama.cpp You can run DNA 1.0 8B Instruct model with `llama.cpp` as follows: 1. Install `llama.cpp`. Please refer to the [llama.cpp repository](https://github.com/ggerganov/llama.cpp) for more details. 2. Download DNA 1.0 8B Instruct model in GGUF format. ```bash huggingface-cli download dnotitia/Llama-DNA-1.0-8B-Instruct-GGUF \ --include "DNA-1.0-8B-Instruct-BF16*.gguf" \ --local-dir . ``` 3. Run the model with `llama.cpp` in conversational mode. ```bash llama-cli -cnv -m ./DNA-1.0-8B-Instruct-BF16.gguf \ -p "You are a helpful assistant, Dnotitia DNA." ``` ### Ollama DNA 1.0 8B Instruct model is compatible with Ollama. You can use it as follows: 1. Install Ollama. Please refer to the [Ollama repository](https://github.com/ollama/ollama) for more details. 2. Create a `Modelfile` for DNA 1.0 8B Instruct. ```text # Model path (choose appropriate GGUF weights) FROM ./DNA-1.0-8B-Instruct-BF16.gguf # Parameter values PARAMETER stop "<|endoftext|>" PARAMETER repeat_penalty 1.0 # PARAMETER num_ctx 131072 # if you need a long context # Chat template TEMPLATE """{{- range $i, $_ := .Messages }} {{- $last := eq (len (slice $.Messages $i)) 1 -}} {{ if eq .Role "system" }}[|system|]{{ .Content }}[|endoftext|] {{ continue }} {{ else if eq .Role "user" }}[|user|]{{ .Content }} {{ else if eq .Role "assistant" }}[|assistant|]{{ .Content }}[|endoftext|] {{ end }} {{- if and (ne .Role "assistant") $last }}[|assistant|]{{ end }} {{- end -}}""" # System prompt SYSTEM """You are a helpful assistant, Dnotitia DNA.""" # License LICENSE """CC BY-NC 4.0""" ``` 3. Convert the model to Ollama. ```bash ollama create dna -f Modelfile ``` 4. Run the model with Ollama. ```bash ollama run dna ```
## Limitations While DNA 1.0 8B Instruct demonstrates strong performance, users should be aware of the following limitations: - The model may occasionally generate biased or inappropriate content. - Responses are based on training data and may not reflect current information. - The model may sometimes produce factually incorrect or inconsistent answers. - Performance may vary depending on the complexity and domain of the task. - Generated content should be reviewed for accuracy and appropriateness.
## License The model is released under the [CC BY-NC 4.0 license](./LICENSE). For commercial usage inquiries, please [Contact us](https://www.dnotitia.com/contact/post-form).
## Citation If you use or discuss this model in your academic research, please cite the project to help spread awareness: ``` @article{dnotitiadna2024, title = {Dnotitia DNA 1.0 8B Instruct}, author = {Jungyup Lee, Jemin Kim, Sang Park, Seungjae Lee}, year = {2024}, url = {https://huggingface.co/dnotitia/DNA-1.0-8B-Instruct}, version = {1.0}, } ```
## Contact For technical support and inquiries: [Contact us](https://www.dnotitia.com/contact/post-form)