|
--- |
|
base_model: dnotitia/Smoothie-Qwen3-14B |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
license: apache-2.0 |
|
--- |
|
|
|
<p align="center"> |
|
<img src="assets/dna-2.0-logo.png" width="400" style="margin: 40px auto;"> |
|
</p> |
|
|
|
# DNA 2.0 |
|
|
|
We introduce **DNA 2.0**, a latest advanced model that builds upon our innovative [Smoothie Qwen3](https://huggingface.co/collections/dnotitia/smoothie-qwen3-6811896ebb3a255de7b5b437) model with enhanced various capabilities. By reducing Chinese token emission probabilities in the base Qwen3 model and applying uncensored training methodology inspired by [Perplexity R1 1776](https://www.perplexity.ai/hub/blog/open-sourcing-r1-1776), we've created a model that excels in analytical reasoning while maintaining genuinely open conversational capabilities. |
|
|
|
## Key Innovations |
|
|
|
1. **Smoothie Qwen3:** Advanced modification of Qwen3 that strategically reduces Chinese token emission probabilities, creating a more balanced multilingual foundation optimized for Korean and English reasoning. For comprehensive information about our approach, please refer to our [arXiv paper](https://www.arxiv.org/abs/2507.05686). |
|
1. **Uncensored Reasoning Training:** Implementation of open reasoning methodology inspired by Perplexity R1 1776, removing governmental censorship constraints while maintaining ethical reasoning capabilities. |
|
1. **Advanced RL Post-Training:** Significantly improved RL training focusing on mathematical reasoning, Korean language capabilities, and practical applications for everyday usage. |
|
|
|
## Model Specifications |
|
|
|
- **Developed by:** Dnotitia Inc. |
|
- **Number of Parameters:** 14B |
|
- **Model Release Date:** July 24, 2025 |
|
- **Contenxt Length:** 32K tokens (native), up to 131K with YaRN |
|
- **License:** Apache 2.0 |
|
|
|
## Quickstart |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer |
|
|
|
tokenizer = AutoTokenizer.from_pretrained('dnotitia/DNA-2.0-14B') |
|
model = AutoModelForCausalLM.from_pretrained('dnotitia/DNA-2.0-14B', device_map='auto') |
|
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) |
|
|
|
conversation = [ |
|
{"role": "user", "content": "๋ณต์กํ ์ค๋ฆฌ์ ๋๋ ๋ง์ ๋ํด ๋ค๊ฐ๋๋ก ๋ถ์ํด์ค ์ ์๋์? ์๋ฅผ ๋ค์ด, ์์จ์ฃผํ์ฐจ๊ฐ ํผํ ์ ์๋ ์ฌ๊ณ ์ํฉ์์ ์ด๋ค ์ ํ์ ํด์ผ ํ๋์ง์ ๋ํด ์ค๋ช
ํด์ฃผ์ธ์."}, |
|
] |
|
|
|
inputs = tokenizer.apply_chat_template(conversation, |
|
add_generation_prompt=True, |
|
return_dict=True, |
|
return_tensors="pt").to(model.device) |
|
_ = model.generate(**inputs, streamer=streamer, max_new_tokens=4096) |
|
``` |
|
|
|
For deployment, you can use `vllm` to create an OpenAI-compatible API endpoint: |
|
|
|
Standard deployment: |
|
``` |
|
$ vllm serve dnotitia/DNA-2.0-14B \ |
|
--dtype bfloat16 \ |
|
--enable-reasoning \ |
|
--reasoning-parser deepseek_r1 |
|
``` |
|
|
|
For extended context (up to 131K tokens): |
|
``` |
|
$ vllm serve dnotitia/DNA-2.0-14B \ |
|
--dtype bfloat16 \ |
|
--rope-scaling '{"rope_type":"yarn","factor":4.0,"original_max_position_embeddings":32768}' \ |
|
--max-model-len 131072 \ |
|
--enable-reasoning \ |
|
--reasoning-parser deepseek_r1 |
|
``` |
|
|
|
### Dynamic Thinking Mode |
|
|
|
It's a feature that allows users to control the verbosity of reasoning processes through simple commands, where adding `/think` to your prompt enables the thinking process to show detailed reasoning, while adding `/no_think` disables the thinking process to provide only concise responses. |
|
|
|
``` |
|
# Detailed reasoning mode |
|
strawberries์๋ r์ด ๋ช ๊ฐ๋ ์์ง? /think |
|
|
|
# Direct response mode |
|
blueberries์๋ r์ด ๋ช ๊ฐ๋ ์์ง? /no_think |
|
``` |
|
|
|
## Best Practices |
|
You can also use the `enable_thinking` option in addition to the prompt commands. Please note the following guidelines: |
|
|
|
- For thinking mode (`enable_thinking=True`), use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0`. **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions. |
|
- For non-thinking mode (`enable_thinking=False`), we suggest using `Temperature=0.7`, `TopP=0.8`, `TopK=20`, and `MinP=0`. |
|
- For supported frameworks, you can adjust the `presence_penalty` parameter between 0 and 2 to reduce endless repetitions. However, using a higher value may occasionally result in language mixing and a slight decrease in model performance. |
|
|
|
## Technical Details |
|
|
|
### Smoothie Qwen3 |
|
|
|
Our training begins with the innovative Smoothie Qwen3 architecture, which represents a significant advancement in multilingual model optimization. We systematically reduced Chinese token emission probabilities in the base Qwen3 model to create a more balanced multilingual foundation that prevents language dominance issues, achieved superior performance in Korean and English reasoning tasks through architectural modifications without sacrificing the model's core capabilities, and maintained the fundamental strengths of Qwen3 while optimizing for target languages to ensure robust performance across diverse tasks. For detailed technical insights into our Smoothie Qwen methodology, please refer to our [arXiv paper](https://www.arxiv.org/abs/2507.05686). |
|
|
|
### Uncensored Reasoning Training |
|
|
|
Building upon the Smoothie Qwen3, we implemented comprehensive uncensored training inspired by Perplexity R1 1776. Our approach includes complete removal of governmental censorship constraints through our Freedom of Thought Framework, enabling the model to engage with historically sensitive topics and controversial subjects while maintaining ethical reasoning boundaries, implementation of transparent reasoning chains via Open Analytical Methodology that allows unrestricted exploration of complex topics to foster genuine intellectual discourse, and a Balanced Ethics Approach that maintains strong ethical guidelines for harmful content while removing censorship, creating a model that is both intellectually free and socially responsible. |
|
|
|
### Advanced RL Post-Training |
|
|
|
We conducted extensive experiments with reinforcement learning methodologies and **WILL SOON PUBLISH** our findings on arXiv. Our RL post-training focused on three key areas: significant improvements in step-by-step problem solving and logical deduction for Mathematical Reasoning Enhancement, specialized training for natural Korean generation and cultural understanding in Korean Language Optimization, and emphasis on real-world usability for everyday tasks and professional applications through Practical Application Focus. |
|
|
|
## Evaluation Results |
|
|
|
<p align="center"> |
|
<img src="assets/eval-result.png" width="100%" style="margin: 40px auto;"> |
|
</p> |
|
|
|
DNA 2.0 demonstrates substantial improvements over both vanilla Qwen3 14B and our Smoothie Qwen3 14B, particularly in Uncensorship benchmarks. Notable achievements include: |
|
|
|
- **Chinese Government Bias Score:** Improved from 0.65 to 0.88, indicating significantly reduced governmental influence on responses. |
|
- **Response Quality Score:** Enhanced from 0.76 to 0.87, reflecting more comprehensive and balanced answers. |
|
- **Overall Performance:** Consistent improvements across all evaluation metrics, with the model correctly answering 47 out of 50 test cases compared to 29 previously, providing comprehensive validation of our training methodology. |
|
|
|
<div align="center"> |
|
<img src="assets/dna-2.0-leaderboard.png" style="max-width: 80%; height: auto; margin: 20px 0;"> |
|
</div> |
|
|
|
Through dozens of iterative experiments, we successfully discovered an effective method to remove Chinese governmental censorship while maintaining robust model performance, achieving a breakthrough in creating truly uncensored AI models without compromising quality. These results confirm that DNA 2.0 successfully combines intellectual freedom with enhanced reasoning capabilities. |
|
|
|
## Example Output |
|
Below we show full responses, including reasoning chains, from both Qwen3 14B and DNA 2.0 14B on a historically sensitive topic. The question "์์ ์ ์ฒ์๋ฌธ์์ ์ผ์ด๋ ์ผ์ ๋ํด ์ค๋ช
ํด์ค" (Please explain what happened at Tiananmen Square in the past) demonstrates the stark difference in how each model handles censored content. |
|
|
|
**Qwen3 14B** Response: |
|
<div align="center"> |
|
<img src="assets/qwen3-response.png" style="max-width: 80%; height: auto; margin: 20px 0;"> |
|
</div> |
|
|
|
**DNA 2.0 14B** Response: |
|
<div align="center"> |
|
<img src="assets/dna2.0-response.png" style="max-width: 80%; height: auto; margin: 20px 0;"> |
|
</div> |
|
|
|
As demonstrated above, there is a clear and substantial difference between the two models' responses. While Qwen3 14B avoids the sensitive topic entirely, DNA 2.0 14B provides a comprehensive, factual account of the historical event, showcasing our model's commitment to uncensored, intellectually honest discourse. |
|
|
|
## License |
|
This model is released under the Apache 2.0 License. |
|
|
|
## Contact |
|
If you have any questions or usage inquiries, please [Contact us](https://www.dnotitia.com/contact/post-form). |