Update README.md
Browse files
README.md
CHANGED
@@ -45,4 +45,26 @@ generated_ids = [
|
|
45 |
]
|
46 |
|
47 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
```
|
|
|
45 |
]
|
46 |
|
47 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
48 |
+
```
|
49 |
+
|
50 |
+
## Reference
|
51 |
+
|
52 |
+
For more detailed information about the model, we encourage you to refer to our paper:
|
53 |
+
|
54 |
+
- **DistilQwen2.5: Industrial Practices of Training Distilled Open Lightweight Language Models**
|
55 |
+
Chengyu Wang, Junbing Yan, Yuanhao Yue, Jun Huang
|
56 |
+
[arXiv:2504.15027](https://arxiv.org/abs/2504.15027)
|
57 |
+
|
58 |
+
You can cite the paper using the following citation format:
|
59 |
+
|
60 |
+
```bibtex
|
61 |
+
@misc{wang2025distilqwen25industrialpracticestraining,
|
62 |
+
title={DistilQwen2.5: Industrial Practices of Training Distilled Open Lightweight Language Models},
|
63 |
+
author={Chengyu Wang and Junbing Yan and Yuanhao Yue and Jun Huang},
|
64 |
+
year={2025},
|
65 |
+
eprint={2504.15027},
|
66 |
+
archivePrefix={arXiv},
|
67 |
+
primaryClass={cs.CL},
|
68 |
+
url={https://arxiv.org/abs/2504.15027}
|
69 |
+
}
|
70 |
```
|