Update README.md
Browse files
README.md
CHANGED
@@ -3,7 +3,6 @@ thumbnail: https://github.com/rinnakk/japanese-pretrained-models/blob/master/rin
|
|
3 |
license: apache-2.0
|
4 |
language:
|
5 |
- ja
|
6 |
-
- en
|
7 |
tags:
|
8 |
- qwen2
|
9 |
- conversational
|
@@ -25,9 +24,13 @@ library_name: transformers
|
|
25 |
This model is an instruction-tuned variant of [rinna/qwen2.5-bakeneko-32b](https://huggingface.co/rinna/qwen2.5-bakeneko-32b), fine-tuned using Chat Vector and Simple Preference Optimization (SimPO).
|
26 |
It adheres to the Qwen2.5 chat format and is designed to deliever superior performance in Japanese language tasks.
|
27 |
|
28 |
-
|
|
29 |
-
| :- | :-
|
30 |
-
|
|
|
|
|
|
|
|
|
|
31 |
|
32 |
* **Model architecture**
|
33 |
|
@@ -55,11 +58,27 @@ It adheres to the Qwen2.5 chat format and is designed to deliever superior perfo
|
|
55 |
- [Toshiaki Wakatsuki](https://huggingface.co/t-w)
|
56 |
- [Kei Sawada](https://huggingface.co/keisawada)
|
57 |
|
|
|
|
|
|
|
|
|
58 |
---
|
59 |
|
60 |
# Benchmarking
|
61 |
|
62 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
63 |
|
64 |
---
|
65 |
|
|
|
3 |
license: apache-2.0
|
4 |
language:
|
5 |
- ja
|
|
|
6 |
tags:
|
7 |
- qwen2
|
8 |
- conversational
|
|
|
24 |
This model is an instruction-tuned variant of [rinna/qwen2.5-bakeneko-32b](https://huggingface.co/rinna/qwen2.5-bakeneko-32b), fine-tuned using Chat Vector and Simple Preference Optimization (SimPO).
|
25 |
It adheres to the Qwen2.5 chat format and is designed to deliever superior performance in Japanese language tasks.
|
26 |
|
27 |
+
| Model Type | Model Name
|
28 |
+
| :- | :-
|
29 |
+
| Japanese Continual Pre-Training Model | Qwen2.5 Bakeneko 32B [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b)
|
30 |
+
| Instruction-Tuning Model | Qwen2.5 Bakeneko 32B Instruct [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct)[[AWQ]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-awq)[[GGUF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gptq-int4)
|
31 |
+
| DeepSeek R1 Distill Qwen2.5 Merged Reasoning Model | DeepSeek R1 Distill Qwen2.5 Bakeneko 32B [[HF]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b)[[AWQ]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-awq)[[GGUF]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gguf)[[GPTQ int8]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gptq-int4)
|
32 |
+
| QwQ Merged Reasoning Model | QwQ Bakeneko 32B [[HF]](https://huggingface.co/rinna/qwq-bakeneko-32b)[[AWQ]](https://huggingface.co/rinna/qwq-bakeneko-32b-awq)[[GGUF]](https://huggingface.co/rinna/qwq-bakeneko-32b-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwq-bakeneko-32b-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwq-bakeneko-32b-gptq-int4)
|
33 |
+
| QwQ Bakeneko Merged Instruction-Tuning Model | Qwen2.5 Bakeneko 32B Instruct V2 [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2)[[AWQ]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-awq)[[GGUF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gptq-int4)
|
34 |
|
35 |
* **Model architecture**
|
36 |
|
|
|
58 |
- [Toshiaki Wakatsuki](https://huggingface.co/t-w)
|
59 |
- [Kei Sawada](https://huggingface.co/keisawada)
|
60 |
|
61 |
+
* **Release date**
|
62 |
+
|
63 |
+
February 13, 2025
|
64 |
+
|
65 |
---
|
66 |
|
67 |
# Benchmarking
|
68 |
|
69 |
+
| Model | Japanese LM Evaluation Harness | Japanese MT-Bench (first turn) | Japanese MT-Bench (multi turn)
|
70 |
+
| :- | :-: | :-: | :-:
|
71 |
+
| [Qwen/Qwen2.5-32B](https://huggingface.co/Qwen/Qwen2.5-32B) | 79.46 | - | -
|
72 |
+
| [rinna/qwen2.5-bakeneko-32b](https://huggingface.co/rinna/qwen2.5-bakeneko-32b) | 79.18 | - | -
|
73 |
+
| [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) | 78.29 | 8.13 | 7.54
|
74 |
+
| [rinna/qwen2.5-bakeneko-32b-instruct](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct) | 79.62 | 8.17 | 7.66
|
75 |
+
| [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2) | 77.92 | 8.86 | 8.53
|
76 |
+
| [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B) | 73.51 | 7.39 | 6.88
|
77 |
+
| [rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b) | 77.43 | 8.58 | 8.19
|
78 |
+
| [Qwen/QwQ-32B](https://huggingface.co/Qwen/QwQ-32B) | 76.12 | 8.58 | 8.25
|
79 |
+
| [rinna/qwq-bakeneko-32b](https://huggingface.co/rinna/qwq-bakeneko-32b) | 78.31 | 8.81 | 8.52
|
80 |
+
|
81 |
+
For detailed benchmarking results, please refer to [rinna's LM benchmark page (Sheet 20250213)](https://rinnakk.github.io/research/benchmarks/lm/index.html).
|
82 |
|
83 |
---
|
84 |
|