Update README.md
Browse files
README.md
CHANGED
@@ -66,6 +66,9 @@ Performance evaluation is ongoing. The model shows promising results in:
|
|
66 |
- Maintaining base model capabilities while achieving linear attention efficiency
|
67 |
- Significantly improved needle-in-haystack task performance compared to pure RWKV architectures
|
68 |
- Competitive performance on standard language modeling benchmarks
|
|
|
|
|
|
|
69 |
|
70 |
## Usage with RWKV-Infer
|
71 |
- **RWKV-Infer** Triton based Hybrid RWKV Inference engine, can be check at: [https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F](https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F)
|
|
|
66 |
- Maintaining base model capabilities while achieving linear attention efficiency
|
67 |
- Significantly improved needle-in-haystack task performance compared to pure RWKV architectures
|
68 |
- Competitive performance on standard language modeling benchmarks
|
69 |
+
- mmlu: 78.39%(Base 82.41%)
|
70 |
+
- gsm8k: 86.88%(Base93.93%) with gentoken=2048
|
71 |
+
- passkey 130k+(Base 500k)
|
72 |
|
73 |
## Usage with RWKV-Infer
|
74 |
- **RWKV-Infer** Triton based Hybrid RWKV Inference engine, can be check at: [https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F](https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F)
|