OpenMOSE commited on
Commit
0f51424
·
verified ·
1 Parent(s): 3a4be1d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -66,6 +66,9 @@ Performance evaluation is ongoing. The model shows promising results in:
66
  - Maintaining base model capabilities while achieving linear attention efficiency
67
  - Significantly improved needle-in-haystack task performance compared to pure RWKV architectures
68
  - Competitive performance on standard language modeling benchmarks
 
 
 
69
 
70
  ## Usage with RWKV-Infer
71
  - **RWKV-Infer** Triton based Hybrid RWKV Inference engine, can be check at: [https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F](https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F)
 
66
  - Maintaining base model capabilities while achieving linear attention efficiency
67
  - Significantly improved needle-in-haystack task performance compared to pure RWKV architectures
68
  - Competitive performance on standard language modeling benchmarks
69
+ - mmlu: 78.39%(Base 82.41%)
70
+ - gsm8k: 86.88%(Base93.93%) with gentoken=2048
71
+ - passkey 130k+(Base 500k)
72
 
73
  ## Usage with RWKV-Infer
74
  - **RWKV-Infer** Triton based Hybrid RWKV Inference engine, can be check at: [https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F](https://github.com/OpenMOSE/RWKV-Infer/wiki/How-to-Running-RWKV-hxa079-models%3F)