Seed-X-PPO-7B — GGUF quantized variants
Important note on quality
This model, by the author's recommendation, should use beam search sampling, which isn't by default supported in llama.cpp and requires a custom sampling implementation. In compression to GGUF (8-2bit) model with knowledge already priorly condensed that much will really take a big hit in quality, especially if you use it with greedy sampling, so don't exect great results with it, without putting some effort into it. I'm releasing it for reasearch and further improvement purposes. Do expect giberish outputs at lower quants especially with greedy sampling.
NOTE: If you'd like to quantize it yourself, you'll notice some necessary files (i.e. tokenizer_config.json
) are missing from the official repo, you can still grab them from the official quants. ;)
Original model: ByteDance-Seed/Seed-X-PPO-7B
This upload contains Seed-X-PPO-7B (Mistral-based multilingual translation across 28 languages: Arabic (ar), Czech (cs), Danish (da), German (de), English (en), Spanish (es), Finnish (fi), French (fr), Croatian (hr), Hungarian (hu), Indonesian (id), Italian (it), Japanese (ja), Korean (ko), Malay (ms), Norwegian Bokmål (nb), Dutch (nl), Norwegian (no), Polish (pl), Portuguese (pt), Romanian (ro), Russian (ru), Swedish (sv), Thai (th), Turkish (tr), Ukrainian (uk), Vietnamese (vi), and Chinese (zh)), converted to GGUF.
Files & sizes
Made with llama.cpp.
File | Size (MB) |
---|---|
Seed-X-PPO-7B.Q2_K.gguf | 2743.19 |
Seed-X-PPO-7B.Q4_K_M.gguf | 4346.45 |
Seed-X-PPO-7B.Q8_0.gguf | 7616.18 |
Prompt format
Translate the following English sentence into Polish:
May the force be with you <pl>
Notice
- Add a target-language tag at the end of the prompt (e.g., , ). This is required because PPO training used language tags.
- No chat template: do not call tokenizer.apply_chat_template and avoid multi-turn chat formatting.
- The model is specialized for multilingual translation.
License
OpenMDW — see the upstream repository for details.
Citation
@misc{cheng2025seedxbuildingstrongmultilingual, title={Seed-X: Building Strong Multilingual Translation LLM with 7B Parameters}, author={Shanbo Cheng and Yu Bao and Qian Cao and Luyang Huang and Liyan Kang and Zhicheng Liu and Yu Lu and Wenhao Zhu and Jingwen Chen and Zhichao Huang and Tao Li and Yifu Li and Huiying Lin and Sitong Liu and Ningxin Peng and Shuaijie She and Lu Xu and Nuo Xu and Sen Yang and Runsheng Yu and Yiming Yu and Liehao Zou and Hang Li and Lu Lu and Yuxuan Wang and Yonghui Wu}, year={2025}, eprint={2507.13618}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2507.13618} }
- Downloads last month
- 16
2-bit
4-bit
8-bit
Model tree for laelhalawani/Seed-X-PPO-7B-GGUF
Base model
ByteDance-Seed/Seed-X-PPO-7B