GGUF

Seed-X-PPO-7B — GGUF quantized variants

Important note on quality

This model, by the author's recommendation, should use beam search sampling, which isn't by default supported in llama.cpp and requires a custom sampling implementation. In compression to GGUF (8-2bit) model with knowledge already priorly condensed that much will really take a big hit in quality, especially if you use it with greedy sampling, so don't exect great results with it, without putting some effort into it. I'm releasing it for reasearch and further improvement purposes. Do expect giberish outputs at lower quants especially with greedy sampling.

NOTE: If you'd like to quantize it yourself, you'll notice some necessary files (i.e. tokenizer_config.json) are missing from the official repo, you can still grab them from the official quants. ;)

Original model: ByteDance-Seed/Seed-X-PPO-7B

This upload contains Seed-X-PPO-7B (Mistral-based multilingual translation across 28 languages: Arabic (ar), Czech (cs), Danish (da), German (de), English (en), Spanish (es), Finnish (fi), French (fr), Croatian (hr), Hungarian (hu), Indonesian (id), Italian (it), Japanese (ja), Korean (ko), Malay (ms), Norwegian Bokmål (nb), Dutch (nl), Norwegian (no), Polish (pl), Portuguese (pt), Romanian (ro), Russian (ru), Swedish (sv), Thai (th), Turkish (tr), Ukrainian (uk), Vietnamese (vi), and Chinese (zh)), converted to GGUF.

Files & sizes

Made with llama.cpp.

File Size (MB)
Seed-X-PPO-7B.Q2_K.gguf 2743.19
Seed-X-PPO-7B.Q4_K_M.gguf 4346.45
Seed-X-PPO-7B.Q8_0.gguf 7616.18

Prompt format

Translate the following English sentence into Polish:
May the force be with you <pl>

Notice

  • Add a target-language tag at the end of the prompt (e.g., , ). This is required because PPO training used language tags.
  • No chat template: do not call tokenizer.apply_chat_template and avoid multi-turn chat formatting.
  • The model is specialized for multilingual translation.

License

OpenMDW — see the upstream repository for details.

Citation

@misc{cheng2025seedxbuildingstrongmultilingual, title={Seed-X: Building Strong Multilingual Translation LLM with 7B Parameters}, author={Shanbo Cheng and Yu Bao and Qian Cao and Luyang Huang and Liyan Kang and Zhicheng Liu and Yu Lu and Wenhao Zhu and Jingwen Chen and Zhichao Huang and Tao Li and Yifu Li and Huiying Lin and Sitong Liu and Ningxin Peng and Shuaijie She and Lu Xu and Nuo Xu and Sen Yang and Runsheng Yu and Yiming Yu and Liehao Zou and Hang Li and Lu Lu and Yuxuan Wang and Yonghui Wu}, year={2025}, eprint={2507.13618}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2507.13618} }

Downloads last month
16
GGUF
Model size
7.51B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

2-bit

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for laelhalawani/Seed-X-PPO-7B-GGUF

Quantized
(12)
this model

Datasets used to train laelhalawani/Seed-X-PPO-7B-GGUF