Update model card for Pref-GRPO: add pipeline tag, library, and correct paper/project/code links

#1
by nielsr HF Staff - opened

This PR improves the model card for UnifiedReward-qwen-7b, clarifying its role as a reward model in the context of the Pref-GRPO: Pairwise Preference Reward-based GRPO for Stable Text-to-Image Reinforcement Learning paper.

Specifically, it addresses the following points:

  • Corrected Paper Link: Updated the paper link in the content to https://huggingface.co/papers/2508.20751 as per the prompt's instruction, which describes the Pref-GRPO framework in which this reward model is used.
  • Updated Project Page Link: Pointed the project page to the more specific https://codegoat24.github.io/UnifiedReward/Pref-GRPO.
  • Added GitHub Link: Included a direct link to the Pref-GRPO code at https://github.com/CodeGoat24/Pref-GRPO.
  • Added Pipeline Tag: Set pipeline_tag: image-text-to-text to accurately reflect the model's input (image, text) and output (text evaluation). This helps users discover the model more easily.
  • Added Library Name: Specified library_name: transformers as the model uses the Hugging Face Transformers library for inference, enabling the "Use in Transformers" widget.
  • Made Sample Usage Runnable: Added import requests to the Python snippet to ensure the example code is runnable out-of-the-box.
  • Updated Citation: Replaced the outdated BibTeX entry with the correct one for the Pref-GRPO paper.
  • Updated Quick Start GitHub Link: Corrected the GitHub link in the Quick Start section to point to the Pref-GRPO repository.

These changes enhance the discoverability, clarity, and completeness of the model card.

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment