Avelina Hadji-Kyriacou

Avelina

AI & ML interests

Trying to squeeze the most performance out of small language models to bring AI inference to the user, and keep personal data out of the cloud.

Recent Activity

Organizations

Social Post Explorers's profile picture Swarms's profile picture

Avelina's activity

replied to their post 6 months ago
view reply

Each attribute should be in the range zero to four, however the included labels are given as is by the reward model which means some values may be outside this range (although only slightly) so it is recommended that you clamp all attributes between zero and four.

We included the unclamped versions because you may want the exact outputs given by the reward model for some specific reason, and if we had clamped these values in the dataset you would be unable to recover them.

posted an update 6 months ago
view post
Post
2198
Hey HF. I just released a new reward modelling dataset: Avelina/UltraSteer-v0

UltraSteer-V0 is a massive collection of single- and multi-turn dialogue with fine-grained reward labels produced by Nvidia's nvidia/Llama2-13B-SteerLM-RM reward model. We have a total of 2.3M labelled sequences taken from high quality datasets with a total of 2.8M labelled turns each containing 9 attributes produced as is from the reward model.

This is still very much an early version of the dataset (but it's fully usable!) and an updated version will be on the way with a full paper.

I would really appreciate if people could take a look at the dataset and suggest any improvements (e.g. more data sources, different cleaning approaches, different label schema, etc) in the community section.
  • 2 replies
ยท
posted an update 10 months ago
view post
Post
1176
Found out my ECCV paper is getting rejected because of a LaTeX compile error :(