Hugging Face Science

company
Activity Feed

AI & ML interests

None defined yet.

Recent Activity

science's activity

fdaudensย 
posted an update 3 days ago
view post
Post
1913
๐ŸŽฅ Just tested Stability AI's Stable Virtual Camera - it turns a single photo into dynamic video with AI-powered camera movements! From static meeting room to cinematic sweeps. ๐Ÿš€

Try it out: stabilityai/stable-virtual-camera
fdaudensย 
posted an update 5 days ago
view post
Post
1812
๐Ÿ”Š Meet Orpheus: A breakthrough open-source TTS model that matches human-level speech with empathy & emotion.
- Available in 4 sizes (150M-3B parameters)
- delivers ultra-fast streaming
- zero-shot voice cloning.
- Apache 2.0 license

canopylabs/orpheus-tts-67d9ea3f6c05a941c06ad9d2
  • 1 reply
ยท
fdaudensย 
posted an update 6 days ago
view post
Post
2202
Want to build useful newsroom tools with AI? Weโ€™re launching a Hugging Face x Journalism Slack channel where journalists turn AI concepts into real newsroom solutions.

Inside the community:
โœ… Build open-source AI tools for journalism
โœ… Get direct help from the community
โœ… Stay updated on new models and datasets
โœ… Learn from other journalistsโ€™ experiments and builds

The goal? Go from โ€œI read about AIโ€ to โ€œI built an AI tool that supercharged my newsroom.โ€ โ€”no more learning in isolation.

Join us! https://join.slack.com/t/journalistson-tnd8294/shared_invite/zt-30vsmhk4w-dZpeMOoxdhCvfNsqtspPUQ (Please make sure to use a clear identityโ€”no teddybear85, for example ๐Ÿ˜‰)

(If you know people who might be interested, tag them below! The more minds we bring in, the better the tools we build.)

fdaudensย 
posted an update 7 days ago
m-ricย 
posted an update 9 days ago
view post
Post
4368
smolagents now support vLLM! ๐Ÿฅณ

As one of the most popular local inference solutions, the community had been asking us to integrate vLLM: after a heavy refactoring of our LLM classes, we've just released smolagents 1.11.0, with a brand new VLLMModel class.

Go try it and tell us what you think!

https://github.com/huggingface/smolagents/blob/45b2c86857b7f7657daaa74e4d17d347e9e2c4a4/src/smolagents/models.py#L497
fdaudensย 
posted an update 11 days ago
view post
Post
798
๐Ÿคฏ Gemma 3's image analysis blew me away!

Tested 2 ways to extract airplane registration numbers from photos with 12B model:

1๏ธโƒฃ Gradio app w/API link (underrated feature IMO) + ZeroGPU infra on Hugging Face in Google Colab. Fast & free.

2๏ธโƒฃ LMStudio + local processing (100% private). Running this powerhouse on a MacBook w/16GB RAM is wild! ๐Ÿš€

Colab: https://colab.research.google.com/drive/1YmmaP0IDEu98CLDppAAK9kbQZ7lFnLZ1?usp=sharing
fdaudensย 
posted an update 12 days ago
view post
Post
1378
Ever wanted 45 min with one of AIโ€™s most fascinating minds? Was with @thomwolf at HumanX Vegas. Sharing my notes of his Q&A with the pressโ€”completely changed how I think about AIโ€™s future:

1๏ธโƒฃ The next wave of successful AI companies wonโ€™t be defined by who has the best model but by who builds the most useful real-world solutions. "We all have engines in our cars, but thatโ€™s rarely the only reason we buy one. We expect it to work well, and thatโ€™s enough. LLMs will be the same."

2๏ธโƒฃ Big players are pivoting: "Closed-source companiesโ€”OpenAI being the firstโ€”have largely shifted from LLM announcements to product announcements."

3๏ธโƒฃ Open source is changing everything: "DeepSeek was open source AIโ€™s ChatGPT moment. Basically, everyone outside the bubble realized you can get a model for freeโ€”and itโ€™s just as good as the paid ones."

4๏ธโƒฃ Product innovation is being democratized: Take Manus, for exampleโ€”they built a product on top of Anthropicโ€™s models thatโ€™s "actually better than Anthropicโ€™s own product for now, in terms of agents." This proves that anyone can build great products with existing models.

Weโ€™re entering a "multi-LLM world," where models are becoming commoditized, and all the tools to build are readily availableโ€”just look at the flurry of daily new releases on Hugging Face.

Thom's comparison to the internet era is spot-on: "In the beginning you made a lot of money by making websites... but nowadays the huge internet companies are not the companies that built websites. Like Airbnb, Uber, Facebook, they just use the internet as a medium to make something for real life use cases."

Love to hear your thoughts on this shift!
  • 1 reply
ยท
thomwolfย 
posted an update 12 days ago
view post
Post
2530
We've kept pushing our Open-R1 project, an open initiative to replicate and extend the techniques behind DeepSeek-R1.

And even we were mind-blown by the results we got with this latest model we're releasing: โšก๏ธOlympicCoder ( open-r1/OlympicCoder-7B and open-r1/OlympicCoder-32B)

It's beating Claude 3.7 on (competitive) programming โ€“a domain Anthropic has been historically really strong atโ€“ and it's getting close to o1-mini/R1 on olympiad level coding with just 7B parameters!

And the best part is that we're open-sourcing all about its training dataset, the new IOI benchmark, and more in our Open-R1 progress report #3: https://huggingface.co/blog/open-r1/update-3

Datasets are are releasing:
- open-r1/codeforces
- open-r1/codeforces-cots
- open-r1/ioi
- open-r1/ioi-test-cases
- open-r1/ioi-sample-solutions
- open-r1/ioi-cots
- open-r1/ioi-2024-model-solutions
eliebakย 
posted an update 12 days ago
view post
Post
1497
Google just dropped an exciting technical report for the brand-new Gemma3 model! ๐Ÿš€ Here are my personal notes highlighting the most intriguing architectural innovations, design choices, and insights from this release:

1) Architecture choices:
> No more softcaping, replace by QK-Norm
> Both Pre AND Post Norm
> Wider MLP than Qwen2.5, ~ same depth
> SWA with 5:1 and 1024 (very small and cool ablation on the paper!)
> No MLA to save KV cache, SWA do the job!

2) Long context
> Only increase the rope in the global layer (to 1M)
> Confirmation that it's harder to do long context for smol models, no 128k for the 1B
> Pretrained with 32k context? seems very high
> No yarn nor llama3 like rope extension

3) Distillation
> Only keep te first 256 logits for the teacher
> Ablation on the teacher gap (tl;dr you need some "patience" to see that using a small teacher is better)
> On policy distillation yeahh (by
@agarwl_
et al), not sure if the teacher gap behave the same here, curious if someone have more info?

4) Others
> Checkpoint with QAT, that's very cool
> RL using improve version of BOND, WARM/WARP good excuse to look at
@ramealexandre
papers
> Only use Zero3, no TP/PP if i understand correctly ?
> Training budget relatively similar than gemma2
  • 1 reply
ยท
clefourrierย 
posted an update 12 days ago
view post
Post
1838
Gemma3 family is out! Reading the tech report, and this section was really interesting to me from a methods/scientific fairness pov.

Instead of doing over-hyped comparisons, they clearly state that **results are reported in a setup which is advantageous to their models**.
(Which everybody does, but people usually don't say)

For a tech report, it makes a lot of sense to report model performance when used optimally!
On leaderboards on the other hand, comparison will be apples to apples, but in a potentially unoptimal way for a given model family (like some user interact sub-optimally with models)

Also contains a cool section (6) on training data memorization rate too! Important to see if your model will output the training data it has seen as such: always an issue for privacy/copyright/... but also very much for evaluation!

Because if your model knows its evals by heart, you're not testing for generalization.
lewtunย 
posted an update 13 days ago
view post
Post
2071
Introducing OlympicCoder: a series of open reasoning models that can solve olympiad-level programming problems ๐Ÿง‘โ€๐Ÿ’ป

- 7B open-r1/OlympicCoder-7B
- 32B open-r1/OlympicCoder-32B

We find that OlympicCoder models outperform Claude 3.7 Sonnet, as well as others over 100x larger ๐Ÿ’ช

Together with the models, we are releasing:

๐Ÿ“ŠCodeForces-CoTs: new dataset of code problems from the most popular competitive coding platform, with R1 traces in C++ and Python open-r1/codeforces-cots

๐Ÿ† IOI'2024: a new benchmark of VERY hard programming problems where even frontier models struggle to match human performance open-r1/ioi

For links to the models and datasets, check out our latest progress report from Open R1: https://huggingface.co/blog/open-r1/update-3
  • 1 reply
ยท
fdaudensย 
posted an update 13 days ago
view post
Post
1765
๐Ÿ”ฅThe Open R1 team just dropped OlympicCoder and it's wild:

- 7B model outperforms Claude 3.7 Sonnet on IOI benchmark (yes, 7B!!)
- 32B crushes all open-weight models tested, even those 100x larger ๐Ÿคฏ

Open-sourcing the future of code reasoning! ๐Ÿš€

Check it out https://huggingface.co/blog/open-r1/update-3
julien-cย 
posted an update 13 days ago
view post
Post
2505
Important notice ๐Ÿšจ

For Inference Providers who have built support for our Billing API (currently: Fal, Novita, HF-Inference โ€“ with more coming soon), we've started enabling Pay as you go (=PAYG)

What this means is that you can use those Inference Providers beyond the free included credits, and they're charged to your HF account.

You can see it on this view: any provider that does not have a "Billing disabled" badge, is PAYG-compatible.
BrigitteTousiย 
posted an update 13 days ago
m-ricย 
posted an update 14 days ago
view post
Post
961
Our new Agentic leaderboard is now live!๐Ÿ’ฅ

If you ever asked which LLM is best for powering agents, we've just made a leaderboard that ranks them all! Built with @albertvillanova , this ranks LLMs powering a smolagents CodeAgent on subsets of various benchmarks. โœ…

๐Ÿ† GPT-4.5 comes on top, even beating reasoning models like DeepSeek-R1 or o1. And Claude-3.7-Sonnet is a close second!

The leaderboard also allows you to show the scores of vanilla LLMs (without any agentic setup) on the same benchmarks: this shows the huge improvements brought by agentic setups. ๐Ÿ’ช

(Note that results will be added manually, so the leaderboard might not always have the latest LLMs)
  • 1 reply
ยท
BrigitteTousiย 
posted an update 14 days ago
view post
Post
3683
Regardless of X being down or not, so glad I can rely on HF Posts for AI news โค๏ธ๐Ÿค—
  • 1 reply
ยท
fdaudensย 
posted an update 15 days ago
view post
Post
5714
Honored to be named among their 12 pioneers and power players in the news industry in the 2025 Tech Trends Report from Future Today Strategy Group.

Incredible group to be part of - each person is doing groundbreaking work at the intersection of AI and journalism. Worth following them all: they're consistently sharing practical insights on building the future of news.

Take the time to read this report, it's packed with insights as always. The news & information section's #1 insight hits hard: "The most substantive economic impact of AI to date has been licensing payouts for a handful of big publishers. The competition will start shifting in the year ahead to separate AI 'haves' that have positioned themselves to grow from the 'have-nots.'"

This AI-driven divide is something I've been really concerned about. Now is the time to build more than ever!

๐Ÿ‘‰ Full report here: https://ftsg.com/wp-content/uploads/2025/03/FTSG_2025_TR_FINAL_LINKED.pdf
  • 2 replies
ยท