Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

Severian 
posted an update 2 days ago
view post
Post
1898
I couldn't watch innocent people get their rights trampled anymore. So I built something to help.

Stories of families torn apart, U.S. citizens detained for hours, people arrested just for speaking Spanish. This isn't the America I believe in.

Instead of doom-scrolling, I spent a few days building FIREWATCH - a free civil rights protection app.

What it does:
• Real-time ICE raid alerts
• Know Your Rights education in 10+ languages
• Secure evidence recording
• Emergency panic button
• Legal hotlines and resources
• 100% private, no tracking

The catch? There isn't one. You just need a free Google API key that stays on your device. Works completely offline.

https://firewatch-ice.vercel.app/

I built this because everyone deserves constitutional protection. The 4th Amendment doesn't have an asterisk.

If this helps one family stay safe, every sleepless night was worth it.

Please share with anyone who needs it.

Stay safe.
  • 1 reply
·
jasoncorkill 
posted an update 1 day ago
view post
Post
2251
"Why did the bee get married?"

"Because he found his honey!"

This was the "funniest" joke out of 10'000 jokes we generated with LLMs. With 68% of respondents rating it as "funny".

Original jokes are particularly hard for LLMs, as jokes are very nuanced and a lot of context is needed to understand if something is "funny". Something that can only reliably be measured using humans.

LLMs are not equally good at generating jokes in every language. Generated English jokes turned out to be way funnier than the Japanese ones. 46% of English-speaking voters on average found the generated joke funny. The same statistic for other languages:

Vietnamese: 44%
Portuguese: 40%
Arabic: 37%
Japanese: 28%

There is not much variance in generation quality among models for any fixed language. But still Claude Sonnet 4 slightly outperforms others in Vietnamese, Arabic and Japanese and Gemini 2.5 Flash in Portuguese and English

We have release the 1 Million (!) native speaker ratings and the 10'000 jokes as a dataset for anyone to use:
Rapidata/multilingual-llm-jokes-4o-claude-gemini
·
fdaudens 
posted an update 1 day ago
view post
Post
1324
You might not have heard of Moonshot AI — but within 24 hours, their new model Kimi K2 shot to the top of Hugging Face’s trending leaderboard.

So… who are they, and why does it matter?

Had a lot of fun co-writing this blog post with @xianbao , with key insights translated from Chinese, to unpack how this startup built a model that outperforms GPT-4.1, Claude Opus, and DeepSeek V3 on several major benchmarks.

🧵 A few standout facts:

1. From zero to $3.3B in 18 months:
Founded in March 2023, Moonshot is now backed by Alibaba, Tencent, Meituan, and HongShan.

2. A CEO who thinks from the end:
Yang Zhilin (31) previously worked at Meta AI, Google Brain, and Carnegie Mellon. His vision? Nothing less than AGI — still a rare ambition among Chinese AI labs.

3. A trillion-parameter model that’s surprisingly efficient:
Kimi K2 uses a mixture-of-experts architecture (32B active params per inference) and dominates on coding/math benchmarks.

4. The secret weapon: Muon optimizer:
A new training method that doubles efficiency, cuts memory in half, and ran 15.5T tokens with zero failures. Big implications.

Most importantly, their move from closed to open source signals a broader shift in China’s AI scene — following Baidu’s pivot. But as Yang puts it: “Users are the only real leaderboard.”

👇 Check out the full post to explore what Kimi K2 can do, how to try it, and why it matters for the future of open-source LLMs:
https://huggingface.co/blog/fdaudens/moonshot-ai-kimi-k2-explained
danielhanchen 
posted an update 3 days ago
jsulz 
posted an update 1 day ago
view post
Post
1354
We've moved over 20PB from Git LFS to Xet on the Hub without downtime or data loss. Having things "just work" on a migration of this scale is about as good as it gets.

Now, we're migrating the rest of the Hub https://huggingface.co/blog/migrating-the-hub-to-xet

But how did we get here?

In the early days of joining Hugging Face, we made a few key design decisions:
* There would be no "hard cut-over" from Git LFS to Xet
* A Xet-enabled repository should be able to contain both Xet and LFS files
* Repository migrations from LFS to Xet can run in the background without disrupting downloads or uploads

These were largely driven by our desire to ensure the community could keep working without interruption.

We cover the infrastructure making this all go in this post, specifically:
* An integral piece of infrastructure known internally as the Git LFS Bridge
* Background content migrations that run around the clock

To skip the wait and join Xet now, sign up here https://huggingface.co/join/xet
YerbaPage 
posted an update 2 days ago
view post
Post
1301
Is your code written by a human or an AI? 🤖

With the rise of AI coding assistants, this question is more critical than ever. Our new tool, DetectCodeGPT, effectively identifies AI-generated code, outperforming SOTA methods with a 7.6% increase in AUC!

How? By analyzing unique stylistic and syntactic patterns in code, not just the text.

👇 Explore more:
Paper (ICSE 2025): https://arxiv.org/html/2401.06461v2
Code: https://github.com/YerbaPage/DetectCodeGPT

#AI #Programming #DeveloperTools #LLM #GenAI
merve 
posted an update 3 days ago
view post
Post
2237
past week had huuuge releases 💗
here's our picks 🔥 find more models, datasets, demos here merve/releases-july-11-68750452c358c98b0fa663f7

> moonshotai/Kimi-K2-Instruct is the new sota LLM with 1T total 32B active parameters 🤯

> HuggingFaceTB/SmolLM3-3B is the new best LM for it's size, offers thinking mode 💭 as well as the dataset HuggingFaceTB/smoltalk2

> Alibaba-NLP/WebSailor-3B is the new agentic LLM for complex browsing

> Google DeepMind released medical vision LMs with an agentic doctor-patient app google/medgemma-release-680aade845f90bec6a3f60c4

> fal released a LoRA to improve details on face images fal/Realism-Detailer-Kontext-Dev-LoRA
SohanVichat 
posted an update 3 days ago
view post
Post
1221
Hey fellow developers and AI enthusiasts,
I am with the team at TeraVera, and we have just launched something we are really excited about — the TeraVera Secure API for AI solutions.
TeraVera API helps protect your data, prevents AI hallucinations, and ensures the integrity of your AI responses. It’s designed to work seamlessly across major platforms like OpenAI/Azure, Google/Gemini, and AWS/Anthropic.
Don’t hesitate to request access to the TeraVera Secure API and ensure your data is never leaked into the AI model. Grab your API key here:
🔗 https://www.teravera.com/api-access-form/
You can also check out more about what we do on our site:
🌐 teravera.com
Read: https://huggingface.co/blog/TeraVera/teravera
We would love to hear your thoughts — feedback, suggestions, or any questions are more than welcome!
Cheers,
Sohan
  • 2 replies
·
GeorgiaArm 
posted an update 1 day ago
view post
Post
1938
Join us in Austin tomorrow for AI Camp’s monthly meetup.
Arm’s Zach Lasiuk and Geremy Cohen will dive into “From Model to Product: Right-Sizing Infrastructure for Real-World Use Cases.”
RSVP here 👉 https://www.aicamp.ai/event/eventdetails/W2025071616
MonsterMMORPG 
posted an update 2 days ago
view post
Post
2614
MultiTalk Levelled Up - Way Better Animation Compared to Before with New Workflows - Image to Video > https://youtu.be/wgCtUeog41g

MultiTalk is greatly upgraded. After doing more than 1 day more research with MultiTalk by using 8x A6000 48 GB GPUs, I have significantly improved the MultiTalk workflows and now I am sharing 4 different category workflows with you. VRAM usages and speeds are same but just better quality and animation. Moreover I am introducing a new app which is image and video comparison sliders. Ultra fast and lightweight. Runs as a html app and no GPU is required.

https://youtu.be/wgCtUeog41g

MultiTalk Full Tutorial With 1-Click Installer - Make Talking and Singing Videos From Static Images > https://youtu.be/8cMIwS9qo4M

By using MeiGen MultiTalk you can generate amazing fully animated real-like videos from given audio input. Not only talking but also animating the body movements is possible. In this video I will show you how to install ComfyUI on Windows and MultiTalk bundle and workflows we prepared with 1-click. Then I will show how to very easily generated amazing videos from these installed workflows. Moreover, I will show our favorite cloud private GPU provider Massed Compute. How to install same there and use it properly. Finally I will show everything on RunPod as well. So whether you are GPU poor or have good GPU, this tutorial covers everything.

https://youtu.be/8cMIwS9qo4M

  • 1 reply
·