hf-inference

Enterprise
non-profit
Activity Feed

AI & ML interests

None defined yet.

Recent Activity

reach-vbΒ  updated a collection 14 days ago
deployed-models
reach-vbΒ  updated a collection 14 days ago
deployed-models
reach-vbΒ  updated a collection 14 days ago
deployed-models
View all activity

hf-inference's activity

jeffboudierΒ 
posted an update about 6 hours ago
jeffboudierΒ 
posted an update 5 days ago
view post
Post
404
Wrapping up a week of shipping and announcements with Dell Enterprise Hub now featuring AI Applications, on-device models for AI PCs, a new CLI and Python SDK... all you need for building AI on premises!

Blog post has all the details: https://huggingface.co/blog/dell-ai-applications
celinahΒ 
posted an update 6 days ago
view post
Post
2118
✨ Today we’re releasing Tiny Agents in Python β€” an MCP-powered Agent in ~70 lines of code 🐍

Inspired by Tiny Agents in JS from @julien-c , we ported the idea to Python and integrated it directly into huggingface_hub β€” with a built-in MCP Client and a Tiny Agents CLI.

TL;DR: With MCP (Model Context Protocol), you can expose tools like web search or image generation and connect them directly to LLMs. It’s simple β€” and surprisingly powerful.

pip install "huggingface_hub[mcp]>=0.32.0"

We wrote a blog post where we show how to run Tiny Agents, and dive deeper into how they work and how to build your own.
πŸ‘‰ https://huggingface.co/blog/python-tiny-agents

  • 1 reply
Β·
reach-vbΒ 
posted an update 9 days ago
view post
Post
3284
hey hey @mradermacher - VB from Hugging Face here, we'd love to onboard you over to our optimised xet backend! πŸ’₯

as you know we're in the process of upgrading our storage backend to xet (which helps us scale and offer blazingly fast upload/ download speeds too): https://huggingface.co/blog/xet-on-the-hub and now that we are certain that the backend can scale with even big models like Llama 4/ Qwen 3 - we;re moving to the next phase of inviting impactful orgs and users on the hub over as you are a big part of the open source ML community - we would love to onboard you next and create some excitement about it in the community too!

in terms of actual steps - it should be as simple as one of the org admins to join hf.co/join/xet - we'll take care of the rest.

p.s. you'd need to have a the latest hf_xet version of huggingface_hub lib but everything else should be the same: https://huggingface.co/docs/hub/storage-backends#using-xet-storage

p.p.s. this is fully backwards compatible so everything will work as it should! πŸ€—
Β·