Dhruv's picture
3 1

Dhruv PRO

dhruv3006
·

AI & ML interests

None yet

Recent Activity

reacted to burtenshaw's post with 🔥 about 2 hours ago
We’re launching a FREE and CERTIFIED course on Agents! We're thrilled to announce the launch of the Hugging Face Agents course on Learn! This interactive, certified course will guide you through building and deploying your own AI agents. Here's what you'll learn: - Understanding Agents: We'll break down the fundamentals of AI agents, showing you how they use LLMs to perceive their environment (observations), reason about it (thoughts), and take actions. Think of a smart assistant that can book appointments, answer emails, or even write code based on your instructions. - Building with Frameworks: You'll dive into popular agent frameworks like LangChain, LlamaIndex and smolagents. These tools provide the building blocks for creating complex agent behaviors. - Real-World Applications: See how agents are used in practice, from automating SQL queries to generating code and summarizing complex documents. - Certification: Earn a certification by completing the course modules, implementing a use case, and passing a benchmark assessment. This proves your skills in building and deploying AI agents. Audience This course is designed for anyone interested in the future of AI. Whether you're a developer, data scientist, or simply curious about AI, this course will equip you with the knowledge and skills to build your own intelligent agents. Enroll today and start building the next generation of AI agent applications! https://bit.ly/hf-learn-agents
reacted to singhsidhukuldeep's post with 👍 12 days ago
Groundbreaking Research Alert: Rethinking RAG with Cache-Augmented Generation (CAG) Researchers from National Chengchi University and Academia Sinica have introduced a paradigm-shifting approach that challenges the conventional wisdom of Retrieval-Augmented Generation (RAG). Instead of the traditional retrieve-then-generate pipeline, their innovative Cache-Augmented Generation (CAG) framework preloads documents and precomputes key-value caches, eliminating the need for real-time retrieval during inference. Technical Deep Dive: - CAG preloads external knowledge and precomputes KV caches, storing them for future use - The system processes documents only once, regardless of subsequent query volume - During inference, it loads the precomputed cache alongside user queries, enabling rapid response generation - The cache reset mechanism allows efficient handling of multiple inference sessions through strategic token truncation Performance Highlights: - Achieved superior BERTScore metrics compared to both sparse and dense retrieval RAG systems - Demonstrated up to 40x faster generation times compared to traditional approaches - Particularly effective with both SQuAD and HotPotQA datasets, showing robust performance across different knowledge tasks Why This Matters: The approach significantly reduces system complexity, eliminates retrieval latency, and mitigates common RAG pipeline errors. As LLMs continue evolving with expanded context windows, this methodology becomes increasingly relevant for knowledge-intensive applications.
View all activity

Organizations

MLX Community's profile picture Hugging Face Discord Community's profile picture

models

None public yet

datasets

None public yet