Unsloth Dynamic 2.0 Quants Collection New 2.0 version of our Dynamic GGUF + Quants. Dynamic 2.0 achieves superior accuracy & SOTA quantization performance. β’ 46 items β’ Updated 1 day ago β’ 195
view article Article Groq on Hugging Face Inference Providers π₯ By sbrandeis and 4 others β’ Jun 16 β’ 42
QwenLong-L1: Towards Long-Context Large Reasoning Models with Reinforcement Learning Paper β’ 2505.17667 β’ Published May 23 β’ 89
Paper2Code: Automating Code Generation from Scientific Papers in Machine Learning Paper β’ 2504.17192 β’ Published Apr 24 β’ 114
TaoAvatar: Real-Time Lifelike Full-Body Talking Avatars for Augmented Reality via 3D Gaussian Splatting Paper β’ 2503.17032 β’ Published Mar 21 β’ 27
view article Article Open-source DeepResearch β Freeing our search agents By m-ric and 4 others β’ Feb 4 β’ 1.28k
O1 Replication Journey -- Part 3: Inference-time Scaling for Medical Reasoning Paper β’ 2501.06458 β’ Published Jan 11 β’ 32
Granite 3.1 Language Models Collection A series of language models with 128K context length trained by IBM licensed under Apache 2.0 license. β’ 9 items β’ Updated May 2 β’ 63
view article Article πΊπ¦ββ¬ LLM Comparison/Test: 25 SOTA LLMs (including QwQ) through 59 MMLU-Pro CS benchmark runs By wolfram β’ Dec 4, 2024 β’ 79
BloombergGPT: A Large Language Model for Finance Paper β’ 2303.17564 β’ Published Mar 30, 2023 β’ 26
π Ichigo v0.4 Collection The experimental family designed to train LLMs to understand sound natively. β’ 3 items β’ Updated Apr 22 β’ 8
OpenCoder Collection OpenCoder is an open and reproducible code LLM family which matches the performance of top-tier code LLMs. β’ 8 items β’ Updated Nov 23, 2024 β’ 85
Llama 3.2 3B & 1B GGUF Quants Collection Llama.cpp compatible quants for Llama 3.2 3B and 1B Instruct models. β’ 4 items β’ Updated Sep 26, 2024 β’ 46
Moshi v0.1 Release Collection MLX, Candle & PyTorch model checkpoints released as part of the Moshi release from Kyutai. Run inference via: https://github.com/kyutai-labs/moshi β’ 15 items β’ Updated Apr 18 β’ 236
Jamba 1.5 Collection The AI21 Jamba family of models are state-of-the-art, hybrid SSM-Transformer instruction following foundation models β’ 2 items β’ Updated Mar 6 β’ 87