DAPO: An Open-Source LLM Reinforcement Learning System at Scale Paper β’ 2503.14476 β’ Published 7 days ago β’ 104
DAPO: An Open-Source LLM Reinforcement Learning System at Scale Paper β’ 2503.14476 β’ Published 7 days ago β’ 104
Phi-3 Collection Phi-3 family of small language and multi-modal models. Language models are available in short- and long-context lengths. β’ 26 items β’ Updated Jan 8 β’ 565
view article Article Cosmopedia: how to create large-scale synthetic data for pre-training Large Language Models Mar 20, 2024 β’ 82
Running 893 893 FineWeb: decanting the web for the finest text data at scale π· Generate high-quality web text data for LLM training
view post Post EVA-CLIP π¦ is the CLIP scaled to the moon! π₯ The new SotA CLIP-like model π Highlights β¨ - Performs better in linear probing- Outperforms in Zero-Shot Image-Text Retrieval- Higher zero-shot accuracy in IN-1K As usual, try it with the notebook I built for you https://colab.research.google.com/drive/1K7DdCORC3x4qyhwhuB4fT4wcfJ_BQLKw?usp=sharing#scrollTo=0ZS_lJ7SK6YsI also built a Space for you to compare the output probabilities to CLIP, seems that EVACLIP is more "sure" of it's results π merve/EVACLIPThe authors have shared 8B checkpoints open with Apache 2.0 license π and it's built on top of transformers, super easy to use! BAAI/EVA-CLIP-8BRead the paper EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters (2402.04252) π β€οΈ 9 9 + Reply