📢 NVIDIA Releases Nemotron-CC-Math Pre-Training Dataset: A High-Quality, Web-Scale Math Corpus for Pretraining Large Language Models 22 days ago • 2
NVIDIA Releases Improved Pretraining Dataset: Preserves High Value Math & Code, and Augments with Multi-Lingual 22 days ago • 2
NVIDIA Releases 3 Million Sample Dataset for OCR, Visual Question Answering, and Captioning Tasks 29 days ago • 72
Llama-NeMoRetriever-ColEmbed: Developer-Focused Guide to NVIDIA's State-of-the-Art Text-Image Retrieval Jul 9 • 4
Nemotron-Personas: Improve AI Training With the First Synthetic Personas Dataset Aligned to Real-World Distributions Jun 10 • 16
Model Optimizer Collection A collection of generative models quantized and optimized with TensorRT Model Optimizer. • 37 items • Updated 4 days ago • 29
Deep Speech 2: End-to-End Speech Recognition in English and Mandarin Paper • 1512.02595 • Published Dec 8, 2015 • 2
MIND: Math Informed syNthetic Dialogues for Pretraining LLMs Paper • 2410.12881 • Published Oct 15, 2024 • 1
Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models Paper • 2504.03624 • Published Apr 4 • 13