More than one training run goes into making a large language model, but developers rarely release the small models and datasets they experiment with during the development process. How do they decide what dataset to use for pretraining or which benchmarks to hill climb on? To empower open exploration of these questions, we release DataDecide—a suite of models we pretrain on 25 corpora with differing sources, deduplication, and filtering up to 100B tokens, over 14 different model sizes ranging from 4M parameters up to 1B parameters (more than 30k model checkpoints in total).
350 Models over Differences in Data in Scale
For each of our 25 datasets and 14 model sizes, we train a model linked below. Each has intermediate checkpoints (uploading after initial release), runs over 3 random seeds. All models finish training at a token to parameter ratio of 100 (e.g., 1B parameters -> 100B tokens).
Load a Model
To load a specific model with HuggingFace:
from hf_olmo import OLMoForCausalLM # pip install ai2-olmo
olmo = OLMoForCausalLM.from_pretrained("allenai/DataDecide-dolma1_7-1B", revision="step69369-seed-default")
Model Description
- Developed by: Allen Institute for AI (Ai2)
- Model type: a Transformer style autoregressive language model.
- Language(s) (NLP): English
- License: The code and model are released under Apache 2.0.
- Contact: Technical inquiries:
[email protected]
. Press:[email protected]
Model Sources
- Repository: https://github.com/allenai/DataDecide
- Paper: https:/allenai.org/paper/datadecide
- Data: https://huggingface.co/datasets/allenai/datadecide
Data
Source / Recipe | Description |
---|---|
Dolma1.7 Original, No code, No math/code, No Reddit, No Flan | A 2.3T-token corpus (Dolma; 1.7 Soldaini et al., 2024) sampling common LM sources for open research. We ablate code, math/code, Reddit, or Flan subsets. |
Dolma1.6++ Original | Dolma 1.6 plus additional sources from Dolma 1.7: RedPajama’s arxiv subset, openwebmath, algebraic stack, flan, starcoder, falcon. |
C4 Original | The C4 dataset (Raffel et al., 2019) as prepared in Dolma 1.7, heuristically filtered from the April 2019 Common Crawl. |
FineWeb-Pro Original | The FineWeb Pro corpus (Zhou et al., 2024), featuring model-driven data cleaning on FineWeb. |
FineWeb-Edu Original | The deduplicated FineWeb-Edu subset of SmoLLM-Corpus (Ben Allal et al., 2024), focused on educational web pages. |
Falcon Original | The Falcon RefinedWeb corpus (Penedo et al., 2023) in Dolma 1.7, derived from Common Crawl through June 2023 and more aggressively filtered/deduplicated than C4. |
Falcon+CC Original, QC 10%, QC 20%, QC Orig 10%, QC Tulu 10% | Falcon and Dolma 1.7’s Common Crawl. We quality filter to top 10% or 20% documents with reproduced or original Li et al. (2024) filter or retrain filter on pre-release version of Tulu-v3 (Lambert et al., 2024). |
DCLM-Baseline Original, QC 7% FW2, QC 7% FW3, QC FW 10%, QC 10%, QC 20% | A SOTA Common Crawl corpus using best ablated deduplication, cleaning heuristics, and quality filter. We quality filter to top 7% of DCLM classified documents and further take 2+ or 3+ scores with FineWeb-edu classifier; or filter to top 3% or 10% with FineWeb-edu classifier; or take top 10% or 20% with reproduced DCLM classifier. |
λ% DCLM-Baseline + 1 – λ% Dolma1.7 | Fractional combinations of Dolma1.7 and DCLM-Baseline mixing different proportions of the two datasets for λ ∈ {25%, 50%, 75%}. |
Evaluation
We evaluate all checkpoints over OLMES suite of 10 multiple choice question answering benchmarks (Gu et al., 2024):
- MMLU (Hendrycks et al., 2021)
- HellaSwag (Zellers et al., 2019)
- ARC-Challenge (Clark et al., 2018)
- ARC-Easy (Clark et al., 2018)
- PIQA (Bisk et al., 2020)
- CommonsenseQA (Talmor et al., 2019)
- Social IQa (Sap et al., 2019)
- OpenBookQA (Mihaylov et al., 2018)
- BoolQ (Clark et al., 2019)
- Winogrande (Sakaguchi et al., 2020)
We release all these evaluations:
- for task-level metric results: https://huggingface.co/datasets/allenai/DataDecide-eval-results
- for instance-level results: https://huggingface.co/datasets/allenai/DataDecide-eval-instances
Hyperparameters
Name | Batch Size | Hidden Dim. | LR | Model size | Heads | Layers | Training steps | Tokens trained |
---|---|---|---|---|---|---|---|---|
4M | 32 | 64 | 1.4e-02 | 3.7M | 8 | 8 | 5,725 | 0.4B |
6M | 32 | 96 | 1.2e-02 | 6.0M | 8 | 8 | 9,182 | 0.6B |
8M | 32 | 128 | 1.1e-02 | 8.5M | 8 | 8 | 13,039 | 0.9B |
10M | 32 | 144 | 1.0e-02 | 9.9M | 8 | 8 | 15,117 | 1.0B |
14M | 32 | 192 | 9.2e-03 | 14.4M | 8 | 8 | 21,953 | 1.4B |
16M | 32 | 208 | 8.9e-03 | 16.0M | 8 | 8 | 24,432 | 1.6B |
20M | 64 | 192 | 8.4e-03 | 19.1M | 8 | 16 | 14,584 | 1.9B |
60M | 96 | 384 | 5.8e-03 | 57.1M | 12 | 16 | 29,042 | 5.7B |
90M | 160 | 528 | 4.9e-03 | 97.9M | 12 | 16 | 29,901 | 9.8B |
150M | 192 | 768 | 4.2e-03 | 151.9M | 12 | 12 | 38,157 | 15.0B |
300M | 320 | 1,024 | 3.3e-03 | 320.0M | 16 | 16 | 45,787 | 30.0B |
530M | 448 | 1,344 | 2.8e-03 | 530.1M | 16 | 16 | 57,786 | 53.0B |
750M | 576 | 1,536 | 2.5e-03 | 681.3M | 16 | 16 | 63,589 | 75.0B |
1B | 704 | 2,048 | 2.1e-03 | 1176.8M | 16 | 16 | 69,369 | 100.0B |
Bias, Risks, and Limitations
Like any base or fine-tuned language model, AI can be prompted by users to generate harmful and sensitive content. Such content may also be produced unintentionally, especially in cases involving bias, so we recommend that users consider the risks when applying this technology. Additionally, many statements from any LLM are often inaccurate, so facts should be verified.
Citation
BibTeX:
@article{MagnussonDataDecide2025,
title={{DataDecide: How to Predict Best Pretraining Data with Small Experiments}},
author={Ian Magnusson and Nguyen Tai and Ben Bogin and David Heineman and Jena Hwang and Luca Soldaini and Akshita Bhagia and Jiacheng Liu and Dirk Groeneveld and Oyvind Tafjord and Noah A. Smith and Pang Wei Koh and Jesse Dodge},
year={2025},
journal={arXiv preprint},
}
Model Card Contact
For errors in this model card, contact [email protected]
- Downloads last month
- 0