Skip to content

allenai/DataDecide

Repository files navigation

image/png

GitHub License Paper URL

More than one training run goes into making a large language model, but developers rarely release the small models and datasets they experiment with during the development process. How do they decide what dataset to use for pretraining or which benchmarks to hill climb on? To empower open exploration of these questions, we release DataDecide—a suite of models we pretrain on 25 corpora with differing sources, deduplication, and filtering up to 100B tokens, over 14 different model sizes ranging from 4M parameters up to 1B parameters (more than 30k model checkpoints in total).

350 Models over Differences in Data in Scale

For each of our 25 datasets and 14 model sizes, we train a model linked below. Each has intermediate checkpoints (uploading after initial release), runs over 3 random seeds. All models finish training at a token to parameter ratio of 100 (e.g., 1B parameters -> 100B tokens).

Dolma1.7 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Dolma1.7 (no code) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Dolma1.7 (no math, code) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Dolma1.7 (no Reddit) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Dolma1.7 (no Flan) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Dolma1.6++ 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
C4 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
FineWeb-Pro 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
FineWeb-Edu 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon+CC 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon+CC (QC 10%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon+CC (QC 20%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon+CC (QC Orig 10%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
Falcon+CC (QC Tulu 10%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC 7%, FW2) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC 7%, FW3) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC FW 3%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC FW 10%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC 10%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline (QC 20%) 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline 25% / Dolma 75% 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline 50% / Dolma 50% 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B
DCLM-Baseline 75% / Dolma 25% 4M 6M 8M 10M 14M 16M 20M 60M 90M 150M 300M 530M 750M 1B

Load a Model

To load a specific model with HuggingFace:

from hf_olmo import OLMoForCausalLM  # pip install ai2-olmo

olmo = OLMoForCausalLM.from_pretrained("allenai/DataDecide-dolma1_7-1B", revision="step69369-seed-default")

Data

Released at https://huggingface.co/datasets/allenai/DataDecide-data-recipes

Source / Recipe Description
Dolma1.7 Original, No code, No math/code, No Reddit, No Flan A 2.3T-token corpus (Dolma; 1.7 Soldaini et al., 2024) sampling common LM sources for open research. We ablate code, math/code, Reddit, or Flan subsets.
Dolma1.6++ Original Dolma 1.6 plus additional sources from Dolma 1.7: RedPajama’s arxiv subset, openwebmath, algebraic stack, flan, starcoder, falcon.
C4 Original The C4 dataset (Raffel et al., 2019) as prepared in Dolma 1.7, heuristically filtered from the April 2019 Common Crawl.
FineWeb-Pro Original The FineWeb Pro corpus (Zhou et al., 2024), featuring model-driven data cleaning on FineWeb.
FineWeb-Edu Original The deduplicated FineWeb-Edu subset of SmoLLM-Corpus (Ben Allal et al., 2024), focused on educational web pages.
Falcon Original The Falcon RefinedWeb corpus (Penedo et al., 2023) in Dolma 1.7, derived from Common Crawl through June 2023 and more aggressively filtered/deduplicated than C4.
Falcon+CC Original, QC 10%, QC 20%, QC Orig 10%, QC Tulu 10% Falcon and Dolma 1.7’s Common Crawl. We quality filter to top 10% or 20% documents with reproduced or original Li et al. (2024) filter or retrain filter on pre-release version of Tulu-v3 (Lambert et al., 2024).
DCLM-Baseline Original, QC 7% FW2, QC 7% FW3, QC FW 10%, QC 10%, QC 20% A SOTA Common Crawl corpus using best ablated deduplication, cleaning heuristics, and quality filter. We quality filter to top 7% of DCLM classified documents and further take 2+ or 3+ scores with FineWeb-edu classifier; or filter to top 3% or 10% with FineWeb-edu classifier; or take top 10% or 20% with reproduced DCLM classifier.
λ% DCLM-Baseline + 1 – λ% Dolma1.7 Fractional combinations of Dolma1.7 and DCLM-Baseline mixing different proportions of the two datasets for λ ∈ {25%, 50%, 75%}.

Evaluation

We evaluate all checkpoints over OLMES suite of 10 multiple choice question answering benchmarks (Gu et al., 2024):

We release all these evaluations:

Hyperparameters

Name Batch Size Hidden Dim. LR Model size Heads Layers Training steps Tokens trained
4M 32 64 1.4e-02 3.7M 8 8 5,725 0.4B
6M 32 96 1.2e-02 6.0M 8 8 9,182 0.6B
8M 32 128 1.1e-02 8.5M 8 8 13,039 0.9B
10M 32 144 1.0e-02 9.9M 8 8 15,117 1.0B
14M 32 192 9.2e-03 14.4M 8 8 21,953 1.4B
16M 32 208 8.9e-03 16.0M 8 8 24,432 1.6B
20M 64 192 8.4e-03 19.1M 8 16 14,584 1.9B
60M 96 384 5.8e-03 57.1M 12 16 29,042 5.7B
90M 160 528 4.9e-03 97.9M 12 16 29,901 9.8B
150M 192 768 4.2e-03 151.9M 12 12 38,157 15.0B
300M 320 1,024 3.3e-03 320.0M 16 16 45,787 30.0B
530M 448 1,344 2.8e-03 530.1M 16 16 57,786 53.0B
750M 576 1,536 2.5e-03 681.3M 16 16 63,589 75.0B
1B 704 2,048 2.1e-03 1176.8M 16 16 69,369 100.0B

Citation

BibTeX:

@article{MagnussonDataDecide2025,
      title={{DataDecide: How to Predict Best Pretraining Data with Small Experiments}},
      author={Ian Magnusson and Nguyen Tai and Ben Bogin and David Heineman and Jena Hwang and Luca Soldaini and Akshita Bhagia and Jiacheng Liu and Dirk Groeneveld and Oyvind Tafjord and Noah A. Smith and Pang Wei Koh and Jesse Dodge},
      year={2025},
      journal={arXiv preprint},
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •