OLMoE-mix-0924 / README.md
Muennighoff's picture
Update README.md
c11c6e1 verified
metadata
task_categories:
  - text-generation
language:
  - en
size_categories:
  - 1B<n<10B
license: odc-by
pretty_name: OLMoE Mix (September 2024)

OLMoE Mix (September 2024)

Dataset Description

OLMoE Mix Logo.

The following data mix was used to train OLMoE-1B-7B, a Mixture-of-Experts LLM with 1B active and 7B total parameters released in September 2024.

The base version of OLMoE-1B-7B can be found at this page, the SFT of OLMoE-1B-7B is available here, and a version combining SFT and DPO is available following this link.

Statistics

Subset Tokens Words Bytes Docs
DCLM Baseline 1.0 3.86 T 3.38 T 16.7 T 2.95 B
Starcoder 101 B 63.9 B 325 B 78.7 M
peS2o
(Dolma)
57.2 B 51.3 B 268 B 38.8 M
Arxiv
(RedPajama v1
via Proof Pile II)
21.1 B 23.5 B 88.8 B 1.55 M
OpenWebMath
(Proof Pile II)
12.7 B 10.2 B 42.4 B 2.91 M
Algebraic Stack
(Proof Pile II)
12.6 B 9.6 B 39.3 B 2.83 M
En Wikipedia +
Wikibooks
(Dolma)
3.69 B 3.16 B 16.2 B 6.17 M
Total 4.07 T 3.53 T 17.4 T 3.08 B

Preprocessing

All subsets were pre-processed to remove documents with a sequence of 32 or more repeated ngrams.

  • a ngram is a span of 1 to 13 tokens, included;
  • tokens are obtained using the model tokenizer;
  • a sequence is a contiguous span of repeated ngrams.

In addition of the above, Starcoder dataset was further processed by removing any document meeting any of the following rules:

  • document is from a repository with fewer than 2 stars on GitHub;
  • the top most frequent word in the document constitutes over 30% of the document;
  • the two most frequent words in the document constitutes over 50% of the document.

Licensing Information

This mix is licensed under Open Data Commons Attribution License (ODC-By) v1.0. By using this dataset, you are bound to licenses and Terms of Services of underlying datasets, which you can access by clicking on the links in the table above.

Citation

@misc{muennighoff2024olmoeopenmixtureofexpertslanguage,
      title={OLMoE: Open Mixture-of-Experts Language Models}, 
      author={Niklas Muennighoff and Luca Soldaini and Dirk Groeneveld and Kyle Lo and Jacob Morrison and Sewon Min and Weijia Shi and Pete Walsh and Oyvind Tafjord and Nathan Lambert and Yuling Gu and Shane Arora and Akshita Bhagia and Dustin Schwenk and David Wadden and Alexander Wettig and Binyuan Hui and Tim Dettmers and Douwe Kiela and Ali Farhadi and Noah A. Smith and Pang Wei Koh and Amanpreet Singh and Hannaneh Hajishirzi},
      year={2024},
      eprint={2409.02060},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2409.02060}, 
}