|
--- |
|
dataset_info: |
|
features: |
|
- name: text |
|
dtype: string |
|
- name: source |
|
dtype: string |
|
- name: filtering_features |
|
dtype: string |
|
- source_other: dump |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_examples: 1594197267 |
|
download_size: 3.3TB |
|
license: odc-by |
|
pretty_name: Zyda |
|
task_categories: |
|
- text-generation |
|
language: |
|
- en |
|
size_categories: |
|
- n>1T |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/*/*/* |
|
- config_name: zyda_no_starcoder |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/*/* |
|
- config_name: zyda_arxiv_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/zyda_arxiv/* |
|
- config_name: zyda_c4-en_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/c4_en/* |
|
- config_name: zyda_peS2o_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/zyda_peS2o/* |
|
- config_name: zyda_pile-uncopyrighted_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/zyda_pile-uncopyrighted/* |
|
- config_name: zyda_refinedweb_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/zyda_refinedweb/* |
|
- config_name: zyda_slimpajama_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_no_starcoder/zyda_slimpajama/* |
|
- config_name: zyda_starcoder_only |
|
data_files: |
|
- split: train |
|
path: data/zyda_starcoder/*/* |
|
--- |
|
# Dataset Card for Zyda |
|
|
|
<!-- Provide a quick summary of the dataset. --> |
|
|
|
Zyda is a 1.3T language modelling dataset created by collecting open and high quality datasets and combining them and performing a uniform filtering and deduplication step. We find that Zyda performs extremely well in ablations and is at least comparable and potentially better to the best openly available datasets available, due to our meticulous post-processing pipeline. We think the best use of Zyda is either as a standalone dataset for language model training up to the 1T scale, or in combination with Fineweb or Dolma for multi-trillion token training. |
|
|
|
|
|
## How to download |
|
|
|
Full dataset: |
|
`datasets.load_dataset("Zyphra/Zyda", split="train")` |
|
|
|
Full dataset without StarCoder: |
|
`datasets.load_dataset("Zyphra/Zyda", name="zyda_no_starcoder", split="train")` |
|
|
|
For downloading individual components put their name in the name arg of `load_dataset()`: |
|
- zyda_arxiv_only |
|
- zyda_c4-en_only |
|
- zyda_peS2o_only |
|
- zyda_pile-uncopyrighted_only |
|
- zyda_refinedweb_only |
|
- zyda_slimpajama_only |
|
- zyda_starcoder_only |
|
|
|
|
|
### Dataset Description |
|
|
|
<!-- Provide a longer summary of what this dataset is. --> |
|
|
|
- **Curated by:** Zyphra |
|
- **Language(s) (NLP):** Primarily English |
|
- **License:** Open Data Commons License |
|
|
|
|
|
## Dataset Structure |
|
|
|
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. --> |
|
|
|
Dataset fields: |
|
- `text`: contains actual text for training |
|
- `source`: component the text is coming from |
|
- `filtering_features`: precomputed values of different features that were used for filtering (converted to json string) |
|
- `source_other`: metadata from the source dataset (converted to json string) |
|
|
|
|
|
|
|
### Source Data |
|
|
|
|
|
Pile Uncopyrighted: https://huggingface.co/datasets/monology/pile-uncopyrighted |
|
|
|
C4-en: https://huggingface.co/datasets/allenai/c4 |
|
|
|
peS2o: https://huggingface.co/datasets/allenai/peS2o |
|
|
|
RefinedWeb: https://huggingface.co/datasets/tiiuae/falcon-refinedweb |
|
|
|
SlimPajama: https://huggingface.co/datasets/cerebras/SlimPajama-627B |
|
|
|
arxiv_s2orc_parsed: https://huggingface.co/datasets/ArtifactAI/arxiv_s2orc_parsed |
|
|
|
StarCoder: https://huggingface.co/datasets/bigcode/starcoderdata |
|
|
|
#### Data Collection and Processing |
|
|
|
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. --> |
|
|
|
[More Information Needed] |
|
|
|
|
|
|
|
|
|
#### Personal and Sensitive Information |
|
|
|
As a language modelling dataset, it likely contains PII which has not been filtered out of the component datasets and which may have been missed by our own filters. |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
As a dataset comprised of open web scrapes, it is likely that it contains biased and toxic content. |
|
|
|
## Citation [optional] |
|
|
|
If you use our dataset to train a model, please cite us at: |
|
|
|
(-/TODO) |
|
|