The Pile
DatasetFreeEleutherAI's 825 GiB diverse training dataset from 22 sources.
Capabilities11 decomposed
multi-domain pretraining corpus assembly
Medium confidenceCombines 22 discrete, curated text datasets (academic papers, books, code, web text, specialized sources) into a single 825 GiB jsonlines corpus compressed with zstandard. The assembly approach prioritizes diversity across domains rather than size maximization, enabling language models trained on this corpus to develop broad cross-domain knowledge and generalization capabilities. Data is provided as-is without documented preprocessing, deduplication, or filtering pipelines, placing responsibility for data cleaning on downstream users.
Pioneered the multi-domain curation approach by intentionally combining 22 diverse, high-quality subsets (academic papers, books, code, web, specialized sources) rather than scraping a single massive web corpus. This architectural choice prioritizes knowledge breadth and domain coverage over raw scale, influencing the design of subsequent open datasets like LAION, RedPajama, and Falcon-Refinedweb.
Broader domain coverage than Common Crawl-only datasets (e.g., C4) and higher quality than raw web scrapes due to curation of academic, code, and book sources; smaller than Falcon-Refinedweb (1.5T tokens) but more carefully curated and widely adopted as a benchmark for model evaluation
cross-domain model evaluation via pile bpb metric
Medium confidenceProvides a standardized evaluation metric (Pile Bits Per Byte, or BPB) that measures language model perplexity across the full 22-subset corpus, enabling comparison of model generalization across diverse text domains. The metric is computed by evaluating a trained model on held-out portions of each subset and aggregating results, producing a single scalar score where lower values indicate better cross-domain performance. This approach surfaces domain-specific weaknesses that single-domain metrics would miss.
Introduced BPB (Bits Per Byte) as a standardized metric for evaluating language model performance across a curated multi-domain corpus rather than a single domain or random web text. This approach surfaces generalization gaps that domain-specific metrics (e.g., code completion accuracy, translation BLEU) would miss, establishing a precedent for multi-domain evaluation in subsequent benchmarks (MMLU, HELM).
More comprehensive than single-domain metrics (e.g., GLUE for NLU, HumanEval for code) because it evaluates across 22 domains simultaneously; more reproducible than web-scale benchmarks (e.g., zero-shot on random web text) due to fixed, curated evaluation set, though leaderboard adoption remains limited due to sparse published results
model-agnostic training data format and integration
Medium confidenceProvides training data in a model-agnostic jsonlines format that integrates with standard ML frameworks (PyTorch, TensorFlow, Hugging Face) without requiring custom preprocessing or format conversion. The jsonlines + zstandard approach enables seamless integration with existing dataloaders, tokenizers, and training pipelines, reducing friction for researchers adopting the dataset. No custom APIs or proprietary tools are required — standard open-source libraries suffice.
Uses standard, framework-agnostic jsonlines + zstandard format that integrates directly with PyTorch, TensorFlow, and Hugging Face without custom preprocessing or proprietary tools. This contrasts with proprietary formats (HDF5, custom binary formats) that require custom loaders, or single-framework datasets that lock users into specific ML libraries.
More portable than proprietary formats because it uses standard jsonlines; more efficient than uncompressed text because zstandard compression reduces storage by ~3-4x; simpler than database formats (SQLite, Parquet) because jsonlines requires no schema definition or query language.
jsonlines-formatted text corpus with zstandard compression
Medium confidenceEncodes the 825 GiB corpus as jsonlines (one JSON object per line, typically with a 'text' field containing raw text) and compresses with zstandard (zstd), a modern compression algorithm offering faster decompression and better compression ratios than gzip. This format choice enables streaming decompression and line-by-line parsing without loading the entire dataset into memory, critical for training pipelines on resource-constrained hardware. The jsonlines structure allows metadata (e.g., source subset, document ID) to be stored alongside text.
Chose zstandard compression over gzip or bzip2, offering ~20% better compression ratios and 5-10x faster decompression speeds, critical for large-scale training pipelines where I/O is a bottleneck. Paired with jsonlines format to enable streaming decompression and line-by-line parsing without materializing the full 825 GiB dataset in memory.
Faster decompression than gzip-compressed datasets (e.g., C4) and more memory-efficient than uncompressed datasets; jsonlines format is more flexible than binary formats (e.g., HDF5, TFRecord) for preserving metadata and enabling ad-hoc analysis, though slightly slower to parse than optimized binary formats
subset-level source attribution and composition transparency
Medium confidenceExplicitly enumerates the 22 constituent subsets of the Pile (academic papers from PubMed and ArXiv, books from Books3 and Gutenberg, code from GitHub, web text from OpenWebText2 and Pile-CC, specialized sources like USPTO patents, Ubuntu IRC, and Stack Exchange) and provides source attribution for each document. This transparency enables users to understand the composition of their training data, audit for potential biases or contamination, and selectively exclude subsets if needed. However, exact composition percentages and subset enumeration are not fully documented.
Pioneered explicit, multi-source composition transparency in large pretraining datasets by publicly naming 22 constituent subsets and their sources, establishing a precedent for data provenance documentation in subsequent datasets (RedPajama, Falcon-Refinedweb). This approach enables auditing and selective subset exclusion, though exact composition percentages remain undocumented.
More transparent than Common Crawl-only datasets (e.g., C4) which provide minimal source attribution; comparable to RedPajama in subset enumeration but less detailed in per-document source labels and composition percentages
academic and specialized text domain coverage
Medium confidenceIncludes curated subsets of academic papers (PubMed, ArXiv), specialized technical sources (USPTO patents, Stack Exchange), and code repositories (GitHub), providing dense coverage of high-signal, domain-specific text that is underrepresented in web-only corpora. These subsets are integrated into the broader corpus at a fixed ratio, ensuring that models trained on the Pile develop specialized knowledge in these domains without requiring separate fine-tuning. The inclusion of academic papers and code is particularly valuable for training models intended for scientific or technical applications.
Intentionally curated academic papers (PubMed, ArXiv) and code (GitHub) as core subsets rather than treating them as incidental web scrape byproducts, establishing a precedent for domain-specific data curation in pretraining. This approach ensures models trained on the Pile develop strong performance on technical and scientific tasks without requiring separate fine-tuning or domain-specific pretraining.
More comprehensive academic and code coverage than web-only datasets (e.g., C4, Common Crawl); comparable to domain-specific datasets (e.g., CodeSearchNet for code, S2ORC for academic papers) but integrated into a single multi-domain corpus for broader generalization
books and long-form text corpus inclusion
Medium confidenceIncorporates two book-focused subsets (Books3 and Gutenberg) providing long-form, narrative text with complex linguistic structures, enabling models to develop strong performance on coherent, multi-paragraph generation and understanding of narrative arcs. Books represent a fundamentally different text distribution than web text (longer documents, more complex grammar, narrative structure) and are valuable for training models intended for creative writing, summarization, or long-context understanding. The inclusion of both contemporary books (Books3) and public-domain classics (Gutenberg) provides temporal and stylistic diversity.
Explicitly includes book-focused subsets (Books3, Gutenberg) as core components rather than incidental web scrape byproducts, recognizing that long-form narrative text develops different linguistic capabilities than short web snippets. This architectural choice influences model performance on coherence, narrative structure, and long-context understanding.
More comprehensive book coverage than web-only datasets (e.g., C4); comparable to book-specific datasets (e.g., BookCorpus) but integrated into a multi-domain corpus for broader generalization rather than domain-specific pretraining
web-scale text corpus with deduplication and quality filtering
Medium confidenceCombines two web-derived subsets (OpenWebText2 and Pile-CC) providing broad coverage of diverse web text while applying quality filtering and deduplication to reduce noise compared to raw Common Crawl. OpenWebText2 is derived from URLs shared on Reddit (a proxy for human-curated quality), while Pile-CC is a filtered subset of Common Crawl. Together, these subsets provide web-scale coverage without the extreme noise and duplication of raw web scrapes, balancing breadth with quality.
Combines Reddit-curated web text (OpenWebText2) with filtered Common Crawl (Pile-CC) rather than relying on raw Common Crawl alone, applying implicit quality filtering through Reddit curation and explicit deduplication/filtering on Pile-CC. This hybrid approach balances web-scale coverage with quality, addressing a key limitation of earlier web-only datasets.
Higher quality than raw Common Crawl (e.g., C4) due to Reddit curation and filtering; broader coverage than Reddit-only datasets; comparable to Falcon-Refinedweb in approach but with less documented filtering methodology
static dataset versioning and reproducibility
Medium confidenceProvides a fixed, immutable 825 GiB snapshot of the Pile corpus, enabling reproducible model training and evaluation across teams and time periods. The static nature ensures that models trained on the Pile in 2021 can be compared directly with models trained in 2024 without worrying about dataset drift or updates. However, no explicit versioning scheme, release notes, or update mechanism is documented, limiting transparency about potential corrections or improvements.
Provides a fixed, immutable snapshot of a large pretraining corpus, establishing a stable benchmark for model evaluation and reproducibility. This approach contrasts with continuously-updated datasets (e.g., Common Crawl) and enables long-term reproducibility, though it sacrifices the ability to correct errors or incorporate new data.
More reproducible than continuously-updated datasets (e.g., Common Crawl, web-scale datasets); less flexible than modular, versioned datasets (e.g., Hugging Face Datasets with explicit version tags) due to lack of documented versioning scheme
citation and attribution framework for multi-source datasets
Medium confidenceProvides formal citation guidance (Gao et al., 2020, arXiv:2101.00027) for the Pile itself and requires attribution to individual component datasets, establishing a precedent for proper data provenance documentation in large pretraining corpora. This framework enables researchers to trace the lineage of their training data and acknowledge the original sources and curators. However, no machine-readable citation metadata or automated attribution tools are provided.
Established a precedent for formal citation and attribution of large multi-source pretraining datasets by providing explicit citation guidance (Gao et al., 2020) and requiring attribution to component datasets. This approach influenced subsequent datasets (RedPajama, Falcon-Refinedweb) to provide similar citation frameworks, though machine-readable metadata and automated tools remain absent.
More transparent than datasets with minimal citation guidance (e.g., early Common Crawl releases); less comprehensive than datasets with machine-readable citation metadata and automated attribution tools (e.g., Hugging Face Datasets with CITATION.cff files)
public reproducibility and open-source model training
Medium confidenceEnables reproducible, open-source language model training by providing a publicly-available, freely-downloadable dataset used to train GPT-NeoX, Pythia, and other open models. The dataset is released under an open license (exact license terms not specified in artifact), allowing researchers and organizations to train models with full transparency and reproducibility. The Pile has influenced the design of subsequent open datasets, establishing a standard for open-source LLM training data.
Provides a large-scale, publicly-available, freely-downloadable pretraining dataset specifically designed for open-source LLM development, enabling full reproducibility and transparency. This contrasts with proprietary datasets (used by OpenAI, Google, Meta) that are not publicly available, or academic datasets that lack the scale and diversity needed for large models. The Pile's influence on subsequent open datasets (C4, RedPajama, etc.) establishes it as a foundational artifact for open-source AI.
More accessible than proprietary datasets (OpenAI, Google) because it is freely available; more comprehensive than earlier open datasets (WikiText, BookCorpus) because it includes 825 GiB across 22 domains; more influential than contemporary datasets because it established design patterns for open-source LLM training data.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with The Pile, ranked by overlap. Discovered automatically through the match graph.
RedPajama v2
30 trillion token web dataset with 40+ quality signals per document.
TxT360
Dataset by LLM360. 10,70,517 downloads.
Language Is Not All You Need: Aligning Perception with Language Models (Kosmos-1)
* ⭐ 03/2023: [PaLM-E: An Embodied Multimodal Language Model (PaLM-E)](https://arxiv.org/abs/2303.03378)
bert-base-uncased
fill-mask model by undefined. 5,92,18,905 downloads.
punctuate-all
token-classification model by undefined. 5,53,415 downloads.
Dolma
Allen AI's 3T token dataset for fully reproducible LLM training.
Best For
- ✓researchers and teams training large language models from scratch with compute budgets >100 GPU-hours
- ✓open-source model developers building alternatives to proprietary LLMs (GPT, Claude, Gemini)
- ✓academic institutions studying language model pretraining and generalization
- ✓model developers and researchers comparing pretraining approaches and dataset compositions
- ✓teams evaluating whether a model trained on their custom dataset generalizes as well as Pile-trained baselines
- ✓benchmark leaderboard maintainers seeking a standardized, reproducible evaluation metric
- ✓ML engineers building training pipelines with PyTorch, TensorFlow, or Hugging Face
- ✓Teams seeking to minimize data engineering overhead when adopting large-scale pretraining datasets
Known Limitations
- ⚠English-only; no multilingual coverage or non-English language support
- ⚠Static snapshot with no versioning, update mechanism, or reproducibility guarantees documented
- ⚠Exact composition percentages and subset enumeration not fully documented; 22 subsets mentioned but only 8-10 named explicitly
- ⚠No documented deduplication strategy; potential for data leakage across subsets or contamination with test sets
- ⚠825 GiB fixed size requires significant storage infrastructure; no streaming or sampling utilities provided for resource-constrained environments
- ⚠Leaderboard contains only 2 published entries (GPT-3, GPT-2) with asterisks indicating 'potential test-set overlap', severely limiting comparative value
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
EleutherAI's seminal 825 GiB English text dataset composed of 22 diverse high-quality subsets. Includes academic papers (PubMed, ArXiv), books (Books3, Gutenberg), code (GitHub), web (OpenWebText2, Pile-CC), and specialized sources (USPTO patents, Ubuntu IRC, Stack Exchange). Designed for training large language models with broad knowledge coverage. Used to train GPT-NeoX, Pythia, and influenced the design of virtually every subsequent open training dataset.
Categories
Alternatives to The Pile
Open-source image generation — SD3, SDXL, massive ecosystem of LoRAs, ControlNets, runs locally.
Compare →Are you the builder of The Pile?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →