The Stack v2
DatasetFree67 TB permissively licensed code dataset across 600+ languages.
Capabilities10 decomposed
permissively-licensed source code dataset curation and aggregation
Medium confidenceAggregates 67 TB of source code from the Software Heritage archive, filtering for permissively licensed repositories (MIT, Apache 2.0, BSD, etc.) across 600+ programming languages. Uses automated license detection and validation to ensure legal compliance for model training. Implements a rigorous deduplication pipeline at file and repository levels to eliminate redundant training data and reduce dataset bloat.
Largest open-source code dataset at 67 TB with automated opt-out governance allowing repository owners to request removal, combined with rigorous deduplication and PII removal pipeline — no other public dataset offers this scale with legal compliance and community control mechanisms
Larger and more legally compliant than GitHub's CodeSearchNet (14M files) or Google's BigQuery public datasets, with explicit opt-out governance vs. implicit inclusion, and covers 600+ languages vs. Codex training data's undisclosed language distribution
opt-out governance and repository exclusion management
Medium confidenceImplements a community-driven opt-out system where repository owners can request removal of their code from the dataset without legal takedown notices. Maintains a registry of excluded repositories and re-applies exclusions during dataset updates. Provides transparent governance documentation and a clear submission process for removal requests, balancing open access with creator rights.
First large-scale code dataset to implement opt-out governance at dataset level rather than relying solely on license compliance, with transparent registry and community submission process — shifts power from dataset creators to code contributors
More respectful of creator autonomy than GitHub Copilot's training approach (no opt-out) or academic datasets (one-time snapshot), and more scalable than individual DMCA takedowns
pii and sensitive data removal pipeline
Medium confidenceAutomated pipeline that scans source code for personally identifiable information (email addresses, API keys, SSH keys, credit card patterns, phone numbers) and removes or redacts them before dataset release. Uses regex patterns, entropy-based detection for secrets, and heuristic rules to identify sensitive data. Operates at file level with configurable sensitivity thresholds to balance data utility against privacy risk.
Combines regex pattern matching, entropy-based secret detection, and heuristic rules in a unified pipeline with configurable sensitivity — more comprehensive than simple regex-only approaches, but trades off false positive rate against security coverage
More thorough than GitHub's secret scanning (which only flags known patterns) because it includes entropy-based detection for unknown secret formats, but less accurate than specialized tools like TruffleHog due to language-agnostic approach
multi-language source code indexing and retrieval
Medium confidenceIndexes 67 TB of source code across 600+ programming languages with language-aware metadata (syntax, file extension, language family). Enables retrieval by language, license, repository, or code patterns. Uses Software Heritage's existing indexing infrastructure as foundation, augmented with language detection and classification. Supports both bulk download and filtered queries for specific language subsets.
Leverages Software Heritage's existing language detection and indexing infrastructure, then augments with BigCode-specific language classification and filtering — avoids reinventing language detection while providing dataset-specific query capabilities
More comprehensive language coverage (600+ languages) than GitHub's Linguist (500+ languages) and more accessible than Software Heritage's raw API because it's pre-filtered for permissive licenses and deduplicated
content-based deduplication at file and repository levels
Medium confidenceRemoves duplicate code files and repositories using content hashing (SHA-256 or similar) and fuzzy matching for near-duplicates. Operates in two stages: exact deduplication via hash matching, then fuzzy matching (e.g., Jaccard similarity or MinHash) to catch semantically identical code with minor formatting differences. Preserves one canonical copy of each unique code pattern while removing redundant training examples.
Two-stage deduplication combining exact hash matching with fuzzy similarity matching (likely MinHash or Jaccard) to catch both identical and near-identical code — more thorough than single-stage approaches but computationally expensive
More aggressive deduplication than CodeSearchNet (which uses simple hash matching) because it catches near-duplicates, but less semantic than clone detection tools (which understand code structure) because it's content-based
software heritage archive integration and version control history access
Medium confidenceIntegrates with Software Heritage's comprehensive archive of 200+ million repositories and their full version control history. Extracts source code snapshots from Software Heritage's Git/Mercurial/SVN repositories, preserving repository metadata (commit history, author info, timestamps). Provides access to code at specific points in time, enabling historical analysis or training on code evolution patterns.
Leverages Software Heritage's universal code archive (200M+ repositories) as data source, providing access to code that would be impossible to collect via GitHub API alone — enables training on archived/deleted repositories and non-GitHub platforms (GitLab, Gitea, etc.)
More comprehensive than GitHub-only datasets because it includes code from GitLab, Gitea, SourceForge, and other platforms archived by Software Heritage; more legally defensible than web scraping because it uses an established, community-maintained archive
license compliance and legal metadata tracking
Medium confidenceTracks and validates SPDX license identifiers for each repository, ensuring only permissively licensed code (MIT, Apache 2.0, BSD, etc.) is included. Maintains license metadata alongside code files, enabling downstream users to verify legal compliance. Implements license hierarchy and compatibility checking to handle dual-licensed or complex licensing scenarios.
Combines automated SPDX detection with manual review and maintains license metadata alongside code, enabling downstream users to verify compliance — more transparent than datasets that simply claim 'permissive licenses' without proof
More legally rigorous than GitHub's CodeSearchNet (which doesn't validate licenses) and more transparent than Codex training data (which doesn't disclose license filtering at all)
dataset versioning and reproducibility tracking
Medium confidenceMaintains versioned snapshots of the dataset (e.g., v2.0, v2.1) with documented changes between versions (new repositories added, deduplication improvements, PII removal updates). Provides checksums and manifests for reproducibility, enabling researchers to cite specific dataset versions and reproduce results. Tracks dataset lineage and transformation history.
Maintains semantic versioning and detailed changelogs for dataset releases, enabling researchers to cite specific versions and understand dataset evolution — more rigorous than one-off dataset releases without versioning
More reproducible than academic datasets that are released once without versioning, and more transparent than commercial datasets (Codex) that don't disclose version history or changes
training data preparation and tokenization for llm fine-tuning
Medium confidenceProvides pre-processed code files formatted for direct use in LLM training pipelines, with optional tokenization using standard tokenizers (GPT-2, GPT-3, Llama, etc.). Includes language-specific formatting (e.g., preserving indentation for Python, handling multi-line strings) and optional code-specific preprocessing (e.g., removing comments, normalizing whitespace). Supports both raw code and tokenized sequences depending on downstream model architecture.
Provides multiple tokenization options and language-aware preprocessing rather than forcing single format, enabling flexibility for different model architectures — more flexible than pre-tokenized datasets but requires more user configuration
More flexible than pre-tokenized datasets (which lock you to specific tokenizer) but less convenient than fully preprocessed datasets; enables experimentation with different tokenizers without re-downloading raw data
training data for starcoder2 and code generation models
Medium confidenceServes as the primary training dataset for StarCoder2 models and other code generation models. Provides high-quality, permissively-licensed, deduplicated code across 600+ languages with repository context. Enables training of state-of-the-art code LLMs that understand diverse programming paradigms, languages, and coding patterns. Documented as essential resource for reproducing StarCoder2 and training similar models.
Curated and published as the official training dataset for StarCoder2 models, providing permissively-licensed, deduplicated, PII-removed code across 600+ languages with repository context and governance
More comprehensive and higher-quality than previous code datasets (CodeSearchNet, GitHub-Code) with rigorous deduplication, PII removal, and licensing compliance; enables training of state-of-the-art code models
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with The Stack v2, ranked by overlap. Discovered automatically through the match graph.
StarCoder Data
783 GB curated code dataset from 86 languages with PII redaction.
StarCoderData
250GB curated code dataset for StarCoder training.
Granite
IBM's enterprise-focused open foundation models.
banned-historical-archives
Dataset by banned-historical-archives. 18,46,708 downloads.
Indicium Tech
Transform raw data into actionable, industry-specific...
Mend.io
AI-powered application security with auto-remediation.
Best For
- ✓ML teams training large code LLMs (10B+ parameters)
- ✓Open-source model developers needing legally defensible training data
- ✓Researchers studying code generation across language families
- ✓Open-source projects concerned about code reuse in commercial models
- ✓Individual developers wanting control over their code's use in AI training
- ✓Organizations building datasets with community trust as a core value
- ✓Teams training models that will generate code in production environments
- ✓Privacy-conscious organizations handling code from diverse contributors
Known Limitations
- ⚠Permissive license filtering excludes GPL and AGPL code, limiting coverage of certain ecosystems (Linux kernel, GNU tools)
- ⚠Deduplication is content-based, not semantic — similar algorithms in different styles may be retained as duplicates
- ⚠License detection relies on heuristics and file headers; edge cases with dual-licensing or custom licenses may be misclassified
- ⚠67 TB dataset requires significant storage infrastructure and bandwidth for download/processing
- ⚠Opt-out is reactive, not proactive — requires developers to actively request removal
- ⚠No guarantee of removal from already-trained models using earlier dataset versions
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
BigCode project's 67 TB dataset of permissively licensed source code from Software Heritage archive covering 600+ programming languages. The largest open code dataset available, used to train StarCoder2 models. Includes full file content, repository metadata, and license information. Follows an opt-out governance model allowing repository owners to exclude their code. Rigorous deduplication and PII removal pipeline. Essential resource for training code generation models.
Categories
Alternatives to The Stack v2
Open-source image generation — SD3, SDXL, massive ecosystem of LoRAs, ControlNets, runs locally.
Compare →Are you the builder of The Stack v2?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →