Dataset Card for Enwiki Dataset
This is an automatically updating dataset containing ~7 million English Wikipedia articles, with expanded templates and converted into Markdown. This dataset was created with the intention to provide a bite-sized, LLM-readable version of Wikipedia for various applications, including RAG.
Dataset Overview
There are two main versions of the dataset:
merged-articles
- Complete Wikipedia dump with all articles merged into a single file.merged-article-chunked
- Articles chunked into ~700 word segments with Markdown-header hierarchical breadcrumbs (Gemini embeddings coming soon!)
The latest version of the dataset was updated on 2025-08-07
.
Blogpost coming soon!