Datasets:

Modalities:
Text
Languages:
English
Libraries:
Datasets
License:
prooflang / README.md
christopherastone's picture
Update README.md
a569c18
---
license: cc-by-4.0
task_categories:
- text-generation
language:
- en
size_categories:
- 1B<n<10B
pretty_name: ProofLang Corpus
dataset_info:
- config_name: proofs
num_bytes: 3197091800
num_examples: 3681901
features:
- name: fileID
dtype: string
- name: proof
dtype: string
- config_name: sentences
num_bytes: 3736579062
num_examples: 38899130
features:
- name: fileID
dtype: string
- name: sentence
dtype: string
download_size: 6933683563
dataset_size: 6933670862
---
# Dataset Card for the ProofLang Corpus
## Dataset Summary
The ProofLang Corpus includes 3.7M proofs (558 million words) mechanically extracted from papers that were posted on [arXiv.org](https://arXiv.org) between 1992 and 2020.
The focus of this corpus is proofs, rather than the explanatory text that surrounds them, and more specifically on the *language* used in such proofs.
Specific mathematical content is filtered out, resulting in sentences such as `Let MATH be the restriction of MATH to MATH.`
This dataset reflects how people prefer to write (non-formalized) proofs, and is also amenable to statistical analyses and experiments with Natural Language Processing (NLP) techniques.
We hope it can serve as an aid in the development of language-based proof assistants and proof checkers for professional and educational purposes.
## Dataset Structure
There are multiple TSV versions of the data. Primarily, `proofs` divides up the data proof-by-proof, and `sentences` further divides up the same data sentence-by-sentence.
The `raw` dataset is a less-cleaned-up version of `proofs`. More usefully, the `tags` dataset gives arXiv subject tags for each paper ID found in the other data files.
* The data in `proofs` (and `raw`) consists of a `paper` ID (identifying where the proof was extracted from), and the `proof` as a string.
* The data in `sentences` consists of a `paper` ID, and the `sentence` as a string.
* The data in `tags` consists of a `paper` ID, and the arXiv subject tags for that paper as a single comma-separated string.
Further metadata about papers can be queried from arXiv.org using the paper ID.
In particular, each paper `<id>` in the dataset can be accessed online at the url `https://arxiv.org/abs/<id>`
## Dataset Size
* `proofs` is 3,094,779,182 bytes (unzipped) and has 3,681,893 examples.
* `sentences` is 3,545,309,822 bytes (unzipped) and has 38,899,132 examples.
* `tags` is 7,967,839 bytes (unzipped) and has 328,642 rows.
* `raw` is 3,178,997,379 bytes (unzipped) and has 3,681,903 examples.
## Dataset Statistics
* The average length of `sentences` is 14.1 words.
* The average length of `proofs` is 10.5 sentences.
## Dataset Usage
Data can be downloaded as (zipped) TSV files.
Accessing the data programmatically from Python is also possible using the `Datasets` library.
For example, to print the first 10 proofs:
```python
from datasets import load_dataset
dataset = load_dataset('proofcheck/prooflang', 'proofs', split='train', streaming='True')
for d in dataset.take(10):
print(d['paper'], d['proof'])
```
To look at individual sentences from the proofs,
```python
dataset = load_dataset('proofcheck/prooflang', 'proofs', split='train', streaming='True')
for d in dataset.take(10):
print(d['paper'], d['sentence'])
```
To get a comma-separated list of arXiv subject tags for each paper,
```python
from datasets import load_dataset
dataset = load_dataset('proofcheck/prooflang', 'tags', split='train', streaming='True')
for d in dataset.take(10):
print(d['paper'], d['tags'])
```
Finally, to look at a version of the proofs with less aggressive cleanup (straight from the LaTeX extraction),
```python
dataset = load_dataset('proofcheck/prooflang', 'raw', split='train', streaming='True')
for d in dataset.take(10):
print(d['paper'], d['proof'])
```
### Data Splits
There is currently no train/test split; all the data is in `train`.
## Dataset Creation
We started with the LaTeX source of 1.6M papers that were submitted to [arXiv.org](https://arXiv.org) between 1992 and April 2022.
The proofs were extracted using a Python script simulating parts of LaTeX (including defining and expanding macros).
It does no actual typesetting, throws away output not between `\begin{proof}...\end{proof}`, and skips math content. During extraction,
* Math-mode formulas (signalled by `$`, `\begin{equation}`, etc.) become `MATH`
* `\ref{...}` and variants (`autoref`, `\subref`, etc.) become `REF`
* `\cite{...}` and variants (`\Citet`, `\shortciteNP`, etc.) become `CITE`
* Words that appear to be proper names become `NAME`
* `\item` becomes `CASE:`
We then run a cleanup pass on the extracted proofs that includes
* Cleaning up common extraction errors (e.g., due to uninterpreted macros)
* Replacing more references by `REF`, e.g., `Theorem 2(a)` or `Postulate (*)`
* Replacing more citations with `CITE`, e.g., `Page 47 of CITE`
* Replacing more proof-case markers with `CASE:`, e.g., `Case (a).`
* Fixing a few common misspellings
## Additional Information
This dataset is released under the Creative Commons Attribution 4.0 licence.
Copyright for the actual proofs remains with the authors of the papers on [arXiv.org](https://arXiv.org), but these simplified snippets are fair use under US copyright law.