|
--- |
|
annotations_creators: |
|
- no-annotation |
|
language: |
|
- en |
|
language_creators: |
|
- found |
|
license: |
|
- mit |
|
multilinguality: |
|
- monolingual |
|
size_categories: |
|
- n<1K |
|
source_datasets: |
|
- original |
|
task_categories: |
|
- text-classification |
|
task_ids: |
|
- text-classification |
|
paperswithcode_id: null |
|
pretty_name: Text360 Sample Dataset |
|
tags: |
|
- text-classification |
|
- arxiv |
|
- wikipedia |
|
dataset_info: |
|
data_files: |
|
train: |
|
- dir1/subdir1/s1.jsonl |
|
- dir2/subdir2/s2.jsonl |
|
config_name: default |
|
--- |
|
|
|
# Dataset Card for Text360 Sample Dataset |
|
|
|
## Dataset Description |
|
|
|
- **Repository:** [Add your repository URL here] |
|
- **Paper:** [Add paper URL if applicable] |
|
- **Point of Contact:** [Add contact information] |
|
|
|
### Dataset Summary |
|
|
|
This dataset contains text samples from two sources (arXiv and Wikipedia) organized in a hierarchical directory structure. Each sample includes a text field and a subset identifier. |
|
|
|
### Data Files Structure |
|
|
|
The dataset maintains its original directory structure: |
|
``` |
|
. |
|
├── dir1/ |
|
│ └── subdir1/ |
|
│ └── sample1.jsonl # Contains arXiv samples |
|
└── dir2/ |
|
└── subdir2/ |
|
└── sample2.jsonl # Contains Wikipedia samples |
|
``` |
|
|
|
### Data Fields |
|
|
|
Each JSONL file contains records with the following fields: |
|
- `text`: string - The main text content |
|
- `subset`: string - Source identifier ("arxiv" or "wikipedia") |
|
|
|
### Data Splits |
|
|
|
All data is included in the train split, distributed across the JSONL files in their respective directories. |
|
|
|
### Example Instance |
|
|
|
```json |
|
{ |
|
"text": "This is a long text sample from arxiv about quantum computing...", |
|
"subset": "arxiv" |
|
} |
|
``` |
|
|
|
## Additional Information |
|
|
|
### Dataset Creation |
|
|
|
The dataset is organized in its original directory structure, with JSONL files containing text samples from arXiv and Wikipedia sources. Each file maintains its original location and format. |
|
|
|
### Curation Rationale |
|
|
|
The dataset was created to provide a sample of text data from different sources for text classification tasks. |
|
|
|
### Source Data |
|
|
|
#### Initial Data Collection and Normalization |
|
|
|
The data was collected from two sources: |
|
1. arXiv papers |
|
2. Wikipedia articles |
|
|
|
#### Who are the source language producers? |
|
|
|
- arXiv: Academic researchers and scientists |
|
- Wikipedia: Community contributors |
|
|
|
### Annotations |
|
|
|
#### Annotation process |
|
|
|
No additional annotations were added to the source data. |
|
|
|
#### Who are the annotators? |
|
|
|
N/A |
|
|
|
### Personal and Sensitive Information |
|
|
|
The dataset does not contain any personal or sensitive information. |
|
|
|
## Considerations for Using the Data |
|
|
|
### Social Impact of Dataset |
|
|
|
This dataset can be used for educational and research purposes in text classification tasks. |
|
|
|
### Discussion of Biases |
|
|
|
The dataset may contain biases inherent to the source materials (arXiv papers and Wikipedia articles). |
|
|
|
### Other Known Limitations |
|
|
|
The dataset is a small sample and may not be representative of all content from the source materials. |
|
|
|
### Dataset Curators |
|
|
|
[Add curator information] |
|
|
|
### Licensing Information |
|
|
|
This dataset is released under the MIT License. |
|
|
|
### Citation Information |
|
|
|
[Add citation information] |
|
|
|
### Contributions |
|
|
|
[Add contribution information] |
|
|
|
### Contact |
|
|
|
[Add contact information] |