applied-ai-018's picture
Add files using upload-large-folder tool
a15b28b verified
- To download
- <space> wget --header="Authorization: Bearer <HF_READ_TOKEN>" https://huggingface.co/datasets/bigcode/the-stack/resolve/main/data/actionscript/train-00000-of-00002.parquet?download=true
- hfd.sh from https://gist.github.com/padeoe/697678ab8e528b85a2a7bddafea1fa4f
- the stack dedup
- the stack smol
- need to accept agreement on HF
- <space> bash hfd.sh bigcode/the-stack-smol --hf_username <username> --hf_token <token> --tool aria2c -x 8 --dataset
- 3 GB in 1.5 minutes
- github, books, wiki, stackexchange, arxiv
- RPv1
- English wikipedia from https://huggingface.co/datasets/wikimedia/wikipedia
- hf c4/realnewslike
- git lfs
```
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/allenai/c4
cd c4
git lfs pull --include "realnewslike/*"
```
- better: aria2c
- (hfd_c4.sh)
- looked at hfd.sh (after `printf "\nStart Downloading lfs files, bash script:\n"`)
- so after `GIT_LFS_SKIP_SMUDGE=1 ...` and `cd`, get `files` like e.g. `git lfs ls-files -I "realnewslike/*"`
- and use later part of hfd.sh
-----
- Amber
- Wikipedia is 2%
- This is based on RPv1
- CC subset of RPv1 changed to RefinedWeb (I think only a subset is open-sourced)
- StarCoderData for code
- Procedure here https://github.com/LLM360/amber-data-prep
- Alibaba Data Juicer
- Filtered RPv1 and other subsets available
- Also, CC snapshots available; dedup with simhash (why RPv2 then?)
- https://github.com/alibaba/data-juicer/blob/main/configs/data_juicer_recipes/README.md
- RP reproduced, some token counts here https://github.com/alibaba/data-juicer/tree/main/configs/reproduced_redpajama
- Semantic scholar peS2o (one file, s2orc val set)
- 51K -> 40K, but most removed with average and max line length filter
- Example: /home1/BharatGPT_Data/data-juicer/demos/process_sci_data/outputs/trace/filter-maximum_line_length_filter.jsonl
- seem OK papers
- (ig use arxiv or create own config for SemanticScholar)
- Issue of huggingface cache - used in home directory
- use ds_cache_dir flag (override config by cmd line arg)
- this DID NOT WORK with mixture... corrected by setting HF_HOME (and maybe TRANSFORMERS_CACHE too) - mixture demo
- I've put these in ~/.bashrc
```
# cache home
export DATA_JUICER_CACHE_HOME="/home1/BharatGPT_Data/data-juicer/.cache"
# cache models
export DATA_JUICER_MODELS_CACHE="/home1/BharatGPT_Data/data-juicer/.cache/models"
# cache assets
export DATA_JUICER_ASSETS_CACHE="/home1/BharatGPT_Data/data-juicer/.cache/assets"
```
- important files
- /home1/BharatGPT_Data/data-juicer/tools/process_data.py
- /home1/BharatGPT_Data/data-juicer/tools/postprocess/data_mixture.py
- both these used with respective demos
- mixer code
- /home1/BharatGPT_Data/data-juicer/data_juicer/format/mixture_formatter.py