# Slurm scripts | |
Mainly here as indicative. Adapt to current traning. | |
- `cpu.slurm` -> for data preprocessing | |
- `gpu.slurm` -> arguments are adapted to maximize the gpu mem of the 8 32GB GPU requested | |
We are using common disk spaces for datasets, caches and experiment dumps: | |
- Experiment dumps -> `$six_ALL_CCFRWORK/experiments` | |
`SCRATCH` disk spaces are wiped regularly (wiping every file that was not accessed in the past 30 days) so we have S3 buckets (https://console.cloud.google.com/storage/browser/bigscience-experiments and https://console.cloud.google.com/storage/browser/bigscience-datasets) as shared storage that is accessible from JZ but from others instances too. | |