Ares was recently equipped with a new filesystem, which will take over the role of the scratch space. The new filesystem is significantly faster and allows for significant advancement in computing efficiency. This advantage comes at the cost of the reduced maximum size of space, don't hesitate to contact the helpdesk if this is a significant issue in your case. To fully utilize the new filesystem, we've updated the data retention policy, please consult the information below. New scratch is currently available in early access mode. Both storage systems, new and old, will function for a certain period so that the migration process can progress smoothly without interruption of work. Instructions on how to use the new filesystem are included below.
Please note that the new filesystem is still being tuned and optimized, and periods of reduced accessibility might occur.
Naming convention
The filesystems were named as follows:
- ascratch - the OLD scratch filesystem, located at the path: /net/ascratch/people/<login name>/
- afscra - the NEW scratch filesystem, located at the path: /net/afscra/people/<login name>/
Space management
You can check the availability of the new scratch space using the "hpc-fs" command:
[ares][plguser@login01 ~]$ hpc-fs Name Used space Limit %QuotaUsage | Files Limit %QuotaUsage ---- ---------- ----- ----------- | ----- ----- ----------- $HOME 4.79GB 10.00GB 47.9% | 391 100000 0.4% $SCRATCH[ascratch] 12.19GB 100.00TB 0.0% | 1010 1000000 0.1% $SCRATCH[afscra] 4KB 12.00TB 0.0% | 1 1000000 0.0% $PLG_GROUPS_STORAGE(............) 11KB 1.00GB 0.0% | 1 0 0.0% ....
Note the $SCRATCH[afscra] entry - this is the new scratch space, while the $SCRATCH[ascratch] is the old one. You can store up to 12TB of data and 1 million of files in the new scratch space.
How to use the new scratch space
By default, the system works in the same way as before and doesn't change the environment. The environment variable $SCRATCH points to /net/ascratch/poeple/<your login>/ directory. If you want to switch to the new filesystem, please complete the following steps:
- Ensure you don't have any running or queued jobs on the cluster.
- Use the "hpc-scratch afscra" command on the login node to switch to the new scratch space.
- Shut down any working processes (screen, tmux, etc.), and log out from the machine.
- Log on to the system, issue the "hpc-scratch" command, and make sure that the output says that you are using SCRATCH with a path starting with /net/afscra/
all done! You are set to use the new scratch space. From now on, the $SCRATCH environment variable points to the new location, and Slurm will create job-specific temporary directories in the new $SCRATCH space.
The hpc-scratch is a utility for displaying information and managing your $SCRATCH environment variable. The old scratch is still available under the full path of /net/ascratch/, and its location is set as the $SCRATCH_OLD environment variable.
Issue reporting
If you encounter any issues while using the new $SCRATCH space, please create a ticket in the helpdesk: https://helpdesk.plgrid.pl/
Data storage policy
$SCRATCH space is crucial for performing efficient computations. To fully utilize the filesystem's potential, a policy allows for the automated deletion of old data. Data is considered old and can be automatically removed from $SCRATCH according to the following rules:
- Files stored in $SCRATCH can be removed after 30 days of not being modified. This point is identical to the old $SCRATCH policy.
- Slurm workdirs, which include $SCRATCHDIR and $TMPDIR for jobs, located in $SCRATCH/slurm_jobdir/<job id>/ can be removed 7 days after the given job is finished.