Ares was recently equipped with a new filesystem, which will take over the role of the scratch space. The new filesystem is significantly faster and allows for significant advancement in computing efficiency. This advantage comes at the cost of the reduced maximum size of space, don't hesitate to contact the helpdesk if this is a significant issue in your case. To fully utilize the new filesystem, we've updated the data retention policy, please consult the information below. New scratch is currently available in early access mode. Both storage systems, new and old, will function for a certain period so that the migration process can go smoothly without interruption of work. Instructions on how to use the new filesystem are included below. In the near future, the new scratch filesystem will be the default, and the old scratch will be decommissioned. If you choose to use the new scratch space please don't use intensively the old one at the same time.
Please note that the new filesystem is still being tuned and optimized, and periods of reduced accessibility might occur.
Naming convention
The filesystems were named as follows:
- ascratch - the OLD scratch filesystem, located at the path: /net/ascratch/people/<login name>/
- afscra - the NEW scratch filesystem, located at the path: /net/afscra/people/<login name>/
Space management
You can check the availability of the new scratch space using the "hpc-fs" command:
[ares][plguser@login01 ~]$ hpc-fs Name Used space Limit %QuotaUsage | Files Limit %QuotaUsage ---- ---------- ----- ----------- | ----- ----- ----------- $HOME 4.79GB 10.00GB 47.9% | 391 100000 0.4% $SCRATCH[ascratch] 12.19GB 100.00TB 0.0% | 1010 1000000 0.1% $SCRATCH[afscra] 4KB 12.00TB 0.0% | 1 1000000 0.0% $PLG_GROUPS_STORAGE(............) 11KB 1.00GB 0.0% | 1 0 0.0% ....
Note the $SCRATCH[afscra] entry - this is the new scratch space, while the $SCRATCH[ascratch] is the old one. You can store up to 12TB of data and 1 million of files in the new scratch space.
How to use the new scratch space
By default, the system works in the same way as before and doesn't change the environment. The environment variable $SCRATCH points to /net/ascratch/poeple/<your login>/ directory. If you want to switch to the new filesystem, please complete the following steps:
- Ensure you don't have any running jobs on the cluster. Queued jobs can be present in the system.
- On the login node, use the "hpc-scratch afscra" command to switch to the new scratch space.
- Terminate any processes on the login node (screen, tmux, etc.), and log out from the machine.
- Log on to the system, execute the "hpc-scratch" command, and make sure that the output says that you are using SCRATCH with a path starting with /net/afscra/
All done! You are set to use the new scratch space. From now on, the $SCRATCH environment variable points to the new location, and Slurm will create job-specific temporary directories in the new $SCRATCH space.
The hpc-scratch is a utility for displaying information and managing your $SCRATCH environment variable. The old scratch is still available under the full path of /net/ascratch/people/<login>/, and its location is set as the $SCRATCH_OLD environment variable.
There is no synchronization between the new and old scratch space. If the old scratch contains any critical data, please move it to one of the other storage spaces.
Reporting issues
If you encounter any issues while using the new $SCRATCH space, please create a ticket in the helpdesk: https://helpdesk.plgrid.pl/
Data storage policy
$SCRATCH space is crucial for performing efficient computations. To fully utilize the filesystem's potential, a policy allows for the automated deletion of old data. Data is considered old and can be automatically removed from $SCRATCH according to the following rules:
- Files stored in $SCRATCH can be removed after 30 days of not being modified. This point is identical to the old $SCRATCH policy.
- Slurm workdirs, which include $SCRATCHDIR and $TMPDIR for jobs, located in $SCRATCH/slurm_jobdir/<job id>/ can be removed 7 days after the given job is finished.