Disclaimer
Ares is still under development and, even despite our best efforts, Ares might experience unscheduled outages or even data loss.
Support
Please get in touch with the PLGrid Helpdesk: https://helpdesk.plgrid.pl/ regarding any difficulties in using the cluster.
For important information and announcements, please follow this page and the messages displayed in the login message.
Machine description
Available login nodes:
- ssh <login>@ares.cyfronet.pl
Note that Ares uses PLGrid accounts and grants, make sure to request the access service called "Ares access" in the PLGrid portal.
Ares is built with Infiniband EDR interconnect and nodes of the following specification:
Partition | Number of nodes | CPU | RAM | Accelerator |
---|---|---|---|---|
plgrid and plgrid-* | 532 | 48 cores, Intel(R) Xeon(R) Platinum 8268 CPU @ 2.90GHz | 192GB | |
plgrid-bigmem | 256 | 48 cores, Intel(R) Xeon(R) Platinum 8268 CPU @ 2.90GHz | 384GB | |
plgrid-gpu-v100 | 9 | 32 cores, Intel(R) Xeon(R) Gold 6242 CPU @ 2.80GHz | 384GB | Tesla V100-SXM2 |
Job submission
Ares is using Slurm resource manager, jobs should be submitted to the following partitions:
Name | Timelimit | Remarks |
---|---|---|
plgrid | 72h | Standard partition. |
plgrid-long | 168h | Used for jobs with extended runtime. |
plgrid-testing | 1h | High priority, testing jobs, limited to 3 jobs. |
plgrid-bigmem | 72h | Jobs using an extended amount of memory. |
plgrid-now | 12h | The highest priority, interactive jobs, limited to 1 running job. |
plgrid-gpu-v100 | 72h | GPU partition. |
Accounts and computing grants
Ares uses a new scheme of naming accounts for CPU and GPU computing grants. CPU only grants are named: grantname-cpu, while GPU accounts use grantname-gpu appropriate suffix. Please mind that sbatch -A grantname
won't work on its own, you need to add the -cpu or -gpu suffix! Available computing grants, with respective account names (allocations), can be viewed by using the hpc-grants
command.
Storage
Available storage spaces are described in the following table:
Location | Location in the filesystem | Purpose |
---|---|---|
$HOME | /net/people/<login> | Storing own applications, configuration files |
$SCRATCH | /net/pr2/scratch/people/<login> | High-speed storage used for short-lived data heavily used in computations. Data present for more than 30 days can be deleted without notice. |
group storage | /net/pr2/projects/plgrid/<group name> | Long term storage, used for data living for the period of computing grant. |
Current usage, capacity and other storage attributes can be checked by issuing the hpc-fs
command.
System utilities
Please use the following commands for interacting with the account and storage management system:
hpc-grants
-
shows available grants, resource allocationshpc-fs
- shows available storagehpc-jobs
- shows currently pending/running jobshpc-jobs-history
- shows information about past jobs
Software
Applications and libraries are available through the modules system, list of available modules can be obtained by issuing the command:
module avail
a module can be loaded by the add command:
module add openmpi/4.1.1-gcc-11.2.0
and the environment can be purged by:
module purge
More information
Ares is following Prometheus' configuration and usage patterns. Prometheus documentation can be found here: https://kdm.cyfronet.pl/portal/Prometheus:Basics