This is an old revision of the document!
Partitions and Accounting
This page explains the partitions available to users and the accounting for each partition. This page assumes knowledge on partition usage and how to submit a job using SLURM. Please refer to the HPC user guide for a general introduction to these topics.
partitions
Compute nodes are grouped into partitions in order to allow the user to select different hardware to run their software on. Each partition includes a subset of nodes with a different type of hardware and a specific maximum wall time.
The partitions can be selected by users via the SLURM option:
#SBATCH --partition=<partition name>
or its short form:
#SBATCH -p <partition name>
The partitions available on Grid.UP are summarised in the table below. For details of the different hardware available on each node, please look at the Grid.UP hardware page.
Partition name | Node type | Smallest possible allocation | Max wall time | Notes |
---|---|---|---|---|
batch | thin compute nodes | 1 cores + 500 MB memory | 5 days | default partition for small jobs |
big | > 16 cores | 16 cores + 16 GB memory | 7 days | only for jobs with > 16 cores or GPU |
ceft | resticted | 1 core + 1 GB memory | 28 days | exclusive for ceft registered users |
lsrelcm | resticted | 1 core + 1 GB memory | 28 days | exclusive for lsre/lcm registered users |
accounting
<tbd>