Slurm Reference

Contents


THIS SITE IS DEPRECATED

We have transitioned to another service and are no longer actively updating this site.

Refer to our new documentation at: hpcdocs.hpc.arizona.edu

This page is under construction! We cannot guarantee completion or accuracy of information herein while in development.

Overview

This page includes many commonly used Slurm options, but it is not an exhaustive list. Please refer to the official Slurm documentation for additional information.

SLURM Output Filename Patterns

SLURM offers ways to make your job's output filenames customizable through the use of character replacements. A table is provided below as a guide with some examples. Variables may be used or combined as desired. Note: character replacements may also be used with other SBATCH directives such as error filename, input filename, and job name.

VariableMeaningExample Slurm Directive(s)Output
%AA job array's main job ID

#SBATCH --array=1-2
#SBATCH -o %A.out
#SBATCH --open-mode=append

12345.out
%aA job array's index number#SBATCH --array=1-2
#SBATCH -o %A_%a.out
12345_1.out
12345_2.out
%JJob ID plus stepid#SBATCH -o %J.out12345.out
%jJob ID#SBATCH -o %j.out12345.out
%NHostname of the first compute node allocated to the job#SBATCH -o %N.outr1u11n1.out
%uUsername#SBATCH -o %u.outnetid.out
%xJobname#SBATCH --job-name=JobName
#SBATCH -o %x.out
JobName.out

Batch Job Directives

Command Purpose
#SBATCH --account=group_nameSpecify the account where hours are charged. Don't know your group name? Run the command "va" to see which groups you belong to
#SBATCH --partition=partition_nameSet the job partition. This determines your job's priority and the hours charged. See Job Partition Requests for additional information
#SBATCH --time=DD-HH:MM:SSSet the job's runtime limit in days, hours, minutes, and seconds. A single job cannot exceed 10 days or 240 hours.
#SBATCH --nodes=N

Allocate N nodes to your job.

For non-MPI enabled jobs, this should be set to "–-nodes=1" to ensure access to all requested resources and prevent memory errors.

#SBATCH --ntasks=N

ntasks specifies the number of tasks (or processes) the job will run. For MPI jobs, this is the number of MPI processes.  Most of the time, you can use ntasks to specify the number of CPUs your job needs. However, in some odd cases you might run into issues. For example, see: Using Matlab

By default, you will be allocated one CPU/task. This can be increased by including the additional directive --cpus-per-task.

The number of CPUs a job is allocated is cpus/task * ntasks, or M*N

#SBATCH --cpus-per-task=M
#SBATCH --mem=Ngb

Select N gb of memory per node. If "gb" is not included, this value defaults to MB. Directives --mem and --mem-per-cpu are mutually exclusive.

#SBATCH --mem-per-cpu=NgbSelect N GB of memory per CPU. Valid values can be found in the Node Types/Example Resource Requests section below. If "gb" is not included, this value defaults to MB.
#SBATCH --gres=gpu:N

Optional: Request N GPUs.

#SBATCH --gres=gpu:ampere:NOptional: Request N A100 GPUs.
#SBATCH --gres=gpu:volta:NOptional: Request N V100s GPUs.
#SBATCH --constraint=hi_memOptional: Request a high memory node (Ocelote and Puma only).
#SBATCH --array=N-MSubmits an array job from indices N to M
#SBATCH --job-name=JobNameOptional: Specify a name for your job. This will not automatically affect the output filename.
#SBATCH -e output_filename.err
#SBATCH -o output_filename.out
Optional: Specify output filename(s). If -e is missing, stdout and stderr will be combined.
#SBATCH --open-mode=appendOptional: Append your job's output to the specified output filename(s). 
#SBATCH --mail-type=BEGIN|END|FAIL|ALLOptional: Request email notifications. Beware of mail bombing yourself.
#SBATCH --mail-user=email@address.xyzOptional: Specify email address. If this is missing, notifications will go to your UArizona email address by default.
#SBATCH --exclusiveOptional: Request exclusive access to node.
#SBATCH --export=VAROptional: Export a comma-delimited list of environment variables to a job. 
#SBATCH --export=all (default)Optional: Export your working environment to your job.
#SBATCH --export=noneOptional: Do not export working environment to your job.

SLURM Environment Variables

VariablePurposeExample Value
$SLURM_ARRAY_JOB_IDJob array's parent ID399124
$SLURM_ARRAY_TASK_COUNTTotal number of subjobs in the array4
$SLURM_ARRAY_TASK_IDJob index number (unique for each job in the array)1
$SLURM_ARRAY_TASK_MAXMaximum index for the job array7
$SLURM_ARRAY_TASK_MINMinimum index for the job array1
$SLURM_ARRAY_TASK_STEPJob array's index step size2
$SLURM_CLUSTER_NAMEWhich cluster your job is running onelgato
$SLURM_CONFPoints to the SLURM configuration file/var/spool/slurm/d/conf-cache/slurm.conf
$SLURM_CPUS_ON_NODENumber of CPUs allocated to target node3
$SLURM_GPUS_ON_NODENumber of GPUs allocated to the target node1
$SLURM_GPUS_PER_NODENumber of GPUs per node. Only set if --gpus-per-node is specified1
$SLURM_JOB_ACCOUNTAccount being chargedgroupname
$SLURM_JOB_GPUSThe global GPU IDs of the GPUs allocated to the job. Only set in batch and interactive jobs.0
$SLURM_JOB_IDYour SLURM Job ID399072
$SLURM_JOB_CPUS_PER_NODENumber of CPUs per node. This can be a list if there is more than one node allocated to the job. The list has the same order as SLURM_JOB_NODELIST3,1
$SLURM_JOB_NAMEThe job's nameinteractive
$SLURM_JOB_NODELISTThe nodes that have been assigned to your jobgpu[73-74]
$SLURM_JOB_NUM_NODESThe number of nodes allocated to the job2
$SLURM_JOB_PARTITIONThe job's partitionstandard
$SLURM_JOB_QOSThe job's QOS/Partitionqos_standard_part
$SLURM_JOB_USERThe username of the person who submitted the jobnetid
$SLURM_JOBIDSame as SLURM_JOB_ID, your SLURM Job ID399072
$SLURM_MEM_PER_CPUThe memory/CPU ratio allocated to the job4096
$SLURM_NNODESSame as SLURM_JOB_NUM_NODES – the number of nodes allocated to the job2
$SLURM_NODELISTSame as SLURM_JOB_NODELIST, The nodes that have been assigned to your jobgpu[73-74]
$SLURM_NPROCSThe number of tasks allocated to your job4
$SLURM_NTASKSSame as SLURM_NPROCS, the number of tasks allocated to your job4
$SLURM_SUBMIT_DIRThe directory where sbatch was used to submit the job/home/u00/netid
$SLURM_SUBMIT_HOSTThe hostname where sbatch was used to submit the jobwentletrap.hpc.arizona.edu
$SLURM_TASKS_PER_NODEThe number of tasks to be initiated on each node. This can be a list if there is more than one node allocated to the job. The list has the same order as SLURM_JOB_NODELIST3,1
$SLURM_WORKING_CLUSTERValid for interactive jobs, will be set with remote sibling cluster's IP address, port and RPC version so that any sruns will know which cluster to communicate with.elgato:foo:0000:0000:000

SLURM Reason Codes

Sometimes, if you check a pending job using squeue, there are some messages that show up under Reason indicating why your job may not be running. Some of these codes are non-intuitive so a human-readable translation is provided below:

ReasonExplanation
AssocGrpCpuLimitThis is a per-group limitation on the number of CPUs that can be used simultaneously by all group members. Your job is not running because this limit has been reached. Check your group's limits using "job-limits <group_name>".

AssocGrpMemLimit

This is a per-group limitation on the amount of memory that can be used simultaneously by all group members. Your job is not running because this limit has been reached. Check your group's limits using "job-limits <group_name>".
AssocGrpCPUMinutesLimitEither your group is out of CPU hours or your job will exhaust your group's CPU hours.
AssocGrpGRESThis is a per-group limitation on the number of GPUs that can be used simultaneously by all group members. Your job is not running because this limit has been reached. Check your group's limits using "job-limits <group_name>".

Dependency

Your job depends on the completion of another job. It will wait in queue until the target job completes.

QOSGrpCPUMinutesLimitThis message indicates that your high priority or qualified hours allocation has been exhausted for the month.
QOSMaxWallDurationPerJobLimit Your job's time limit exceeds the max allowable and will never run. Generally, the limit is 10 days or 240 hours. To see an individual job's limits, run "job-limits <group_name>".

Nodes_required_for_job_are_

DOWN,_DRAINED_or_reserved_

or_jobs_in_higher_priority_

partitions

This very long message simply means your job is waiting in queue until there is enough space for it to run
PriorityYour job is waiting in queue until there is enough space for it to run.
QOSMaxCpuPerUserLimitThis is a per-user limitation on the number of CPUs that you can use simultaneously among all of your jobs. Your job is not running because this limit has been reached. Check your user limits using "job_limits <group_name>".
ReqNodeNotAvail, Reserved for maintenanceYour job's time limit overlaps with an upcoming maintenance window. Run "uptime_remaining" to see when the system will go offline. If you remove and resubmit your job with a shorter walltime that does not overlap with maintenance, it will likely run. Otherwise, it will remain pending until after the maintenance window.

Resources

Your job is waiting in queue until the required resources are available.