Skip to content

ULHPC Slurm QoS

Quality of Service or QoS is used to constrain or modify the characteristics that a job can have. This could come in the form of specifying a QoS to request for a longer run time or a high priority queue for a given job.

To select a given QoS with a Slurm command, use the --qos=<QoS> option (available onlt in long form):

srun|sbatch|salloc|sinfo|squeue... [--partition=<partition>] --qos=<QoS> [...]

The default QoS of your jobs depends on your account and affiliation. Normally, the --qos=<QoS> directive does not need to be set for most jobs

We favor in general cross-partition QoS, mainly tied to priority level (low \rightarrow urgent). A special preemptible QoS exists for best-effort jobs and is named besteffort.

Available QoS's

QoS (partition) Prio GrpTRES MaxTresPJ MaxJobPU MaxWall
besteffort (*) 1 300 50-00:00:00
low (*) 10 4
normal (*) 100 50
long (*) 100 node=24 node=2 4 14-00:00:00
debug (interactive) 150 node=50 10
high (*) 200 50
urgent (*) 1000 100
wide (*) 100 node=160 10 0-02:00:00

List QoS Limits

Use the sqos utility function to list the existing QOS limits.

List current ULHPC QOS limits with sqos

$ sqos
# sacctmgr show qos  format="name%20,preempt,priority,GrpTRES,MaxTresPerJob,MaxJobsPerUser,MaxWall,flags"
                Name    Preempt   Priority       GrpTRES       MaxTRES MaxJobsPU     MaxWall                Flags
-------------------- ---------- ---------- ------------- ------------- --------- ----------- --------------------
              normal besteffort        100                                   100                      DenyOnLimit
          besteffort                     1                                   300 50-00:00:00            NoReserve
                 low besteffort         10                                     4                      DenyOnLimit
                high besteffort        200                                    50                      DenyOnLimit
              urgent besteffort       1000                                   100                      DenyOnLimit
               debug besteffort        150       node=50                      10                      DenyOnLimit
                long besteffort        100       node=24        node=2         4 14-00:00:00 DenyOnLimit,Partiti+
               admin besteffort       1000                                                            DenyOnLimit
                wide besteffort        100                    node=160        10    02:00:00          DenyOnLimit

What are the possible limits set on ULHPC QoS?

At the QoS level, the following elements are composed to define the resource limits for our QoS:

  • Limits on Trackable RESources TRES - a resource (nodes, cpus, gpus, etc.) tracked for usage or used to enforce limits against, in particular:
    • GrpTRES: The total count of TRES able to be used at any given time from all jobs running from the QoS; if this limit is reached new jobs will be queued but only allowed to run after resources have been relinquished from this group.
    • MaxTresPerJob: the maximum size in TRES any given job can have from the QoS.
  • MaxJobsPerUser: The maximum number of jobs a user can have running at a given time.
  • MaxWall[DurationPerJob]: The maximum wall clock time any individual job can run for in the given QoS.

As explained in the Limits section, there are basically three layers of Slurm limits, from least to most priority:

  1. None
  2. Partitions
  3. Account associations: Root/Cluster -> Account (ascending the hierarchy) -> User
  4. Job/Partition QoS

Last update: March 13, 2025