Checking Job status

To check your job(s) status, using myqueue command.

For example,

hpcuser@x3002c0s7b0n0:~>myqueue JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON) 29745 compute sander.M hpcuser R 7:33 1 lanta-c-065 29746 compute sleep hpcuser R 0:09 1 lanta-c-065

Where,

JOBID

job ID of a given job

PARTITION

running partition of a given job

NAME

name of a given job

USER

user who run the given job

ST

state of a given job (R: running, PD: Pending)

TIME

running time of a given job

NODE

number of nodes used by a given job

NODELIST(REASON)

list of node(s) used by a given job (Reason why given is in PD (Pending) state. 


NODELIST (REASON)

Resources

Refers to the job being the first in the queue waiting to be executed. The job will run automatically when the system has sufficient resources as specified in the script submitted for execution.

Priority

Refers to the job being in the queue, waiting to be executed in another order. It will be executed after jobs with the "Resources" status and when the system has enough resources as specified in the script submitted for execution.

QOSGrpBillingMinutes

Means the job has not entered the queuing system because there are not enough service units hour (SHr) available for the job to run (you can check the SU status of the project using the sbalance command). The job's status will automatically change to another state when there are enough SHr available for that particular job to run.

FAQ: Why is the job still in the QOSGrpBillingMinutes state when there are available SHr?

The behavior of Slurm considers whether there are enough remaining SU to use all the resources specified in the script. For example, if you specify 1 GPU node 4 GPU cards (-p gpu -N 1 --gpus=4) for a duration of 1 day (-t 24:00:00), there must be more than 72 SHr (3 x 24) in the account to avoid getting into the QOSGrpBillingMinutes state. Therefore, it is essential to adjust the resource usage time appropriately for each job, especially when there are limited remaining SU.

PartitionTimeLimit

Refers to the job not entering the queuing system because the specified time limit for resource usage is greater than the defined value. In this case, you must cancel the job (scancel), modify the script used for submission (-t), and submit the job again. The maximum time limit in each partition can be checked using the sinfo command.

MaxCpuPerAccount

Means the job has not entered the queuing system because the number of CPUs/Cores exceeds the policy defined for job submission. The job's status will automatically change to another state when the number of CPUs/Cores running in the system decreases.

MaxJobPerAccount

Means the job has not entered the queuing system because the number of jobs exceeds the policy defined for job submission. The job's status will automatically change to another state when the number of jobs running in the system decreases.