Check status of submitted jobs¶
To monitor jobs you have submitted to the queue you can run the
command. The below example assumes your username is
$ qstat job-ID prior name user state submit/start at queue slots ja-task-ID -------------------------------------------------------------------------------------- 123456 5.76204 exampleA abc123 r 01/09/2020 23:10:10 all.q@sdx11 5 16 123456 5.76204 exampleA abc123 r 01/09/2020 23:54:11 all.q@sdx18 5 17 123450 5.10725 exampleB abc123 r 01/09/2020 12:05:39 all.q@sdx2 1 123461 5.00001 exampleC abc123 qw 01/09/2020 15:03:24 8
Your job will be in one of the following states when running
r- Job is currently running.
qw- Job has been submitted and is waiting in the queue.
hqw- Job has been submitted but is being held in the queue.
Eqw- Job has been submitted but an error is preventing job from starting.
A job could be in
hqw state because of a job dependency or a hold applied by
ITS Research if there is an issue with the job. If you are unsure why your job
Eqw state, you may
contact us for further information.
To see the status of a particular job (e.g. job 999) you can run the
qstat command with the
qstat -j 999
To see the resources you are currently requesting you can run the
command with the
More details on the
qstat command are available in the
Jobs with error statuses¶
qstat output may include jobs with a status of Eqw. This
indicates that an error occurred - not an error within the job itself, but one
that prevented the job from being started. Typically, this may be because the
user ran out of file space.
The jobs with errors can be deleted from the queue using:
qdel <job ID>
Alternatively, if the cause of the error has been cleared and the jobs need to run, the error state can be cleared using:
qmod -cj <job ID>
Checking where my job is in the queue¶
We provide some additional commands to display current activity of the cluster
and the queues in a more readable format than the standard
nodestatus will show the current core and memory usage of all of
the nodes. Note that some nodes are on queues that may not be available to all
users - these are coloured blue in the list of nodes.
gives a detailed view of jobs running on each node, plus the number of slots
used, and the latest finish time of the job based on the maximum requested
runtime. Below demonstrates two of the most commonly used options.
-N option allows you to inspect the jobs running on a selected node. For
nodestatus -N sdx1 Node cores memory used/total used/available/total sdx1 (36/36) (40/142/384) abc123 all.q 4008031 Thu, 22 Jun 2017 15:20:39 1 xyz126 all.q 4012201 Mon, 26 Jun 2017 17:08:21 1 xyz126 all.q 4012202 Tue, 27 Jun 2017 05:02:36 1 abc985 all.q 4015392 Sat, 24 Jun 2017 16:03:05 8 hij208 all.q 4019133 Wed, 21 Jun 2017 13:48:18 1
-T options will display a summary or running jobs on a given
node type (i.e. nxv) or group type (serial, parallel or gpu). For example:
nodestatus -T gpu Node cores memory used/total used/available/total nxg1 (32/32) (10/3/256) abc123 all.q 606585 Wed, 13 Mar 2019 17:54:14 16 abc123 all.q 606588 Wed, 13 Mar 2019 17:54:21 16 nxg2 (32/32) (8/242/256) xyz126 all.q 618546 Fri, 15 Mar 2019 15:57:34 32 nxg3 (32/32) (10/3/256) xyz126 all.q 605681 Wed, 13 Mar 2019 17:53:59 16 abc985 all.q 605683 Wed, 13 Mar 2019 17:54:07 16 nxg4 (32/32) (10/3/256) hij208 all.q 607750 Thu, 14 Mar 2019 03:31:36 16 hij208 all.q 607751 Thu, 14 Mar 2019 03:31:38 16 sbg1 (32/32) (17/131/384) hij208 all.q 605789 Wed, 13 Mar 2019 20:57:06 16 hij208 all.q 605792 Thu, 14 Mar 2019 11:00:03 16 sbg2 (32/32) (114/248/384) abc123 all.q 618755 Thu, 14 Mar 2019 14:43:22 16 xyz126 all.q 617451 Wed, 13 Mar 2019 18:20:54 16
memory used is the actual RAM in use on the node at that moment. The
memory available value is the approximate amount of RAM available
for use by the scheduler. Note that the sum of these values may not equal the
total memory on the node - for example, if a job has requested memory but not
nodestatus command usage
nodestatus -h will provide a summary of available options.
showqueue is another useful command that shows all of the jobs waiting to
be run, in the order of priority. It is useful for inspecting typical wait
times for different job sizes.
Note that some jobs may not be running because they are restricted by resource
showqueue -F gives additional detail on each queued job, such as the
total RAM requested, and resource quotas. While most users won't hit core or
memory quotas, they can be inspected using the
Jobs which are coloured yellow in the 'Submission Time' field have been queuing for 24 hours. Jobs coloured in red have been queuing for 7 days.
Should your job remain in the queue for a long period of time, please inspect
the output of
showqueue -F and check your job submission carefully.
Typically, long wait times are caused by large resource requests, exclusive
node access or errors in your job submission script. Please contact
ITS Research Support with a job id
number for assistance regarding a specific queued job.
showqueue command usage
showqueue -h will provide a summary of available options.
You can request email notifications of a change in status of your jobs by adding the following code to the "Grid Engine options" section of your submission script:
#$ -m bea # Send email at the beginning and end of the job and if aborted #$ -M email@example.com # The email address to notify
If you do not add the
-M line you will be emailed at the address
that is registered with us (usually your QMUL email address).
Check where jobs are running¶
hosts_for_job utility script displays information about which host(s)
a serial or parallel job is running on. Passing the
-s switch will print
a list of hostnames only.
run inside a UGE job
hosts_for_job utility must be run inside a UGE job therefore, you
must execute the script interactively in a QLOGIN job or within your job
script for batch submissions.
For example, to print full details about a running serial job:
$ hosts_for_job Job 457773 is running in the smp PE using 8 core(s) on host sdx6
For example, to print hostnames only for a running parallel job:
$ hosts_for_job -s sdv2 sdv3 sdv4 sdv14