Skip to content

Tier 2 HPC facilities

We have access to a number of Tier 2 clusters which are among the TOP500 list of the world's most powerful computer systems. If you are running multi-node parallel jobs you may benefit from access to these, please contact us to see if your job is appropriate and to organise access.

Who can access these clusters?

QMUL Academics may apply to use a cluster free of charge if:

  • they are performing predominantly EPSRC-funded research
  • jobs are an appropriate size for the Tier 2 service (the QMUL Apocrita service is sufficient for many users) e.g. parallel jobs running over multiple nodes
  • jobs are well-tested and known to run successfully
  • the scope and size of the work is stated in advance as part of an application to use the cluster (or by using the initial 50,000 CPUhour allocation to determine specific resource requirements)
  • the work fits the designated application areas for the cluster
  • you notify us if you don't think you will be able to use the CPUhour allocation within the agreed time frame (so that the unused hours can be allocated to other users)

Project allocations

Typically, new projects will be granted 50,000 CPUhours for benchmarking and job sizing. After this, to obtain resource allocation for your project, you will need to provide a short description of your project, along with job sizes and a commitment to use the resources within the agreed time-frame.

QMUL receive an allocation to use within a given accounting period on each cluster, which is divided among the various projects according to their requirements. At the end of each accounting period the balances are reset.

CPUhours

A CPUhour is the amount of work done by a single core in one hour. For accounting purposes, you need to calculate the cumulative total over all cores that your job runs on. If your job runs for one hour on ten 24-core nodes, the CPU time used is 240 CPUhours. Part-used nodes are counted as using all of the cores, since jobs are granted exclusive access to nodes.

Thomas - Hub in Materials and Molecular Modelling

Host Institution Cores Nodes RAM/Node Scheduler Wallclock Accounting period
UCL 17,280 720 128GB SGE 48hrs 3 months

Thomas is funded by EP/P020194/1 and is designed for materials and molecular modelling. QMUL receive 4 Million CPUhours for each 3 month accounting period.

Further information, including available software packages, is available on the UCL wiki. They also include some example job submission scripts.

Users are given 50GB of home space, and 200GB of scratch space.

Maximum job size is 864 cores. Typical job sizes are between 2-5 nodes

Athena - HPC Midlands Plus

Host Institution Cores Nodes RAM/Node Scheduler Wallclock Accounting Period
Loughborough 14,336 512 128GB Slurm 100hrs 6 months

Athena is funded by EP/P020232/1.

QMUL receive 7.4 Million CPUhours for each 6 month accounting period on Athena.

Documentation is available on the Midlands Plus pages, including a Quick Start guide.

JADE - Joint Academic Data science Endeavour

Host Institution Nodes GPUs per node Scheduler EPSRC Grant
Oxford 22 NVidia DGX-1 8 Nvidia P100 Slurm EP/P020275/1

JADE is a GPU cluster designed for machine learning and molecular dynamics applications. The Nvidia DGX-1 nodes run optimised versions of Caffe, Tensorflow, Theano and Torch for machine learning. More information is available in the JADE documentation