GPU Usage

From HPC users
Revision as of 08:31, 16 March 2020 by Thole (talk | contribs) (→‎Introduction)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search

Introduction

Since we got 12 dedicated GPU nodes (mpcg[001-009]) containing one NVIDIA Tesla P100 each and four additional nodes (mpcb[001-004]) containing two GTX 1080 each, its possible to run your jobs with one or multiple associated GPUs. The usage might not be self-explanatory, we created this guide to help you get everything set up and working properly.

How to request a GPU

In order to use GPUs for your job, you will have to request a Generic resource (GRES). You can do that by adding the following line to your job script:

#SBATCH --gres=gpu:1

This will request one GPU per requested node. Suitable nodes will be automatically chosen by SLURM but you have to select a partition with GPU nodes (see below).

Of course, it is possible to request more than one GPU per node, however, currently GPU nodes have only one or two GPUs each. With the following line of code you will request 2 GPUs per requested node:

#SBATCH --gres=gpu:2

As mentioned above, you will also have to select a partition with GPU nodes by adding the following line to your job script:

#SBATCH --partition=mpcg.p

This will allow you to use one (or more) of the Tesla P100 cards. Alternatively, you could also add the line

#SBATCH --partition=mpcb.p

which has some nodes with GTX 1080 cards.