Versionen im Vergleich

Schlüssel

  • Diese Zeile wurde hinzugefügt.
  • Diese Zeile wurde entfernt.
  • Formatierung wurde geändert.

...

Program build and execution

  • Each node of the GPU A100 system is a combination of a host CPU and their four attached device GPUs. There is a wide range of software to support this hardware.
  • We recommend to use the GPU A100 login nodes for program build. If a program build needs for the presence of CUDA drivers, compilation is possible on a compute node within a slurm job session, too.
  • We restrict our presentation to examples. For that, please visit our manual on

Job monitoring

A running job can be monitored interactively, directly on each of the compute nodes. Once you know the names of the job nodes you can login and monitor the host CPU as well as the GPUs.

...

Software and environment modules

  • Login and compute nodes of the A100 GPU partition are running under Rocky Linux (currently version 8.6).
  • Software for the A100 GPU partition provided by NHR@ZIB can be found using the module command, see Quickstart.
  • Please note the presence of the sw.a100 environment module. It controls the software selection for the GPU A100 partition.
Codeblock
languagetext
titleExample: Show the currently available software and access compilers
bgnlogin1 $ module avail
...
bgnlogin1 $ module load gcc
...
bgnlogin1 $ module list
Currently Loaded Modulefiles:
 1) HLRNenv   2) sw.a100   3) slurm   4) gcc/11.3.0(default)

(Glühbirne) Please note the presence of the sw.a100 environment module. When loaded, environment modules are shown for software installed for the NVidia A100 GPU partition. This is the default setting on the A100 GPU login and compute nodes.

...

Using the batch system

The GPU nodes are available via partitions of the batch system slurm.

...