Table of Contents
Inhalt |
---|
Help and answers
For questions, please contact the support crew support@nhr.zib.de.
Login
Login authentication is possible via SSH keys only. Please visit our tutorial SSH Login.
...
Codeblock | ||||
---|---|---|---|---|
| ||||
office $ ssh -i $HOME/.ssh/id_rsa_nhr nhr_username@blogin.nhr.zib.de Enter passphrase for key '...': blogin1 $ |
File systems
Each complex has the following file systems available. More information about Quota, usage, and best pratices are available here. Hints for data transfer are given here.
...
Info |
---|
Hints for fair usage of the shared WORK ressource: Metadata Usage on WORK |
Software and environment modules
The webpage Software gives you information about available software on the NHR systems.
...
To avoid conflicts between different compilers and compiler versions, builds of most important libraries are provided for all compilers and major release numbers.
Program build
Here only a brief introduction to program building using the intel compiler is given. For more detailed instructions, including important compiler flags and special libraries, refer to our webpage Compilation Guide.
Examples for building a program on the Atos system
To build executables for the Atos system, call the standard compiler executables (icc, ifort, gcc, gfortran) directly.
...
Codeblock | ||||
---|---|---|---|---|
| ||||
module load intel icc -qopenmp -o hello.bin hello.c |
MPI, Communication Libraries, OpenMP
We provide several communication libraries:
...
OpenMP support is available with the compilers from Intel and GNU.
Using the batch system
To run your applications on the systems, you need to go through our batch system/scheduler: Slurm. The scheduler uses meta information about the job (requested node and core count, wall time, etc.) and then runs your program on the compute nodes, once the resources are available and your job is next in line. For a more in depth introduction, visit our Slurm documentation.
...
- Interactive job execution
- Job script execution
Resource specification
To request resources, there are multiple flags to be used when submitting the job.
Parameter | Default Value | |
---|---|---|
# tasks | -n # | 1 |
# nodes | -N # | 1 |
# tasks per node | --tasks-per-node # | |
partition | -p <name> | standard96/medium40 |
Timelimit | -t hh:mm:ss | 12:00:00 |
Interactive jobs
Anker | ||||
---|---|---|---|---|
|
For using compute resources interactively, e.g. to follow the execution of MPI programs, the following steps are required. Note that non-interactive batch jobs via job scripts (see below) are the primary way of using the compute resources.
...
Codeblock | ||
---|---|---|
| ||
blogin1 ~ $ salloc -t 00:10:00 -p standard96:test -N2 --tasks-per-node 24 salloc: Granted job allocation [...] salloc: Waiting for resource configuration salloc: Nodes bcn[1001,1003] are ready for job # To get a shell on one of the allocated nodes at the Berlin complex/Lise (not required for Göttingen/Emmy) blogin1 ~ $ srun --pty --interactive --preserve-env ${SHELL} bcn1001 ~ $ srun hostname | sort | uniq -c 24 bcn1001 24 bcn1003 bcn1001 ~ $ exit # Exit a second time for Berlin/Lise blogin1:~ > exit salloc: Relinquishing job allocation [...] |
Job scripts
Please go to our webpage CPU partition "Lise" for more details about job scripts. For introduction, standard batch system jobs are executed applying the following steps:
...
Erweitern | |||||||
---|---|---|---|---|---|---|---|
| |||||||
Requesting 1 large node with 96 CPUs (physical cores) for 20 minutes, and then using 192 hyperthreads
|
Job Accounting
The webpage Accounting and NPL gives you more information about job accounting.
...