Versionen im Vergleich

Schlüssel

  • Diese Zeile wurde hinzugefügt.
  • Diese Zeile wurde entfernt.
  • Formatierung wurde geändert.

tocContent

Inhalt

Code execution

After creation of

submit the slurm job script to execute the binary on compute nodes.

Kein Format
> sbatch myjobscipt.slurm
Submitted batch job 8028673
> ls slurm-8028673.out
slurm-8028673.out

Partition for CPU CLX

The compute nodes of the CPU cluster of system Lise are organised via the following slurm Slurm partitions.

Partition name

Node
numberWalltime
count

CPU

Main memory (GB)

Max. nodes
per job

Max. jobs per user (running/ queued)

per user
Wall time limit (hh:mm:ss)Remark
standard96
cpu-clx
1204
948Cascade 9242362512

128 / 500

12:00:00default
partition
standard96
cpu-clx:test
32
16 dedicated
+128 on demand
362 161 / 50001:00:00test nodes with higher priority but
lower walltimelarge96287478128 
less wall time
cpu-clx:ssd50362
128/50012:00:00
fat memory nodeslarge96:test2 dedicated
+2 on demand
local 2TB SSD for IO
cpu-clx:large28747
2
8
1
128 / 500
10
12:00:00fat memory
test
nodes
with higher priority but lower walltimelarge96:shared2 dedicated7471128 / 50048:00:00fat memory nodes for data pre- and postprocessinghuge96

blogin1-2.nhr.zib.de
cpu-clx:huge215221128 / 50024:00:00

very fat memory nodes for data pre- and

postprocessing

post-processing

See Slurm usage how to pass the 12h walltime wall time limit with job dependencies.

Which partition to choose?

If you do not request a partition, your job will be placed in the default partition, which is standard96.

The default partitions are The default partition cpu-clx is suitable for most calculations. The :test partitions are, as the name suggests, intended for shorter and smaller test runs. These have a higher priority and a few dedicated nodes, but are limited in time and number of nodesprovide only limited resources. Shared nodes are suitable for pre- and postprocessingpost-processing. A job running on a shared node is only accounted for its core fraction (cores of job / all cores per node). All non-shared nodes are exclusive to one job only at a time.

The available home/local-ssd/work/perm storages file systems are discussed in Storage under File Systems.

An For an overview of all Slurm partitions and node statuses is provided bystatus of nodes: sinfo -r
To see For detailed information about a particular nodes type: scontrol show node <nodename>

Charge rates for accounting

Charge rates for the slurm partitions you find in Slurm partitions can be found under Accounting.

Fat-Tree Communication Network of Lise

See OPA Fat Tree network of Lise

List of CPUs

...


Short nameLink to manufacturer specificationsWhere to findUnits per node

Cores per unit

Clock speed
[GHz]

Cascade 9242Intel Cascade Lake Platinum 9242 (CLX-AP)CPU partition "Lise"2482.3
Cascade 4210Intel Cascade Lake Silver 4210 (CLX)blogin[1-
6
8]2102.2
Tesla A100NVIDIA Tesla A100 40GB and 80GB

GPU A100 partition

4

432/6912*

...