Versionen im Vergleich

Schlüssel

  • Diese Zeile wurde hinzugefügt.
  • Diese Zeile wurde entfernt.
  • Formatierung wurde geändert.

...

The OS migration phase starts with a small number of two login nodes and one rack of 96 compute nodes already running Rocky Linux 9 for testing, while the other nodes remain available under CentOS 7. After an initial testing period, a major fraction of nodes will be switched to the new OS Rocky Linux 9 to allow for general job production under Rocky Linux 9the new OS. In the last step, the few nodes still under CentOS 7 will be converted to Rocky Linux 9, too. During the migration phase the use of Rocky Linux 9 "clx" compute nodes will be free of charge.

...

nodes

CentOS 7

Rocky Linux 9

login

blogin[1-6]

blogin[7-8]

compute (384 GB RAM)

932852

1696

compute (768 GB RAM)

32

0

compute (1536 GB RAM)

2

0

...

CentOS 7

Rocky Linux 9

TMPDIR=/scratch/tmp/$USER

(undefined, local /tmp is used)

SLURM_MPI_TYPE=pmi2

SLURM_MPI_TYPE=pmix

(undefined)

I_MPI_PMI_LIBRARY=<path-to>/libpmix.so

(undefined, defaults to psm2)

FI_PROVIDER=opx

(undefined)

NHRZIB_ARCH=clx

(undefined)

NHRZIB_OS=el9

(undefined)

NHRZIB_TARGET=clx.el9

SLURM

...

partitions

CentOS 7

Rocky Linux 9

standard96

cpu-clx

standard96:test

cpu-clx:test

standard96:ssd

cpu-clx:ssd

large96

cpu-clx:large

large96:test

large96:shared

huge96

cpu-clx:huge

...

  • For users of SLURM’s srun job launcher:
    Open MPI 5.x has dropped support for the PMI-2 APAPI, it solely depends on PMIx to bootstrap MPI processes. For this reason the environment setting was changed from SLURM_MPI_TYPE=pmi2 to SLURM_MPI_TYPE=pmix, so binaries linked against Open MPI can be started as usual “out of the box” using srun mybinary. For the case of a binary linked against Intel-MPI, this works too when a recent version (≥2021.11) of Intel-MPI has been used. If an older version of Intel-MPI has been used, and relinking/recompiling is not possible, one can follow the workaround for PMI-2 with srun as described in the Q&A section below. Switching from srun to mpirun instead should also be considered.

...

  • log in to an already migrated login node and get familiar with the new environment

  • check self-compiled software for continued operability

  • relink/recompile software as needed

  • adapt and test job scripts and workflows

  • submit test jobs to the "cpu-clx:test" SLURM partition

  • read the Q&A section and ask for support in case of further questions, problems, or software requests (support@nhr.zib.de)

...