Software Modules
This article describes how to find and use pre-installed software on JURECA.
Basic module usage
Loading a module sets environment variables to give you access to a specific set of software and its dependencies. We use a hierarchical organization of modules to ensure that you get a consistent software stack, e.g., all built with the same compiler version or all relying on the same implementation of MPI. Please note that the organization of Software on JURECA, JUWELS, JUSUF and JEDI is identical but the installed software packages may differ.
What this means on JURECA is that there are multiple compilers and MPI runtimes available.
As a JURECA user, your first task would be to load the desired compiler.
The compilers available, as well as some other compiler independent tools, can be listed with the module avail
command:
[user@system ~]$ module avail
----------------------------- Core packages ------------------------------
AMD-uProf/3.4.502
AMD-uProf/3.5.671 (D)
ARMForge/21.1.2
[...]
zarr/2.10.1
zsh/5.8
zstd/1.5.0
---------------------------------- Compilers ----------------------------------
GCC/11.2.0 Intel/2021.4.0 (L) NVHPC/22.3 (g,D)
GCCcore/.11.2.0 (H,L) NVHPC/22.1 (g)
--------------------------- Production Stages ---------------------------
Stages/2020 (S) Stages/2022 (S,L,D)
-------------------- User-based install configuration --------------------
UserInstallations/easybuild
Where:
g: built for GPU
L: Module is loaded
Aliases: Aliases exist: foo/1.2.3 (1.2) means that "module load foo/1.2" will load foo/1.2.3
D: Default Module
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
Once you have chosen a compiler you can load it with module load <compiler>
:
[user@system ~]$ module load Intel
You can verify which modules you have loaded with module list
:
[user@system ~]$ module list
Currently Loaded Modules:
1) Stages/2022 (S) 4) binutils/.2.37 (H)
2) GCCcore/.11.2.0 (H) 5) StdEnv/2022
3) zlib/.1.2.11 (H) 6) Intel/2021.4.0
Where:
H: Hidden Module
Note that the module environment loads the dependencies that are needed, even if they are hidden.
Loading the Intel compiler will give you access to a set of software compatible with your selection, that again can be listed with module avail
:
[user@system ~]$ module avail
----------- System MPI runtimes available for Intel compilers ------------
IntelMPI/2021.4.0 ParaStationMPI/5.5.0-1-mt (g)
OpenMPI/4.1.1 (g) ParaStationMPI/5.5.0-1 (g,D)
OpenMPI/4.1.2 (g,D)
------------- System packages compiled with Intel compilers --------------
METIS/5.1.0 (D) libxc/5.1.7 libxsmm/1.16.3
----------------------------- Core packages ------------------------------
AMD-uProf/3.4.502
AMD-uProf/3.5.671 (D)
ARMForge/21.1.2
[...]
zarr/2.10.1
zsh/5.8
zstd/1.5.0
------------------------------- Compilers -------------------------------
GCC/11.2.0 Intel/2021.4.0 (L) NVHPC/22.3 (g,D)
GCCcore/.11.2.0 (H,L) NVHPC/22.1 (g)
--------------------------- Production Stages ---------------------------
Stages/2020 (S) Stages/2022 (S,L,D)
-------------------- User-based install configuration --------------------
UserInstallations/easybuild
Where:
g: built for GPU
L: Module is loaded
Aliases: Aliases exist: foo/1.2.3 (1.2) means that "module load foo/1.2" will load foo/1.2.3
D: Default Module
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
Among these newly available modules, the most important ones are the MPI runtimes (which appear at the top of the available software).
Loading an MPI runtime will again give you access to software built on top of that runtime.
Please note that when loading a module where multiple versions are available, the default one is the one with a (D)
at its side.
[user@system ~]$ module load ParaStationMPI
[user@system ~]$ module avail
------------------------ ParaStationMPI settings -------------------------
mpi-settings/CUDA mpi-settings/plain mpi-settings/UCX (L,D)
---- System packages compiled with ParaStationMPI and Intel compilers ----
ABINIT/9.6.2 SUNDIALS/6.1.0
ADIOS2/2.7.1 Scalasca/2.6
ARPACK-NG/3.8.0 Score-P/7.1
[...]
SCOTCH/6.1.2 netcdf4-python/1.5.7 (D)
SIONfwd/1.0.1 sprng/1
SIONlib/1.7.7 sprng/5-10062021 (D)
--------------------- Settings for software packages ---------------------
UCX-settings/RC-CUDA UCX-settings/DC UCX-settings/UD
UCX-settings/RC (L,D) UCX-settings/plain
UCX-settings/DC-CUDA UCX-settings/UD-CUDA
----------- System MPI runtimes available for Intel compilers ------------
IntelMPI/2021.4.0 ParaStationMPI/5.5.0-1-mt (g)
OpenMPI/4.1.1 (g) ParaStationMPI/5.5.0-1 (g,L,D)
OpenMPI/4.1.2 (g,D)
------------- System packages compiled with Intel compilers --------------
METIS/5.1.0 (D) libxc/5.1.7 libxsmm/1.16.3
----------------------------- Core packages ------------------------------
AMD-uProf/3.4.502
AMD-uProf/3.5.671 (D)
ARMForge/21.1.2
[...]
zarr/2.10.1
zsh/5.8
zstd/1.5.0
------------------------------- Compilers -------------------------------
GCC/11.2.0 Intel/2021.4.0 (L) NVHPC/22.3 (g,D)
GCCcore/.11.2.0 (H,L) NVHPC/22.1 (g)
--------------------------- Production Stages ---------------------------
Stages/2020 (S) Stages/2022 (S,L,D)
-------------------- User-based install configuration --------------------
UserInstallations/easybuild
Where:
g: built for GPU
L: Module is loaded
Aliases: Aliases exist: foo/1.2.3 (1.2) means that "module load foo/1.2" will load foo/1.2.3
D: Default Module
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
Sometimes, as a user, you simply want to find out which modules you have to load to enable the loading of a particular software package or application.
module spider
can help you with that task.
It will look in the whole hierarchy and report back with specific module combinations to enable the loading of that package:
[user@system ~]$ module spider gromacs
----------------------------------------------------------------------------
GROMACS:
----------------------------------------------------------------------------
Description:
GROMACS is a versatile package to perform molecular dynamics, i.e.
simulate the Newtonian equations of motion for systems with hundreds
to millions of particles. It is primarily designed for biochemical
molecules like proteins and lipids that have a lot of complicated
bonded interactions, but since GROMACS is extremely fast at
calculating the non-bonded interactions (that usually dominate
simulations) many groups are also using it for research on
non-biological systems, e.g. polymers.
Versions:
GROMACS/2019.1
GROMACS/2019.3
GROMACS/2020.4-plumed
GROMACS/2020.4
GROMACS/2021.2
GROMACS/2021.4-plumed
GROMACS/2021.4
----------------------------------------------------------------------------
For detailed information about a specific "GROMACS" package (including how to load the modules)
use the module's full name.
Note that names that have a trailing (E) are extensions provided by other modules.
For example:
$ module spider GROMACS/2021.4
----------------------------------------------------------------------------
Currently there are more than 700 packages installed per Stage (see Stages).
To keep a clean and uncluttered view, a significant number of these packages (mostly helper libraries) are hidden.
If you want to see them you can do it with module --show-hidden avail
:
[user@software ~]$ module --show-hidden avail
------------------------ ParaStationMPI settings -------------------------
mpi-settings/CUDA mpi-settings/plain mpi-settings/UCX (L,D)
---- System packages compiled with ParaStationMPI and Intel compilers ----
ABINIT/9.6.2 SUNDIALS/6.1.0
ADIOS2/2.7.1 Scalasca/2.6
ARPACK-NG/3.8.0 Score-P/7.1
[...]
SCOTCH/6.1.2 netcdf4-python/1.5.7 (D)
SIONfwd/1.0.1 sprng/1
SIONlib/1.7.7 sprng/5-10062021 (D)
--------------------- Settings for software packages ---------------------
UCX-settings/RC-CUDA UCX-settings/DC UCX-settings/UD
UCX-settings/RC (L,D) UCX-settings/plain
UCX-settings/DC-CUDA UCX-settings/UD-CUDA
----------- System MPI runtimes available for Intel compilers ------------
IntelMPI/2021.4.0 ParaStationMPI/5.5.0-1-mt (g)
OpenMPI/4.1.1 (g) ParaStationMPI/5.5.0-1 (g,L,D)
OpenMPI/4.1.2 (g,D)
------------- System packages compiled with Intel compilers --------------
METIS/5.1.0 (D) libxc/5.1.7 libxsmm/1.16.3
------------------------- System side compilers -------------------------
AOCC/3.1.0 AOCC/3.2.0 (D) Clang/13.0.1
----------------------------- Core packages -----------------------------
ACTC/.1.1 (H)
AMD-uProf/3.4.502
AMD-uProf/3.5.671 (D)
ANTLR/.2.7.7 (H)
[...]
zfp/.0.5.5 (H)
zlib/.1.2.11 (H)
zlib/.1.2.11 (H,L)
zsh/5.8
zstd/1.5.0
------------------------------- Compilers -------------------------------
GCC/11.2.0 Intel/2021.4.0 (L) NVHPC/22.3 (g,D)
GCCcore/.11.2.0 (H,L) NVHPC/22.1 (g)
--------------------------- Production Stages ---------------------------
Stages/2020 (S) Stages/2022 (S,L,D)
----------------- User-based install configuration -----------------
UserInstallations/easybuild
Where:
g: built for GPU
L: Module is loaded
Aliases: Aliases exist: foo/1.2.3 (1.2) means that "module load foo/1.2" will load foo/1.2.3
D: Default Module
H: Hidden Module
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
Available compilers
JURECA has 3 major compilers available: GCC, Intel and NVHPC. With these compilers we build full toolchains (MPIs, math libraries, applications, etc). Additionally, AOCC and Clang are also available.
The table shows the particular compiler, MPI and basic mathematical libraries (BLAS, LAPACK, FFTW, ScaLAPACK) combinations that have been made available on JURECA. Note that at the moment Intel MKL is the primary math library, but the deployment of BLIS and LibFLAME is planned for the near future.
Compiler |
MPI |
Math library |
---|---|---|
GCC |
BullMPI |
Intel MKL |
GCC |
OpenMPI |
Intel MKL |
GCC |
ParaStationMPI |
Intel MKL |
Intel |
IntelMPI |
Intel MKL |
Intel |
OpenMPI |
Intel MKL |
Intel |
ParaStationMPI |
Intel MKL |
Intel |
ParaStationMPI-mt 2 |
Intel MKL |
NVHPC |
ParaStationMPI |
Intel MKL |
NVHPC |
OpenMPI |
Intel MKL |
MPI runtimes
JURECA has 2 major MPI runtimes available, ParaStationMPI
and OpenMPI
. Both are CUDA-aware (ie: they can directly communicate buffers placed in GPU memory).
Additionally, IntelMPI
and BullMPI
(with CUDA-awareness) are available on JURECA-DC.
All MPI runtimes load a default MPI-settings
module (mpi-settings
in older stages), that configures the runtime for most users. JSC provides a set of these modules that accomodate a few different use cases. Note that MPI runtimes are highly configurable, and the configuration modules provided are non-exhaustive.
These are the modules available per runtime. Note that older stages can have slighty different options.
For
ParaStationMPI
there are 2 possibilities:MPI-settings/UCX
: This module configures the runtime to run in JURECA. It implies usingUCX
as communication library. This is the default. If you wish to use the GPUs on this part, please load the next module.MPI-settings/CUDA
: This module is the same asMPI-settings/UCX
, but it makes sure that theUCX
module is loaded with the CUDA transports enabled and disables the shared memory plugin in pscom (in favour of the equivalent inUCX
).
For
OpenMPI
there are 4 possibilities:MPI-settings/UCX
: This module configures the runtime to run in JURECA. It implies usingUCX
as communication library. Extra options can be examined withompi_info -a
. This is the default module. If you wish to use the GPUs please load the next module.MPI-settings/CUDA
: This module is the same asMPI-settings/UCX
, but it makes sure that theUCX
module is loaded with the CUDA transports enabled, effectively enabling CUDA-awareness.MPI-settings/UCX-UCC
: This module is equivalent toMPI-settings/UCX
, with the difference thatUCC
is prioritized and used by default.MPI-settings/CUDA-UCC
: This module is equivalent toMPI-settings/CUDA
, with the difference thatUCC
is prioritized and used by default.
For
BullMPI
the situation is the same as forOpenMPI
, with the exception thatUCC
is not available:MPI-settings/UCX
: This module configures the runtime to run in JURECA. It implies usingUCX
as communication library. Extra options can be examined withompi_info -a
. This is the default module. If you wish to use the GPUs please load the next module.MPI-settings/CUDA
: This module is the same asMPI-settings/UCX
, but it makes sure that theUCX
module is loaded with the CUDA transports enabled, effectively enabling CUDA-awareness.
For
IntelMPI
there is one possibility:MPI-settings/UCX
: This is the default, which enables sets thelibfabric
provider tomlx
, ie: communication will happen via theUCX
library.
UCX
is by far the most used communication library in all our MPIs. UCX
is also highly configurable. That’s why we provide also modules for configuring UCX
. These modules are also non-exhaustive. The recommended default is loaded automatically for each MPI. The list of modules is:
UCX-settings/RC-CUDA
enables the accelerated RC (Reliable Connected) and CUDA transports. RC is the default transport and the recommended for most cases.UCX-settings/RC
enables the accelerated RC (Reliable Connected) transport. Select this if you see warnings regarding initializing CUDA transports in nodes without GPUs. RC is the default transport and the recommended for most cases.UCX-settings/UD-CUDA
enables the accelerated UD (Unreliable Datagram) and CUDA transports. UD has a lower memory footprint than RC, and could be recommended for medium size simulations.UCX-settings/UD
enables the accelerated UD (Unreliable Datagram) transport. UD has a lower memory footprint than RC, and could be recommended for medium size simulations.UCX-settings/DC-CUDA
enables the DC (Dynamically Connected) and CUDA transports. DC is a relatively new transport, and it has not been tested exhaustively in our systems. It’s memory footprint is low, and could be recommended for very large simulations.UCX-settings/DC
enables the DC (Dynamically Connected) transport. DC is a relatively new transport, and it has not been tested exhaustively in our systems. It’s memory footprint is low, and could be recommended for very large simulations.UCX-settings/plain
basically disables the restriction of transports, soUCX
can decide on its own which transport should be used. Use this if you would like to rely onUCX
’s heuristics. This is equivalent to unloading theUCX-settings
module.
To see which options are enabled in each MPI runtime or UCX
you can type ml show MPI-settings
or ml show UCX-settings
. To see a full list of UCX
options you can type ucx_info -c -f
Warning
Since 2023-08-30 the UCX-settings/*CUDA
modules also set UCX_RNDV_FRAG_MEM_TYPE=cuda
. This enables the GPU to initiate transfers of CUDA managed buffers.
This can have a large speed-up in case Unified Memory (cudaMallocManaged()
) is used, as staging of data is avoided. Our testing indicates performance gains,
especially when communicating between the GPUs in a node, since NVLink is used effectively. We don’t expect any performance degradation for other codes; if you do notice
worse performance, please contact sc@fz-juelich.de
GPUs and modules
Software with specific GPU support are marked with a (g)
at their side when listing modules.
They can be reached loading the compilers listed in the table of the previous section.
Finding software packages
There are 3 commands that are the main tools to locate software on JURECA:
module avail
module spider <software>
module key <keyword or software>
Normally, the first 2 are enough.
Occasionally, module key
can be necessary to look for keywords or packages bundled in a single module.
An example would be numpy
, which is included in the SciPy-Stack
module.
In the example below, the module environment looks for all occurrences of numpy
in the description of the modules.
That helps to locate SciPy-Stack
.
[user@system ~]$ module key numpy
----------------------------------------------------------------------------
The following modules match your search criteria: "numpy"
----------------------------------------------------------------------------
SciPy-Stack: SciPy-Stack/2020-Python-3.8.5, SciPy-Stack/2021-Python-3.8.5
SciPy Stack is a collection of open source software for scientific computing
in Python.
SciPy-bundle: SciPy-bundle/2021.10
Bundle of Python packages for scientific software
[...]
Additionally, the complete list of software installed can be checked online in the JURECA module browser.
Stages
JURECA will go through major scientific software updates once a year in November, at the same time that new projects start their allocation time. We call these updates Stages. During these stage switches, the available software will be updated to the latest stable releases. Typically this will require that user applications are recompiled. In such cases, there are two possible solutions:
Load the new versions of the required dependency modules and recompile.
Load the old Stage.
To load the old Stage, users should use these commands:
[user@system ~]$ module use $OTHERSTAGES
[user@system ~]$ module load Stages/2019a
Then the old software view will become available again as before the stage switch. In the example above the desired Stage was 2019a, but as new stage transitions happen more possibilities will be available.
Stages Changelog
The changes of the stages are documented in the changelog of stages.
Scientific software at JSC
JSC provides a significant amount of software installed on its systems. In Scientific Application Software you can find an overview of what is supported and how to use it.
Requesting new software
It is possible to request new software to be installed in JURECA. To do that please send an email to sc@fz-juelich.de, describing which software and version you need. Please note that this will be done on a “best effort” basis and might have limited support.
Installing your own software with EasyBuild
Starting with the 2023 stage, it is possible to install software with [EasyBuild](https://easybuild.io/) in a directory of your choice, leveraging the software that is already available on the system. For that, there are 2 steps you have to follow:
1. First you have to specify a installation directory. You can choose between one of the two possibilities: a) exporting USERINSTALLATIONS b) creating a symbolic link in ~/easybuild Depending on your needs, you might want to have that in a private directory, or in a directory shared with other users or members of a project. Please take into account that the link should not point to a directory in $HOME, as the quota in $HOME is small and the filesystem is not intended to be used this way. Well suited is for example the filesystem $PROJECT or $USERSOFTWARE (you need to apply for a data project in order to use it).
Load the UserInstallations module. That will load EasyBuild and configure it in a way that allows it to reuse the stack already in place. You can use this to install versions of the software that you might need, that are not available in the general stack.
If you would like to use this feature before the 2023 stage is made the default, you can do it simply by typing ml use /p/software/$SYSTEMNAME/userinstallations
If you want to follow the same approach in other systems, you can have a copy of our EasyBuild setup in https://github.com/easybuilders/JSC. After cloning it in a location of your choice, you would have to configure EasyBuild appropriately to leverage it. A good starting point would be to take a look at ml show UserInstallations and adapt variables as required.
[user@system ~]$ export USERINSTALLATIONS=/p/project/yourproject/${USER}
[user@system ~]$ ml UserInstallations
- Loading default (typically latest) EasyBuild module
- Enabling bash tab completion for EasyBuild and EasyConfigs
(You can enable easyconfig autocomplete from robot
search path in addition by setting JSC_EASYCONFIG_AUTOCOMPLETE)
Found $USERINSTALLATIONS in the environment, using that for installation. To override this and do
a personal installation set the environment variable PREFER_USER=1 and reload this module.
** LOADING USERSPACE DEVELOPER CONFIGURATION **
Preparing the environment for software installation via EasyBuild into userspace leveraging stage 2022
- Adding our license servers to LM_LICENSE_FILE
- Giving priority to JSC custom Toolchains (EASYBUILD_INCLUDE_TOOLCHAINS)
- Giving priority to JSC custom EasyBlocks (EASYBUILD_INCLUDE_EASYBLOCKS)
- Giving priority to JSC custom easyconfigs (EASYBUILD_ROBOT)
- Allowing searching of distribution easyconfigs (EASYBUILD_SEARCH_PATHS)
- To keep module view clean, hiding some dependencies (EASYBUILD_HIDE_DEPS)
- Using JSC EasyBuild hooks (EASYBUILD_HOOKS)
- Setting module classes to include side compilers (EASYBUILD_MOODULECLASSES)
- Setting EASYBUILD_PARALLEL to 8
- Setting EASYBUILD_OPTARCH to GCCcore:march=haswell -mtune=haswell
- Setting EASYBUILD_CUDA_COMPUTE_CAPABILITIES to 7.0,6.0
- Using shared group installation, therefore expanding dependency searching
- To allow collaboration in the development process, all installations are
group-writable! (EASYBUILD_GROUP_WRITABLE_INSTALLDIR)
Note: If you wish to submit software builds to slurm with the '--job' flag you will need to set environment
variables to configure job submission, see
https://slurm.schedmd.com/sbatch.html#lbAJ
for details.
To use a module that you or a college has installed in a group readable directory, you need to either export USERINSTALLATIONS or create a symbolic link in ~/easybuild pointing at the correct directory. Module path expansion when loading a module will allow you to load the modules installed with UserInstallations. If you have problems to load the modules compiled with UserInstallations, please try to reload the already loaded modules in your environment. 1 This can for example be done with ml update.
In case of questions or for more advanced use cases please to not hesitate and reach out to sc@fz-juelich.de.
- 1
This can happen if you set USERINSTALLATIONS after you have loaded the modules that expand the path. The modules that expand the path are, for example, compiler modules (like GCC or GCCcore) or MPI modules (like OpenMPI or ParaStationMPI), as they build the base level of our EasyBuild installation (the toolchain).