Configuration

Hardware Configuration

  • 48 standard compute nodes
    • 4× NVIDIA GH200 Grace-Hopper Superchip (see also node image)
      • CPU: NVIDIA Grace (Arm Neoverse-V2), 72 cores at 3.1 GHz base frequency; 120 GB LPDDR5X memory at 512 GB/s (8532 MHz)

      • GPU: NVIDIA Hopper H100, 132 multiprocessors, 96 GB HBM3 memory at 4 TB/s

      • NVIDIA NVLink-C2C CPU-to-GPU link at 900 GB/s

      • TDP: 680 W (for full GH200 superchip)

    • NVLink 4 GPU-to-GPU link, 300 GB/s between pairs of GPUs (150 GB/s per direction)

    • Network: 4× InfiniBand NDR200 (Connect-X7)

  • 1 login nodes
    • NVIDIA GH200 Grace-Hopper Superchip (differences highlighted)
      • CPU: NVIDIA Grace (Arm Neoverse-V2), 72 cores at 3.1 GHz base frequency; 480 GB LPDDR5X memory at 384 GB/s (6400 MHz)

      • GPU: NVIDIA Hopper H100, 132 multiprocessors, 96 GB HBM3 memory at 4 TB/s

      • NVIDIA NVLink-C2C CPU-to-GPU link at 900 GB/s

      • TDP: 900 W (for full GH200 superchip)

    • Infiniband NDR (Connect-X7)

    • 100 Gigabit Ethernet external connection

    • local disk for operating system (1× 960 GB NVME)

  • 13824 CPU cores

  • 200 Gbit/s network connection to JUST

_images/jupiter-node-design--jedi.svg

Node diagram of the 4× NVIDIA GH200 node design of JUPITER Booster / JEDI. Links and bandwidths are added.

Software Overview

  • RHEL 9 distribution

  • JUPITER Management Stack
    • Eviden Smart Management Center xScale

  • Scientific Software (EasyBuild)
    • GCC

    • NVIDIA HPC Compiler (_NVHPC_)

    • OpenMPI

  • IBM Storage Scale (GPFS) parallel file system