Skip to content

List of metrics

Here we list the current metrics for each section and detail their meaning.

  • Usage: Average core usage over the runtime of the job per node (y-axis) and per core (x-axis) of the node.

Warning

The abscissa in this graph are the cores instead of the timestamp, and it includes both "Physical" cores (first half) as well as the "Logical" ones (second-half).

  • Usage: 1-min average usage of the CPU across all cores in a node. As the CPU can use simultaneous multithreading (SMT) using physical and logical cores, the value can go up to 200%.

  • Physical Cores Used: Numbers of "Physical cores" with usage above 25% in the last minute. The "Physical cores" represents the first half of the node.

  • Logical Cores Used: Numbers of "Logical cores" with usage above 25% in the last minute. The "Logical cores" represents the second half of the node.

  • Load: The CPU load is a measure of the amount of computational work that a computer system performs (1 means a load of 1 core on average). It is a quantity provided by Linux in three numbers: 1-, 5- and 15-min average loads. The load is then collected by LLview via Slurm.

Note

In the job reports, the CPU Load is obtained from Slurm, which at JSC contains the 1-min load average.

  • Memory Usage: Amount of allocated RAM memory (in GiB) in the node.

Note

In the job reports, the Memory Usage graphs (both for CPU and GPU) is scaled by default from 0 up to the memory limit of the partition. A swich between Job and System limits can be found on the interactive reports.

Danger

Some system processes may use up to a few GiB of memory on the system, so it is better to plan for 10-15GiB less than the maximum amount.

  • Utilization: Percent of time over the past sample period during which one or more kernels was executing on the GPU.

Warning

The Utilization graph reflect the usage of at least one kernel on the GPU - it does not contain information of how much occupied it is.

  • Memory Usage: Amount of memory (in GiB) used on the device by the context.
  • Power: Current power (in Watts) being used by the GPU.
  • Temperature: Current Temperature (in Celsius) on a given GPU.

Warning

Note that high temperatures may trigger slow down of the GPU frequency (see examples of High Temperature / GPU Throttling).

  • Clk Throttle Reason: Information about factors that are reducing the frequency of clocks. These are:

    1. GpuIdle - Nothing is running on the GPU and the clocks are dropping to Idle state.
    2. AppClkSet - GPU clocks are limited by applications clocks setting.
    3. SwPwrCap - SW Power Scaling algorithm is reducing the clocks below requested clocks because the GPU is consuming too much power.
    4. HWSlowDown - HW Slowdown (reducing the core clocks by a factor of 2 or more) is engaged. This is an indicator of:
                    * Temperature being too high
                    * External Power Brake Assertion is triggered (e.g. by the system power supply)
                    * Power draw is too high and Fast Trigger protection is reducing the clocks
    5.  SyncBoost - This GPU has been added to a Sync boost group with nvidia-smi or DCGM in order to maximize performance per watt. All GPUs in the sync boost group will boost to the minimum possible clocks across the entire group. Look at the throttle reasons for other GPUs in the system to see why those GPUs are holding this one at lower clocks.
    6.  SwThermSlDwn - SW Thermal Slowdown. This is an indicator of one or more of the following:
                       * Current GPU temperature above the GPU Max Operating Temperature
                       * Current memory temperature above the Memory Max Operating Temperature
    7.  HwThermSlDwn - HW Thermal Slowdown (reducing the core clocks by a factor of 2 or more) is engaged. This is an indicator of:
                       * Temperature being too high
    8.   PwrBrakeSlDwn - Power brake throttle to avoid that given racks draw more power than the facility can safely provide.
    

Note

The Clk Throttle Reason graphs are not shown when no throttling was ever active for the job.

  • StreamMP Clk: Current frequency in MHz of SM (Streaming Multiprocessor) clock. The frequency may be slowed down for the reasons given above.
  • Memory Usage Rate: Percent of time over the past sample period during which global (device) memory was being read or written.
  • Memory Clk: Current frequency of the memory clock, in MHz.
  • Performance State: The current performance state for the GPU. States range from P0 (maximum performance) to P12 (minimum performance).

Note

The Performance State graphs are only shown when it differs from the default value of 0.

  • PCIE TX: The GPU-centric transmission throughput across the PCIe bus (in GiB/s) over the past 20ms.
  • PCIE RX: The GPU-centric receive throughput across the PCIe bus (in GiB/s) over the past 20ms.

Warning

The PCIE TX and PCIE RX graphs only include throughput via PCIe bus, i.e., between GPU and CPU.

  • Read: Average read data rate (in MiB/s) in the last minute.
  • Write: Average write data rate (in MiB/s) in the last minute.
  • Open/Close Operations: Average operation rate (in operations/s) in the last minute.
  • Data Input: Average data input throughput (in MiB/s) in the last minute.
  • Data Output: Average data output throughput (in MiB/s) in the last minute.
  • Packet Input: Average package input throughput (in pkt/s) in the last minute.
  • Packet Output: Average package output throughput (in pkt/s) in the last minute.

Attention

The Interconnect values refer to input and output transfers to/from a given node, so it does not include communications within the node itself. However, I/O data is also included in the transferred data in or out of a node.