- Infos im HLRS Wiki sind nicht rechtsverbindlich und ohne Gewähr -
- Information contained in the HLRS Wiki is not legally binding and HLRS is not responsible for any damages that might result from its use -

HPE Hawk

From HLRS Platforms
Revision as of 11:35, 16 February 2020 by Hpcbk (talk | contribs)
Jump to navigationJump to search
Warning: HAWK is currently in the set up phase. For more details about about the timing please see the Hawk installation schedule.
Warning: Please have in mind that the system is currently under construction. Hence, modifications might occur and the observed performance can vary.




Introduction


Troubleshooting



Hardware

Node/Processor

Compute nodes as well as login nodes are equipped with

 AMD EPYC 7742 64-Core Processor

detailed information will be provided later. Please check for additional infos AMD Rome 7742 With respect to node and processor details cf. here.


Interconnect

Hawk deploys an Infiniband HDR based interconnect with a 9-dimensional enhanced hypercube topology. Please refer to here with respect to the latter. Infiniband HDR has a bandwidth of 200 Gbit/s and a MPI latency of ~1.3us per link. The full bandwidth of 200 Gbit/s can be used if communicating between the 16 nodes connected to the same node of the hypercube (cf. here). Within the hypercube, the higher the dimension, the less bandwidth is available. Topology aware scheduling is used to exclude major performance fluctuations. This means that larger jobs can only be requested with defined node numbers (64, 128, 256, 512, 1024, 2048 and 4096) in regular operation. This restriction ensures optimal system utilization while simultaneously exploiting the network topology. Jobs with a node number of < 128 nodes are processed in a special partition. Jobs over 4096 nodes are processed at special times.


Filesystem


Access

Login-Node: hawk-tds-login1.hww.hlrs.de

Note: Access to the Hawk TDS is possible now on request. In case you have early access, we ask you to provide us with your experience regarding usage and performance (approximately half a page) once a month.



Module environment

cf. here


Compiler

cf. here


MPI

Tuned MPI: In order to use the MPI implementation provided by HPE, please load the Message Passing Toolkit (MPT) module mpt (not ABI-compatible to other MPI implementations) or hmpt (ABI-compatible to MPICH-derivatives).

User Guide: For detailed information cf. the HPE Message Passing Interface (MPI) User Guide.

Performance optimization: With respect to MPI performance optimization by means of tuning environment variables please cf. Tuning of MPT


Libraries

cf. here



Batch System

cf. here



Disk storage

Home directories as well as workspaces are handled in the same way as on Hazel Hen, so please cf. Storage Description regarding details.


Pre- and post processing

Within HLRS simulation environment special nodes for pre- and post processing tasks are available. Such nodes could be requested via the batch system (follow this link for more info). Available nodes are

  table... 
    4 nodes 2 TB Memory 2 Socket AMD ...x TB local storage   shared usage model
    1 Node  4 TB Memory 2 Socket AMD    x TB local storage   shared usage model


more specialized nodes e.g. graphics, vector, DataAnalytics, ... are available in the Vulcan cluster


Manuals

Processor:


MPI:


Batch system: