- Infos im HLRS Wiki sind nicht rechtsverbindlich und ohne Gewähr -
- Information contained in the HLRS Wiki is not legally binding and HLRS is not responsible for any damages that might result from its use -

HPE Hawk: Difference between revisions

From HLRS Platforms
Jump to navigationJump to search
Line 44: Line 44:
'''Performance optimization''': With respect to MPI performance optimization by means of tuning environment variables please cf.  
'''Performance optimization''': With respect to MPI performance optimization by means of tuning environment variables please cf.  
[https://kb.hlrs.de/platforms/upload/Tuning_of_MPT.pdf Tuning of MPT]
[https://kb.hlrs.de/platforms/upload/Tuning_of_MPT.pdf Tuning of MPT]
'''Interconnect topology''': Hawk deploys an Infiniband HDR based interconnect with a 9-dimensional enhanced hypercube topology. Please refer to [https://kb.hlrs.de/platforms/upload/Interconnect_topology.pdf here] with respect to the latter. Infiniband HDR has a bandwidth of 200 Gbit/s and a MPI latency of ~1.3us per link. The full bandwidth of 200 Gbit/s can be used if communicating between the 16 nodes connected to the same node of the hypercube (cf. [https://kb.hlrs.de/platforms/upload/Interconnect_topology.pdf here]). Within the hypercube, the higher the dimension, the less bandwidth is available.


<br>
<br>

Revision as of 13:19, 4 February 2020


If your job does not start, please have in mind the time-dependent limitations according to Batch System!

This Page is under construction!

The information below applies to the Test and Development System (TDS) which is similar to the future Hawk production system. Please have in mind that this is a system under construction. Hence modifications might occur without announcement and stuff may not work as expected from time to time!


Hardware

Node/Processor

With respect to node and processor details cf. here.


Interconnect

Filesystem


Access

Login-Node: hawk-tds-login1.hww.hlrs.de

Note: Access to the Hawk TDS is possible now on request. In case you have early access, we ask you to provide us with your experience regarding usage and performance (approximately half a page) once a month.



Module environment

cf. here


Compiler

cf. here


MPI

Tuned MPI: In order to use the MPI implementation provided by HPE, please load the Message Passing Toolkit (MPT) module mpt (not ABI-compatible to other MPI implementations) or hmpt (ABI-compatible to MPICH-derivatives).

User Guide: For detailed information cf. the HPE Message Passing Interface (MPI) User Guide.

Performance optimization: With respect to MPI performance optimization by means of tuning environment variables please cf. Tuning of MPT


Libraries

cf. here



Batch System

cf. here



Disk storage

Home directories as well as workspaces are handled in the same way as on Hazel Hen, so please cf. Storage Description regarding details.



Manuals

Processor:

MPI: