- Infos im HLRS Wiki sind nicht rechtsverbindlich und ohne Gewähr -
- Information contained in the HLRS Wiki is not legally binding and HLRS is not responsible for any damages that might result from its use -

HPE Hawk: Difference between revisions

From HLRS Platforms
Jump to navigationJump to search
(Rome node up with CentOS 8)
(39 intermediate revisions by one other user not shown)
Line 4: Line 4:
This Page is under construction!
This Page is under construction!


== Best Practises for Software Installation ==
The information below applies to the Test and Development System (TDS) which is similar to the future Hawk production system. Please have in mind that this is a system under construction. Hence modifications might occur ''without'' announcement and stuff may not work as expected from time to time!
Best practices for software installation on Hawk is describe on a separate (internal) wiki page [https://kb.hlrs.de/staff/index.php/Hawk_installation Software Installation on Hawk].
 


== Access ==
== Access ==


Login-Node: hawk-tds-login2.hww.hlrs.de
Login-Node: hawk-tds-login1.hww.hlrs.de
{{note|text=Access to the Hawk TDS is limited to support staff at the moment. Please check the [[Hawk installation schedule]] for details about the start of user access.}}
{{note|text=Access to the Hawk TDS is possible now on request. In case you have early access, we ask you to provide us with your experience regarding usage and performance (approximately have a page) once a month.}}


== Batch System ==
== Tutorial slides ==


[[Batch_System_PBSPro_(Hawk)|Batch System PBSPro (Hawk)]]
[http://fs.hlrs.de/~hpcbjdic/Hawk/Hawk_Slides.pdf Tutorial Slides from HLRS Results & Review Workshop 2019]


<br>


== MPI ==
== Manuals ==
 
* [https://developer.amd.com/wp-content/resources/56305_SOG_3.00_PUB.pdf Software Optimization Guide for AMD EPYC Rome Processors]
In order to use the MPI implementation provided by HPE, please load the Message Passing Toolkit (MPT) module ''mpt'' (not ABI-compatible to other MPI implementations) or ''hmpt'' (ABI-compatible to MPICH-derivatives).
* [https://developer.amd.com/wp-content/resources/56255_3_03.PDF Open-Source Register Reference for AMD EPYC Rome Processors (especially documenting available hardware performance counters)]
For detailed information see the [http://www.hpe.com/support/mpi-ug-036 HPE Message Passing Interface (MPI) User Guide].
* [https://www.amd.com/system/files/TechDocs/32035.pdf Compiler Usage Guidelines for AMD64 Platforms]
* [https://www.amd.com/system/files/TechDocs/47414_15h_sw_opt_guide.pdf Software Optimization Guide for AMD Family 15h]
* [https://developer.amd.com/wordpress/media/2013/12/55723_SOG_Fam_17h_Processors_3.00.pdf Software Optimization Guide for AMD Family 17h (Zen & Zen2) Processors]


<br>
<br>


== Test cases best practices ==
== Compiler ==


Test cases will help to identify and determine the scaling/ performance behavior of the new system.
[[Compiler(Hawk)|Compiler]]
Ideally, those test cases can be compared to other systems as well to get a full picture.


To do:
<br>
* Definition of a best practice guideline on how to set up a correct test case
** Only measurement of time-stepping loop or equivalent excluding the initialization phase or cleanup
** Well defined measure of computational progress (e.g. LUPS, DoF-UPS, Iterations/s or Flop/s)
** Ideally, the test case is mostly automated with scripts and does also the evaluation on top with a meaningful result file


== MPI ==


== TODO ==
In order to use the MPI implementation provided by HPE, please load the Message Passing Toolkit (MPT) module ''mpt'' (not ABI-compatible to other MPI implementations) or ''hmpt'' (ABI-compatible to MPICH-derivatives).
For detailed information see the [http://www.hpe.com/support/mpi-ug-036 HPE Message Passing Interface (MPI) User Guide].
Furthermore, please us the compiler wrappers mpif77 / mpif90 / mpif08 / mpicc / mpicxx.


* 2019-08-22, niethammer@hlrs.de: missing pbs headers (tm.h, ...)
<br>
* 2019-08-18, dick@hlrs.de: (exuberant) ctags missing on frontend, probably available from RHEL repository
* 2019-08-18, dick@hlrs.de: manpages are missing on the frontend
* 2019-08-15, niethammer@hlrs.de: need more explanation on how ''omplace'' works for pinning in the context of SMT (numbering of cores?)
* 2019-08-15, niethammer@hlrs.de: how to run correctly scripts/wrappers with mpirun? (executes script only once per node, but MPI application if called inside multiple times)


* 2019-08-15, niethammer@hlrs.de: missing commands:
== Libraries ==
**''resize'' (likely coming with xterm)


* 2019-08-15, niethammer@hlrs.de:
[[Libraries(Hawk)|Libraries]]
** MPT = Message Passing Toolkit
** MPI from the mpt module uses the SGI ABI, MPI from the hmpt module uses the MPICH ABI
** for the MPI compiler wrappers to detect the correct compiler please set MPICC_CC, MPICXX_CXX, MPIF90_F90, MPIF08_F08 to the corresponding compiler commands (2019-08-15, dick@hlrs.de: done)
** Should e.g. applications using cae/platform_mpi use perfboost?


* 2019-08-07, dick@hlrs: hmpt is ABI-compatible with MPICH-derivativs, but not so mpt
<br>
** user should know about this!
** @HPE: is hmpt a MPICH-derivative, but not so mpt?


* 2019-08-07, dick@hlrs: unclear that (h)mpt provides MPI lib -> call it "mpi/hmpt" and "mpi/mpt" instead?
== Disk storage ==


* 2019-08-07, dick@hlrs: remove MPI delivered with RHEL
Home directories as well as workspaces are handled in the same way as on Hazel Hen, so please cf. [[CRAY_XC40_Disk_Storage | Storage Description ]] regarding details.


* 2019-08-07, dick@hlrs: Intel loads gcc module -> use (LD_)LIBRARY_PATH intead
<br>


* 2019-08-07, dick@hlrs: be careful: cc points to /usr/bin/gcc!
== Batch System ==


* 2019-08-14, khabi/offenhaeuser@hlrs.de: How to pin OpenMP-threads in hybrid jobs (naive approach pins 2 threads to _same_ core instead of two different), i.e.: how to do aprun -d?
[[Batch_System_PBSPro_(Hawk)|Batch System]]


<br>
<br>


== Training ==
== Best Practises for Software Installation ==
 
Best practices for software installation on Hawk is described on a separate (internal) wiki page [https://kb.hlrs.de/staff/index.php/Hawk_installation Software Installation on Hawk].
There will be internal (i.e. HLRS staff only) trainings on the following topics (tentative):
 
* HPE Performance MPI
** [https://terminplaner4.dfn.de/qTagd9VodYiy89mj survey to set date]
** topics available via the above link
** target audience: user support staff, internal users of the system
 
* Processor
** [https://terminplaner4.dfn.de/COB6iw5DAFFyDtwe survey to set date]
** (tentative) schedule available via the above link
** target audience: user support staff, internal users of the system
 
* Workload Management PBSpro for end users
** probably one day in week 2019-11-11 tp 2019-11-15
** target audience: user support staff, internal users of the system
 
* Cluster and System Administration using HPCM
** target audience: admin staff
 
* Infiniband-Administration and Tuning
** target audience: admin staff
 
* Lustre and Storage Administration
** target audience: admin staff
 
* Workload Management PBSpro Administration and Tuning
** target audience: admin staff

Revision as of 13:55, 20 November 2019

Hawk is the next generation HPC system at HLRS. It will replace the existing HazelHen system. The installation is planed to take place in Q4 2019. For more detailed information see the Hawk installation schedule.

This Page is under construction!

The information below applies to the Test and Development System (TDS) which is similar to the future Hawk production system. Please have in mind that this is a system under construction. Hence modifications might occur without announcement and stuff may not work as expected from time to time!


Access

Login-Node: hawk-tds-login1.hww.hlrs.de

Note: Access to the Hawk TDS is possible now on request. In case you have early access, we ask you to provide us with your experience regarding usage and performance (approximately have a page) once a month.


Tutorial slides

Tutorial Slides from HLRS Results & Review Workshop 2019


Manuals


Compiler

Compiler


MPI

In order to use the MPI implementation provided by HPE, please load the Message Passing Toolkit (MPT) module mpt (not ABI-compatible to other MPI implementations) or hmpt (ABI-compatible to MPICH-derivatives). For detailed information see the HPE Message Passing Interface (MPI) User Guide. Furthermore, please us the compiler wrappers mpif77 / mpif90 / mpif08 / mpicc / mpicxx.


Libraries

Libraries


Disk storage

Home directories as well as workspaces are handled in the same way as on Hazel Hen, so please cf. Storage Description regarding details.


Batch System

Batch System


Best Practises for Software Installation

Best practices for software installation on Hawk is described on a separate (internal) wiki page Software Installation on Hawk.