- Infos im HLRS Wiki sind nicht rechtsverbindlich und ohne Gewähr -
- Information contained in the HLRS Wiki is not legally binding and HLRS is not responsible for any damages that might result from its use -

Vulcan: Difference between revisions

From HLRS Platforms
Jump to navigationJump to search
No edit summary
No edit summary
(15 intermediate revisions by 3 users not shown)
Line 1: Line 1:
The NEC Cluster is part of a heterogenious computing platform.
The NEC Cluster is part of a heterogenious computing platform.
{{Warning
{{Warning
| text = Begin Production October 1st
| text = Hardware Upgrade took place on 2024-05-24<br>
 
Some of the compute nodes and network infrastructure of vulcan has been replaced by up to date hardware.
Due to the outdated software version on the Laki cluster,
and the upcoming installation of the new cluster named Vulcan,
we would like to inform you how we plan to migrate the
production from the old Cluster to Vulcan:
* All active projects on Laki will automatically get access to Vulcan.
* Login to Vulcan will be possible starting from October 1st  2018
* The Lustre workspace filesystem and the users Home filesystems will be identical on both clusters (Laki and Vulcan).
* Compute nodes will be moved from Laki to Vulcan. The following time schedule is planned:
** 10/01-17: start moving special node types (graphics nodes, large memory nodes, ...)
** 10/08 :start moving SandyBridge nodes to Vulcan
** 10/10: start moving Haswell (hsw128gb10c) nodes to Vulcan
** 10/15: start moving Haswell (hsw128gb12c) nodes to Vulcan
* Final shutdown of Laki (all queued jobs on Laki will be deleted): October 31st
* The operating system on Vulcan will be based on CentOS 7.
* The batch system will be PBSPro
* accounting of the cluster nodes will be changed from core hours (Laki) to node hours (Vulcan).
}}
}}


{| class="wikitable" border="1" cellpadding="2"
|+'''Replacement Overview:'''
|-
|'''node_type'''||'''historical node number'''||'''current node number'''
|-
|''aurora''|| 8 || 8
|-
|''clx-21''|| 8 || 8
|-
|''clx-25''        || 96 ||        96
|-
|<font color=red>''clx-ai''</font>        ||  4 ||          <font color=red>0</font>
|-
|<font color=red>''hsw128gb20c''</font>  || 84 ||          <font color=red>0</font>
|-
|<font color=red>''hsw128gb24c''</font>  || 152 ||          <font color=red>0</font>
|-
|<font color=red>''hsw256gb20c''</font>  || 4 ||          <font color=red>0</font>
|-
|<font color=red>''hsw256gb24c''</font>  || 16 ||          <font color=red>0</font>
|-
|<font color=red>''k20xm''</font>        ||  3 ||          <font color=red>0</font>
|-
|''p100''          ||  3 ||          3
|-
|''skl''          || 68 ||        72
|-
|''smp''          ||  2 ||          1
|-
|''visamd''        ||  6 ||          6
|-
|''visnv''        ||  2 ||          2
|-
|<font color=red>''visp100''</font>      || 10 ||          <font color=red>0</font>
|-
|''zen256gb32c''  ||  3 ||          3 <sup>(1)(2)</sup>
|-
|''zen512gb48c-ai'' || 10 ||        10 <sup>(1)(3)</sup>
|-
|<font color=green>''genoa''</font>          || 0 ||        <font color=green>60</font> <sup>(4)(5)</sup>
|-
|<font color=green>''genoa-a30''</font>      || 0 ||        <font color=green>24</font> <sup>(4)(6)</sup>
|-
|<font color=green>''genoa-smp''</font>      || 0 ||          <font color=green>2</font> <sup>(4)(7)</sup>
|-
|}
<sup>
(1) academic usage only<br>
(2) 2x AMD Epic 7302 Rome, 3.0GHz base, 32 cores total, 256GB DDR4, 3.5TB NVMe<br>
(3) 2x AMD Epyc 7642 Rome, 2.3GHz base, 96 cores total, 512GB DDR4, 1.8TB NVMe, 8x AMD Instinct Mi50 with 32GB<br>
(4) new nodes, node_type not yet fixed<br>
(5) 2x AMD Epyc 9334 Genoa, 2.7GHz base, 64 cores total, 768GB DDR5<br>
(6) 2x AMD Epyc 9124 Genoa, 3.0GHz base, 32 cores total, 768GB DDR5, 3.8TB NVMe, 1x Nvidia A30 with 24GB HBM2e<br>
(7) 2x AMD Epyc 9334 Genoa, 2.7GHz base, 64 cores total, 3072GB DDR5<br>
</sup>
----
----


Line 65: Line 105:
* [[NEC_Cluster_Using_MPI | MPI Usage examples ]]
* [[NEC_Cluster_Using_MPI | MPI Usage examples ]]
* [[Graphic_Environment | Graphic environment]] (Description how to use the graphic hardware via VirtualGL and/or VNC)
* [[Graphic_Environment | Graphic environment]] (Description how to use the graphic hardware via VirtualGL and/or VNC)
* [[Big_Data,_AI_Aplications_and_Frameworks|Big Data, AI Applications and Frameworks]]
* [[NEC_Cluster_NUMA_Tuning | NUMA Tuning ]] What to consider when using a NUMA machine
* [[NEC_Cluster_NUMA_Tuning | NUMA Tuning ]] What to consider when using a NUMA machine
* [[Performance Analysis Tools ]]
|}
|}
</div>
</div>
Line 78: Line 120:
<!--* [[NEC_Nehalem_Program_Development | Program Development]]-->
<!--* [[NEC_Nehalem_Program_Development | Program Development]]-->
* [[CAE_utilities|CAE Utilities]]
* [[CAE_utilities|CAE Utilities]]
* [[CAE_howtos|CAE HOWTOs]]
* [[MKL | MKL Fortran Interfaces ]]
* [[MKL | MKL Fortran Interfaces ]]
* [[FFTW | FFTW library usage ]]
* [[FFTW | FFTW library usage ]]

Revision as of 19:23, 29 May 2024

The NEC Cluster is part of a heterogenious computing platform.

Warning: Hardware Upgrade took place on 2024-05-24
Some of the compute nodes and network infrastructure of vulcan has been replaced by up to date hardware.


Replacement Overview:
node_type historical node number current node number
aurora 8 8
clx-21 8 8
clx-25 96 96
clx-ai 4 0
hsw128gb20c 84 0
hsw128gb24c 152 0
hsw256gb20c 4 0
hsw256gb24c 16 0
k20xm 3 0
p100 3 3
skl 68 72
smp 2 1
visamd 6 6
visnv 2 2
visp100 10 0
zen256gb32c 3 3 (1)(2)
zen512gb48c-ai 10 10 (1)(3)
genoa 0 60 (4)(5)
genoa-a30 0 24 (4)(6)
genoa-smp 0 2 (4)(7)

(1) academic usage only
(2) 2x AMD Epic 7302 Rome, 3.0GHz base, 32 cores total, 256GB DDR4, 3.5TB NVMe
(3) 2x AMD Epyc 7642 Rome, 2.3GHz base, 96 cores total, 512GB DDR4, 1.8TB NVMe, 8x AMD Instinct Mi50 with 32GB
(4) new nodes, node_type not yet fixed
(5) 2x AMD Epyc 9334 Genoa, 2.7GHz base, 64 cores total, 768GB DDR5
(6) 2x AMD Epyc 9124 Genoa, 3.0GHz base, 32 cores total, 768GB DDR5, 3.8TB NVMe, 1x Nvidia A30 with 24GB HBM2e
(7) 2x AMD Epyc 9334 Genoa, 2.7GHz base, 64 cores total, 3072GB DDR5


Introduction


Troubleshooting



Cluster Documentation


Programming / Utilities



Help for Wiki Usage