- Infos im HLRS Wiki sind nicht rechtsverbindlich und ohne Gewähr -
- Information contained in the HLRS Wiki is not legally binding and HLRS is not responsible for any damages that might result from its use -

NEC Cluster cacau introduction: Difference between revisions

From HLRS Platforms
Jump to navigationJump to search
 
(4 intermediate revisions by 2 users not shown)
Line 4: Line 4:


== Hardware and Architecture ==
== Hardware and Architecture ==
The HWW Xeon EM64T cluster platform consists of one front node for interactive access
The HWW Xeon EM64T cluster platform consists of one front node for interactive access
(cacau.hww.de) and several nodes for execution of parallel programs. The cluster consists of 210 dual socket and 2 quad socket nodes with 3.2GHz/3.0GHz/2.4GHz Xeon EM64T CPU's + 2/8/16/128 GByte memory on the nodes and two 2way frontend node with 2 Xeon EM64T 3.2GHz CPU's + 6GByte memory. Additionally a RAID system with 8 TByte and a GPFS with 15 TByte is available. The local disks of each node (58 GByte) serves as scratch disks. 2 nodes are installed with 128GB memory and a fast local disk with 1.7TB.
(cacau.hww.de) and several nodes for execution of parallel programs. The cluster consists of 210 dual socket and 2 quad socket nodes with 3.2GHz/3.0GHz/2.4GHz Xeon EM64T CPU's + 2/8/16/128 GByte memory on the nodes and two 2way frontend node with 2 Xeon EM64T 3.2GHz CPU's + 6GByte memory. Additionally a RAID system with 8 TByte and a GPFS with 15 TByte is available. The local disks of each node (58 GByte) serves as scratch disks. 2 nodes are installed with 128GB memory and a fast local disk with 1.7TB.


'''Features:'''
'''Features:'''
* Cluster of 210 dual SMPs nodes NEC servers with 2/8/16/128 GByte memory  
* Cluster of 210 dual SMPs nodes NEC Express 5800 120Re-1 servers with 2/8/16/128 GByte memory  
* Frontend node is a 2way NEC Express5800/120Rg-2 server with 6GByte memory  
* Frontend node is a 2way NEC Express5800/120Rg-2 server with 6GByte memory  
* Node-Node interconnect Voltaire Infiniband(Switch:ISR9288) Network + Gigabit Ethernet  
* Node-Node interconnect Voltaire Infiniband(Switch:ISR9288) Network + Gigabit Ethernet  
* Disk 8 TByte home/shared scratch + 1.2 TByte local scratch + 15 TByte GPFS parallel Filesystem  
* Disk 8 TByte home/shared scratch + 1.2 TByte local scratch + 15 TByte GPFS parallel Filesystem  
* Batch system: Torque, Maui scheduler  
* Batch system: Torque, Maui scheduler  
* Operating System: Tao Linux release 1 (Mooch Update 2)  
* Operating System: Scientific Linux SL release 5.2 (Boron), Kernel: 2.6.18-92.1.6.el5 (x86_64)
* NEC HPC Linux software packages  
* NEC HPC Linux software packages  
* Intel Compilers  
* Intel Compilers  
Line 42: Line 43:
!width="30"|number
!width="30"|number
|-
|-
| 1 || 2GB || 3.2 GHz || 2 || 80GB || <font color=red>-</font> || mem2gb || infiniband || noco001-075, noco109-204 || 172
| 1 || 2GB || 3.2 GHz || <small>2*1=</small> 2 || 80GB || <font color=red>-</font> || mem2gb || infiniband || noco001-075, noco109-204 || 172
|-
|-
| 2 || 8GB || 3.0 GHz || 4 || 160GB || <font color=red>workq</font> || - || infiniband || noco075-106 || 32
| 2 || 8GB || 3.0 GHz || <small>2*2=</small> 4 || 160GB || <font color=red>workq</font> || - || infiniband || noco075-106 || 32
|-
|-
| 3 || 8GB || 3.2 GHz || 2 || 80GB || <font color=red>-</font> || mem8gb || infiniband || noco205-208 || 4
| 3 || 8GB || 3.2 GHz || <small>2*1=</small> 2 || 80GB || <font color=red>-</font> || mem8gb || infiniband || noco205-208 || 4
|-
|-
| 4 || 16GB || 3.2 GHz || 2 || 80GB || <font color=red>workq</font> || mem16gb || infiniband || noco209-210 || 2
| 4 || 16GB || 3.2 GHz || <small>2*1=</small> 2 || 80GB || <font color=red>workq</font> || mem16gb || infiniband || noco209-210 || 2
|-
|-
| 5 || 128GB || 2.4 GHz || 8 || 1.7TB || <font color=red>pp</font> || - || GigE || pp2 - pp3 || 2
| 5 || 128GB || 2.4 GHz || <small>2*4=</small> 8 || 1.7TB || <font color=red>pp</font> || - || GigE || pp2 - pp3 || 2
|}
|}


==Access==
==Access==
The only way to access cacau.hww.de (frontend node of NEC Cluster) from outside HWW net is through ssh. Information on how to set up ssh can be found on our webserver at [http://www.hlrs.de/hw-access/platforms/ssh.html ssh].
The only way to access cacau.hww.de (frontend node of NEC Cluster) from outside HWW net is through ssh. Information on how to set up ssh can be found on our webserver at [[Secure_Shell_ssh | Secure Shell (ssh)]].
 


==Usage==
==Usage==
Line 156: Line 155:
# DISPLAY
# DISPLAY
<ul>
<ul>
For X11 applications on interactive batch jobs you need to set the DISPLAY variable. Please see subsection: DISPLAY: X11 applications on interactive Batch Jobs 2.1.1
For X11 applications on interactive batch jobs you need to set the DISPLAY variable. Please see subsection: [[Cacau_Batch_System#DISPLAY:_X11_applications_on_interactive_batch_jobs | DISPLAY: X11 applications on interactive Batch Jobs]].
</ul>
</ul>
</ul>
</ul>

Latest revision as of 08:55, 8 May 2009

This platform serves the following purpose. It enables development and computation of parallel programs on the Intel Xeon processors with Intel EM64T Technology. The two major parallel programming standards MPI and OpenMP are supported. Please note that you must limit the execution time of your jobs during daytime to guarantee the short turn around times that are necessary for development.

Hardware and Architecture

The HWW Xeon EM64T cluster platform consists of one front node for interactive access (cacau.hww.de) and several nodes for execution of parallel programs. The cluster consists of 210 dual socket and 2 quad socket nodes with 3.2GHz/3.0GHz/2.4GHz Xeon EM64T CPU's + 2/8/16/128 GByte memory on the nodes and two 2way frontend node with 2 Xeon EM64T 3.2GHz CPU's + 6GByte memory. Additionally a RAID system with 8 TByte and a GPFS with 15 TByte is available. The local disks of each node (58 GByte) serves as scratch disks. 2 nodes are installed with 128GB memory and a fast local disk with 1.7TB.

Features:

  • Cluster of 210 dual SMPs nodes NEC Express 5800 120Re-1 servers with 2/8/16/128 GByte memory
  • Frontend node is a 2way NEC Express5800/120Rg-2 server with 6GByte memory
  • Node-Node interconnect Voltaire Infiniband(Switch:ISR9288) Network + Gigabit Ethernet
  • Disk 8 TByte home/shared scratch + 1.2 TByte local scratch + 15 TByte GPFS parallel Filesystem
  • Batch system: Torque, Maui scheduler
  • Operating System: Scientific Linux SL release 5.2 (Boron), Kernel: 2.6.18-92.1.6.el5 (x86_64)
  • NEC HPC Linux software packages
  • Intel Compilers
  • Voltaire MPI
  • Switcher/Module


Peak Performance: 	        3.9 TFLOP/s
Cores/node: 	                4
Memory: 	                1 TB
Shared Disk: 	                24 TB 
Local Disks/node: 	        80 GB
Number of Nodes: 	        212
Node-node data transfer rate: 	10 Gbps(Full bisectional: 20Gbps) infiniband


Short overview of installed compute nodes
Type memory Freq cores Disk PBS Queue PBS properties Interconnect nodes number
1 2GB 3.2 GHz 2*1= 2 80GB - mem2gb infiniband noco001-075, noco109-204 172
2 8GB 3.0 GHz 2*2= 4 160GB workq - infiniband noco075-106 32
3 8GB 3.2 GHz 2*1= 2 80GB - mem8gb infiniband noco205-208 4
4 16GB 3.2 GHz 2*1= 2 80GB workq mem16gb infiniband noco209-210 2
5 128GB 2.4 GHz 2*4= 8 1.7TB pp - GigE pp2 - pp3 2

Access

The only way to access cacau.hww.de (frontend node of NEC Cluster) from outside HWW net is through ssh. Information on how to set up ssh can be found on our webserver at Secure Shell (ssh).

Usage

The frontend node cacau.hww.de is intended as single point of access to the entire cluster. Here you can set your environment, move your data, edit and compile your programs and create batch scripts. Interactive usage like run your program which leads to a high load is NOT allowed on the frontend node cacau.hww.de. The compute nodes for running parallel jobs are available only through the Batch system installed on the frontend node cacau.hww.de!

HOME directories

All user HOME directories for every compute node of the cluster are located on the master node cacau.hww.de. The compute nodes have the HOME directories mounted via NFS. On every node of the cluster the path to your HOME is the same. The filesystem space on HOME is limited by a quota of 50MB! Please note the Filesystempolicy! Default startup files (.profile, .cshrc,...) for your environmental settings can be found in: /usr/local/skel Only the default .profiles and the commands module or switcher support the HWW cluster features like MPI, Compiler settings,...(see Program Development and Environment Settings).

SCRATCH directories

Local scratch

When allocating nodes using the batch queuing system (Torque), the system creates your own scratch area on each of the allocated nodes. The path to this local scratch area is stored in the environment variable SCRDIR (echo $SCRDIR) in your batch job shell. After your batch jobs are finished, the $SCRDIR will be removed automatically.

Global scratch

Another scratch you can get are global space on shared filesystems. There are 2 globel shared filesystems available on cacau:

  • default
    It's a filesystem which is available via NFS on all cacau cluster nodes and on the cacau frontend system
    IBM GPFS filesystem shared globaly on different HWW Clusters To use it on cacau compute nodes, you need to create a file named '.gpfs' in your HOME directory (touch $HOME/.gpfs). The GPFS filesystem need some of the compute nodes memory. If you are short in memory on those nodes and you didn't need this filesystem, then please delete $HOME/.gpfs. If no such file found in your HOME, then the GPFS modules will not be loaded on the compute nodes.

You are responsible to obtain it from the system. To get access to this global scratch filesystems you have to use the workspace mechanism.

Environment Settings

In order to use some software features like special MPI versions, or Compilers, you have to perform some environmental settings.

  • Environment Settings using command switcher
    You can make some software settings with the command switcher. This command helps you to define a user default for a special software and defines your environment accordingly. After modifying your environment defaults and to make your new settings valid, you have to login again!
    switcher <tag> --show [--system or --user]

    This shows you the current system or user default for a certain tag.

    switcher --list

    This shows you all available tags.

    switcher <tag> --list

    This shows you all available names for tag.

    switcher mpi = voltaire_gcc

    This will set you a new MPI default.

  • Environment Settings using command module
    The environmental setting using this methode will not be saved and will be lost for a new session. A new session (login, new job) will have the default environment defined in your .profile (see above). The Cluster system uses modules in the user environment to support multiple versions of software, such as compilers, and to create integrated software packages. As new versions of the supported software become available, they are added automatically to the programming environment, while earlier versions are retained to support legacy applications. By specifying the module to load, you can choose the default version of an application, or another version. Modules also porvide a simple mechanism for updating certain environment variables, such as PATH, MANPATH, and LD_LIBRARY_PATH.
    The following topics describe the fundamentals of using the modules environment.
    • to invoke the module command, type:
      module option args
      module help modulecommand
      The help command will provide more detailed information on the specified module. Without argument modulecommand you will get online help for the module command.
      module avail
      The avail option displays all the modules that are available on the system. Where there is more than one version of a module, the default version is denoted by (default).
      module list
      The list option displays all the modules that are currently loaded into your user environment.
      module add / module load modulename
      The add option and the load option have the same function - to load the specified module into your user environment.
      module rm / module unload modulename
      The rm option and the unload option have the same function - to unload the specified module from your user environment. Before loading a module that replaces another version of the same package, you should always unload the module that is to be replaced.
      module display modulename
      The display option shows the changes that the specified module will make in your environment, for example, what will be added to the PATH and MANPATH environment variables.
      module switch modulename/currentversionmodulename/newversion
      The switch option replaces the currently loaded version of a module with a different version. When the new verion is loaded, the man page for the specified software will also be updated.


    • using $HOME/.modulerc
      This file can be used to load or to define your own environment during each login. An example looks like this:
      #%Module1.0#
      
      set version 1.0
      module load use.own
      

      The module use.own will add $HOME/privatemodules to the list of directories that the module command will search for modules. Place your own module files here. This module, when loaded, will create this directory if necessary.

      see also:

       man module
      


  • Special environmental settings:


Filesystem Policy

IMPORTANT! NO BACKUP!! There is NO backup done of any user data located on HWW systems. The only protection of your data is the redundant disk subsystem. This RAID system (Raid5) is able to handle a failure of one component (e.g. a single disk or a controller). There is NO way to recover inadvertently removed data. Users have to backup critical data on their local site!

The homedirectory of each user is available on all nodes. Each node has a local /scratch directory (which is much faster than the NFS-mounted home-directory) that should be used as temporary file space. In your batch job session an environment $SCRDIR will be set on all allocated nodes. This $SCRDIR is your local scratch directory located in /scratch.


Support / Feedback

Please report all problems to:


  • System Administrators
    • Thomas Beisel
      Phone: +49-711-685-87220
      e-Mail: beisel@hlrs.de
    • Bernd Krischok
      Phone: +49-711-685-87221
      e-Mail: krischok@hlrs.de
    • Danny Sternkopf
      Phone: +49-711-685-87256
      e-Mail: dsternkopf@hpce.nec.com


  • Applications
    • Martin Bernreuther