Help:CSDMS HPCC: Difference between revisions

From CSDMS
m (layout)
No edit summary
Line 36: Line 36:


==== Request an HPCC account ====
==== Request an HPCC account ====
The CSDMS HPCC is free available for all CSDMS members. To get an account on the HPCC you have to:
The HPCC is for members who participate in one of the three ways outlined below.
# Be a [[Organization#Join|member]] one of the Working (WG) or Focus Research Groups (FRG),
# Provide the community beforehand with [[Models questionnaire|metadata of the model]] you want to run on the HPCC,
# Make the source code free available to the community, either by submitting your code to the CSDMS repository [mailto:csdms@colorado.edu csdms@colorado.edu] or by submitting it to one of the free accessible online repositories ([http://sourceforge.net/ sourceforge], [http://code.google.com/opensource/ googlecode], etc).  


Once you met the above requirements you can request an [[HPCC account request | '''CSDMS HPCC account''']].
# They have submitted code into the CSDMS Repository, to either run their models in advance of science, or to advance their developing modeling efforts that will ultimately become part of the Repository.  Provide the community beforehand with [[Models questionnaire|metadata of the model]] you want to run on the HPCC
 
# The HPCC is also for members who wish to apply compliant CSDMS models developed by others within the CSDMS framework, to help them advance their science.   
 
# The HPCC is for members who wish to experiment with new data systems in support CSDMS models, or to develop visualizations of the model runs.
 
Once you meet the above requirements you can request an [[HPCC account request | '''CSDMS HPCC account''']].
   
   
Your HPCC guest account will be valid for ''one year''. You will receive an email as soon as your account expires. Your data (model, source code, simulations, etc) will be removed from the HPCC if you don't extend your account (by email to [mailto:csdms@colorado.edu CSDMS@colorado.edu]). Unfortunately, we have to charge a fee if data needs to be recovered after an account expires.
Your HPCC guest account will be valid for ''one year''. You will receive an email as soon as your account expires. Your data (model, source code, simulations, etc) will be removed from the HPCC if you don't extend your account (by email to [mailto:csdms@colorado.edu CSDMS@colorado.edu]). Unfortunately, we have to charge a fee if data needs to be recovered after an account expires.

Revision as of 14:02, 10 March 2009

The CSDMS High Performance Computing Cluster


Hardware

Altix XE.jpg

Our CSDMS HPCC System is an SGI Altix XE 1300 with integrated 512 x 3.0GHz/12M/1600MHz/80W E5472 processors, using non-blocking Infiniband Interconnect with 1.152TB of memory, with one head node, 28 compute nodes, 4 compute nodes with heavy memory, associated infrastructure, 72TB/7200RPM/SATA Raid storage, web server 4 x 2.33GHz/8GB RAM E5420 processor.
The CSDMS-HPCC (≈ 6Tflops) is configured with two HPC approaches:

  1. massive shared memory among fewer processors,
  2. the more typical parallel configuration each running Linux Red Hat with Fortran, C and C++ compilers.

This system offers to potential CSDMS researchers a state of the art HPC, once their code can be scaled up to take advantage of the capaboilities of these systems.

The CSDMS system will be tied in to the larger, overarching CU-wide Front Range Computing Consortium. This supercomputer will consist of 10 Sun Blade 6048 Modular System racks, nine deployed to form a tightly integrated computational plant, and the remaining rack to serve as a GPU-based accelerated computing system. Each computational rack is composed of four integrated Sun Blade 6000 series chassis units, containing 12 blades connected by an internal quad-data-rate InfiniBand fabric; a 24-port Network Expansion Module (NEM) provides external connectivity. Each blade is composed of two dual-socket boards with each socket containing a quad-core Intel Nehalem-EP processor clocked at an expected frequency of ≈ 3.3 GHz. Each dual-socket board has eight 2 GB DDR-3 DIMMS for a total of 16 GB of RAM (2 GB per core). Each rack contains 192 processors (768 cores), for a total of 7680 cores (including the accelerated computing rack), with a peak performance exceeding 10 Teraflops/s for aggregate system peak performance of 101.4 Teraflop/s. The storage solution will consist of a high-performance Lustre file system built on 12 to 16 Sun Thumper- 2 storage server chassis. Each of the Thumper-2 units has 48 internal disks, an 8x PCI-Express IB Host Bus Adapter (HBA), and utilizes the Sun ZFS block allocation layer to provide in excess of 800 Megabytes per second per Thumper.
Using 1 Terabyte disks, the total raw capacity is between 576 and 768 Terabytes. The remaining computational rack provides the accelerated computing component using NVIDIA Tesla 870 GPU technology. The Tesla 870 GPU system is a 1U chassis containing four 128-simultaneous thread GPUs and 6 Gigabytes of RAM accessible at 76.8 Gigabytes per second. The entire system will utilize a standard Linux-based software stack, vendor-supplied IB-based MPI, and the Coordinated TerraGrid Software and Services. In addition, the Grid environment will provide access to NCAR’s mass storage system. 4.10


Software

Compute Nodes:

Free stuff:
  1. Python Imaging Library (PIL)
  2. Udunits
  3. netcdf
  4. openmotif
  5. Eclipse
  6. Subclipse
  7. Pgplot
  8. HDF5
  9. Matlab
  10. Eucalyptus
  11. Languages:
    1. Python 2.6, 3000 (along with numpy and scipy where possible)
    2. Java 1.6
    3. GNU compilers > 4.0 (gcc, g++, gfortran)

Request an HPCC account

The HPCC is for members who participate in one of the three ways outlined below.

  1. They have submitted code into the CSDMS Repository, to either run their models in advance of science, or to advance their developing modeling efforts that will ultimately become part of the Repository. Provide the community beforehand with metadata of the model you want to run on the HPCC
  1. The HPCC is also for members who wish to apply compliant CSDMS models developed by others within the CSDMS framework, to help them advance their science.
  1. The HPCC is for members who wish to experiment with new data systems in support CSDMS models, or to develop visualizations of the model runs.

Once you meet the above requirements you can request an CSDMS HPCC account.

Your HPCC guest account will be valid for one year. You will receive an email as soon as your account expires. Your data (model, source code, simulations, etc) will be removed from the HPCC if you don't extend your account (by email to CSDMS@colorado.edu). Unfortunately, we have to charge a fee if data needs to be recovered after an account expires.

Access

Once you have an account you can access the CSDMS HPCC with any secure-shell (SSH) application (primarily ssh, scp, sftp) from workstations located in the CU Internet domain (*.colorado.edu) or from workstations connected to the colorado.edu domain through a virtual private network (VPN) connection. A VPN account will automatically be created for users outside the colorado.edu domain.

Displaying of the graphical desktop of the HPCC master-control node to your personal workstation is possible through SSH Tunneling X Windows software. This might require prior installation and configuration of software on your workstation. See information below on how to operate the graphical desktop for Mac and for Windows operating systems.

SSH Tunneling X Windows for Mac OSX

You will need X11 to tunnel X Windows to a Mac. Fortunately, Mac OSX comes with X11. If you're using an older version of OSX, download X11 from the apple site.
Open X11, select Applications and then Terminal. In the terminal type:

ssh -Y beach.colorado.edu -l <your_username>

Type <password> and that's it. Now you can test the Tunneling by for example typing <matlab>.

SSH Tunneling X Windows for Windows

Install Xming on your windows machine. Needs more info