Research Computing

From WolfTech
(Redirected from High Performance Computing)
Jump to navigation Jump to search

We've all been there. You know the moment: sitting at your computer just wondering when it would finish the current task so you could move on with your life. If you have a new computer, this probably does not happen to you very often. That is, of course, unless you wish to run complex mathematical operations or simulations on your computer. To effectively run such software, one must rely upon the power of super computers.

Simply put, a super computer is a computer with access to vast amounts of resources. This could mean more processors, faster processor speed, more memory, or larger hard drive space. Here at NC State, there are three main high performance computing (HPC) facilities. They are the ITD's High Performance Center, LSF, and Wolfgrid.


OIT Supercomputers

In July 2003, The NC State Information Technology Division put into production an IBM p690 acquired from the now defunct North Carolina Supercomputing Center. The system harnesses the power of 32 Power4 1.3 Ghz processors and has 128GB of shared memory, with access to over 10TB of storage space.

New to NCSU is the IBM Blade Center #1; it makes use of 32 Xeon 2.4 Ghz processors and 48GB of distributed memory, meaning each processor has 3GB. It has access to 1.2TB of hard drive space with a 1Gbps connection to mass storage. For more information on running jobs on this computer, please visit the Getting Started with the IBM Blade Center page.

A third super computer is being assembled. In all, it will consist of 64 Xeon 2.8Ghz processors. This computer will have 256Gb of distributed memory. There will be a 2.56Tb hard drive for storage. And, as soon as possible, Myrinet and a 1Gbps connection with a larger mass storage device.

For information on accessing these computers, visit ITD's High Performance Computer page. Current software packages already installed or licensed include:

  • Amber 7 - A suite of molecular simulation programs
  • Ansys - A design simulation program for structural/physical analysis
  • ARPS - Atmospheric modeling and prediction software
  • PAUP - Software used to track and make inferences on evolutionary trees
  • NCAR - A utility for drawing various maps and diagrams

Condor

The Condor Project is currently under way within the ECE department. This is an effort to bring the Condor software to NC State. Condor software provides a specialized workload management system for a network of computers, allowing for the unused clock cycles of computers on a network to be exploited.

When a computer goes idle, the Condor software will recognize this and look to its queue to see if there are any jobs waiting to be run. If it finds jobs, they are run, while the computer stays idle. When, however, someone comes to and uses the computer, Condor instructs itself to set a checkpoint and stops processing the data.

Condor at NC State is still in developmental stages, but hopefully will be available soon. Both a Windows and Red Hat Linux solution are being developed. More information will be made available as soon as the project nears completion.

Condor

LSF

Need info on LSF @ NCState.

CSC ARC Cluster

1728 cores on 108 compute nodes integrated by Advanced HPC. All machines are 2-way SMPs with AMD Opteron 6128 (Magny Core) processors with 8 cores per socket (16 cores per node). For details, including how to get access: http://moss.csc.ncsu.edu/~mueller/cluster/arc/

Macintosh Wolfgrid

Wolfgrid *was* a community supercomputer that allows any recent apple computer to connect to Wolfgrid and will then process jobs that you submit to it. You can also become part of Wolfgrid by letting your computer become a processing node (an agent). You can set your computer to only act as an agent when the screen saver comes on so it will not disrupt your use of the computer. Wolfgrid puts the power in your hands, where you have a supercomputer grid working for you.

This particular project seems to have fallen by the wayside. Though if there is interest, we may be able to revive it. We're investigating the potential requirements, investment, and interest from the academic/research community. Any comments are welcome.