What is it? The Richmond Supercomputing Cluster History The Supercomputing Cluster at the University of Richmond is used to perform computationally intensive running simulations.
Download ReportTranscript What is it? The Richmond Supercomputing Cluster History The Supercomputing Cluster at the University of Richmond is used to perform computationally intensive running simulations.
What is it? The Richmond Supercomputing Cluster History The Supercomputing Cluster at the University of Richmond is used to perform computationally intensive running simulations and data analysis. It consists of one master node and 52 slave nodes of which 49 are dual 1.4 GHz Athlon class and 3 2.0 GHz Athlon Class. The cluster project at Richmond began in the spring of 1999. Professors Mike Vineyard and Jerry Gilfoyle realized the need for an increase in computing power to support their research in nuclear physics at the Thomas Jefferson National Accelerator Facility (JLab). Using funds from their Department of Energy (DOE) research grant and other monies from the University, they put together a prototype computing cluster in the summer of 1999. That prototype (shown below) is no longer operational. Details on Using the Supercomputer The Supercomputing Cluster runs on the Linux operating system and uses the Beowulf system for managing batch jobs. Programs are written in Fortan, Perl, or C++ and are submitted to the master node. The master node then sends the commands to the slave nodes and What we are currently working on? Out-of-Plane Measurements of the Structure Functions of the Deuteron Research cluster prototype consisting of 12 nodes. Working with SpiderWulf. Students Sherwyn Granderson (left) monitors the status of the cluster while Rusty Burrell and Kuri Gill (right) review some results from We now know that particles called quarks are the basis for the recent data analysis. atoms, molecules, and atomic nuclei that form us and our world. Nevertheless, how these quarks actually combine to form that matter Why we need it... is still a mystery. The deuteron is an essential testing ground for any theory of the nuclear force because it is the simplest nucleus in Although initial analysis of data taken in CLAS is done at Jefferson Lab, we nature. In this project we use the unique capabilities of the CLAS analyze this data more deeply at the University of Richmond. Performing this detector at Jefferson Lab to make some of the first measurements of second pass analysis requires us to be able to store full data sets from a little-known electric and magnetic properties of the deuteron. In the particular running period, so that we don’t have to repeatedly move data on past scattering experiments like those done at JLab were confined to and off the system. The set we are currently working on is about 1 Terabyte reactions where the debris from the collision was in the same plane (1,000 Gigabytes), and past data sets have been even larger. The computing (usually horizontal) as the incoming and outgoing projectile (an cluster has a maximum capacity of 4.4Terabytes. electron in this case). With CLAS we can now measure particles that are scattered out of that plane and are sensitive to effects that have been often ignored up to now. These measurements will open a new window into the atomic nucleus. For more, see “Nuclear Physics at UR” poster. University of Richmond Undergraduate Education Students who choose to participate in an undergraduate research project develop a deeper understanding of not only nuclear physics, but how to go about solving a daunting, computational challenge to develop a new application. One vital aspect to studying nuclear physics is scientific computing. Along with pure physics instruction, students are given access to the supercomputer and learn how to use it. They learn about new operating systems and programming languages, as well as the networking structure and usage of the supercomputing cluster. ENIAC: The world’s first electronic digital computer, 1945 SpiderWulf: UR’s supercomputing cluster, 2002-present In addition to disk space, considerable computing power is necessary to perform not only data analysis but also precise simulations of CLAS in order to produce publication-quality measurements and results. For example, running a typical simulation on our prototype cluster developed in 1999 (see “History”) would take about 39 days. This is dar too long to expect the analysis to be complete in a reasonable amount of time. On our current cluster, this same job can be run over a weekend or even overnight. The current Research cluster was assembled by LinuxLabs in the winter of 2001-2002 and arrived in February 2002. After a commissioning phase, the cluster became fully operational in fall, 2002. The funds for the project came from a successful grant application by Dr. Gilfoyle and Dr. Vineyard to the Major Research Instrumentation Program of the National Science Foundation for $175,000. The current configuration consists of 52, 512-MByteRAM, 1.4 GHz Athlon remote nodes and one master. Each node has a single 18-GByte disk. It is supported by 3 TBytes of space in three RAID disks and communicating by another 1.4 GHz Athlon fileserver. The Power of SpiderWulf Often in the world of computing (and specifically supercomputing), a measure of FLOPS (floating point operations per second) is used to compare computer performances. In order to see where our cluster stands, we measured our machine and compared it to the top supercomputers in the world since 1993. The # Year Name Location # of processors GFLOPS results are shown below. 1. 1993 CM-5/1024 Los Alamos National Laboratory 1,024 59.7 2. 1996 SR2201/1024 University of Tokyo 1,024 232.4 3. 1999 ASCI Blue-Pacific SST IBM SP604e Lawrence Livermore 5,808 National Laboratory (LLNL) 2144 4. 2001 ASCI White, SP Power3 375Mhz Lawrence Livermore National Laboratory 8,192 7,304 5. 2001 SpiderWulf University of Richmond 53 35.99 6. 2006 eServer Blue Gene Solution DOE/NNSA/LLNL 131,072 280,600 The data points are GigaFLOPS per processor, as many of these supercomputers contain hundreds or thousands of processors. The red points represent the top supercomputer of each year, and the blue point represents where our cluster fell in the year that it was purchased. For further comparison, we measured the FLOPS of a typical desktop PC, and compared to our measure of the cluster….