- Best iPhone, iPad Business Apps for 2014
- 14 Tech Conventions You Should Attend in 2014
- 10 Desktop Apps to Power Your Windows PC
- How to Add New Job Skills Without Going Back to School
Network World - Peer inside one of the country's premiere university computational research centers and what will you find?
That's one key message James Leylek, executive director of the Clemson University Computational Center for Mobility Systems (CU-CCMS), wants to deliver about the unusual place he oversees. The CU-CCMS provides a massive computational infrastructure for private sector companies in the automotive, aviation and energy industries.
On a more operational level, a second key message is no matter how much compute capacity in this data center, CU-CCMS doesn't stand a chance of meeting its mission without the InfiniBand network underlying it.
Leylek's brainchild is designed to balance industry's need for fast, timely results delivered on budget against the academic requirements of the master's and doctorate candidates conducting the research. The private sector and academia had long been butting heads over the need for speed vs. the academic mindset of slow, deliberate research leading to the publishing of scholarly papers, Leylek says.
Leylek, who has worked on both sides of the fence, wanted to find a way to put these tensions to rest. He envisioned a research center run under the auspices of Clemson, but staffed by career engineers rather than students. The engineers would develop unique mathematical models that would go above and beyond what any company could get from commercial software packages for attacking big physics challenges.
Three years ago, Leylek's vision edged toward reality. With matching dollars from Clemson's home state of South Carolina, which sought investments in high-tech, knowledge-based activities, the university committed to the idea. The engineering team immediately began designing a "balanced" data center infrastructure, Leylek says.
As part of its balancing act, the CU-CCMS computational infrastructure would need to be capable of dealing with extremes. "We knew we would have a mix of software that could only make use of a single CPU for simulations to nearly linearly scaling software that could run on up to 10,000 CPUs and everything in between," Leylek says.
"Imagine hundreds of millions of billions of equations solved hundreds of thousands of times -- that's what we needed to do," Leylek says.
CU-CCMS engineers work their math magic via a 35-teraflop computational grid composed of 43 Sun Blade 6000 Modular Systems, Sun Fire servers and Sun StorageTek systems. The grid comprises 3,440 processing cores, hundreds of gigabytes of RAM and 14 terabytes of storage, Leylek says.
While 35 teraflops (1 trillion floating point operations per second) may not seem like much oomph in the supercomputer community, CU-CCMS engineers actually have far more crunching power than peers elsewhere, Leylek says. At CU-CCMS, each engineer has access to a dedicated 8.75 teraflops "around the clock, no strings attached." Compare this with the University of Texas at Austin's Texas Advanced Computer Center, which divvies up 579.4 teraflops among 1,500 users. That's roughly 0.4 teraflop per user, he explains.