Today the Ohio Supercomputer Center (OSC) announced plans to boost scientific and industrial discovery and innovation with a powerful new supercomputer from Dell. To be deployed later this year, the new system is part of a $9.7 million investment that received approval from the State Controlling Board in January.
In this WGRZ video, researchers describe supercomputing at the Center for Computational Research at the University of Buffalo. “The Center’s extensive computing facilities, which are housed in a state-of-the-art 4000 sq ft machine room, include a generally accessible (to all UB researchers) Linux cluster with more than 8000 processor cores and QDR Infiniband, a subset (32) of which contain (64) NVidia Tesla M2050 “Fermi” graphics processing units (GPUs).”
Dell in Austin is seeking an HPC Benchmarking Principal Engineer in our Job of the Week.
In this video from the Dell booth at SC15, Addison Snell from Intersect360 Research discusses why HPC is now important to a broader group of use cases, and dug deep into overviews of HPC for research, life sciences and manufacturing. Participants learned more about why HPC, Big Data, and Cloud are converging.
Dr. Lewey Anton reports on who’s moving on up in High Peformance Computing. Familiar names in this edition include: Sharan Kalwani, John Lee, Jay Muelhoefer, Brian Sparks, and Ed Turkel. And be sure to let us know of HPC folks in new positions!
National Center for Supercomputing Applications (NCSA) has a private sector program (PSP) which works with the smaller companies to help them adopt HPC technologies based on the expertise acquired over the past quarter century. By working with these organizations, NCSA can help them to determine the Return on Investment (ROI) of using more computing power to solve real world problems than is possible on smaller, less capable systems.
In this video from SC15, Dell’s Onur Celebioglu discusses why HPC is now important to a broader group of use cases. He also provides an overview of HPC for research, life sciences and manufacturing. Participants learned more about why HPC, big data and cloud are converging, and how Dell solves challenges in our HPC engineering lab and through collaborative work with other leading technology partners and research institutions.
In this video, Tommy Minyard from TACC describes how Dell helped develop the 9.6 Petaflop Stampede supercomputer for scientific computing. “The Texas Advanced Computing Center supports the University of Texas System and National Science Foundation researchers with the newest version of their Stampede High Performance Computing cluster. Stampede uses Dell PowerEdge servers, Intel Xeon processors and the new Dell Networking H-Series switches and adapters based on Intel Omni-Path Architecture. These newly implemented technologies greatly reduce latency, optimize traffic flow, and give Stampede a peak performance of 10 petaflops.”
In many HPC environments, the storage system is an afterthought. While the main focus is on the CPU’s the selection and implementation of the storage hardware and software is critical to an efficient and productive overall HPC environment. Without the ability to move data quickly into and out of the CPU system, the HPC users would not be able to obtain the performance that is expected.
In this TACC podcast, Niall Gaffney from the Texas Advanced Computing Center discusses the Wrangler supercomputer for data-intensive computing. “We went to propose to build Wrangler with (the data world) in mind. We kept a lot of what was good with systems like Stampede, but then added new things to it like a very large flash storage system, a very large distributed spinning disc storage system, and high speed network access to allow people who have data problems that weren’t being fulfilled by systems like Stampede and Lonestar to be able to do those in ways that they never could before.”