“Today, scalable compute and storage systems suffer from data bottlenecks that limit research, product development, and constrain application services. ConnectX-5 will help unleash business potential with faster, more effective, real-time data processing and analytics. With its smart offloading, ConnectX-5 will enable dramatic increases in CPU, GPU and FPGA performance that will enhance effectiveness and maximize the return on data centers’ investment.”
In this WKAR podcast, Andy Keen from Michigan State University describes MSU’s new Laconia supercomputer. The $3-million system is part of a cluster that’s run by their ICER High Performance Computing Center. “It’s an exciting time for the computational sciences at MSU,” said iCER Director Kennie Merz. “Our researchers are poised to tackle transformative problems ranging from the discovery of new drugs to understanding the universe we live in.”
In this Programming Throwdown podcast, Mark Harris from Nvidia describes Cuda programming for GPUs. “CUDA is a parallel computing platform and programming model invented by NVIDIA. It enables dramatic increases in computing performance by harnessing the power of the graphics processing unit (GPU). With millions of CUDA-enabled GPUs sold to date, software developers, scientists and researchers are finding broad-ranging uses for GPU computing with CUDA.”
In this TACC podcast, Ari Kahn from the Texas Advanced Computing Center and Eddie Garcia from Cloudera describe a recent Hackathon in Austin designed to tackle data challenges in the fight against the Zika virus. The Texas Advanced Computing Center provided time on the Wrangler data intensive supercomputer as a virtual workspace for the Zika hackers.
Getting started with HPC can be a challenge for SMEs, but managing a cluster doesn’t have to be a struggle. IBM’s Platform Computing group has been helping users to stand up and run clusters efficiently for years. Now, with the recently announced IBM Platform LSF Suites for Workgroups and HPC, the company has made it easier than ever to get kick the tires on High Performance Computing. “So basically, we would give you all the tools that would allow you to easily migrate from a loose collection of work stations to a small cluster environment. And we would handle the bare metal provisioning and then installing the software that you need really to manage your workload.”
In this podcast, the Radio Free HPC team looks at the news highlights for the week leading up to Friday the 13th of May, 2016. Highlights include a 25 Petaflop Fujitsu supercomputer coming to Japan, an OpenPOWER Summit coming to Europe, and fighting the Zombie Apocalypse with HPC.
In this podcast, the Radio Free HPC team recaps the ASC16 Student Cluster Competition in China and the 2016 MSST Conference in Santa Clara. Dan spent a week in Wuxi interviewing ASC16 student teams, he came back impressed with the Linpack benchmark tricks from the team at Zhejiang University, who set a new student LINPACK record with 12.03 TFlop/s. Meanwhile, Rich was in Santa Clara for the MSST conference, where he captured two days of talks on Mass Storage Technologies.
In this TACC Podcast, Jorge Salazar looks at how researchers are using the Stampede supercomputer to shed light on the microscale world of colloidal gels — liquids dispersed in a solid medium as a gel. “Colloidal gels are actually soft solids, but we can manipulate their structure to produce ‘on-demand’ transitions from liquid-like to solid-like behavior that can be reversed many times,” Zia said. Zia is an Assistant Professor of Chemical and Biomolecular Engineering at Cornell University.
In this slidecast, Gabor Samu from IBM describes the newly available IBM Platform LSF Suites for Workgroup and HPC. Designed to make it much easier to “kick the tires” on LSF, the new suites can help you configure install, maintain, and job manage HPC clusters with a single download. “The new IBM Platform LSF Suites are packages that include more than IBM Platform LSF, they provide additional functionalities designed to simplify HPC for users, administrators and the IT organization.”
In this TACC podcast, Joe Stubbs from the Texas Advanced Computing Centter describes potential benefits to scientists of open container platform Docker in supporting reproducibility, NSF-funded Agave API. “As more scientists share not only their results but their data and code, Docker is helping them reproduce the computational analysis behind the results. What’s more, Docker is one of the main tools used in the Agave API platform, a platform-as-a-service solution for hybrid cloud computing developed at TACC and funded in part by the National Science Foundation.”