MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Podcast: Strengthening High Performance Computing with OpenHPC

OpenHPC is a collaborative, community effort that initiated from a desire to aggregate a number of common ingredients required to deploy and manage High Performance Computing Linux clusters including provisioning tools, resource management, I/O clients, development tools, and a variety of scientific libraries.

Podcast: HPE and Intel Alliance for High Performance Computing & Big Data

In this Chip Chat podcast, Bill Mannel, Vice President and General Manager for HPC and Big Data from Hewlett Packard Enterprise (HPE) describes the High Performance Computing Alliance between HPE and Intel. He highlights how the two companies are developing innovative solutions based on Intel Scalable System Framework (Intel SSF) and are working to enhance HPC solutions while engaging customers directly in centers of excellence (COEs) located in Grenoble, France and Houston, Texas. Bill also emphasizes how HPE compute solutions are experiencing incredible momentum in government, commercial and academic market verticals and that HPE is receiving excellent results from the integration of HPE Apollo products and Intel HPC technology.

2016 Predictions from Radio Free HPC

In this podcast, the Radio Free HPC team makes their tech predictions for 2016. Will secure firmware be the key differentiator for HPC vendors? Will this be the year of FPGAs? And could we see a 100 Petaflop machine on the TOP500 before the year ends?

Podcast: Alan Gara on the Intel Scalable System Framework

In this Intel Chip Chat podcast, Alan Gara describes how Intel’s Scalable System Framework (SSF) is meeting the extreme challenges and opportunities that researchers and scientists face in high performance computing today. He explains that SSF incorporates many different Intel technologies including; Intel Xeon and Phi processors, Intel Omni-Path Fabrics, silicon photonics, innovative memory technologies, and efficiently integrates these elements into a broad spectrum of system solutions optimized for both compute and data-intensive workloads. Mr. Gara emphasizes that this framework has the ability to scale from very small HPC systems all the way up to exascale computing systems and meets the needs of users in a very scalable and flexible way.

Podcast: Top 10 Ways Intel Will Drive HPC Democratization in 2016

The use of High Performance Computing continues to grow in the enterprise and beyond. In this podcast, James Reinders from Intel describes how Intel will continue to drive HPC democratization in 2016. “At Intel, our passion to help drive the democratization of HPC is exemplified by many things. Here is my list of ten things which caught my attention as being most significant as we enter 2016.”

Podcast: XSEDE Powers Search for How the Earth was Formed

In this NICS podcast, Scott Gibson discusses how the Earth was formed with Wladimir Lyra, an assistant professor at California State University at Northridge, Department of Physics and Astronomy. Lyra is using XSEDE supercomputing resources to explore this age-old question.

Podcast: Societal Impact of Earthquake Simulations at Extreme Scale

In this TACC podcast, the Thomas Jordan from the University of Southern California describes how he uses the computational resources of XSEDE, the Extreme Science and Engineering Discovery Environment, to model earthquakes and help reduce their risk to life and property. Dr. Jordan was invited to speak at SC15 on the Societal Impact of Earthquake Simulations at Extreme Scale.

Podcast: Big Data on the Wrangler Supercomputer

In this TACC podcast, Niall Gaffney from the Texas Advanced Computing Center discusses the Wrangler supercomputer for data-intensive computing. “We went to propose to build Wrangler with (the data world) in mind. We kept a lot of what was good with systems like Stampede, but then added new things to it like a very large flash storage system, a very large distributed spinning disc storage system, and high speed network access to allow people who have data problems that weren’t being fulfilled by systems like Stampede and Lonestar to be able to do those in ways that they never could before.”

Radio Free HPC Looks at Why it Pays to Upgrade Your Home Network

In this podcast, the Radio Free HPC team looks at why it pays to upgrade your home networking gear.

Podcast: Supercomputing the Deep Earth with the Gordon Bell Prize Winners

In this podcast, Jorge Salazar from TACC interviews two winners of the 2015 ACM Gordon Bell Prize, Omar Ghattas and Johann Rudi of the Institute for Computational Engineering and Sciences, UT Austin. As part of the discussion, Ghattas describes how parallelism and exascale computing will propel science forward.