The move to network offloading is the first step in co-designed systems. A large amount of overhead is required to service the huge number of packets required for modern data rates. This amount of overhead can significantly reduce network performance. Offloading network processing to the network interface card helped solve this bottleneck as well as some others.
Paul Messina presented this talk at the 2016 Argonne Training Program on Extreme-Scale Computing. The U.S. Department of Energy has been charged with carrying out that role in an initiative called the Exascale Computing Project (ECP). Messina has been tapped to lead the project, heading a team with representation from the six major participating DOE national laboratories: Argonne, Los Alamos, Lawrence Berkeley, Lawrence Livermore, Oak Ridge and Sandia. The project program office is located at Oak Ridge.
Indiana University plans to unveil three new HPC resources at a launch event on Sept 1: Jetstream, Big Red II+, and Diet. “With these new systems, IU continues to provide our researchers the leading-edge computational tools needed for the scale of today’s research problems,” said Brad Wheeler, IU vice president for IT and CIO. “Each of these systems is quite distinct in its purpose to meet the needs of our researchers and students.”
Thomas Schulthess presented this talk at the MVAPICH User Group. “Implementation of exascale computing will be different in that application performance is supposed to play a central role in determining the system performance, rather than just considering floating point performance of the high-performance Linpack benchmark. This immediately raises the question as to what the yardstick will be, by which we measure progress towards exascale computing. I will discuss what type of performance improvements will be needed to reach kilometer-scale global climate and weather simulations. This challenge will probably require more than exascale performance.”
Seven women who work in IT departments at research institutions around the country have been selected to help build and operate the high performance SCinet conference network at SC16. The announcement came from the Women in IT Networking at SC program, also known as WINS.
We’d like to invite our readers to participate in our new HPC Customer Experience Survey. It’s an effort to better understand our readers and what is really happening out there in the world of High Performance Computing. “This survey should take less than 10 minutes to complete. All information you provide will be treated as private and kept confidential.”
“Galaxies are complex—many physical processes operate simultaneously, and over a huge range of scales in space and time. As a result, accurately modeling the formation and evolution of galaxies over the lifetime of the universe presents tremendous technical challenges. In this talk I will describe some of the important unanswered questions regarding galaxy formation, discuss in general terms how we simulate the formation of galaxies on a computer, and present simulations (and accompanying published results) that the Enzo collaboration has recently done on the Blue Waters supercomputer. In particular, I will focus on the transition from metal-free to metal-enriched star formation in the universe, as well as the luminosity function of the earliest generations of galaxies and how we might observe it with the upcoming James Webb Space Telescope.”
Karl Schulz from Intel presented this talk at the 4th Annual MVAPICH User Group meeting. “Today, many supercomputing sites spend considerable effort aggregating a large suite of open-source projects on top of their chosen base Linux distribution in order to provide a capable HPC environment for their users. This presentation will introduce a new, open-source HPC community (OpenHPC) that is focused on providing HPC-centric package builds for a variety of common building-blocks in an effort to minimize duplication, implement integration testing to gain validation confidence, incorporate ongoing novel R&D efforts, and provide a platform to share configuration recipes from a variety of sites.”
In this podcast, the Radio Free HPC team looks at why it’s so difficult for new processor architectures to gain traction in HPC and the datacenter. Plus, we introduce a new regular feature for our show: The Catch of the Week.
In this video, students describe their learning experience at the 2016 PRACE Summer of HPC program in Barcelona. “The PRACE Summer of HPC is a PRACE outreach and training program that offers summer placements at top HPC centers across Europe to late-stage undergraduates and early-stage postgraduate students. Up to twenty top applicants from across Europe will be selected to participate. Participants spend two months working on projects related to PRACE technical or industrial work and produce a report and a visualization or video of their results.”