Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

SC17 Invited Talk Preview: High Performance Machine Learning

Over at the SC17 Blog, Brian Ban begins his series of SC17 Session Previews with a look at a talk on High Performance Big Data. “Deep learning, using GPU clusters, is a clear example but many Machine Learning algorithms also need iteration, and HPC communication and optimizations.”

Interview: Dr. Thomas Sterling on New Approaches to Efficient Supercomputing

In this video from KAUST, Professor Thomas Sterling, Professor of Intelligent Systems Engineering at Indiana University, shares his thoughts on new approaches to energy efficient supercomputing. “Our technical strategy focuses on the research and development of advanced technologies for extreme-scale computing and future exascale systems, including the following key elements: Execution Models; Runtime Systems; Graph Processing; Programming Interfaces; Compilers, Libraries, and Languages; Systems Architecture (Architecture, Power/Energy, Fault Tolerance, Networking), and Extreme Scale Applications and Visualization.”

Thomas Sterling Presents: HPC Runtime System Software for Asynchronous Multi-Tasking

Thomas Sterling presented this Invited Talk at SC16. “Increasing sophistication of application program domains combined with expanding scale and complexity of HPC system structures is driving innovation in computing to address sources of performance degradation. This presentation will provide a comprehensive review of driving challenges, strategies, examples of existing runtime systems, and experiences. One important consideration is the possible future role of advances in computer architecture to accelerate the likely mechanisms embodied within typical runtimes. The talk will conclude with suggestions of future paths and work to advance this possible strategy.”

OpenSFS Reinvents Itself with a Focus on Lustre Users

Today OpenSFS, the nonprofit organization dedicated to the success of the Lustre file system, announced organizational changes to give its users—the researchers and scientists who use Lustre every day—more of a say in how OpenSFS moves forward. “Now the time has come for those who rely most on Lustre—its users—to guide OpenSFS into the future,” said Stephen Simms, OpenSFS president and manager, high performance file systems at Indiana University.

Indiana University to Launch Three New HPC Systems

Indiana University plans to unveil three new HPC resources at a launch event on Sept 1: Jetstream, Big Red II+, and Diet. “With these new systems, IU continues to provide our researchers the leading-edge computational tools needed for the scale of today’s research problems,” said Brad Wheeler, IU vice president for IT and CIO. “Each of these systems is quite distinct in its purpose to meet the needs of our researchers and students.”

XDMoD Tool to Measure Impact of Campus Cyberinfrastructure

Thanks to a grant from the National Science Foundation, Indiana University is developing an online service that will make it easier for university administrators to understand the importance of funding related to IT systems based at their institutions. The novel functionality will be available as a module for Open XDMoD (XD Metrics on Demand), which was developed by the University at Buffalo Center for Computational Research (CCR).

OpenSFS Releases Lustre 2.8.0 for LUG 2016 Conference

Today the Open Scalable File Systems (OpenSFS) community announced the release of Lustre 2.8.0, the fastest and most scalable parallel file system. OpenSFS, founded in 2010 to advance Lustre development, is the premier non-profit organization promoting the use of Lustre and advancing its capabilities through coordinated releases of the Lustre file system.

Thomas Sterling Weighs in on the OpenHPC Community

Over at the OpenHPC Blog, Thomas Sterling from Indiana University describes why the Crest Project has joined the OpenHPC Community: “We want to make a specific contribution. By associating ourselves with an emergent framework, in which we could benefit from the work of many different people interested in different things but under a unifying guidance of scaffolding interfaces, we were able to achieve our objectives in low cost, HPC for end users. If, and I have to say if, OpenHPC does this right, you will provide that framework.”

Video: Jetstream to Make HPC a Breeze for Researchers

In this video, researchers describe how the Jetstream project at Indiana University. Jetstream is a user-friendly cloud environment designed to give researchers access to interactive computing and data analysis resources on demand, whenever and wherever they want to analyze their data. It will provide a library of virtual machines designed to do discipline specific scientific analysis. Software creators and researchers will also be able to create their own customized virtual machines or their own private computing system within Jetstream.

Dell Panel on the Convergence of HPC, Big Data, and Cloud

“Modeling and simulation have been the primary usage of high performance computing (HPC). But the world is changing. We now see the need for rapid, accurate insights from large amounts of data. To accomplish this, HPC technology is repurposed. Likewise the location where the work gets done is not entirely the same either. Many workloads are migrating to massive cloud data centers because of the speed of execution. In this panel, leaders in computing will share how they, and others, integrate tradition and innovation (HPC technologies, Big Data analytics, and Cloud Computing) to achieve more discoveries and drive business outcomes.”