Tony Hey from the Science and Technology Facilities Council presented this talk at The Digital Future conference in Berlin. “Increasingly, scientific breakthroughs will be powered by advanced computing capabilities that help researchers manipulate and explore massive datasets. The speed at which any given scientific discipline advances will depend on how well its researchers collaborate with one another, and with technologists, in areas of eScience such as databases, workflow management, visualization, and cloud computing technologies.”
In this video from PASC16, Andrew Lumsdaine from Indiana University gives his perspectives on the conference. “The PASC16 Conference, co-sponsored by the Association for Computing Machinery (ACM) and the Swiss National Supercomputing Centre (CSCS), brings together research across the areas of computational science, high-performance computing, and various domain sciences.”
Today SIGHPC announced the first-ever recipients of the ACM SIGHPC/Intel Computational and Data Science Fellowship. The fellowship is funded by Intel and was announced at the high performance computing community’s SC conference in November of last year. Established to increase the diversity of students pursuing graduate degrees in data science and computational science, the fellowship […]
“We present a procedure of implementing the intermediate profiling for openQCD code that will enable the global reduction of the cost of profiling and optimizing this code commonly used in the lattice QCD community. Our approach is based on well-known SimGrid simulator, which allows for fast and accurate performance predictions of the codes on HPC architectures. Additionally, accurate estimations of the program behavior on some future machines, not yet accessible to us, are anticipated.”
In this video from ISC 2016, Dave Sundstrom from Hewlett Packard Enterprise describes the newly enhanced HPE Software Stack for High Performance Computing. “The HPE Core HPC Software Stack is a complete software set for the creation, optimization, and running of HPC applications. It includes development tools, runtime libraries, a workload scheduler, and cluster management, integrated and validated by Hewlett Packard Enterprise into a single software set. Core HPC Stack uses the included HPC Cluster Setup Tool to simplify and speed the installation of an HPC cluster built with HPE servers.”
Olaf Weber from SGI presented this talk at LUG 2016. “In collaboration with Intel, SGI set about creating support for multiple network connections to the Lustre filesystem, with multi-rail support. With Intel Omni-Path and EDR Infiniband driving to 200Gb/s or 25GB/s per connection, this capability will make it possible to start moving data between a single SGI UV node and the Lustre file system at over 100GB/s.”
Today NERSC announced plans to host a new, data-centric event called Data Day. The main event will take place on August 22, followed by a half-day hackathon on August 23. The goal: to bring together researchers who use, or are interested in using, NERSC systems for data-intensive work.
At ISC 2016, HPE introduced new high-performance computing solutions that aim to accelerate HPC adoption by enabling faster time-to-value and increased competitive differentiation through better parallel processing performance, reduced complexity, and faster deployment time. “By combining these latest advancements in Intel Scalable System Framework with the scalability, flexibility and manageability of the HPE Apollo portfolio, customers will gain new levels of performance, efficiency and reliability. In addition, customers will be able to run HPC applications in a massively parallel manner with minimal code modification.”
Jeff Layton from Amazon presented this talk at the AWS Public Sector Summit. “Learn how to build HPC clusters on the fly, leverage Amazon’s Spot market pricing to minimize the cost of HPC jobs, and scale HPC jobs on a small budget, using all the same tools you use today, and a few new ones too.”
The first Joint International Workshop on Parallel Data Storage and Data Intensive Scalable Computing Systems (PDSW-DISCS’16) has issued its Call for Papers. As a one day event held in conjunction with SC16, the workshop will combine two overlapping communities to to address some of the most critical challenges for scientific data storage, management, devices, and processing infrastructure. To learn more, we caught up with workshop co-chairs Dean Hildebrand (IBM) and Shane Canon (LBNL).