“The impending end of traditional MOSFET scaling has sparked research into preserving HPC performance improvements through alternative computational models. To better shape our strategy, we need to understand where each technology is headed and where it will be in a span of 20 years. This workshop brings together experts who develop or use promising technologies to present the state of their work, and spark a discussion on the promise and detriments of each approach.”
Applications that can take advantage of the new vectorization capabilities of the Intel Xeon Phi processor will show tremendous performance gains. “When considering vectorization, there are different tools that can assist the developer in determining where to look further. The first is to look at the optimization reports that are generated by the Intel compiler and then to also use the Vector Analyzer that can give specific advice on what to do to get more vectorization from the code.”
“Over two days we’ll delve into a wide range of interests and best practices – in applications, tools and techniques and share new insights on the trends, technologies and collaborative partnerships that foster this robust ecosystem. Designed to be highly interactive, the open forum will feature industry notables in keynotes, technical sessions, workshops and tutorials. These highly regarded subject matter experts (SME’s) will share their works and wisdom covering everything from established HPC disciplines to emerging usage models from old-school architectures and breakthrough applications to pioneering research and provocative results. Plus a healthy smattering of conversation and controversy on endeavors in Exascale, Big Data, Artificial Intelligence, Machine Learning and much much more!”
Today Asetek announced that its server pump has achieved 200 million hours of reliable operation in real world use. Installed at end-user locations as diverse as Singapore and Norway, Asetek pumps have run fault-free for the equivalent of 22,000 years. “To date, we have not had a single server pump failure at any of our data center installations around the world,” said Mette Nørmølle, Vice President of Engineering. “Our low-pressure architecture is the key to enabling a cost-effective solution that is relied on by data centers demanding unrivaled performance and maximum uptime.”
“I’m pleased to have the opportunity to lead this important Council,” said Dr. J. Michael McQuade of United Technologies Corporation, who will serve as the first Chair of the ECP Industry Council. “Exascale level computing will help industry address ever more complex, competitively important problems, ones which are beyond the reach of today’s leading edge computing systems. We compete globally for scientific, technological and engineering innovations. Maintaining our lead at the highest level of computational capability is essential for our continued success.”
“Billed as an exposition into ‘The Future of Cloud HPC Simulation,’ the event brought together experts in high-performance computing and simulation, cloud computing technologists, startup founders, and VC investors across the technology landscape. In addition to product demonstrations with Rescale engineers, including the popular Deep Learning workshop led by Mark Whitney, Rescale Director of Algorithms, booths featuring ANSYS, Microsoft Azure, Data Collective, and Microsoft Ventures offered interactive sessions for attendees.”
HPC4Mfg will host their first annual High Performance Computing for Manufacturing Industry Engagement Day on March 2-3 in San Diego. With a theme of “Spurring Innovation in U.S. Manufacturing Through Advanced Computing,” the conference will bring together representatives from U.S. manufacturing, national laboratories, universities, and consortiums to discuss the recent advancements in manufacturing realized through the application of HPC and how leveraging HPC expertise through public-private partnerships has lowered the risk of adoption.
Argonne has selected 10 computational science and engineering research projects for its Aurora Early Science Program starting this month. Aurora, a massively parallel, manycore Intel-Cray supercomputer, will be ALCF’s next leadership-class computing resource and is expected to arrive in 2018. The Early Science Program helps lay the path for hundreds of other users by doing actual science, using real scientific applications, to ready a future machine. “As with any bleeding edge resource, there’s testing and debugging that has to be done,” said ALCF Director of Science Katherine Riley.
In-Memory Computing can accelerate traditional applications by using a memory first design. Applicable to a wide range of domains, In-Memory Computing and In-Memory Data Grids take advantage of the latest trends in computer systems technology. “In-memory computing is designed to address some of the most critical and real-time task requirements today. This include real-time fraud detection, biometrics and border security and financial risk analytics. All of these use cases require very low latency access to data from very large amounts of data, which results in faster and more accurate decisions.”
The IEEE International Workshop on High-Performance Interconnection Networks in the Exascale and Big-Data Era (HiPINEB) has published its program agenda. Held in conjunction with the IEEE HPCA 2017 Conference, HiPINEB takes place Feb 5, in Austin, Texas.