Today IBM unveiled a series of new servers designed to help propel cognitive workloads and to drive greater data center efficiency. Featuring a new chip, the Linux-based lineup incorporates innovations from the OpenPOWER community that deliver higher levels of performance and greater computing efficiency than available on any x86-based server. “Collaboratively developed with some of the world’s leading technology companies, the new Power Systems are uniquely designed to propel artificial intelligence, deep learning, high performance data analytics and other compute-heavy workloads, which can help businesses and cloud service providers save money on data center costs.”
Hailing from Norway, big-memory appliance maker Numascale has been a fixture at the ISC conference since the company’s formation in 2008. At ISC 2016, Numascale was noticeably absent from the show and the word on the street was that the company was retooling their NumaConnect™ technology around NVMe. To learn more, we caught up with Einar Rustad, Numascale’s CTO.
Today Movidius announced that the company is being acquired by Intel. As a deep learning startup, Movidius’s mission is to give the power of sight to machines.
Today, the National Geospatial-Intelligence Agency and NSF released 3-D topographic maps that show Alaska’s terrain in greater detail than ever before. Powered by the Blue Waters supercomputer, the maps are the result of a White House Arctic initiative to inform better decision-making in the Arctic. “We can’t live without Blue Waters now,” said Paul Morin, head of the University of Minnesota’s Polar Geospatial Center. “The supercomputer itself, the tools the Blue Waters team at NCSA developed, the techniques they’ve come up with in using this hardware. Blue Waters is changing the way digital terrain is made and that is changing how science is done in the Arctic.”
“We have enhanced Bright Cluster Manager 7.3 so our customers can quickly and easily deploy new deep learning techniques to create predictive applications for fraud detection, demand forecasting, click prediction, and other data-intensive analyses,” said Martijn de Vries, Chief Technology Officer of Bright Computing. “Going forward, customers using Bright to deploy and manage clusters for deep learning will not have to worry about finding, configuring, and deploying all of the dependent software components needed to run deep learning libraries and frameworks.”
“Bridges has enabled early scientific successes, for example in metagenomics, organic semiconductor electrochemistry, genome assembly in endangered species, and public health decision-making. Over 2,300 users currently have access to Bridges for an extremely wide range of research spanning neuroscience, machine learning, biology, the social sciences, computer science, engineering, and many other fields.”
Paul Messina presented this talk at the 2016 Argonne Training Program on Extreme-Scale Computing. “The President’s NSCI initiative calls for the development of Exascale computing capabilities. The U.S. Department of Energy has been charged with carrying out that role in an initiative called the Exascale Computing Project (ECP). Messina has been tapped to lead the project, heading a team with representation from the six major participating DOE national laboratories: Argonne, Los Alamos, Lawrence Berkeley, Lawrence Livermore, Oak Ridge and Sandia. The project program office is located at Oak Ridge.
Indiana University plans to unveil three new HPC resources at a launch event on Sept 1: Jetstream, Big Red II+, and Diet. “With these new systems, IU continues to provide our researchers the leading-edge computational tools needed for the scale of today’s research problems,” said Brad Wheeler, IU vice president for IT and CIO. “Each of these systems is quite distinct in its purpose to meet the needs of our researchers and students.”
Coming in the second half of 2016: The HPE Apollo 6500 System provides the tools and the confidence to deliver high performance computing (HPC) innovation. The system consists of three key elements: The HPE ProLiant XL270 Gen9 Server tray, the HPE Apollo 6500 Chassis, and the HPE Apollo 6000 Power Shelf. Although final configurations and performance are not yet available, the system appears capable of delivering over 40 teraflop/s double precision, and significantly more in single or half precision modes.
George Slota presented this talk at the Blue Waters Symposium. “In recent years, many graph processing frameworks have been introduced with the goal to simplify analysis of real-world graphs on commodity hardware. However, these popular frameworks lack scalability to modern massive-scale datasets. This work introduces a methodology for graph processing on distributed HPC systems that is simple to implement, generalizable to broad classes of graph algorithms, and scales to systems with hundreds of thousands of cores and graphs of billions of vertices and trillions of edges.”