Researchers using the JUQUEEN supercomputer have computed the small mass difference between protons and neutrons. The fact that the neutron is slightly more massive than the proton is the reason why atomic nuclei have exactly those properties that make our world and ultimately our existence possible. The fact that the neutron is slightly more massive […]
This week at the OpenPOWER Summit in San Jose, the OpenPOWER Foundation showed off real hardware for the first time with 13 systems including a a prototype HPC server from IBM and new microprocessor customized for China. Built collaboratively by OpenPOWER members, the new solutions exploit the POWER architecture to provide more choice, customization and performance to customers, including hyperscale data centers.
In this episode, the Radio Free HPC team wraps up the GPU Technology Conference. The theme of the show this year was Deep Learning, a topic that is heating up the market for GPUs with challenges like image recognition and self-driving cars. As a sister conference, the OpenPOWER Summit this week in San Jose showcased the first OpenPower hardware, including a prototype HPC server from IBM that will pave the way to the two IBM/Nvidia/Mellanox Coral supercomputers expected in 2017.
“I came to IBM via the acquisition of Platform Computing. There’s also been other IBM assets around HPC, namely GPFS. What’s been the evolution of those items as well and how they really come together under this concept of software-defined infrastructure, and how we’re now taking these capabilities and expanding them into other initiatives that have sort of bled into the HPC space.”
Clusters that are purchased for specific applications tend not to be flexible as workloads change. What is needed is an infrastructure that can expand or contract as the workload changes. IBM, a recognized leader in High Performance Computing is applying its expertise in both HPC and Cloud computing to bring together the technologies to create the HPC Cloud.
“Imagine an entire IT infrastructure controlled not by hands and hardware, but by software. One in which application workloads such as big data, analytics, simulation and design are serviced automatically by the most appropriate resource, whether running locally or in the cloud. A Software Defined Infrastructure enables your organization to deliver IT services in the most efficient way possible, optimizing resource utilization to accelerate time to results and reduce costs.”
“A new approach is needed to help clients address the cost and complexity driven by tremendous data growth,” said Tom Rosamilia, Senior Vice President, IBM Systems. “Traditional storage is inefficient in today’s world where the value of each piece of data is changing all the time. IBM is revolutionizing storage with our Spectrum Storage software that helps clients to more efficiently leverage their hardware investments to extract the full business value of data.”