The Department of Energy’s Oak Ridge National Laboratory will add its computational know-how to the battle against cancer through several new projects recently announced at the White House Cancer Moonshot Summit. “ORNL brings to the table our world-class resources in high-performance computing, including the Titan supercomputer, as well as leading experts in the data sciences and neutron analysis, to the fight against cancer,” said ORNL’s Joe Lake, deputy for operations at the HDSI.”
Tony Hey from the Science and Technology Facilities Council presented this talk at The Digital Future conference in Berlin. “Increasingly, scientific breakthroughs will be powered by advanced computing capabilities that help researchers manipulate and explore massive datasets. The speed at which any given scientific discipline advances will depend on how well its researchers collaborate with one another, and with technologists, in areas of eScience such as databases, workflow management, visualization, and cloud computing technologies.”
This week the White House Office of Science and Technology Policy released the Strategic Plan for the NSCI Initiative. “The NSCI strives to establish and support a collaborative ecosystem in strategic computing that will support scientific discovery and economic drivers for the 21st century, and that will not naturally evolve from current commercial activity,” writes Altaf Carim, William Polk, and Erin Szulman from the OSTP in a blog post.
“It’s great to have these incredible servers and incredible processors, but if you don’t have the people to run them – if you don’t have the people that are passionate about supercomputing, we would never get there from here.”Behind all of this magnificent technology are the fantastic faculty, researchers, interns, our corporate partners that are part of this, the National Science Foundation, there are people behind all of the success of the TACC. I think that’s the point we can never forget.”
“With our latest innovations incorporating Intel Xeon Phi processors in a performance and density optimized Twin architecture and 100Gbps OPA switch for high bandwidth connectivity, our customers can accelerate their applications and innovations to address the most complex real world problems.”
Olaf Weber from SGI presented this talk at LUG 2016. “In collaboration with Intel, SGI set about creating support for multiple network connections to the Lustre filesystem, with multi-rail support. With Intel Omni-Path and EDR Infiniband driving to 200Gb/s or 25GB/s per connection, this capability will make it possible to start moving data between a single SGI UV node and the Lustre file system at over 100GB/s.”
This week, the Helix Nebula Science Cloud (HNSciCloud) launched a €5.3 million tender for the establishment of a European hybrid cloud platform. The purpose of the platform is to support the deployment of high-performance computing and big-data capabilities for scientific research. “The European Cloud Initiative will unlock the value of big data by providing world-class supercomputing capability, high-speed connectivity and leading-edge data and software services for science, industry and the public sector,” said Günther H. Oettinger, Commissioner for the Digital Economy and Society.
Today NERSC announced plans to host a new, data-centric event called Data Day. The main event will take place on August 22, followed by a half-day hackathon on August 23. The goal: to bring together researchers who use, or are interested in using, NERSC systems for data-intensive work.
At ISC 2016, HPE introduced new high-performance computing solutions that aim to accelerate HPC adoption by enabling faster time-to-value and increased competitive differentiation through better parallel processing performance, reduced complexity, and faster deployment time. “By combining these latest advancements in Intel Scalable System Framework with the scalability, flexibility and manageability of the HPE Apollo portfolio, customers will gain new levels of performance, efficiency and reliability. In addition, customers will be able to run HPC applications in a massively parallel manner with minimal code modification.”
Jeff Layton from Amazon presented this talk at the AWS Public Sector Summit. “Learn how to build HPC clusters on the fly, leverage Amazon’s Spot market pricing to minimize the cost of HPC jobs, and scale HPC jobs on a small budget, using all the same tools you use today, and a few new ones too.”