Michael Resch from HLRS gave this rousing talk at the HPC User Forum. “HLRS supports national and European researchers from science and industry by providing high-performance computing platforms and technologies, services and support. Supercomputer Hazel Hen, a Cray XC40-system, is at the heart of the HPC system infrastructure of the HLRS. With a peak performance of 7.42 Petaflops (quadrillion floating point operations per second), Hazel Hen is one of the most powerful HPC systems in the world (position 8 of TOP500, 11/2015) and is the fastest supercomputer in the European Union. The HLRS supercomputer, which was taken into operation in October 2015, is based on the Intel Haswell Processor and the Cray Aries network and is designed for sustained application performance and high scalability.”
Rangan Sukumar from ORNL presented this talk at the HPC User Forum in Tucson. “ORiGAMI is a tool for discovering and evaluating potentially interesting associations and creating novel hypothesis in medicine. ORiGAMI will help you “connect the dots” across 70 million knowledge nuggets published in 23 million papers in the medical literature. The tool works on a ‘Knowledge Graph’ derived from SEMANTIC MEDLINE published by the National Library of Medicine integrated with scalable software that enables term-based, path-based, meta-pattern and analogy-based reasoning principles.”
Hadoop and Spark clusters have a reputation for being extremely difficult to configure, install, and tune, but help is on the way. The good folks at Cluster Monkey are hosting a crash course entitled Apache Hadoop with Spark in One Day. “After completing the workshop attendees will be able to use and navigate a production Hadoop cluster and develop their own projects by building on the workshop examples.”
Today ISC 2016 announced that five renowned experts in computational science will participate in their new Distinguished Speaker series. Topics will include exascale computing efforts in the US, the next supercomputers in development in Japan and China, cognitive computing advancements at IBM, and quantum computing research at NASA.
In this video from the HPC User Forum in Tucson, Saul Gonzalez Martirena from NSF provides an update on the NSCI initiative. “As a coordinated research, development, and deployment strategy, NSCI will draw on the strengths of departments and agencies to move the Federal government into a position that sharpens, develops, and streamlines a wide range of new 21st century applications. It is designed to advance core technologies to solve difficult computational problems and foster increased use of the new capabilities in the public and private sectors.”
“Cavium ThunderX has significant differentiation in the 64-bit ARM market as Cavium is the first ARMv8 vendor to deliver dual socket support with full ARMv8.1 implementation and significant advantage in CPU cores with 48 cores per socket. In addition, ThunderX supports large memory capacity (512GB per socket, 1TB in a 2S system) with excellent memory bandwidth and low memory latency. In addition, ThunderX includes multiple 10 GbE / 40GbE network interfaces delivering excellent IO throughput. These features enable ThunderX to deliver the core performance & scale out capability that the HPC market requires.”
If you are in the Northwest and you happen to like surf and turf, have I got a deal for you! Dell is hosting a series of Big Data lunch events in Seattle and Portland at the end of April. On April 26, Dell brings the event to Blueacre Seafood in Seattle. In Portland, lunch is on April 27 at the mighty Fogo de Chao, a Brazilian steak house for the Where’s the Beef? crowd. They’re also coming to Flemings in Salt Lake City on April 28.
In this video from the 2016 GPU Technology Conference, Jason Pai from Supermicro describes the new 1028GQ-TRT SuperServer. With support for up to four Nvidia Tesla K80 GPUs, the 1U superserver offers extreme compute density in 1U of rack space. “From HPC to Deep Learning and Big Data Analytics, denser, more powerful GPU solutions have become a necessity in order to service the next generation of GPU-accelerated applications. At GTC, Supermicro demonstrated how these applications have progressed, and how its GPU solutions are influencing this evolution.”
According to the latest Intersect360 Research site census data, of the 50 most popular application packages mentioned by HPC users, 34 have offer GPU support, including 9 of the top 10. As is evident from the number of GPU-accelerated applications available in areas such as chemical research, physics, structural analysis, and visualization, the use of this accelerator technology has become well established in the HPC user community. “According to the latest Intersect360 Research site census data, of the 50 most popular application packages mentioned by HPC users, 34 have offer GPU support, including 9 of the top 10. As is evident from the number of GPU-accelerated applications available in areas such as chemical research, physics, structural analysis, and visualization, the use of this accelerator technology has become well established in the HPC user community.”
Parallel file systems have become the norm for HPC environments. While typically used in high end simulations, these parallel file systems can greatly affect the performance and thus the customer experience when using analytics from leading organizations such as SAS. This whitepaper is an excellent summary of how parallel file systems can enhance the workflow and insight that SAS Analytics gives.