Today Altair announced that eleven international customers participated in the company’s recent HPC Cloud Challenge. The contest was set up to demonstrate the benefits of leveraging the cloud for large-scale design exploration in the area of computer-aided engineering. Organizations of all sizes from manufacturing and academic fields participated in the Challenge, utilizing Altair technologies in structural, CFD and design studies, and expressed great satisfaction with the program overall.
The Extreme Science and Engineering Discovery Environment (XSEDE), a five-year project supported by the US National Science Foundation, has awarded 324 million cpu hours, valued at $16.2 million, to 150 research projects throughout the US.
Michael Resch from HLRS gave this rousing talk at the HPC User Forum. “HLRS supports national and European researchers from science and industry by providing high-performance computing platforms and technologies, services and support. Supercomputer Hazel Hen, a Cray XC40-system, is at the heart of the HPC system infrastructure of the HLRS. With a peak performance of 7.42 Petaflops (quadrillion floating point operations per second), Hazel Hen is one of the most powerful HPC systems in the world (position 8 of TOP500, 11/2015) and is the fastest supercomputer in the European Union. The HLRS supercomputer, which was taken into operation in October 2015, is based on the Intel Haswell Processor and the Cray Aries network and is designed for sustained application performance and high scalability.”
In this slidecast, Gilad Shainer from Mellanox describes the advantages of InfiniBand and the company’s off-loading network architecture for HPC. “The path to Exascale computing is clearly paved with Co-Design architecture. By using a Co-Design approach, the network infrastructure becomes more intelligent, which reduces the overhead on the CPU and streamlines the process of passing data throughout the network. A smart network is the only way that HPC data centers can deal with the massive demands to scale, to deliver constant performance improvements, and to handle exponential data growth.”
Rangan Sukumar from ORNL presented this talk at the HPC User Forum in Tucson. “ORiGAMI is a tool for discovering and evaluating potentially interesting associations and creating novel hypothesis in medicine. ORiGAMI will help you “connect the dots” across 70 million knowledge nuggets published in 23 million papers in the medical literature. The tool works on a ‘Knowledge Graph’ derived from SEMANTIC MEDLINE published by the National Library of Medicine integrated with scalable software that enables term-based, path-based, meta-pattern and analogy-based reasoning principles.”
Today Intersect360 Research published a new research report on the Hyperscale market. “This report provides definitions, segmentations, and dynamics of the hyperscale market and describes its scope, the end-user applications it touches, and the market drivers and dampers for future growth. It is the foundational report for the Intersect360 Research hyperscale market advisory service.”
Matt Vaughan from TACC presented this talk at the HPC User Forum. “Jetstream is the first user-friendly, scalable cloud environment for XSEDE. The system enables researchers working at the “long tail of science” and the creation of truly customized virtual machines and computing architectures. It has a web-based user interface integrated with XSEDE via Globus Auth. The architecture is derived from the team’s collective experience with CyVerse Atmosphere, Chameleon and Quarry. The system also fosters reproducible, sharable computing with geographically isolated clouds located at Indiana University and TACC.”
In this video, Siraj Rival from Twilio presents a quick tutorial on How to Build a Neural Net in 4 Minutes. Siraj describes himself as the Bill Nye of Computer Science.
“This talk will present RDMA-based designs using OpenFabrics Verbs and heterogeneous storage architectures to accelerate multiple components of Hadoop (HDFS, MapReduce, RPC, and HBase), Spark and Memcached. An overview of the associated RDMAenabled software libraries being designed and publicly distributed as a part of the HiBD project.”
Hadoop and Spark clusters have a reputation for being extremely difficult to configure, install, and tune, but help is on the way. The good folks at Cluster Monkey are hosting a crash course entitled Apache Hadoop with Spark in One Day. “After completing the workshop attendees will be able to use and navigate a production Hadoop cluster and develop their own projects by building on the workshop examples.”