In this special guest feature from Scientific Computing World, Darren Watkins from Virtus Data Centres explains the importance of building a data centre from the ground up to support the requirements of HPC users – while maximizing productivity, efficiency and energy usage. “The reality for many IT users is they want to run analytics that –with the growth of data – have become too complex and time critical for normal enterprise servers to handle efficiently.”
In this video from LUG 2016 in Portland, Steve Simms from Indiana University presents: Lustre 101 – A Quick Overview. Now in its 14th year, the Lustre User Group is the industry’s primary venue for discussion and seminars on the Lustre parallel file system and other open source file system technologies.
Mark Seamans from SGI presented this talk at the HPC User Forum in Tucson. “As the trusted leader in high performance computing, SGI helps companies find answers to the world’s biggest challenges. Our commitment to innovation is unwavering and focused on delivering market leading solutions in Technical Computing, Big Data Analytics, and Petascale Storage. Our solutions provide unmatched performance, scalability and efficiency for a broad range of customers.”
“Single Root I/O Virtualization (SR-IOV) technology has been steadily gaining momentum for high-performance interconnects such as InfiniBand. SR-IOV can deliver near native performance but lacks locality-aware communication support. This talk presents an efficient approach to build HPC clouds based on MVAPICH2 over OpenStack with SR-IOV. We discuss the high-performance design of virtual machine-aware MVAPICH2 library over OpenStack-based HPC Clouds with SR-IOV. A comprehensive performance evaluation with micro-benchmarks and HPC applications has been conducted on an experimental OpenStack-based HPC cloud and Amazon EC2. The evaluation results show that our design can deliver near bare-metal performance.”
“Cycle Computing software leverages cloud resources to make computation in the cloud productive at any scale, by orchestrating workflows, managing data, balancing cloud options, and enabling users in a secure, controlled way. Our software works with public cloud providers like Amazon Web Services, Google Cloud Platform, and Microsoft Azure, as well as with internal and private cloud environments.”
Today the University of Iceland unveiled a new supercomputer that will boost research in a range of scientific areas. Manufactured by Lenovo, the cluster was funded by the Research Infrastructure Fund Iceland with matching funds from the University of Iceland, Reykjavik University.
Leo Reiter from Nimbix presented this deck at the HPC User Forum. “Nimbix is a pure high performance computing cloud built for volume, speed and simplicity. We give people the tools and the processing power to solve their biggest, toughest problems. We give you the freedom to imagine new possibilities, to test the limits of reality, and to model the future. For most workloads, Nimbix is far less expensive than building, running and maintaining your own supercomputer. It’s also more efficient at spinning up, executing, completing the job and delivering your results — which saves you time and money. And our user-friendly platform means you invest less in development and infrastructure.”
In this video from the 2016 GPU Technology Conference, Greg Schmidt from Hewlett Packard Enterprise describes the new Apollo 6500 server. “With up to eight high performance NVIDIA GPU cards designed for maximum transfer bandwidth, the HPE Apollo 6500 System is purpose-built for deep learning applications. Its high ratio of GPUs to CPUs, dense 4U form factor and efficient design enable organizations to run deep learning recommendation algorithms faster and more efficiently, significantly reducing model training time and accelerating the delivery of real-time results, all while controlling costs.”
In this special guest feature from Scientific Computing World, ISC’s Nages Sieslack highlights a convergence of technologies around HPC, a focus of the ISC High Performance conference, which takes place June 19-23 in Frankfurt. “In addition to the theme of convergent HPC technologies, this year’s conference will also offer two days of sessions in the industry track, specially designed to meet the interests of commercial users. Our focus is Industrie 4.0, a German strategic initiative conceived to take a leading role in pioneering industrial IT, which is currently revolutionizing engineering in the manufacturing sector.”
In this RCE Podcast, Marcel Kornacker from Cloudera describes the Impala project. Impala brings scalable parallel database technology to Hadoop, enabling users to issue low-latency SQL queries to data stored in HDFS and Apache HBase without requiring data movement or transformation. Impala is integrated with Hadoop to use the same file and data formats, metadata, security and resource management frameworks used by MapReduce, Apache Hive, Apache Pig and other Hadoop software.