In this TACC podcast, Joe Stubbs from the Texas Advanced Computing Centter describes potential benefits to scientists of open container platform Docker in supporting reproducibility, NSF-funded Agave API. “As more scientists share not only their results but their data and code, Docker is helping them reproduce the computational analysis behind the results. What’s more, Docker is one of the main tools used in the Agave API platform, a platform-as-a-service solution for hybrid cloud computing developed at TACC and funded in part by the National Science Foundation.”
“Research computational workflows consist of several pieces of third party software and, because of their experimental nature, frequent changes and updates are commonly necessary thus raising serious deployment and reproducibility issues. Docker containers are emerging as a possible solution for many of these problems, as they allow the packaging of pipelines in an isolated and self-contained manner. This presentation will introduce our experience deploying genomic pipelines with Docker containers at the Center for Genomic Regulation (CRG). I will discuss how we implemented it, the main issues we faced, the pros and cons of using Docker in an HPC environment including a benchmark of the impact of containers technology on the performance of the executed applications.”
In this special guest feature from Scientific Computing World, Dr Bruno Silva from The Francis Crick Institute in London writes that new cloud technologies will make the cloud even more important to scientific computing. “The emergence of public cloud and the ability to cloud-burst is actually the real game-changer. Because of its ‘infinite’ amount of resources (effectively always under-utilized), it allows for a clear decoupling of time-to-science from efficiency. One can be somewhat less efficient in a controlled fashion (higher cost, slightly more waste) to minimize time-to-science when required (in burst, so to speak) by effectively growing the computing estate available beyond the fixed footprint of local infrastructure – this is often referred to as the hybrid cloud model. You get both the benefit of efficient infrastructure use, and the ability to go beyond that when strictly required.”
“With Docker v1.9 a new networking system was introduced, which allows multi-host network- ing to work out-of-the-box in any Docker environment. This talk provides an introduction on what Docker networking provides, followed by a demo that spins up a full SLURM cluster across multiple machines. The demo is based on QNIBTerminal, a Consul backed set of Docker Images to spin up a broad set of software stacks.”
“UberCloud specializes in running HPC workloads on a broad spectrum of infrastructures, anywhere from national centers to public Cloud services. This session will be review of the learnings of UberCloud Experiments performed by industry end users. The live demonstration will cover how to achieve peak simulation performance and usability in the Cloud and national centers, using fast interconnects, new generation CPU’s, SSD drives and UberCloud technology based on Linux containers.”
Today ThinkParQ announced that the complete BeeGFS parallel file system is now available as open source. Developed specifically for performance-critical environments, the BeeGFS parallel file system was developed with a strong focus on easy installation and high flexibility, including converged setups where storage servers are also used for compute jobs. By increasing the number of servers and disks in the system, performance and capacity of the file system can simply be scaled out to the desired level, seamlessly from small clusters up to enterprise-class systems with thousands of nodes.
Over at the UberCloud, Wolfgang Gentzsch writes that, despite the ever increasing complexity of CAE tools, hardware, and system components engineers have never been this close to ubiquitous CAE as a common tool for every engineer.
Bright Cluster Manager Version 7.2 is out today, a new release that “extends insight, integration, and ease-of-use for managing clustered and cloud-based IT infrastructures.” The new release incorporates a wide range of new features and significantly enhanced monitoring capabilities. “Bright Computing has always prided itself on upgrading its product offerings to respond to new technological trends and user feedback,” said Martijn de Vries, Chief Technology Officer of Bright Computing. “The enhancements we have made in Version 7.2 address recent technology trends, such as the rapid adoption of containers to drive IT efficiency, and support our customers’ ongoing need to stay on top of their dynamic, complex, and converging IT infrastructures.”
On December 18, Oracle closed an agreement to acquire StackEngine, an Austin-based startup focused on Docker automation. All StackEngine employees will be joining Oracle as part of Oracle Public Cloud.
Last week at SC15, NEC Corporation announced that the Flemish Supercomputer Center (VSC) has selected an LX-series supercomputer. With a peak performance of 623 Teraflops, the new system will be the fastest in Belgium, ranking amongst the top 150 biggest and fastest supercomputers in the world. Financed by the Flemish minister for Science and Innovation in Belgium, the infrastructure will cost 5.5 million Euro.