Today Rogue Wave Software announced it is working with IBM to help make open source software (OSS) support more available. This will help provide comprehensive, enterprise-grade technical support for OSS packages. “With our ten-year history in open source, organizations can feel confident in our ability to resolve issues,” said Richard Sherrard, director of product management at Rogue Wave Software. “We have tier-3 and 4 enterprise architects that offer round-the-clock support for entire ecosystems. We are long-standing experts when it comes to OSS and proud to be working with IBM.”
Today D-Wave Systems announced details of its most advanced quantum computing system, featuring a new 2000-qubit processor. The announcement is being made at the company’s inaugural users group conference in Santa Fe, New Mexico. The new processor doubles the number of qubits over the previous generation D-Wave 2X system, enabling larger problems to be solved and extending D-Wave’s significant lead over all quantum computing competitors. The new system also introduces control features that allow users to tune the quantum computational process to solve problems faster and find more diverse solutions when they exist. In early tests these new features have yielded performance improvements of up to 1000 times over the D-Wave 2X system.
“As more organizations turn to high performance computing to process large data sets, demand is growing for scalable and secure data centre solutions. The source, availability and reliability of the power grid infrastructure is becoming a critical factor in a data centre site selection decision,” said Jeff Monroe, CEO at Verne Global. “Verne Global is able to deliver EI a forward-thinking path for growth with a solution that combines unparalleled costs savings with operational efficiencies to support their data-intensive research.”
Today Quantum Corp. announced that two of Europe’s premier research institutions are using the company’s StorNext workflow storage as the foundation for managing their growing data and enabling a range of scientific initiatives. “With the StorNext platform, we have removed barriers to research,” Thomas Disper, CISO and Head of IT, Max Planck Institute for Chemistry. “It allows us to provide a lot more capacity quickly and easily. We don’t need to give research teams data limits, and storage for new projects can be ready in an afternoon.”
Over at the ANSYS Blog, Tony DeVarco writes that the company worked with SGI to break a world record for HPC scalability. “Breaking last year’s 129,024 core record by more than 16,000 cores, SGI was able to run the ANSYS provided 830 million cell gas combustor model from 1,296 to 145,152 CPU cores.This reduces the total solver wall clock time to run a single simulation from 20 minutes for 1,296 cores to a mere 13 seconds using 145,152 cores and achieving an overall scaling efficiency of 83%.”
The National Computational Infrastructure in Canberra, Australia’s national advanced computing facility, is the first Australian institution to deploy the latest generation of Intel Xeon Phi processors, formerly code named Knights Landing. “NCI is leading efforts in the scientific community to tune applications for Intel Xeon Phi processors,” explains Dr Muhammad Atif, NCI’s HPC Systems and Cloud Services Manager. “We have identified a large number of applications that will benefit from this hardware and software paradigm, including those applications in the domains of computational physics, computational chemistry and climate research.”
Today Cadence Design Systems announced several important deliveries in its collaboration with TSMC to advance 7nm FinFET designs for mobile and high-performance computing platforms. Working together, Cadence and TSMC have developed some of the first design IP offerings for the 7nm process, offering early IP access to protocols that are optimized for and most relevant to mobile and HPC applications.
This week Minimal Metrics announced an early-adopter program for PerfMiner, which uses lightweight, and pervasive performance data collection technology, automates its collection, and mines the data for key performance indicators. These indicators were developed through Minimal Metrics’ extensive experience tuning HPC and enterprise application performance, presented in an audience-specific, drill-down hierarchy that provides accountability for site productivity down to the performance of individual application threads.
Today Verne Global announced Volkswagen is moving more than 1 MW of high performance computing applications to the company’s datacenter in Iceland. The company will take advantage of Verne Global’s hybrid data center approach – with variable resiliency and flexible density – to support HPC applications in its continuous quest to develop cutting-edge cars and automotive technology.
In this RCE Podcast, Brock Palen and Jeff Squyres speak with Gregory Kurtzer about Singularity, a container solution for HPC and research environments. “Singularity allows a non-privileged user to “swap out” the operating system on the host for one they control. So if the host system is running RHEL6 but your application runs in Ubuntu, you can create an Ubuntu image, install your applications into that image, copy the image to another host, and run your application on that host in it’s native Ubuntu environment.”