In a perfect world, there would be one version of all compilers, libraries, and profilers. To make things even easier, hardware would never change. However, technology marches forward, and such a world does not exist. Software tool features are updated, bugs are fixed, and performance is increased. Developers need these improvements but at the same time must manage these differences.
Today Cray announced that the Swiss National Supercomputing Centre (CSCS) has installed a Cray CS-Storm cluster supercomputer to power the operational numerical weather forecasts run by the Swiss Federal Office of Meteorology and Climatology (MeteoSwiss). This is the first time a GPU-accelerated supercomputer has been used to run production numerical weather models for a major national weather service.
HPC developers want to write code and create new applications. The advanced nature of HPC often requires that this process be associated with specific hardware and software environment present on a given HPC resource. Developers want to extract the maximum performance from HPC hardware and at the same time not get mired down in the complexities of software tool chains and dependencies.
The AMD graphics cards are uniquely equipped with AMD Multiuser GPU technology embedded into the GPU delivering consistent and predictable performance,” said Sean Burke, AMD corporate vice president and general manager, Professional Graphics. “When these AMD GPUs are appropriately configured to the needs of an organization, end users get the same access to the GPU no matter their workload. Each user is provided with the virtualized performance to design, create and execute their workflows without any one user tying up the entire GPU.”
HPC and Beer have always had a certain affinity ever since the days when Cray Research would include a case of Leinenkugel’s with every supercomputer. Now, Brian Caulfield from Nvidia writes that a Pennsylvania startup is using GPUs and Deep Learning technologies to enable brewers to make better beer.
“I will describe a decade-long, multi-disciplinary, multi-institutional effort spanning neuroscience, supercomputing and nanotechnology to build and demonstrate a brain-inspired computer and describe the architecture, programming model and applications. I also will describe future efforts in collaboration with DOE to build, literally, a “brain-in-a-box”. The work was built on simulations conducted on Lawrence Livermore National Laboratory’s Dawn and Sequoia HPC systems in collaboration with Lawrence Berkeley National Laboratory.”
The Distributed European Computing Initiative (DECI) in Europe has issued its 13th Call for Proposals for HPC Compute Resources. “Administered by PRACE, DECI enables European researchers to obtain access to the most powerful national (Tier-1) computing resources in Europe regardless of their country of origin or employment and to enhance the impact of European science and technology at the highest level.”
In this video from the SF Big Analytics Meetup, Bryan Catanzaro from Baidu presents: Why is HPC so important to AI? “We built Deep Speech because we saw the opportunity to re-conceive speech recognition in light of the new capabilities afforded by Deep Learning, to take advantage of even larger datasets to solve even harder problems.”