In this episode of This Week in HPC, Michael Feldman and Addison Snell from Intersect360 Research discuss a recent National Research Council report that indicates the NSF is falling behind in supercomputing. After the break, they look at new Big Data Appliances from Cray and SGI.
This Week in HPC: Supercomputing Future Uncertain for NSF, and Cray and SGI Unveil Big Data Appliances
“If you have something that’s not optimal on a single laptop, that’s no big deal, but if you’re on a large national resource [an HPC system], first you’re preventing other researchers from using the machines, and second, you’re wasting taxpayer money,” Biros said. “Efficient computing is about maximizing science per dollar.”
An innovative artist named Janet Echelman is using HPC to build amazing floating sculptures. “She builds and installs some of the most beautiful and permanent outdoor installations to be found anywhere. This is because her installations are made of a special netting that flows and billows with the wind, yet is strong and stable enough to weather the forces of nature year round. In 2011, Janet explained her journey to creating such lovely sculptures in her TED talk, seen now by over 1.2 million people around the world. I invite you to become one of them.”
“High performance computing (HPC) is inextricably linked to innovation, fueling breakthroughs in science, engineering, and business. HPC is viewed as a cost-effective tool for speeding up the R&D process, and two-thirds of all US-based companies that use HPC say that “increasing performance of computational models is a matter of competitive survival.”
In this podcast, the Radio Free HPC team discuss the possibility of a future where the Big 3 (Amazon, Google, and Microsoft) figure out that Cloud is not profitable and pull the plug. If that Cloud Apocalypse sounds far fetched, a look at recent AWS revenue numbers may prompt you to stock up your bomb shelter.
“Slagter remarked that a cloud environment meant at least three actors had both practical and legal responsibility in keeping data private and secure: the cloud provider itself was responsible for the physical security of the building where the servers were located as well as the security protocols used; the ISV had responsibility for the security of the application that was being run; and the customer had to have a set of security policies and procedures governing who had access to the portal into the cloud and who was licensed, within the customer’s own company, to use the application software and access the data.”
In this special guest feature, Penguin Computing Chief Technology Officer Phil Pokorny writes that Open Compute Project has great potential to spark HPC innovation. In just three and a half years since Facebook announced the Open Compute Project (OCP), the cooperative industry effort is showing exceptional progress toward delivering significantly more efficient hardware into the […]
Over at the Ansys Blog, Wim Slagter has posted the second segment of his blog post six common myths about HPC for engineering simulation. “I would be really ignorant to state that it is easy to deploy and manage a HPC cluster. As a matter of fact, sizing, building, integrating, provisioning and supporting a cluster infrastructure requires highly specialized IT expertise that is often lacking users of engineering simulation software.”
“The independent software vendors, which make the programs that scientists and engineers would like to use in the cloud, are not sure how they can license their software for such an environment. As Felix Wolfheimer of CST remarked, in a moment of candor: “There is a lot of fear in the sales department about opening up the licensing model’ so that software licenses will be flexible enough for use in the cloud.”