In this video from the HPC Advisory Council Switzerland Conference, Alexander Moskovsky from RSC presents: Leading Energy Efficient Supercomputing to Exascale.
Watch to get the latest on the Coral supercomputer coming to ORNL. “ORNL’s supercomputing program has grown from humble beginnings to deliver some of the most powerful systems in the world. On the way, it has helped researchers deliver practical breakthroughs and new scientific knowledge in climate, materials, nuclear science, and a wide range of other disciplines.”
This week at the OpenPOWER Summit in San Jose, the OpenPOWER Foundation showed off real hardware for the first time with 13 systems including a a prototype HPC server from IBM and new microprocessor customized for China. Built collaboratively by OpenPOWER members, the new solutions exploit the POWER architecture to provide more choice, customization and performance to customers, including hyperscale data centers.
In this video, Tesla does a field test of a software update that will bring powerful auto-steering functionality to its Model S fleet.
“Deep Learning has transformed many important tasks, including speech and image recognition. Deep Learning systems scale well by absorbing huge amounts of data to create accurate models. The computational resources afforded by GPUs have been instrumental to this scaling. However, as Deep Learning has become more mainstream, it has generated some hype, and has been linked to everything from world peace to evil killer robots. In this talk, Dr. Ng will help separate hype from reality, and discuss potential ways that Deep Learning technologies can benefit society in the short and long term.”
This week insideHPC will be streaming live keynotes from the GPU Technology Conference in San Jose. Today’s keynote will feature Google Senior Fellow Jeff Dean. “Google has built large-scale computer systems for training neural networks, and then applied these systems to a wide variety of problems that have traditionally been very difficult for computers.”
In this video from the 2015 OFS Developer’s Workshop, Katie Antypas from LBNL describes preparations for the Cori supercomputer. “We need to emphasize here that the Knights Landing processor is self-hosted, and so that means it’s not an accelerator. It’s not a coprocessor and the particular kernel processor that will be having for NERSC-8, will have more than 60 cores and it will have multiple hardware threads for the core. That’s a lot, right? Having 60 cores per node with multiple hardware threads. That a significant increase from both our Hopper and Edison system, which has 24 cores each.”