In this video, Siraj Rival from Twilio presents a quick tutorial on How to Build a Neural Net in 4 Minutes. Siraj describes himself as the Bill Nye of Computer Science.
The hpc-ch Forum on Intra- and Inter-Site Networking has posted its Call for Participation. Hosted by the University of Zurich, the event will take place Thursday, May 19, 2016.
“This talk will present RDMA-based designs using OpenFabrics Verbs and heterogeneous storage architectures to accelerate multiple components of Hadoop (HDFS, MapReduce, RPC, and HBase), Spark and Memcached. An overview of the associated RDMAenabled software libraries being designed and publicly distributed as a part of the HiBD project.”
Hadoop and Spark clusters have a reputation for being extremely difficult to configure, install, and tune, but help is on the way. The good folks at Cluster Monkey are hosting a crash course entitled Apache Hadoop with Spark in One Day. “After completing the workshop attendees will be able to use and navigate a production Hadoop cluster and develop their own projects by building on the workshop examples.”
Today ISC 2016 announced that five renowned experts in computational science will participate in their new Distinguished Speaker series. Topics will include exascale computing efforts in the US, the next supercomputers in development in Japan and China, cognitive computing advancements at IBM, and quantum computing research at NASA.
If you are in the Northwest and you happen to like surf and turf, have I got a deal for you! Dell is hosting a series of Big Data lunch events in Seattle and Portland at the end of April. On April 26, Dell brings the event to Blueacre Seafood in Seattle. In Portland, lunch is on April 27 at the mighty Fogo de Chao, a Brazilian steak house for the Where’s the Beef? crowd. They’re also coming to Flemings in Salt Lake City on April 28.
In this video from the 2016 GPU Technology Conference, Jason Pai from Supermicro describes the new 1028GQ-TRT SuperServer. With support for up to four Nvidia Tesla K80 GPUs, the 1U superserver offers extreme compute density in 1U of rack space. “From HPC to Deep Learning and Big Data Analytics, denser, more powerful GPU solutions have become a necessity in order to service the next generation of GPU-accelerated applications. At GTC, Supermicro demonstrated how these applications have progressed, and how its GPU solutions are influencing this evolution.”
Parallel file systems have become the norm for HPC environments. While typically used in high end simulations, these parallel file systems can greatly affect the performance and thus the customer experience when using analytics from leading organizations such as SAS. This whitepaper is an excellent summary of how parallel file systems can enhance the workflow and insight that SAS Analytics gives.
Addison Snell from Intersect360 Research presented this talk at the Switzerland HPC Conference. “Based on updated research studies, Addison Snell of Intersect360 Research will present on forward-looking topics for HPC and Hyperscale markets. With an expanding look at hyper- scale, Intersect360 Research will describe the size and influence of the market, including evolving standards like Open Compute Project, OpenStack, and Beiji/Scorpio. Intersect360 Research has also investigated users’ plans for evaluating competing processing and interconnect options, including Xeon, Xeon Phi, GPU, FPGA, POWER, ARM, InfiniBand, and OmniPath.”
Today Penguin Computing announced Open Compute Project (OCP)-based systems that reinforce both its continued collaboration with NVIDIA and new options in Penguin Computing’s Magna family of OpenPOWER-based servers. “Customers benefit when we partner with exceptional organizations like NVIDIA, the OpenPOWER Foundation and Open Compute Foundation in developing our systems,” said Jussi Kukkonen, Director Product Management, Penguin Computing. “An essential part of our mission is to provide customers with form factor flexibility, choice of architecture and peak performance, which are all hallmarks of Penguin Computing.”