Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Job of the Week: Cloud Performance Architect at Oracle in Seattle

Oracle in Seattle is seeking a Cloud Performance Architect in our Job of the Week. “As a Cloud Performance Architect, you will be a leading contributor in improving performance of Oracle’s latest Cloud Services Technologies. You will take an active role in the definition and evolution of standard practices and procedures. Additionally, you will be responsible for defining and developing software for tasks associated with the developing, designing and debugging of software applications or operating systems. If you have a passion for improving and creating high performance software products, this is the place where you can make a difference.”

The New Nvidia HGX-1: GPU Power for Machine Learning at Hyperscale

Rob Ober from Nvidia describes the company’s new HGX-1 reference platform for GPU computing. “Powered by NVIDIA Tesla GPUs and NVIDIA NVLink high-speed interconnect technology; the HGX-1 comes as AI workloads—from autonomous driving and personalized healthcare to superhuman voice recognition—are taking off in the cloud.”

How DDN Technology Speeds Machine Learning at Scale

Today DDN announced that large commercial machine learning programs in manufacturing, autonomous vehicles, smart cities, medical research and natural-language processing are overcoming production scaling challenges with DDN’s large-scale, high-performance storage solutions. “The high performance and flexible sizing of DDN systems make them ideal for large-scale machine learning architectures,” said Joel Zysman, director of advanced computing at the Center for Computational Science at the University of Miami. “With DDN, we can manage all our different applications from one centrally located storage array, which gives us both the speed we need and the ability to share information effectively. Plus, for our industrial partnership projects that each rely on massive amounts of instrument data in areas like smart cities and autonomous vehicles, DDN enables us to do mass transactions on a scale never before deemed possible. These levels of speed and capacity are capabilities that other providers simply can’t match.”

Exxact Powers HPC & Deep Learning at the GPU Technology Conference

In this video from the 2017 GPU Technology Conference, Jason Chen from Exxact Corporation describes the company’s innovative GPU solutions for HPC & Deep Learning workloads. “Exxact Corporation is a global, value-added supplier of computing products and solutions. We provide services in IT distribution and system integration for HPC, Big Data, Cloud, and AV applications and markets, while servicing many of world’s top national labs, research universities, online retailers, and Fortune 100 & 1000 companies.”

RoCE Initiative Launches Online Product Directory

Today the RoCE Initiative at the InfiniBand Trade Association announced the availability of the RoCE Product Directory. The new online resource is intended to inform CIOs and enterprise data center architects about their options for deploying RDMA over Converged Ethernet (RoCE) technology within their Ethernet infrastructure.

Behind the Curtain of Backblaze Hard Drive Stats

In this video from the MSST 2017 Mass Storage Conference, Andrew Klein from Backblaze shares the results from their latest disk drive reliability study. “For the last four years, Backblaze has collected and reported on the failure rates and SMART stats of the hard drives in use in our data centers. Currently we have over 80,000 drives ranging from 3 to 8TB. Let’s take a look at what we learned over the years about hard drives, including failure rates, by model, and the ability to predict drive failure before it happens.”

Supermicro Showcases Powerful GPU Solutions at GTC 2017

In this video from GTC 2017, Rodolfo Campos from Supermicro describes the company’s innovative solutions for GPU-accelerated computing. “Leveraging our extensive portfolio of GPU solutions, customers can massively scale their compute clusters to accelerate their most demanding deep learning, scientific and hyperscale workloads with fastest time-to-results, while achieving maximum performance per watt, per square foot, and per dollar.”

Eric Barton Joins DDN as CTO for Software-Defined Storage

Today DDN appointed Eric Barton as the company’s chief technology officer for software-defined storage. In this role, Barton will lead the company’s strategic roadmap, technology architecture and product design for DDN’s newly created Infinite Memory Engine business unit. Barton brings with him more than 30 years of technology innovation, entrepreneurship and expertise in networking, distributed systems and storage software.

Transaction Processing Performance Council Launches TPCx-HS Big Data Benchmark

Today the Transaction Processing Performance Council (TPC) announced the immediate availability of TPCx-HS Version 2, extending the original benchmark’s scope to include the Spark execution framework and cloud services. “Enterprise investment in Big Data analytics tools is growing exponentially, to keep pace with the rapid expansion of datasets,” said Tariq Magdon-Ismail, chairman of the TPCx-HS committee and staff engineer at VMware. “This is leading to an explosion in new hardware and software solutions for collecting and analyzing data. So there is enormous demand for robust, industry standard benchmarks to enable direct comparison of disparate Big Data systems across both hardware and software stacks, either on-premise or in the cloud. TPCx-HS Version 2 significantly enhances the original benchmark’s scope, and based on industry feedback, we expect immediate widespread interest.”

Kimberly Keeton from HPE Presents: Memory-Driven Computing

Kimberly Keeton from HPE gave this talk at the MSST Conference. “Data growth and data analytics requirements are outpacing the compute and storage technologies that have provided the foundation of processor-driven architectures for the last five decades. This divergence requires a deep rethinking of how we build systems, and points towards a memory-driven architecture, where memory is the key resource and everything else, including processing, revolves around it.