Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Apply Now for Summer of HPC 2017 in Barcelona

“The PRACE Summer of HPC is an outreach and training program that offers summer placements at top High Performance Computing centers across Europe to late-stage undergraduates and early-stage postgraduate students. Up to twenty top applicants from across Europe will be selected to participate. Participants will spend two months working on projects related to PRACE technical or industrial work and produce a report and a visualization or video of their results.”

Allinea Adds CUDA 8 Support for GPU Developers

Today Allinea Software announces availability of its new software release, version 6.1, which offers full support for programming parallel code on the Pascal GPU architecture, CUDA 8 from Nvidia. “The addition of Allinea tools into the mix is an exciting one, enabling teams to accurately measure GPU utilization, employ smart optimization techniques and quickly develop new CUDA 8 code that is bug and bottleneck free,” said Mark O’Connor, VP of Product Management at Allinea.

ArrayFire v3.4 Parallel Computing Library Speeds Machine Learning

Today ArrayFire released the latest version of their ArrayFire open source library of parallel computing functions supporting CUDA, OpenCL, and CPU devices. ArrayFire v3.4 improves features and performance for applications in machine learning, computer vision, signal processing, statistics, finance, and more.

Tutorial: GPU Performance Nuggets

In this video from the 2016 Blue Waters Symposium, GPU Performance Nuggets – Carl Pearson and Simon Garcia De Gonzalo from the University of Illinois present: GPU Performance Nuggets. “In this talk, we introduce a pair of Nvidia performance tools available on Blue Waters. We discuss what the GPU memory hierarchy provides for your application. We then present a case study that explores if memory hierarchy optimization can go too far.”

AMD Boltzmann Initiative Promotes HPC Freedom of Choice

AMD’s motivation for developing these open-source GPU tools is based on an opportunity to remove the added complexity of proprietary programming frameworks to GPU application development. “If successful, these tools – or similar versions – could help to democratize GPU application development, removing the need for proprietary frameworks, which then makes the HPC accelerator market much more competitive for smaller players. For example, HPC users could potentially use these tools to convert CUDA code into C++ and then run it on an Intel Xeon co-processor.”

CocoLink Using Consumer GPUs for Deep Learning

CocoLink, a subsidiary of Seoul National University, in collaboration with Orange Silicon Valley, has upgraded its KLIMAX 210 server with 20 of the latest GeForce 1080 GPUs – with the eventual goal of scaling the single 4U rack to more than 200 teraflops.

Video: Lorena Barba Keynote at PYCON 2016

In this video from PYCON 2016 in Portland, Lorena Barba from George Washinton University presents: Beyond Learning to Program, Education, Open Source Culture, Structured Collaboration, and Language. “PyCon is the largest annual gathering for the community using and developing the open-source Python programming language.”

Podcast: CUDA Programming for GPUs

In this Programming Throwdown podcast, Mark Harris from Nvidia describes Cuda programming for GPUs. “CUDA is a parallel computing platform and programming model invented by NVIDIA. It enables dramatic increases in computing performance by harnessing the power of the graphics processing unit (GPU). With millions of CUDA-enabled GPUs sold to date, software developers, scientists and researchers are finding broad-ranging uses for GPU computing with CUDA.”

Video: UPC++ Parallel Programming Extension

In this video from the 2016 OpenFabrics Workshop, Zili Zheng from LBNL presents: UPC++. “UPC++ is a parallel programming extension for developing C++ applications with the partitioned global address space (PGAS) model. UPC++ has demonstrated excellent performance and scalability with applications and benchmarks such as global seismic tomography, Hartree-Fock, BoxLib AMR framework and more. In this talk, we will give an overview of UPC++ and discuss the opportunities and challenges of leveraging modern network features.”

Brookhaven Lab is the latest Nvidia GPU Research Center

Today Nvidia announced that Brookhaven National Laboratory has been named a 2016 GPU Research Center. “The center will enable Brookhaven Lab to collaborate with Nvidia on the development of widely deployed codes that will benefit from more effective GPU use, and in the delivery of on-site GPU training to increase staff and guest researchers’ proficiency,” said Kerstin Kleese van Dam, director of CSI and chair of the Lab’s Center for Data-Driven Discovery.