MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

NAG Optimizes C and C++ Algorithms for ARM-based Cavium ThunderX Processors

The Numerical Algorithms Group (NAG) has engineered NAG C Library algorithms to execute efficiently on Cavium ThunderX ARMv8-A based Workload Optimized Processors. Preliminary results, announced at ISC 2016, show excellent scaling across 96 cores of ThunderX in a dual socket configuration.

OpenHPC Establishes Leadership & Releases Initial Software Stack

Today the Linux Foundation announced a set of technical, leadership and member investment milestones for OpenHPC, a Linux Foundation project to develop an open source framework for High Performance Computing environments. “The OpenHPC community has quickly paved a path of collaborative development that is highly inclusive of stakeholders invested in HPC-optimized software,” said Jim Zemlin, executive director, The Linux Foundation. “To see OpenHPC members include the world’s leading computing labs, universities, and hardware experts, illustrates how open source unites the world’s leading technologists to share technology investments that will shape the next 30+ years of computing.”

NAG Rolls Out Software Modernization Service

Today the Numerical Algorithms Group (NAG) has announced the NAG Software Modernization Service. The new service solves the porting and performance challenges faced by customers wishing to use the capabilities of modern computing systems, such as multi-core CPUs, GPUs and Xeon Phi. NAG HPC software engineering experts modernize the code to enable portability to appropriate architectures, optimize for performance and assure robustness.

1998 ATLAS Paper Wins SC16 Test of Time Award

SC16 has announced the winner of their Test of Time Award. This year the winning paper “Automatically Tuned Linear Algebra Software” by Clint Whaley and Jack Dongarra. The paper, which has received hundreds of citations with new citations still appearing, is about ATLAS – an autotuning, optimized implementation of the Basic Linear Algebra Subprograms (BLAS).

Video: Speeding Up Code with the Intel Distribution for Python

David Bolton from Slashdot shows how ‘embarrassingly parallel’ code can be sped up over 2000x (not percent) by utilizing Intel tools including the Intel Python compiler and OpenMP. “The Intel Distribution for Python* 2017 Beta program is now available. The Beta product adds new Python packages like scikit-learn, mpi4py, numba, conda, tbb (Python interfaces to Intel Threading Building Blocks) and pyDAAL (Python interfaces to Intel Data Analytics Acceleration Library). “

Intel Developer Summer Workshops Coming to Stanford

Intel is offering a 4-part summer series of developer training workshops at Stanford University to introduce high performance computing tools.

Video: Using GPUs for Electromagnetic Simulations of Human Interface Technology

Chris Mason from Acceleware presented this talk at GTC 2016. “This session will focus on real life examples including an RF powered contact lens, a wireless capsule endoscopy, and a smart watch. The session will also outline the basics of the subgridding algorithm along with the GPU implementation and the development challenges. Performance results will illustrate the significant reduction in computation times when using a localized subgridded mesh running on an NVIDIA Tesla GPU.”

HSA 1.1 Specification Adds Multi-Vendor Architecture Support

The Heterogeneous System Architecture (HSA) Foundation has released the HSA 1.1 specification, significantly enhancing the ability to integrate open and proprietary IP blocks in heterogeneous designs. The new specification is the first to define the interfaces that enable IP blocks from different vendors to communicate, interoperate and collectively compose an HSA system.

Podcast: CUDA Programming for GPUs

In this Programming Throwdown podcast, Mark Harris from Nvidia describes Cuda programming for GPUs. “CUDA is a parallel computing platform and programming model invented by NVIDIA. It enables dramatic increases in computing performance by harnessing the power of the graphics processing unit (GPU). With millions of CUDA-enabled GPUs sold to date, software developers, scientists and researchers are finding broad-ranging uses for GPU computing with CUDA.”

Allinea Tools to Power NEC Supercomputer at VSC in Belgium

The Flemish Supercomputer Center (VSC) is planning the deployment of a new NEC cluster that will represent Belgium’s largest investment in HPC to date. To help VSC unleash the potential of the system, Allinea software tools will be used to speed up code performance. “We are delighted to be supporting VSC in providing better education to its users around code efficiency,” said David Lecomber, CEO and Founder of Allinea. “The fact of the matter is, without visibility of code performance, researchers cannot get the full value from HPC. By appreciating how their code makes a difference to project delivery, researchers can achieve more for less cost. By underlining this best practice, VSC’s approach is one that is refreshing and makes great economic sense.”