Today EXTOLL in Germany released its new TOURMALET high-performance network chip for HPC. “The key demands of HPC are high bandwidth, low latency, and high message rates. The TOURMALET PCI-Express gen3 x16 board shows an MPI latency of 850ns and a message rate of 75M messages per second. The message rate value is CPU-limited, while TOURMALET is designed for well above 100M msg/s.”
Today Italy’s E4 Computer Engineering announced plans to showcase of new NVIDIA GPU-accelerated OpenPOWER servers at ISC 2016 in Frankfurt. “For this edition of ISC16, we wanted to reinforce the message that E4 is a company that actively engages and pursues new technologies’ paths with the aim to deliver leading-edge solutions for a number of demanding environments,” said Piero Altoè, Marketing and BDM Manager, E4 Computer Engineering. “Our priority is to collaborate with organizations such as OpenPOWER Foundation and true visionaries like NVIDIA in order to obtain powerful, scalable and affordable solutions for a number of complex applications and contribute to the development of technologies that have a huge impact on many aspects of our lives.”
For Universities and Colleges that have a traditional infrastructure, adding new programs and applications is a huge endeavor. The IT staff needs to determine if all of the hardware meets the installation requirements and how to deploy these new programs on different models of desktops and notebooks. With a VDI environment that utilizes simple boot-up devices that connect to virtual desktops on the school’s server, the IT staff doesn’t have to worry about the age and capability of each individual PC when installing new software.
While all users of HPC technology want the fastest performance available, price and power consumption always seem to come into play, whether in the initial planning or at a later time. Standard performance measures exist that may or may not relate to an end user’s application mix, but it is important to understand the various benchmark results that go into determining the performance of a CPU, a server or an overall cluster.
Today NVIDIA unveiled a comprehensive global program to support the innovation and growth of startups that are driving new breakthroughs in artificial intelligence and data science. “The NVIDIA Inception Program provides unique tools, resources and opportunities to the waves of entrepreneurs starting new companies, so they can develop products and services with a first-mover advantage.”
Today One Stop Systems announced that its the GPUltima product line now employs Bright Computing’s HPC Cluster Manager software. Bright Computing is a provider of comprehensive software solutions for provisioning and managing HPC clusters. Where conventional computer cluster systems use CPUs as the primary data processor, the GPUltima employs numbers of GPU cards, providing 10 times the performance by adding thousands more cores,” said Steve Cooper, CEO of One Stop Systems. “The GPUltima is completely ‘application-ready’, configured and tested to the customer’s specifications, so that the customer can begin processing immediately. The unique cluster management and monitoring software and the service and support packages that accompany the GPUltima make this a user-friendly system that allows the customer to begin his work without having to configure the cluster.”
Chris Mason from Acceleware presented this talk at GTC 2016. “This session will focus on real life examples including an RF powered contact lens, a wireless capsule endoscopy, and a smart watch. The session will also outline the basics of the subgridding algorithm along with the GPU implementation and the development challenges. Performance results will illustrate the significant reduction in computation times when using a localized subgridded mesh running on an NVIDIA Tesla GPU.”
The MVAPICH User Group (MUG) meeting has issued its Call for Presentations. The event takes place August 15-17 in Columbus, Ohio.
The International Workshop on Communication Architectures at Extreme Scale has published its Advance Agenda. Now in its second year, Exacom 2016 will be held in conjunction with ISC 2016 in Frankfurt on Thursday, June 23, 2016.
In this Programming Throwdown podcast, Mark Harris from Nvidia describes Cuda programming for GPUs. “CUDA is a parallel computing platform and programming model invented by NVIDIA. It enables dramatic increases in computing performance by harnessing the power of the graphics processing unit (GPU). With millions of CUDA-enabled GPUs sold to date, software developers, scientists and researchers are finding broad-ranging uses for GPU computing with CUDA.”