In this video, ITIF hosts a hearing on the The Vital Importance of High-Performance Computing to U.S. Competitiveness and National Security. Their recently published report urges U.S. policymakers to take decisive steps to ensure the United States continues to be a world leader in high-performance computing.
“I have been collecting massive amounts of data from my own body over the last ten years, which reveals detailed examples of the episodic evolution of this coupled immune-microbial system. An elaborate software pipeline, running on high performance computers, reveals the details of the microbial ecology and its genetic components. A variety of data science techniques are used to pull biomedical insights from this large data set. We can look forward to revolutionary changes in medical practice over the next decade.”
“Scientific code developers have increasingly been adopting software processes derived from the mainstream (non-scientific) community. Software practices are typically adopted when continuing without them becomes impractical. However, many software best practices need modification and/or customization, partly because the codes are used for research and exploration, and partly because of the combined funding and sociological challenges. This presentation will describe the lifecycle of scientific software and important ways in which it differs from other software development. We will provide a compilation of software engineering best practices that have generally been found to be useful by science communities, and we will provide guidelines for adoption of practices based on the size and the scope of the project.”
Pat McGarry from Ryft presented this talk at the HPC User Forum in Tucson. “Years in the making, the Ryft ONE combines two proven innovations in hardware and software to optimize compute, storage and I/O performance: the Ryft Hybrid FPGA/x86 Compute Platform, which leverages a massively parallel bitwise computing architecture and the Ryft Algorithmic Primitives (RAP) Library.
Peter Bojanic presented this talk at LUG 2016 in Portland. “At LUG 2016, Seagate announced it will incorporate Intel Enterprise Edition for Lustre (IEEL), a big data software platform, into its market-leading ClusterStor storage architecture for high-performance computing. The move will strengthen Seagate’s HPC data storage product line and provide customers with an additional choice of Lustre parallel file systems to help drive advancements in the HPC and big data market.”
Researchers are using the Magnus supercomputer at the Pawsey Centre to explore the mysteries of two shipwrecks involved in Australia’s greatest naval disaster. “The process of generating 3D models from the photographs we’ve taken is very computationally intensive. The time it would take to process half a million photographs using our conventional techniques, using our standard computers, would take about a thousand years, so we needed to do something to bring that time down to something achievable.”
Tejas Karmarker from Microsoft Azure presented this talk at the HPC User Forum. “The performance and scalability of a world-class supercomputing center is now available to everyone, on demand in the cloud. Run your Windows and Linux HPC applications using high performance A8 and A9 compute instances on Azure, and take advantage of a backend network with MPI latency under 3 microseconds and non-blocking 32 Gbps throughput. This backend network includes remote direct memory access (RDMA) technology on Windows and Linux that enables parallel applications to scale to thousands of cores. Azure provides you with high memory and HPC-class CPUs to help you get results fast. Scale up and down based upon what you need and pay only for what you use to reduce costs.”
Rob Peglar from Micron presented this talk at the 2016 MSST Conference. “The growing demands of mobile computing and data centers continue to drive the need for high-capacity, high-performance NAND flash technology. With planar NAND nearing its practical scaling limits, delivering to those requirements has become more difficult with each generation. Enter our 3D NAND technology, which uses an innovative process architecture to provide 3X the capacity of planar NAND technologies while providing better performance and reliability. System designers who build products like laptops, mobile devices and servers can take advantage of 3D NAND’s unprecedented performance to meet the rising data movement needs for businesses and consumers.”
Bo Ewald from D-Wave Systems presented this talk at the HPC User Forum in Tucson. “D-Wave continues to advance the state-of-the-art of quantum computing at a rapid pace, with a number of impressive application results, and the release of their 1000 qubit D-Wave 2X system is another major milestone in the industry,” said Earl Joseph, IDC program vice president for HPC. “Complementing today’s high performance computing systems, quantum computers will likely become an important tool to solve important problems that can’t be solved today.”
In this video, Oklahoma State Director of HPC Dana Brunson describes how the Cowboy supercomputer powers research. “High performance computing is often used for simulations that may be too big, too small, too fast, too slow, too dangerous or too costly, another thing it’s used for involves data. So you may remember the human genome project it took nearly a decade and cost a billion dollars, these sorts of things can now be done over the weekend for under a thousand dollars. Our current super computer is named Cowboy and it was funded by a 2011 National Science Foundation Grant and it has been serving us very well.”