This week Berkeley Lab joined the OpenMP Architecture Review Board (ARB), a group of leading hardware and software vendors and research organizations creating the standard for the most popular shared-memory parallel programming model in use today.
In this video, Ken Kraft from Intel describes his experiences as a participant in SC Conference Mentor-Protege program. “The SC15 conference is broadening its mission to build a strong and diverse HPC student community at both the undergraduate and graduate levels, through professional development programs, opportunities to learn from mentors, and engagement with SC’s technical sessions.”
Bridges is a uniquely capable supercomputer designed to help researchers facing challenges in Big Data to work more intuitively. Called Bridges, the new system will consist of tiered, large-shared-memory resources with nodes having 12TB, 3TB, and 128GB each, dedicated nodes for database, web, and data transfer, high-performance shared and distributed data storage, Hadoop acceleration, powerful new CPUs and GPUs, and a new, uniquely powerful interconnection network.
There is always different levels of importance assigned to various data files in a computer system, specifically a very large system that is storing petabytes of data. In order to maximize the use of the highest speed storage, Hierarchical Storage Management (HSM) was developed to move and store data within easy use of users, yet at the appropriate speed and price.
According to IDC, SGI has shipped approximately 8 percent of of all the Hadoop servers in production today. In fact, did you know that SGI introduced the word “Big Data” to supercomputing in 1996? Jorge Titinger, SGI President and CEO, shares SGI’s history in helping to design, develop, and deploy Hadoop clusters. (NOTE: Straw was substituted for actual hay to avoid any potential allergic reactions.)
CHPC Communications Manager Noxolo Moyake describes how the South African Center for High Performance Computing is fostering the use of supercomputing in the region. “CHPC will be a critical resource for tackling the big scientific challenges of the continent, such as the Square Kilometer Array (SKA), climate research, and data-intensive applications in the big data regime.”
“Geoscientists need tools to allow them to rapidly develop algorithms that run fast on accelerators, while at the same time deliver portability and improve productivity. They demand a single source code, with no need to maintain multiple code paths, using a high-level approach that presents a low learning curve. OpenACC provides directives-based approaches to rapidly accelerating applications for GPUs and other parallel architectures. This talk will serve as an introduction to programming with OpenACC 2.0.”
Because the requirements in terms of energy, emissions, and fuel efficiency have not been globally standardized, vehicle manufacturers are making use of increasingly sophisticated simulation to get new vehicles developed as quickly and cheaply as possible while making sure their newest models meet these standards.
“In 2015, the HPCC will begin plans to acquire a new machine that will singularly increase our available compute power at MSU. With MSU’s continuing commitment to iCER, the HPCC, the creation of a new department focused on the computational sciences, and a new Datacenter on the horizon, the future of HPC at MSU is bright.”