“At ISC 2015, Numascale announced record-breaking results from a shared memory system running the McCalpin STREAM Benchmark, a synthetic benchmark program that measures sustainable memory bandwidth and the corresponding computation rate for simple vector kernels. Numascale’s cache coherent shared memory system, which was targeted for big data analytics, reached 10.06 TBytes/second for the Scale function.”
Today the ISC Cloud & Big Data conference announced that Professor Peter V. Coveney from University College London will be their keynote speaker. Coveney will discuss the current state-of-the-art in the development of personalized medicine, at the event, which takes place Sept. 28-30 in Frankfurt, Germany.
Daniel Gutierrez, Managing Editor, of insideBIGDATA has put together a terrific Guide to Scientific Research. The goal of this paper is to provide a road map for scientific researchers wishing to capitalize on the rapid growth of big data technology for collecting, transforming, analyzing, and visualizing large scientific data sets.
“Supercomputing has reached a level of maturity and capability where many areas of science and engineering are not only advancing rapidly due to computing power, they cannot progress without it. I will illustrate examples from NCSA’s Blue Waters supercomputer, and from major data-intensive projects including the Large Synoptic Survey Telescope, and give thoughts on what will be needed going forward.”
Satoshi Matsuoka from the Tokyo Institute of Technology discusses Big Data at the NCSA Blue Waters Symposium. “The trend towards convergence is not only strategic however but rather inevitable as the Moore’s law ends such that sustained growth in data capabilities, not compute, will advance the capacity and thus the overall capacities towards accelerating research and ultimately the industry.”