“It’s great to have these incredible servers and incredible processors, but if you don’t have the people to run them – if you don’t have the people that are passionate about supercomputing, we would never get there from here.”Behind all of this magnificent technology are the fantastic faculty, researchers, interns, our corporate partners that are part of this, the National Science Foundation, there are people behind all of the success of the TACC. I think that’s the point we can never forget.”
“With our latest innovations incorporating Intel Xeon Phi processors in a performance and density optimized Twin architecture and 100Gbps OPA switch for high bandwidth connectivity, our customers can accelerate their applications and innovations to address the most complex real world problems.”
Olaf Weber from SGI presented this talk at LUG 2016. “In collaboration with Intel, SGI set about creating support for multiple network connections to the Lustre filesystem, with multi-rail support. With Intel Omni-Path and EDR Infiniband driving to 200Gb/s or 25GB/s per connection, this capability will make it possible to start moving data between a single SGI UV node and the Lustre file system at over 100GB/s.”
Deep learning is a method of creating artificial intelligence systems that combine computer-based multi-layer neural networks with intensive training techniques and large data sets to enable analysis and predictive decision making. A fundamental aspect of deep learning environments is that they transcend finite programmable constraints to the realm of extensible and trainable systems. Recent developments in technology and algorithms have enabled deep learning systems to not only equal but to exceed human capabilities in the pace of processing vast amounts of information.
Today the FlyElephant team announced the release of the FlyElephant 2.0 platform for High Performance Computing. Versioin 2.0 enhancements include: internal expert community, collaboration on projects, public tasks, Docker and Jupyter support, a new file storage system and work with HPC clusters.
This week, the Helix Nebula Science Cloud (HNSciCloud) launched a €5.3 million tender for the establishment of a European hybrid cloud platform. The purpose of the platform is to support the deployment of high-performance computing and big-data capabilities for scientific research. “The European Cloud Initiative will unlock the value of big data by providing world-class supercomputing capability, high-speed connectivity and leading-edge data and software services for science, industry and the public sector,” said Günther H. Oettinger, Commissioner for the Digital Economy and Society.
In this CCTV video, Alberto Alonso describes his new supercomputer, Breogan, that has the ability to modernize Mexico’s antiquated stock market.
Since it was launched in February, the Breogan computer has generated $475,000 for his company GACS. The algorithm it uses is much faster than the computers in Mexico’s stock exchange. “What makes the Breogan computer so unique, is that it finds attractive opportunities in the market and it buys and sells automatically when it sees a trading opportunity.”
Today NERSC announced plans to host a new, data-centric event called Data Day. The main event will take place on August 22, followed by a half-day hackathon on August 23. The goal: to bring together researchers who use, or are interested in using, NERSC systems for data-intensive work.
Today Seagate unveiled a first-of-its kind, high-capacity drive that can help data centers more easily accommodate exponential data growth, while still maintaining high levels of computing power and performance. The two terabyte (TB) version of its Nytro XM1440 M.2 non-volatile memory express (NVMe) Solid State Drive (SSD) is the highest-capacity, enterprise-class M.2 NVMe SSD available today, making it well suited for demanding enterprise applications that require fast data access, capacity and processing. “This latest version of the Nytro XM1440 M.2 NVMe SSD is the first of its kind and pushes the boundaries for enterprises, so they don’t have to sacrifice speed for access and availability. They can continue to scale operations in an efficient manner and still get the most value from their data, but without the extra overhead.”
At ISC 2016, HPE introduced new high-performance computing solutions that aim to accelerate HPC adoption by enabling faster time-to-value and increased competitive differentiation through better parallel processing performance, reduced complexity, and faster deployment time. “By combining these latest advancements in Intel Scalable System Framework with the scalability, flexibility and manageability of the HPE Apollo portfolio, customers will gain new levels of performance, efficiency and reliability. In addition, customers will be able to run HPC applications in a massively parallel manner with minimal code modification.”