The SGI UV system is uniquely suited for bioinformatics and genomics by providing the computational capabilities and global shared memory architecture needed for even the most demanding sequencing and analytic tasks, including post sequencing and other data intensive workflows.
“Large scale HPC IO is usually done either with a file per process or to a single shared file. Single shared file IO does not scale well in Lustre compared to file per process. This presentation from Cray’s Patrick Farrell will give details, examine the reasons for this, and explore existing and potential solutions. Group locks and a new feature, lock ahead, will be discussed in the context of strided IO.”
Today RSC Group announced the installation of a MVS-10P MP Petastream supercomputer at the Joint Supercomputer Center of the Russian Academy of Sciences (JSCC RAS). As is the first project in the Commonwealth of Independent States based on the latest Intel Xeon Phi 7120D coprocessors, the supercomputer weighs in at #20 of the Top50 list of the most powerful computing systems in Russia and CIS.
“Pleiades, one of the world’s most powerful supercomputers, represents NASA’s state-of-the-art technology for meeting the agency’s supercomputing requirements, enabling NASA scientists and engineers to conduct modeling and simulation for NASA missions. Powered by Lustre, this distributed-memory SGI ICE cluster is connected with InfiniBand in a dual-plane hypercube technology.”
Today the Oak Ridge Leadership Computing Facility (OLCF) announced a significant step forward in the delivery of Summit, the next-generation supercomputer IBM will deliver in 2017 to greatly advance the nation’s energy and science agenda. Following a formal call for proposals and a review process, ORNL revealed 13 science application readiness projects selected as part of the Center for Accelerated Application Readiness (CAAR) program.
“The second generation of SDSC’s Data Oasis Lustre storage is coming online to support Comet, a new XSEDE cluster targeted at the long tail of science. The servers have been designed with Lustre on ZFS in mind, and also update the network to use bonded 40GbE interfaces. The raw storage totals 7.7 PB and are again based on commodity hardware provided by Aeon Computing, maintaining our focus on cost.”