Rangan Sukumar from ORNL presented this talk at the HPC User Forum in Tucson. “ORiGAMI is a tool for discovering and evaluating potentially interesting associations and creating novel hypothesis in medicine. ORiGAMI will help you “connect the dots” across 70 million knowledge nuggets published in 23 million papers in the medical literature. The tool works on a ‘Knowledge Graph’ derived from SEMANTIC MEDLINE published by the National Library of Medicine integrated with scalable software that enables term-based, path-based, meta-pattern and analogy-based reasoning principles.”
Matt Vaughan from TACC presented this talk at the HPC User Forum. “Jetstream is the first user-friendly, scalable cloud environment for XSEDE. The system enables researchers working at the “long tail of science” and the creation of truly customized virtual machines and computing architectures. It has a web-based user interface integrated with XSEDE via Globus Auth. The architecture is derived from the team’s collective experience with CyVerse Atmosphere, Chameleon and Quarry. The system also fosters reproducible, sharable computing with geographically isolated clouds located at Indiana University and TACC.”
The hpc-ch Forum on Intra- and Inter-Site Networking has posted its Call for Participation. Hosted by the University of Zurich, the event will take place Thursday, May 19, 2016.
“Over the last years the OFA community has shown the potential of using high performance networks (InfiniBand) to boost the performance of virtualized cloud environments, however, the network reconfiguration challenges still continue to exist. In this session we present the work we have been doing on InfiniBand subnet management and routing, in the context of dynamic cloud environments. This work includes, but not limited to, techniques in order to provide better management scalability when virtual machines are live migrating, tenant network isolation in multi-tenant environments, and fast performance-driven network reconfiguration.”
“This talk will present RDMA-based designs using OpenFabrics Verbs and heterogeneous storage architectures to accelerate multiple components of Hadoop (HDFS, MapReduce, RPC, and HBase), Spark and Memcached. An overview of the associated RDMAenabled software libraries being designed and publicly distributed as a part of the HiBD project.”
Today ISC 2016 announced that five renowned experts in computational science will participate in their new Distinguished Speaker series. Topics will include exascale computing efforts in the US, the next supercomputers in development in Japan and China, cognitive computing advancements at IBM, and quantum computing research at NASA.
“At Verne Global we combine low power costs, free cooling for 365 days of the year and optimized infrastructure to reduce the total costs of ownership for your data center. Our customers have achieved savings of more than 80% on power alone. The only place in the world where your data center can operate with 100% sustainable green power is at Verne Global’s Icelandic campus. Powered by geothermal and hydro-electric sources, your IT power costs will be stable and predictable for up to 20 years.”
In this video, Steve Hebert from Nimbix presents: The Internet of Machines. “The good folks at Nimbix have also posted full a set video presentations from their recent Developer Summit. The event brought together the best and brightest minds building the next generation of cloud computing applications. The invigorating discussions span topics from rendering and simulation to big data and machine learning, and everything in between.”
In this podcast, the Radio Free HPC team recaps the GPU Technology Conference, which wrapped up last week in San Jose.
Since Rich is traveling around in some desert somewhere, Dan and Henry go it alone and discuss the new Pascal (P1000) GPU, NVIDIA’s new server, and what happened at the concurrent OpenPOWER conference.”
In this video from the HPC User Forum in Tucson, Saul Gonzalez Martirena from NSF provides an update on the NSCI initiative. “As a coordinated research, development, and deployment strategy, NSCI will draw on the strengths of departments and agencies to move the Federal government into a position that sharpens, develops, and streamlines a wide range of new 21st century applications. It is designed to advance core technologies to solve difficult computational problems and foster increased use of the new capabilities in the public and private sectors.”