“Moab Viewpoint is the next generation of Adaptive Computing’s admin portal. This enhanced Web-based graphical user interface enables easy viewing of workload— status, reporting on resource utilization and other system metrics. The Moab Viewpoint Portal plays an instrumental role in ensuring SLAs are met — a key component of Adaptive Computing’s Big Workflow vision — by allowing HPC administrators to maximize uptime and prove services were delivered and resources were allocated fairly.”
“We received an overwhelmingly positive response to the new Moab features during SC14, so we¹re very excited to make the new features generally available. In a competitive computing landscape where enterprises need to accelerate insights, Moab matters,” said Rob Clyde, CEO of Adaptive Computing. “Automating workload workflows is imperative to shorten the timeline to discovery, and this latest version of Moab represents a huge step forward in helping enterprises achieve that. We are excited to reveal our latest innovations and continue driving competitive advantage for our customers.”
Moab 8.1 systems management software includes a revamped Web-based user interface with bolstered reporting and tracking capabilities that give greater insight into the job states, workloads and nodes of a HPC system; massive performance gains and improvements in scale; and system improvements to achieve elastic computing to expand to other resources as workloads demand.
“Please join me in welcoming a new era — an era where HPC is the center of the business universe. It’s about keeping it simple and creating an ecosystem that adapts as demands dictate. Take the ease-of-use and collaboration of the cloud, couple that with the horsepower of HPC and extract the data necessary for the business to make game changing decisions. That’s the recipe for success and HPC is the cornerstone. HPC, take your rightful place in the data center.”
David Beer writes that the NFL plans to equip players on the field with radio-frequency identification (RFID) tags that will provide a flood of data for tracking and simulation. “HPC is also well suited to handle the different use cases that will arise from the different kinds of data analysis that people will want to run. For example, some people may well want to develop a simulation complete with graphics to represent what happened and to show different wrinkles on how the play might be run.”
Ralph H. Castain from Intel presented this talk at the Adaptive Computing booth at SC13. “The solution allows customers to leverage both their HPC and big data investments in a single platform, as opposed to operating them in siloed environments. The convergence between big data and HPC environments will only grow stronger as organizations demand data processing models capable of extracting the results required to make data-driven decisions.”