“Companies are awash in data, but struggle to take advantage of it to build better predictive models and gain deeper insights,” says David Rich, MATLAB marketing director, MathWorks. “With R2016b, we’ve lowered the bar to allow domain experts to work with more data, more easily. This leads to improved system design, performance, and reliability.”
Hailing from Norway, big-memory appliance maker Numascale has been a fixture at the ISC conference since the company’s formation in 2008. At ISC 2016, Numascale was noticeably absent from the show and the word on the street was that the company was retooling their NumaConnect™ technology around NVMe. To learn more, we caught up with Einar Rustad, Numascale’s CTO.
Big data applications represent a fast-growing category of high-value applications that are increasingly employed by business and technical computing users. However, they have exposed an inconvenient dichotomy in the way resources are utilized in data centers. To learn more down load this white paper.
2013 has been an exciting year for the field of Statistics and Big Data, with the release of the new R version 3.0.0. We discuss a few topics in this area, providing toy examples and supporting code for configuring and using Amazon’s EC2 Computing Cloud. There are other ways to get the job done, of course. But we found it helpful to build the infrastructure on Amazon from scratch, and hope others might find it useful, too.
This paper presents our approach to research and development in relation to four applications in which utilization of simulations in super-large-scale computation systems is expected to serve useful purposes.
This paper introduced development work being undertaken at Cambridge to create a new state of the art omics analysis hardware and software platform utilizing the open source software framework called OpenCB
This paper will look at how remote 3D virtual desktop technology can be applied in two usage modes within the scientific and technical computing arena . First, as an HPC remote visualisation platform deployed to provide real time graphical access to large scale data sets stored within the HPC data center . Second, to provide a full 3D capable virtual workstation solution as a replacement of traditional fixed under-desk workstation solutions .
When it comes to generating increasingly larger data sets and stretching the limits of high performance computing (HPC), the field of genomics and next generation sequencing (NGS) is in the forefront.
The major impetus for this data explosion began in 1990 when the U.S. kicked off the Human Genome Project, an ambitious project designed to sequence the three billion base pairs that constitute the complete set of DNA in the human body. Eleven years and $3 billion later the deed was done. This breakthrough was followed by a massive upsurge in genomics research and development that included rapid advances in sequencing using the power of HPC. Today an individual’s genome can be sequenced overnight for less than $1,000.
Analytics is becoming increasingly important for industrial companies. With good results too – a simple and brief web search on the topic will reveal case after case of $1 million plus in cost savings. On the other side of that coin is the burden traditional analytics approaches have taken on IT organizations.
This paper is the first to explore a recent breakthrough with the introduction of the High Performance Computing (HPC) industry’s first Intelligence Community Directive (ICD) 503 (DCID 6/3 PL4) certified compliant and secure scale-out parallel file system solution, Seagate ClusterStorTM Secure Data Appliance, which is designed to address government and business enterprise need for collaborative and secure information sharing within a Multi-Level Security (MLS) framework at Big Data and HPC Scale.