Mellanox Benchmarks Ceph on 100Gb Ethernet

Print Friendly, PDF & Email

MellanoxToday Mellanox announced a set of 100Gb/s network storage performance benchmark results for Red Hat Ceph Storage and Red Hat Gluster Storage.

Mellanox is investing heavily in delivering the benefit of high speed networks to Red Hat Ceph Storage and Red Hat Gluster Storage customers,” said Kevin Deierling, vice president of marketing at Mellanox Technologies. “We see a groundswell of interest in using software-defined storage and in deploying Red Hat Ceph Storage and Red Hat Gluster Storage with 10, 40, 56 and 100Gb/s Ethernet network speeds in order to maximize performance and support flash storage and new workloads such as SQL databases and Big Data.”

Mellanox partnered with Red Hat and its partners, including SanDisk, Scalable Informatics, Supermicro, and Storage Foundry, to show the benefits of using fast, low-latency 10, 40, 56, and 100Gb Ethernet networks for Red Hat Ceph Storage clusters. These solutions allow larger and faster Red Hat Ceph Storage daemon (OSD) servers when using all-flash configurations or large numbers of hard drives in each server to perform beyond standard 10Gb networking, making Red Hat Ceph Storage deployments high performing and cost effective.

Red Hat Ceph Storage is well suited for archival and rich media and cloud infrastructure workloads, such as OpenStack. As enterprise and cloud customers evaluate and deploy Red Hat Ceph Storage on faster servers, with more flash, and for more demanding users, they are looking to faster networking solutions to increase performance.

  • Mellanox, Red Hat, Seagate, and Supermicro are also in an ongoing Red Hat Ceph Storage benchmarking project to demonstrate the performance with various combinations of flash and hard drives. The first phase proved the following benefits of using 40GbE instead of 10GbE networks:
  • Increases large read performance up to 100 percent for Red Hat Ceph Storage servers with 72 drives and up to 20 percent for Red Hat Ceph Storage servers with 36 disks
  • Reduces read and write latency up to 50 percent
  • Provides optimal price performance for write-intensive workloads on large-capacity Red Hat Ceph Storage servers

Scalable Informatics has been collaborating with Mellanox and Red Hat using 100GbE networking on an all-flash based Unison Red Hat Ceph Storage appliance. They have demonstrated 8GB/s+ reads from disk between a single appliance and client using Red Hat Ceph Storage. This combination provides an incredibly dense, high performance Red Hat Ceph Storage SSD solution with simplified networking.

In addition, Remote Direct Memory Access (RDMA) – supporting both RoCE and InfiniBand – is now available as a technology preview in the Ceph Hammer community release and has recently been enhanced in Red Hat Gluster Storage 3.1, having been first made available in Red Hat Gluster Storage since release 3.0.3 in January 2015. Mellanox testing shows RDMA accelerates Red Hat Ceph Storage large block random read throughput up to 23 percent and Red Hat Ceph Storage small block random read IOPs up to 45 percent. In addition, RDMA increases Red Hat Gluster Storage large-file read throughput up to 18 percent.

SanDisk has significantly improved Red Hat Ceph Storage performance on flash with our InfiniFlash all-flash storage system, demonstrating a fourfold to tenfold increase in Red Hat Ceph Storage IOPS,” said Gary Lyng, sr. director of solutions and product marketing, enterprise solutions at SanDisk. “By working with Mellanox and leveraging its 40GbE networking, we can help enable Red Hat Ceph Storage for new workloads and offer a high performance solution at an unprecedented cost structure. We look forward to continuing our work with Red Hat, Mellanox and the open source community to help increase Red Hat Ceph Storage performance and expanding workloads even further.”

Testing of an all-flash configuration with Mellanox 100GbE networking and our SSDs based Unison Ceph Appliance let us achieve a sustained 8+ gigabytes per second from a single server to a single client,” said Russell Nordquist, systems engineering manager at Scalable Informatics, a leading Hyper-converged appliance vendor headquartered in Plymouth, Michigan, that sells storage and analytics appliances to the financial services, pharmaceutical, and research customers. “Our appliances are designed to support incredibly high IO rates, and leveraging Mellanox’s 100GbE networking allows us to serve the data at twice the speed as before over a single link. We have customers now deploying Unison Ceph appliances using 100Gb Ethernet and flash drives to handle bigger and faster workloads beyond the traditional archive and object storage.”

Red Hat Gluster Storage is purpose-built as a scale-out file store with a straightforward architecture suitable for public, private, and hybrid cloud environments. Red Hat Gluster Storage offers mature NFS, SMB and HDFS file interfaces and is well suited for enterprise virtualization, analytics, enterprise sync and share, and rich media workloads.

To scale to the needs of enterprise IT departments and web-scale service providers, software-defined storage can be deployed with advanced networking capabilities that provide the requisite high throughput, IOPS and low latency,” said Brent Compton, director, ecosystem, Storage and Big Data, Red Hat. “Our benchmark testing shows that certain Red Hat Ceph Storage configurations are faster or more cost effective using 40Gb Ethernet.”

Sign up for our insideHPC Newsletter.