When it comes to advanced scientific and computational research in Australia, the leading organization is the National Computational Infrastructure (NCI). NCI was tasked to form a national research cloud, as part of a government effort to connect eight geographically distinct Australian universities and research institutions into a single national cloud system.
NCI decided to establish a high-performance cloud, based on Mellanox 56Gb/s Ethernet solutions. NCI, home to the Southern Hemisphere’s most powerful supercomputer, is hosted by the Australian National University and supported by three government agencies: Geoscience Australia, the Bureau of Meteorology, and the Commonwealth Scientific and Industrial Research Organisation (CSIRO).
Whether executing data analytics or modeling scientific phenomena like climate change, marine biodiversity, and human genomics, researchers are empowered to share knowledge and data and to rapidly, securely, and cost-effectively deploy and access software applications without the burden of operating their own computer servers, instead drawing from the shared computational capabilities of a cloud environment.
Cloud computing relies on converged infrastructure and shared services to create economies of scale. While today’s data centers are growing to handle the world’s exponential data growth, budgets cannot grow at the same exponential pace. Therefore, the scalability of a cloud is paramount.
For a cloud to efficiently scale across a high number of nodes and provide high performance without bottlenecks, the ultimate in today’s interconnect technology and network hardware is required. A cloud such as NCI’s must offer:
Exceptionally high bandwidth to enable maximum throughput of data
When NCI searched for an interconnect provider for its high-performance cloud, it was important that their choice be a cloud vendor that has significant experience with scalable environments. Mellanox quickly emerged as the ideal partner.
Mellanox offered its new CloudX platform, and it demonstrated the performance value, scalability, and savings of using 56Gb Ethernet in a cloud environment. Mellanox switches provide higher bandwidth per unit, saving on both physical footprint and IT costs; plus, the higher bandwidth future proofs the infrastructure for even greater demand in data transfer rates.
With 224 nodes running 56GbE across 22 36-port switches (8 aggregation spines and 14 Top-of-Rack leafs), Mellanox connects 3200 cores in a non-blocking architecture.
NCI recognized that as part of the OpenStack community, Mellanox’s RDMA solution is integrated into both the Neutron (communication) and Cinder (storage) aspects of OpenStack, and is at the development forefront of plugins and support for related software and middleware development packages.
By implementing Mellanox’s end-to-end Ethernet solution, NCI has shown significant gains in performance, scalability, and manageability.
Benchmarking tests of the cloud have shown remarkable results:
Mellanox’s CloudX offered the ideal components to install a high-performance and scalable cloud environment at NCI. Mellanox provided NCI with convergence of its primary requirements: higher bandwidth, lower latency, lower costs, smaller physical footprint, fewer cables, easier management and scalability, support for more users, and future proofing.
Mellanox offered the open, transparent infrastructure that NCI sought by providing flexible infrastructure coupled with open source software and ongoing participation in the OpenStack community.
Most of all, CloudX clearly placed above the competition with its superior performance as a scalable high-performance cloud, providing high bandwidth, exceptionally low latency, and an increased number of VMs for the same compute power.
Author: Brian Klaff is a Senior Technical Communicator at Mellanox. Prior to Mellanox, Brian served as Director of Technical & Marketing Communications for ExperTeam, Ltd. He has also spent time as a Technical Communications Manager at Amdocs Israel, as a Product Marketing Manager at Versaware Technologies, and as a consultant specializing in mobile telecommunications for Mercer Management Consulting. Brian holds a BA in Economics & Near Eastern Studies from Johns Hopkins University.