Author: Jamie Engelhard, CTO
Contributor: Michael Rocchino, IT Support Specialist
Hopefully, you are returning to or continuing this blog series after reading parts 1, 2 and 3. If so, you have already read all the background info and are anxious to jump right into the load testing results for the Hyper-converged appliances we tested. If not, you may want to consider starting at the beginning so you know how and what we tested. Without further ado, let’s get down to the data!
Just as we did in part 3 with VMware Horizon and Citrix XenDesktop, we will compare the results produced by Login VSI using identical testing workloads and session launch timing across two different products. However, in this case we are using the same VDI software solution (VMware Horizon) and evaluating the results on two different hyper-converged hardware platforms. The two products being compared are the Nutanix NX-3460 and the Dell EVO:RAIL appliance for virtual desktops.
We introduced the Nutanix NX-3460 in part 2 so now let’s compare it to the EVO:RAIL appliance from Dell. Below are the hardware specs for a single node in each appliance:
|Category||Component||Nutanix NX-3060||Dell EVO:RAIL|
|CPU||Processor||Intel E5-2680 v2||Intel E5-2620 v2|
|Cores per CPU||10||6|
|Base Clock Speed||2.8 GHz||2.1 GHz|
|Cache||25 MB||15 MB|
|Max Memory Speed||1866 MHz||1600 MHz|
|Max Memory Bandwidth||59.7 GB/s||51.2 GB/s|
|Memory||Module Size & Type||16 GB DDR3||16 GB DDR3|
|Module Speed||1866 MHz||1600 MHz|
|Networking||10 Gb Interface Qty||2||2|
|1 Gb Interface Qty||2||0|
|Disk||SSD Drives||2 x 400 GB||1 x 480 GB|
|SAS HDD Drives||0||4 (3 x 1.2 TB Data, 1 x 300 GB Boot)|
|SATA HDD Drives||4 x 1 TB||0|
|Total Hot Storage Tier||768 GB||480 GB|
|Total Cold Storage Tier||4 TB||3.6 TB|
Each appliance contains 4 nodes, below are the aggregated specs for each device:
|Category||Component||Nutanix NX-3460||Dell EVO:RAIL|
|CPU||Total Core Count||80||48|
|Memory||Total RAM||1024 GB||768 GB|
|Max RAM (with Upgrade)||2048 GB||N/A|
|Disk||Total Hot Storage Tier||3 TB||1.9 TB|
|Total Warm Storage Tier||16 TB||14.4 TB|
As you can see, from a pure hardware standpoint, the Nutanix unit has a substantial advantage in every category. So, why didn’t we test a more powerful EVO:RAIL device? The answer is that EVO:RAIL is not available in a more powerful configuration, no matter which OEM provides it. Processor, memory, disk and all other core components are governed by the EVO:RAIL specifications. At the time of the evaluation, this was the only configuration available for VDI workloads.
So, using the equipment provided by each vendor let’s see how we fared on the Login VSI density test.
We used Login VSI to determine VSImax for both platforms. In case you’re not familiar with Login VSI, refer back to part 3 of this blog series for an explanation of the test parameters and the metrics below.
Below are the parameters for the two tests. Because we expected the EVO:RAIL to be able to handle fewer sessions based on the specs, we started with a number that we thought was easily achievable and planned to run additional rounds of testing with gradually higher session counts. Launch interval and workload were kept identical.
|Setting||Nutanix Test||EVO:RAIL Test|
|Launch Time||33 Minutes||20 Minutes|
|Launch Interval||6 Seconds||6 Seconds|
|Workload||Knowledge Worker||Knowledge Worker|
Below are the results – as it turned out, we did not need to run additional rounds of testing with EVO:RAIL, because we achieved VSImax on our very first test.
|Login VSI Baseline||788 ms||974 ms||Nutanix is 24% faster|
|Login VSI Threshold||1788 ms||1975 ms||Nutanix is 10% faster|
|Login VSI VSImax||311||180||Nutanix is 73% higher|
Login VSI VSImax graphs for each solution are included below. The VSImax is indicated by the red “X”.
Below is a comparison of the Nutanix and EVO:RAIL VSI Index over time as sessions are added. This graph captures the relative performance of the solutions perfectly. You can see that Nutanix (in green) starts with a better (lower) baseline score and the response time increases in a predictable and nearly linear fashion until we get above 300 sessions. EVO:RAIL has a slower baseline response time, and while the performance curve is nearly linear for the first 150 sessions, it is at a much steeper slope than the Nutanix curve. The two performance slopes run parallel in the very beginning, but they start to skew noticeably in favor of Nutanix after 25 sessions or so. After reaching 150 sessions, the response time on EVO:RAIL takes a dramatic turn upward and hits the VSImax quickly.
Below are the key measurements of system utilization during the Login VSI tests for each platform. The first graph is an aggregate across the entire cluster, with individual hosts / nodes following that.
The KPI that is most telling and we believe explains the primary performance difference between the two solutions is the disk throughput and response time. Let’s zoom in and compare the two.
As you can see, EVO:RAIL was generating 500 to 1,000 times more disk throughput during the test, which resulted in significantly increased latency. Let’s also take a closer look at CPU.
Below are the CPU utilization graphs for both appliances throughout the test, Nutanix at left.
What is noteworthy is that the EVO:RAIL appliance peaked at just over 50% of CPU, despite the fact that the processors are significantly less powerful than those in the Nutanix. So while we started out with the belief that EVO:RAIL would be hampered by its limited computing capacity, it appears that the substantially higher disk IO and latency were the true determining factors, creating a bottleneck long before the CPU. This indicates that the VSAN architecture does not scale nearly as well as the Nutanix Distributed File System (NDFS) under heavy IOPS conditions.
We should mention that it is a good thing that the Nutanix comes with those higher performing processors because the Nutanix Controller VMs (CVMs) are directly responsible for a significant portion of the CPU activity shown above. The CPU and RAM allocated to and used by the CVM on each node could probably host another 10 desktops were it available for the desktop workload. The same software that is responsible for the superior scalability of NDFS requires significant resources to run, so this overhead needs to be considered when sizing the solution. Even with that caveat, the Nutanix density numbers and storage performance under significant load are superb.
We hope this series has been informative and helpful for anyone considering VDI software solutions or a hyper-converged platform for VDI or any other virtual workload. We learned a tremendous amount in completing the testing and analysis and we are very glad to be able to share it with you here. In the case of VMware Horizon we were pleasantly surprised to see the recent strides made in terms of bandwidth consumption and the unexpected density benefits over Citrix XenDesktop. As for Nutanix vs. EVO:RAIL, the testing reinforced our belief that Nutanix is the more flexible and higher performing SDS solution, and also revealed some specific performance bottlenecks that VMware needs to address within the VSAN architecture. VSAN 6.0 was released shortly after we completed our testing, so perhaps we will have to have a rematch soon!
Jamie Engelhard, CTO – Helient Systems LLC – Jamie is an industry leading systems architect, business analyst, and hands-on systems engineer. As one of the founding partners of Helient Systems, Jamie brings an immense range of IT knowledge and focuses on several key areas such as architecting fully redundant Virtual Desktop Infrastructure solutions, Application integration and management, Document Management Systems, Microsoft Networking, Load Testing and end-user experience optimization. Jamie has provided technical, sales, and operational leadership to several consulting firms and scores of clients over his 20 year career in IT. If you have any questions or need more information, please contact firstname.lastname@example.org.