Ultimate ESX Virtrualization Contest Update

A Winnar is Me!

Wow I won! But in all honesty, I already have all 3 prizes, so in good conscious I could not accept the 3rd place prize of Mastering VMware 5 by Scott Lowe.

I already have Veeam Backup & Replication as well as Veeam One for monitoring. Great products that I can’t recommend highly enough.

 I have the complete Train Signal course load via there wonderful online training system. I convinced my boss that instead of just going to a class here or there, that the entire department benefits far greatly from a go at your own pace style of training, as well as, offering a humongous catalog of classes that if bought individually would have costs thousands. So for the low, low, price of $999 you get more CBT than you can shake a stick at.

Lastly, I had pre-ordered Scott’s Mastering vSphere 5 new book before it came out.  So hopefully the new winner doesn’t have it and can enjoy it.
Still, this was a great contest with really great prizes. Thanks to Vldan Congrats to all who won.



							
Posted in Virtualization, VMWare | Leave a comment

Some more benchmark results with IO Analyzer

Ok so digging a little deeper into how I had my config setup and running some more tests today. Wanted to see peaks for a single VM instead of several, even though I believe that multiple machines running multiple types of workloads is the better validation.

Made some changes to the test setup. One single VM with RDM mapped and eagerzeroedthick (per recomendations from Jinpyo Kim who is answering questions on the IO Analyzer site.

So a single VM running at MAX IO is hitting a sustained 19.2k IOPS, and from most of what I’ve seen 20k IOPS for a single VM is about peak. I’ve not seen much in the way of higher results from anyone else (though I could be wrong) that show a single FC connected VM breaching 20k IOPS for sustained periods. Now running on Fusion-IO cards or something of the like, yeah you will see much higher performance numbers, which is to be expected.

I get independent confirmation from Veeam Monitor as well as the XIV GUI.

Posted in Storage, Storage & Virtualization, VMWare | Leave a comment

Playing around with VMware IO Analyzer

About a year ago I was at an EMC event in Orange County, I think it was one of the traveling roadshow type events and they had some hands on labs for a bunch of their different product lines. The labs were setup in VMware and at one of the labs had an IO tool that would let you create workloads and run them. At the end of the lab I saw a blurb about IO  Analyzer. Now I wasn’t familiar with the product so when I got back to the office, I tried googling for it, no luck. But I did take a picture of the screen (can’t find that now) and it had a contact email for the beta program for the IO Analyzer, so I shot a note over to see if I could get a copy of the program.

About a day later I got an email from Bobbie Morrison who is a Program Manager at VMWare with an invite to the Beta program. Schwing! I’m not sure many people were aware of the program. There was not a lot of feedback in the beta group page, but I did post a few questions about the results I was getting and got responses fairly quickly. Still, I had not seen anyone around the various VMware blogs post about it, so I quietly ran tests utilizing the beta and in general it gave me some excellent feedback about my storage performance and expectations.

Prior to this, I had been curious about the performance levels I could get from Virtual Machines. There was a pretty long standing thread in the VMTN where users were posting informal performance metrics, and up until that point I had been using the custom ICF scripts from that with a standard windows VM that I had built just for testing. The results were OK, but it would have been nice to have something pre-canned that would do this. So from my standpoint, the IO Analyzer was a really awesome tool.

Once I got the files it was pretty simple to configure and deploy. The steps are pretty simple. Import the OVF template, add a disk, setup the network, login and you’re off and running. After running a few test MaxIO/MaxThroughput tests I decided I wanted to hammer our storage to see what kind of performance I could get. My first tests were very similar in results to what I posted a while back.  I did notice that the results I got from when I first ran the tool were far better than results from the same batch of systems several months later. I’m not 100% sure as to why performance was about 20% lower, but I believe it might have to do with storage vmotion of the VM’s onto a data store with a far different configuration.

Enter the now.

So today I saw a tweet from Wade Homes (@wholmes) who is a VCDX and works as a Solution Architect at VMware mentioning that the VMware IO Analyzer was public release. So, I’ve spent the morning rebuilding 5 VIOA boxes and running some tests.

The setup is as follows:

  • 5 hosts:  3 IBM X3690 X5’s with dual 8core procs, 128GB RAM, Emulex 8GB FC with 10GbE, 2 IBM 3850M2 quad 4core procs with 128GB RAM, Emulex 8GB FC and 10GbE
  • 2 IBM 24-B (brocade silk worm 300) FC Switches zoned between all ESXi Hosts and the XIV FC SAN Array
  • 5 VM IO Analyzer virtual machines, one deployed to each host all running on the same VMFS datastore, each with its own 8GB disk volume
  • 1 XIV Gen2 full frame with 8 4GB FC connections to two fabrics on the IBM 24-B
  • 1 VMFS Datastore 1MB block size no SOIC configured
So that’s the setup, nothing too fancy. With the beta, I would have to run each of the VMIO systems from an individual portal, but the new version lets you control multiple analyzers from the same single page. So lets test this sucker out.
First round, Max IO with 5 workloads one on each host.
This is from the XIV GUI which shows the IOPS for each of the 5 VM’s running with an aggregate of around 59k IOPS for all 5 workloads.
For some independent confirmation, I looked at what Veeam Monitor was showing me. (Love me some Veeam)
 
So we can see that Veeam Monitor is showing roughly the same results as what the XIV TOP Tools GUI shows.  I can also look on the XIV management GUI to see the total IOPS being generated on the array during the test period:
And the results from the IO Analyzer (which shows the ESXTOP output). One thing I’ve noticed is that when I run multiple workloads from a single pane, that the results get skewed a bit in their placement. Not a big deal, but it does make reading the grouped chart a little messy.
Next, lets try a Max Throughput test
Same setup, but now we will run all 5 workloads with the MAX Throughput configuration.
1441 MBps! Groovy.
Veeam Monitor shows the same results.
As do the results from the XIV Top Tools and the IO Analyzer
Ok finally a mixed workload of 5 different workers:
I run multiple disparate workloads on our storage array, what I like most about the IO Analyzer is the ability to run multiple workloads to anticipate how the array will react in a given situation. I don’t run Video on Demand, but if I had to, what kind of results would I get and would the performance be sufficient? That’s the real value I see in using this tool, the ability to model and gauge expectations.
So for this last run, its 5 disparate workloads.
  •  OLTP @ 8K
  • Exchange 2007
  • SQL @ 64K
  • Web Server @ 8k
  • Video on Demand @ 8k
XIV GUI results:
Veeam (this is a 4 hour view)
IO Analyzer results:
Busy morning 🙂
So this is just a pretty hasty setup and test to show what you can do. I’ve noticed a few things, if I run multiple workloads from the same console, it will skew the results columns. The data is right, its just out of whack on the results display. The original instructions forgot the step of having to add an additional disk after you create the IO Analyzer VM (if you don’t do that, it will run off the VM’s memory and skew your results. So far, not a lot changed from the beta to final release product except the ability to save your configurations, which I think is nice.
I have a few ESXi hosts that are connected to Equallogic iSCSI arrays and I will run similar tests with that setup as well in a few days. If you have any questions give leave me a comment.
Posted in Storage & Virtualization, VMWare, XIV Storage | 8 Comments

Opening a can of worms

The limitation of the Kindle Fire are being discussed on G+ by Doug Hazelman and Rick Vanover of Veeam. Good stuff.

Posted in Social Media | Leave a comment

Ultimate ESX Virtrualization Contest!

Great idea for a contest with some really great prizes from the good folks over at ESXVirtualization. I’m a big sucker for contests and prizes. Also, I’m a big fan of the Veeam products and have been using them for a few years now. Veeam 6 is coming out soon and should really up the game when it comes to backup/replication/monitoring.

Also, I just signed my group up for the Train Signal Complete Training Library, which gives you all of their content, online for an annual fee. Honestly you can’t beat $999 per user (lower with orders of 5 or more) considering how much training materials are available.

And what VMware admin worth his weight doesn’t already have (or should have) Scott Lowe’s Mastering VSphere 5. 

So there you go, one great site for VMware knowledge with 3 great prizes.

 

 

 

 

 

 

 

Posted in VMWare | Leave a comment

On array auto-tiering

A good discussion about auto-tiering popped up at ARS. That said, I responded as such:

I’m honestly not sold on the on array tiering, at least right now though I can see its usfullness for certain types of workloads.

With what I’ve seen from all the array manufacturers there is no real prediction ability of the arrays to move your cyclical hot data up prior to need. Example, month end transactions are when I would need a lot of data that was cold during the month to move to my hot tier for the batch processes that will take place when we close the month. I know this may be specific to my workflow, but by the time the data has moved into the hot space, the processing could be over, and then that data will be cold again. I kind of see that as a waste of cycles.

I’ve had several discussions about chunk sizes and the efficiency of certain sizing and the amount of processor utilization it takes to move that data around and track it. 1 MB, 1GB, 256MB or 64k. That’s a lot of data flying around the array from tier to tier and it takes a significant amount of overhead to facilitate those operations. I’m sure todays processors are able to do there’s operations far easier than they used to, but its still overhead that may or may not have real world benefit.

I think if a storage system is designed properly you can mitigate the need for the secret sauce. Every vendor is going to do it differently, and then claim that theirs is the best. Same with de-dupe. In the end, how much benefit are you really achieving, and is it reducing TCO? I’m not sure I can answer that yet.

I think a lot of my experience comes from having an array that doesn’t do tiering, but then again, doesn’t need it. RAID has been around for a long time, and honestly I think that its long time we move past it. The limitations of arrays with multiple RAID sets have forced us to simply implement schemes to mitigate its short comings, and in turn, the storage vendors increase the costs to the end users when in all actuality its still JBOD.

I think when a company comes around and breaks the mold, or introduces something outside the status quo, they will be the targets of a tremendous amount of FUD. I’ve seen it with Equallogic when they first came on the scene and offered all the goodies (thin provision, replication, etc) for a moderately low price. The same with 3PAR when they introduced thin provisioning (which everyone quickly followed suit), and XIV with their non-RAID grid architecture, and to an extent Isilon with their scale out NAS (even EMC was smart enough to see the writing on that wall), and I’d thrown Tintri in there as well with their approach to storage for VMWare environments.

The status quo fears change. Many times to their own detriment.

Posted in Storage & Virtualization | Leave a comment

Speaking of Steve Foskett

Excited to see him speak as the keynote at Storage Decisions “Storage for Virtual Environments” one day seminar in San Diego on Dec, 13th.  I like to attend these events when I can because it’s usually a good place to find a lot of different storage vendors in one area and pick their brains, get information on product roadmaps, etc.

Posted in Storage & Virtualization | Leave a comment

VAAI Primitives List

Stephen Foskett put together a nice list of all the VAAI Primitives here.

VMware’s introduced the “vStorage APIs for Array Integration” (VAAI) in vSphere 4.1, and block-heads like me went nuts. We’ve been trying to integrate storage and servers for decades, and VMware’s APIs finally allowed this to work in truly seamless fashion. But the world of VAAI is a thicket of bizarre naming and puzzling functionality. Some VAAI primitives are ignored or even hidden! Let’s take a look at the complete list.

A nice to have all in one listing of the primitives for both ESXi 4.1 and ESXi 5.0. Really the changes to the storage API’s along with Storage DRS and other storage related upgrades are the only real compelling reasons right now for me to migrate to ESXi 5. I standardized on ESXi instead of ESX early on,  well I rebuilt the entire environment once I found out that ESX was going away, and I’m glad I did it sooner than later. It will definitely make it easier on me for the upgrade process.

VMWareTips has a pretty comprehensive list of all the changes. Most of this is old news, but I’ve been playing catch up as I have purposely tried to shy away from ESXi 5 because I’m taking the VCP410 later this month and just don’t want all the confusion in my brain.

Posted in Storage & Virtualization, VMWare, vSphere 5 | Leave a comment

Finding the sweetspot with ESXi5 Licensing

As most of you know, VMware decided to change their licensing model for vSphere 5 just prior to VMWorld 2011 this year. The results were less than positive. I think the initial thread in VMTN forums was close to 133k views  by the time VMware decided to change the initial scheme and upgrade the entitlements across the board.

A little history:

If you’re not familiar with the new licensing entitlement breakdown, the revised edition is here.

I think customers were right to be alarmed at the initial low entitlements per socket. The documentation that came out with the original entitlements pointed to consolidation ratios of around 5:1 which was the case for say 2006, but today many customers are achieving ratios of 35:1 and higher depending on the workload being virtualized. In my organization I have roughly 30:1 consolidation ratios per host, and expect to reach 45:1 with additional per host memory upgrades in the near future. The increased entitlement went a long way to helping quell the uprising that resulted.

As an FYI, Rynardt Spies of VirtualVCP.com @rynardtspies has a very excellent vsphere licensing calculator that was very helpful in determining pricing.

Birth of a Virtual Environment:

For the record, I have learned a great deal in the last 2 years when it comes to virtualization and host sizing. My initial VMware hosts were dual socket dual core IBM x3650’s with 12GB of RAM running ESXi 3.5. I had migrated off of Windows Virtual Server 2005 and onto “free” ESXi 3.5 as a proof of concept exercise to prove to my manager that VMware was going to be the better virtualization play, as well as to show that I could do the work myself without having to engage outside consultants. My organization is very cost conscious and it is difficult to get budget for many of the IT based projects that we take on. Regardless of the huge ROI we could incur with virtualization, there was serious pushback from upper IT management and a great deal of skepticism and fear.

Currently I manage a small virtualization environment and we are primarily an IBM shop. I have 3 Production Data Centers, separated out by host type and location. Our first true production cluster was 2 IBM 3850M2 Quad Socket Quad Core systems with 128GB of RAM. We use 8GB FC for Data Store connectivity and 10GB Ethernet for LAN.  At the time I purchased these systems, they were the largest x86 machines in our environment, eclipsed only by our Power5 and iSeries.

As we started P2V’ing more and more systems, and as requests for new servers came in the need for additional hosts to support the ever growing environment became clear. At this time the new Nehelam chipset was being finalized and we opted to wait until the IBM X5 series of servers was released before we purchased new hosts. A second production cluster was introduced with 2 Socket 8 core x3690 X5 servers with 128GB of RAM each.  We were very pleased with the performance of the new X5 servers, and noticed that they were far more efficient than our older 4 socket xSeries servers.

As with many virtualized environments, we are memory constrained, to the point that deploying additional VM’s will have to wait until something changes. This graph above was the impetus for this post.

Now when I graph on memory utilization per socket the real issue becomes exposed.

 

As you can see, some servers perform better than others when it comes to memory utilization per socket. For me the the idea is to get as much bang for your buck as you can. The 3690 systems are currently utilizing 46GB of memory per socket, vs the 4Socket 3850’s at 13. Part of this is higher socket density, but when you look at CPU utilization, the higher socket count systems ultimately become less efficient and far more costly than 2 socket systems which outperform them.

Now given the changes to the VMware licensing, my goal is to utilize the full memory entitlement per socket on my ESXi hosts.  I want to be able to deploy enough VM’s per host to eat up the entire 192GB vRAM entitlement for a two socket host, and allow for N+1 redundancy in the cluster, without incurring a vTAX cost.  Thus, the sweetspot.

  • „  4 ESXi Hosts with 256 GB of RAM each
  • „  192 GB Active on each system takes full advantage of Enterprise + Licensing
  • „  Spare 64GB Capacity on each host provide for N+1 with 4 Hosts
  • „  4 Hosts provides for approximately 192 Standard Servers at 4GB of RAM each
  • „  Consolidation Ratio of 48:1 (best case scenario, but not likely with real world workloads)

 

Arriving at this conculsion:

I’m no excel wizard, but I’ve tried my best to determine exactly where the price point was for our organization when it came to getting the most out of our existing hosts, and planning for future host purchases.

Here I took 4 different hosts/memory configurations to see at what point do I come to the sweetspot where I can support N+1 utilizing the full memory entitlement  for each socket without a tremendous amount of surplus memory being wasted. The idea is to get zero excess memory while utilizing the full licensed amount during a host failure.

 

I had to choose 512GB of RAM per host as my upper limit because that is the max for the X5 servers without going to an external memory tray.

For my IBM systems, the prices reflect what I’ve paid in the past and what I’ve been quoted in the future, so your prices of course will vary and this is by no means 100% accurate but it does give a good indication of what I would expect to pay given our current environment.

The red line is the where N+1 comes into effect. A 3 host cluster will need 384GB of RAM per host in order to utilize the full licensed vRAM per socket and sustain a loss of a single host. A 4 host cluster needs 256GB as does a 5 Host system.  It’s not until we reach a 6 hosts with 384GB of RAM per machine that we get the biggest impact with the ability to lose half the cluster and still have enough capacity in each of the remaining systems to run all of the licensed Virtual Machines. Still that’s a lot of un-used physical memory.

For me, the sweet spot will be 4 hosts at 256GB of RAM. This equates to a perfect amount of RAM allocated to result in no excess memory. If I go to 5 hosts, I will have 64GB of spare memory not utilized by actual VM’s, but 4 hosts leaves me with 0. This for me is an exercise at trying to be efficient with memory utilization that supports the loss of a host for a short period of time, but allowing all VM’s to run. I know some shops have different requirements, but for me being able to run all of my hosts at 75% memory utilization that equates to the amount licensed is a good fit. That 25% overhead is simply buying me the +1 without the cost of buying and licensing an entire extra host.

So if I have 4 hosts with 256GB of RAM and each host is running at licensed vRAM max of 192GB then the 4 node cluster will still have enough physical memory capacity available to support the licensed vRAM entitlement for the physical servers. That is, unless I don’t understand what I’ve been lead to believe so far in regards to the licensing scheme for vSphere5.

Now I could be totally wrong on all of this and have missed some glaringly obvious flaw so, if you see something that I am missing please let me know. I have excel spreadsheets that back most of this, but I’m afraid it might not make a lot of sense to anyone but myself, but I will post them soon.

 

Posted in VMWare, vSphere 5 | 4 Comments

VMWare will be moving towards a consumption based pricing model.

VMware CEO Maritz: Get Ready For Consumption-Based Pricing

Quote:
“We are going to have to move towards more of a consumption-based model. This is where we are going,” Maritz said at the event Thursday, as reported by Computerworld UK. “We are trying to keep the licensing stable for as long as we can, but in 10 years from now, things will have changed quite radically.”

Maritz isn’t known as a spinmeister, but the frankness of his comments caught the attention of several VMware partners. Consumption-based pricing is a percolating issue in the IT industry, and one that will likely fuel future friction between vendors and their customers and partners.

Consensus opinion among the partners CRN spoke with is that the vSphere 5 vRAM controversy has caused Maritz to re-evaluate the way he sets expectations for customers.

“He’s never been that forthright and direct about this issue,” said Keith Norbie, vice president and CTO at Nexus Information Systems, a Minnetonka, Minn.-based partner. “However, anyone who thinks the industry isn’t going to this kind of model is kidding themselves.”

I think most astute watchers have seen this coming. It will be interesting to see the upcoming battle between Hyper-V3 and vSphere. Microsoft is close to giving Hyper-V away (uh no anti-trust issues there :rolleys: ) Still given improvements that Hyper-V is promising, the impact on the SMB space will be pretty large. I can’t quite think that VMWare would be willing to give up that part of the market, especially since its still lucrative.

That said, I guess we will have to wait to see what the future holds when/if this change takes place. I for one was not too happy with the initial licensing changes in vSphere5, at least until the Ent+ memory entitlement was doubled to a more respectable level. One has to wonder about the impact on many of the third party companies that are dependant on VMWare for their livelyhood and how changes to the licesning structure will affect their bottom lines as well. I could envision a seachange in licesning since so many companies are tied to that per-socket scheme.

Microsoft will remain the wildcard in all of this. If they continue with their current Hyper-V licensing scheme, it could conceivably eat into a very large chunk of the SMB market, and Hyper-V 3 is supposed to be far more mature than its predecessors. Still, I think for the Enterprise, VMWare will be the de-facto standard for the time to come.

Posted in Hyper-V, VMWare, vSphere 5 | Leave a comment