vMotion Enhancements in vSphere 6
VMware announced several vMotion enhancements in vSphere 6, ranging from “finally” to “interesting”.
vMotion across virtual switches. Finally. The tricks you had to use previous were absolutely bizarre.
vMotion across routed networks. Finally someone learned how to spell routing. What really bothers me about this one is that vMotion across routed networks worked forever (probably relying on proxy ARP), it just wasn't supported. I was always wondering what the real reason for the lack of support was – maybe they had to implement VRF-like functionality to ensure vMotion traffic uses a different routing table than iSCSI or NFS traffic.
vMotion across vCenter servers. This one is a clear illustration of how stupid the long-distance vMotion ideas were. If you wanted to do vMotion across multiple data centers, they had to use a single vCenter, making them a single management-plane failure domain (not to mention the minor challenge of losing control of all but one data center if the DCI link fails).
Long-distance vMotion, which now tolerates 100 msec RTT. As expected, it took approximately 10 femtoseconds before a VMware EVP started promoting vMotion between East- and West Coast (details somewhere in the middle of this blog post).
Note to VMware: just because you fixed your TCP stack (which is good), long-distance vMotion makes absolutely no more sense than it did before… not that I would ever expect some people promoting it to understand the nuances of why that’s so.
Thx
Your blog posts and webinars are absolute gold.
I understand why long-distance vMotion doesn't make sense - large layer 2 extension/traffic hair-pinning/storage issues/extending failure domain, etc. The question is, how do we get around it? What is the alternative - assuming we are not going to get better designed applications in the near future? What would you recommend for geographical redundancy/DR?
Looking forward to chapter 9 of your "Data Center Design Case Studies" book!
What about disaster avoidance? I understand the objections about Layer 2 connections between datacenters, but not about long-distance vMotion and desire to provide continuity to non-redundantly designed applications.
Suppose we do a long-distance vMotion across vSwitches, and have no true L2 extension, but say an OSPFD instance on the VM announcing its routed /32, or a Proxy-Arp gateway instead of a true L2 extension.
Maybe long-distance vMotion has great uses, but we aren't sufficiently imaginative to give them fair consideration.
We started a discussion yesterday at sigs but I assume this infrastructure is broken, right? http://wahlnetwork.com/wn/wp-content/uploads/2014/08/vmotion-graphic-650x401.png
Previously you could use VXLAN to implement overlay L2 VM segment, but were supposed to have L2 connectivity between vMotion interfaces of vSphere hosts.
That was pretty restrictive and limited, especially if you are considering disaster recovery or avoidance to be a goal.
What happens if your one DVswitch gets corrupted, or a script kiddie hacks into vCenter and sends Mass force-poweroff+Delete VM commands? Now both clusters are down!
Your vCenter should probably be pretty well isolated from the Script Kiddies. If they can get into your vCenter they can probably get into your core routers and "reload" or get into your PDUs and power everything off.