vSwitch in Multi-chassis Link Aggregation (MLAG) environment
Yesterday I described how the lack of LACP support in VMware’s vSwitch and vDS can limit the load balancing options offered by the upstream switches. The situation gets totally out-of-hand when you connect an ESX server with two uplinks to two (or more) switches that are part of a Multi-chassis Link Aggregation (MLAG) cluster.
Let’s expand the small network described in the previous post a bit, adding a second ESX server and another switch. Both ESX servers are connected to both switches (resulting in a fully redundant design) and the switches have been configured as a MLAG cluster. Link aggregation is not used between the physical switches and ESX servers due to lack of LACP support in ESX.
VMware vSwitch does not support LACP
This is very old news to any seasoned system or network administrator dealing with VMware/vSphere: the vSwitch and vNetwork Distributed Switch (vDS) do not support Link Aggregation Control Protocol (LACP). Multiple uplinks from the same physical server cannot be bundled into a Link Aggregation Group (LAG, also known as port channel) unless you configure static port channel on the adjacent switch’s ports.
When you use the default (per-VM) load balancing mechanism offered by vSwitch, the drawbacks caused by lack of LACP support are usually negligible, so most engineers are not even aware of what’s (not) going on behind the scenes.
Intelligent Redundant Framework (IRF) – Stacking as Usual
When I was listening to the Intelligent Redundant Framework (IRF) presentation from HP during the Tech Field Day 2010 and read the HP/H3C IRF 2.0 whitepaper afterwards, IRF looked like a technology sent straight from Data Center heavens: you could build a single unified fabric with optimal L2 and L3 forwarding that spans the whole data center (I was somewhat skeptical about their multi-DC vision) and behaves like a single managed entity.
No wonder I started drawing the following highly optimistic diagram when creating materials for the Data Center 3.0 webinar, which includes information on Multi-Chassis Link Aggregation (MLAG) technologies from numerous vendors.
Interesting links (2011-01-23)
Interesting links keep appearing in my Inbox. Thank you, my Twitter friends and all the great bloggers out there! Here’s this week’s collection:
Cores and more Cores… We don’t need them! More is not always better.
Emulating WANs with WANem. I wanted to blog about WANem for at least three years, now I don’t have to; like always, Stretch did an excellent job.
New Year's Resolutions for Geeks Like Me. The ones I like most: Deploy a pure IPv6 subnet somewhere; Put something in public cloud. And there’s “find a mentor” ... more about that in a few months ;)
VPN Network Design: Selecting the Technology
After all the DMVPN-related posts I’ve published in the last days, we’re ready for the OSPF-over-DMVPN design challenge, but let’s step back a few more steps and start from where every design project should start: deriving the technical requirements and the WAN network design from the business needs.
Do I Need a VPN?
Whenever considering this question, you’re faced with a buy-or-build dilemma. You could buy MPLS/VPN (or VPLS) service from a Service Provider or get your sites hooked up to the Internet and build a VPN across it. In most cases, the decision is cost-driven, but don’t forget to consider the hidden costs: increased configuration and troubleshooting complexity, lack of QoS over the Internet and increased exposure of Internet-connected routers.
Configuring OSPF in a Phase 2 DMVPN network
Cliffs Notes version (details in the DMVPN webinar):
- Configure ip nhrp multicast-map for the hub on the spoke routers. Otherwise, the spokes will not send OSPF hellos to the hub.
- Use dynamic NHRP multicast maps on the hub router, or the spokes will not receive its OSPF hellos.
- Use broadcast network type on all routers.
DMVPN Phase 2 Fundamentals
Phase 2 DMVPN in a nutshell:
- Multipoint GRE tunnels on all routers.
- NHRP is used for dynamic spoke registrations (like with Phase 1 DMVPN), but also for on-demand resolution of spoke transport addresses.
- Traffic between the spokes initially flows through the hub router until NHRP resolves the remote spoke transport IP address and IKE establishes the IPSec session with it.
- The IP next-hop address for any prefix reachable over DMVPN must be the egress router (hub or spoke). From the routing perspective, Phase 2 DMVPN subnet should behave like a LAN.
- Multicast packets (including routing protocol hello packets and routing updates) are exchanged only between the hub and the spoke routers.
- Routing adjacencies are established only between the hub and the spoke routers unless you use statically configured neighbors.
For more details watch the DMVPN webinar webinar.
OSPF Configuration in Phase 1 DMVPN Network
This is how you configure OSPF in a Phase 1 DMVPN network (read the introductory post and Phase 1 DMVPN fundamentals first):
Remember:
- Use point-to-multipoint network type on the hub router to ensure the hub router is always the IP next hop for the DMVPN routes.
- Use point-to-multipoint network type on the spoke routers to ensure the OSPF timers match with the hub router.
- The DMVPN part of your network should be a separate OSPF area; if at all possible, make it a stub or NSSA area.
- If absolutely needed, use OSPF LSA flood filter on the hub router and a static default route on the spokes.
For more information, watch the DMVPN Technology and Configuration webinar.
DMVPN Phase 1 Fundamentals
Phase 1 DMVPN in a nutshell:
- Point-to-point GRE tunnel on spoke routers
- Multipoint GRE tunnel on the hub router.
- All the DMVPN traffic (including the traffic between the spokes) flows through the hub router.
- On the spoke routers, the hub router must be the IP next-hop for all destinations reachable through the DMVPN subnet (including other spokes).
- Multicast packets (including routing protocol hello packets and routing updates) are exchanged only between the hub and the spoke routers.
Sometimes You Need to Step Back and Change Your Design
A few days ago I received the following e-mail from one of my readers:
I am trying presently to put in place a DMVPN solution running OSPF. I was wondering if you ever saw a solution with dual hub dual cloud design with OSPF working in practice because since I started I have issue with asymmetric routing because of the OSPF functionality.
Actually, I did… and exactly the same setup is included in the tested router configurations you get with the DMVPN: from Basics to Scalable Networks webinar. While there are many things that can go wrong with DMVPN, I’ve never heard about asymmetric routing problems, so I started to investigate what’s actually going on.