Recently I activated Tanzu with NSX-T in my homelab. After some hurdles in the planning phase, the configuration worked fine and also north-south routing worked flawlessly. My edge nodes established BGP peering with the physical router and advertised new routes. New segments are immediately available without further configuratiom on the router.
One feature that distinguishes my lab from a production environment is that it doesn’t run 24/7. After the work is done, the whole cluster is shut down and the system is powered off. An idle cluster makes a lot of noise and consumes unnecessary energy.
Recently I booted the lab and observed that no communication with the router or DNS server was possible from my NSX segments. A perfect case for troubleshooting.
First I checked the Geneve tunnels between the transport nodes. Here everything was fine and every transport node was able to communicate with every other transport node. The root cause was quickly located in the edge nodes. Neither a reboot of the edges nor a vMotion to another host did improve the situation.
The Edges weren’t completely offline. They were administrable using the management network. Traceroute was working via T1 and T0 service routers up to the fastpath interface fp-eth0. From there, no packets were forwarded.
The interface fp-eth0 is connected to the distributed port group “Edge-Trunk” on vSwitch VDS-NSX. A quick check in the vSphere client showed that the uplink ports of both edges were blocked. Not in the “down” state, but blocked.
At this point, I would ask a customer what he has changed. But I am very sure that I did not make any changes to the system or the configuration. Yes, they all say that 😉
SDDC-Manager is the central management tool in a vCloud Foundation (VCF) environment. You can add workload domains, import clusters to workload domains (WLD) or add Kubernetes namespaces. For every task there’s workflow in the GUI of SDDC-Manager.
Currently, as of version VCF 4.0.1, it is not possible to add a cluster with more than two uplinks and more than one vdSwitch to a WLD. If you try to do that in the GUI, you can only define one dvSwitch with two uplinks.
I will briefly describe the setup process here. First of all, the approx. 6 GB image file of the appliance must be loaded from VMware Downloads (login required). The appliance needs to be deployed into an existing cluster via the “Deploy OVF Template” wizard of vSphere-Client.
Deployment of the Platform Appliance (Collector)
There’s some naming confusion. The collector appliance is now called “platform” appliance. This makes it a bit difficult to find if you search for the collector in the download portal. 😉