Troubleshooting NSX in a VCF environment is mostly about quickly choosing the right layer and the right verification step. A good mental model is a funnel:
This keeps you from guessing and also matches how exam scenarios are written: you’re rewarded for structured narrowing, not random actions.
Most connectivity failures can be categorized by where the forwarding decision is made:
A simple “packet walk” story helps: you don’t need every internal detail—just be able to narrate the expected hop sequence and identify where the flow could be blocked or misrouted.
Scenario A: “Some VMs work, others don’t”
This often points to a partial failure: a subset of hosts, a subset of transport nodes, or a policy scope problem. A fast approach:
Scenario B: “Internal works, external doesn’t”
If east-west is fine but north-south is broken:
Scenario C: “Intermittent drops”
Intermittency is a clue. Common causes:
Scenario D: “It looks healthy, but traffic fails”
Health dashboards can show components up while intent/realization is off:
In this Parent, the exam typically expects you to:
A consistent troubleshooting funnel—scope → layer → smallest proof test → deeper validation—turns complex NSX scenarios into manageable steps. Next (in the Deep stage), you’ll turn the gaps captured in Base into advanced decision patterns, exam traps, and more detailed troubleshooting flows that reflect real multi-domain, multi-site VCF environments.
Exam scenarios often include a tempting distraction: “open the most detailed NSX screen.” The better move is to pick the tool that proves your top hypothesis with the fewest steps.
Treat tools as hypothesis filters:
A high-score “best next step” answer names: (a) the hypothesis, (b) the proof test, (c) the tool that can show it fastest.
Virtual machines on the same logical segment but different hosts cannot communicate. What is the first component that should be verified?
Verify connectivity between the TEP interfaces of the transport nodes.
Overlay networking relies on Geneve tunnels between TEP interfaces on transport nodes. When two virtual machines on different hosts communicate, their traffic is encapsulated and sent through these tunnels. If the underlay network cannot provide IP reachability between TEP addresses, the overlay tunnel cannot form and traffic will fail. Troubleshooting should begin by verifying TEP IP reachability, VLAN configuration, and routing in the underlay network. Tools such as host connectivity checks or ping tests between TEP interfaces help identify whether the issue lies in the overlay configuration or the physical network infrastructure.
Demand Score: 93
Exam Relevance Score: 96
A BGP neighbor on a Tier-0 gateway remains in the Idle state. What is the most likely cause?
The Tier-0 gateway cannot establish TCP connectivity with the BGP neighbor.
BGP sessions rely on a TCP connection (port 179) between two routing peers. If the session remains in the Idle state, the devices are unable to establish this connection. Common causes include incorrect neighbor IP addresses, routing issues preventing reachability, firewall rules blocking TCP 179, or misconfigured Autonomous System numbers. Administrators should verify IP connectivity between the Edge uplink interface and the physical router, ensure VLAN and routing configurations are correct, and confirm that the BGP parameters match on both devices.
Demand Score: 91
Exam Relevance Score: 95
External connectivity to workloads stops after one Edge node fails. What configuration should be verified?
Verify that the Edge cluster is configured with proper high availability or ECMP routing.
Edge nodes host centralized services such as routing and NAT. If high availability is not properly configured, the failure of one Edge node can interrupt North-South traffic. In Active-Standby configurations, the standby Edge should automatically take over services. In Active-Active deployments, ECMP routing distributes traffic across multiple Edge nodes. Administrators should confirm that the Tier-0 gateway is correctly associated with multiple Edge nodes and that failover mechanisms are functioning as expected. Misconfigured HA settings can prevent automatic failover and lead to service disruption.
Demand Score: 88
Exam Relevance Score: 92
Logical switch connectivity works within a host but fails across hosts. What configuration issue is most likely responsible?
The underlay network does not support the required MTU size for Geneve encapsulation.
When overlay packets traverse the physical network, Geneve encapsulation increases the packet size. If the physical network is configured with the default 1500 MTU, packets may be fragmented or dropped when they exceed this limit. This commonly results in communication failures between hosts while local communication within the same host continues to function. Troubleshooting should include verifying MTU configuration on ESXi hosts, physical switches, and network interfaces to ensure consistent support for larger frame sizes.
Demand Score: 86
Exam Relevance Score: 90
Routes learned from the physical network are not visible in the NSX routing table. What should be checked?
Verify that route redistribution and BGP configuration are correctly configured on the Tier-0 gateway.
The Tier-0 gateway learns routes from external routers using dynamic routing protocols such as BGP. If the routing table does not display expected routes, administrators should verify BGP neighbor status, route advertisement settings, and redistribution policies. Misconfigured route filters or disabled redistribution settings can prevent routes from being installed in the NSX routing table. Proper validation of routing configuration ensures that workloads can reach external networks and receive inbound traffic from external sources.
Demand Score: 87
Exam Relevance Score: 91
An overlay network works for some hosts but not others in the same cluster. What is a likely cause?
One or more hosts may not be correctly configured as transport nodes.
All ESXi hosts that participate in NSX overlay networking must be prepared as transport nodes. If a host is missing the NSX virtual switch configuration or does not have a properly configured TEP interface, it will not participate in Geneve tunnel creation. As a result, workloads running on that host may be isolated from the rest of the overlay network. Administrators should verify host preparation status, transport zone membership, and TEP configuration to ensure all hosts are correctly integrated into the NSX networking fabric.
Demand Score: 89
Exam Relevance Score: 93