Install, Configure, Administrate the VMware Solution Detailed Explanation
1. Definition & mental model
This domain is the “make it real” phase: you take the storage decisions (vSAN ESA / vSAN OSA, stretched vs single-site, principal vs supplemental, external datastore choices) and perform the actual deployment and configuration steps that make a Workload Domain cluster usable—and supportable on Day 2.
A practical mental model:
- Deploy = stand up a cluster/domain that can store VMs safely (baseline networking, storage enablement, initial validation).
- Configure services = turn on storage capabilities (Storage Policy Based Management (SPBM) policies, encryption, file, iSCSI, protection, capacity sharing).
- Administer Day 2 = keep it healthy through change (maintenance, device replacement, capacity pressure, balancing, recovery planning, and routine verification).
2. Key concepts & data flows
Deployment flavors you must distinguish
- Standard vSAN cluster in a Workload Domain: the baseline HCI storage pool for the cluster.
- vSAN Stretched Cluster in a Workload Domain: one cluster across two sites (two failure domains) plus a witness role; designed for site-level resilience.
- vSAN 2-Node cluster: a small footprint cluster that typically depends on a witness component to avoid split decisions.
- Workload Domain with supported (non-vSAN) storage: vSAN is not the principal datastore; storage is delivered via NFS/iSCSI/FC/NVMe-oF and consumed as datastores.
“Who talks to whom” in day-to-day storage operations
- SPBM and vCenter Server express storage intent (policies) and evaluate compliance. VMs “consume” storage through policies more than through manual placement.
- ESXi Hosts are always in the data path; inconsistencies at host level (networking, access control, pathing) show up as “some hosts can’t see storage.”
- Skyline Health for vSAN (and standard vSphere health/performance views) is where many “is this normal?” questions are answered before you troubleshoot deeper.
Certificates / authentication / trust at a Base level (config impact)
- vSAN Encryption adds a trust dependency: the cluster must trust and reach the key provider (KMS). If the trust chain or connectivity is broken, encryption workflows fail or become unsafe to operate.
- Non-vSAN datastores have their own trust gates:
- NFS exports (who can mount)
- iSCSI CHAP and target access control (who can log in)
- FC zoning and LUN masking (who can see the LUN)
If trust/access is misaligned, symptoms often look identical to “network problems,” so you need to remember to check access control early.
Basic sizing & placement decisions (how they surface during deployment)
- Small clusters (2-node / small host counts) are more sensitive to maintenance windows and availability constraints.
- Stretched clusters are sensitive to latency and failure-domain correctness; witness placement matters for stability.
- Capacity sharing (vSAN HCI Mesh / cross-cluster capacity sharing) changes “where capacity comes from,” so you must be clear about which cluster is the storage provider and which cluster is the consumer.
3. Typical deployment and operations scenarios
Deploy a vSAN Cluster within a VCF Workload Domain
Common operational flow you should recognize:
- Bring up the Workload Domain cluster, ensure the vSAN network is consistent, then enable vSAN (ESA/OSA choice is a design decision you’re implementing).
- Validate: cluster health baseline, datastore presence, and initial policy compliance.
- Prove: a simple VM placement and storage policy assignment behaves as expected.
Deploy a vSAN Stretched Cluster within a VCF Workload Domain
What makes it different is not “more vSAN,” it’s two-site semantics:
- Define two site fault domains and the witness role.
- Validate site awareness and witness connectivity.
- Confirm expected behavior for “site impairment” vs “host impairment” (this becomes crucial on Day 2).
Deploy a vSAN 2-Node Cluster
The story here is simplicity with strict constraints:
- Two data nodes can’t “break ties” by themselves; you plan for a witness role.
- Operational focus is verifying that the witness is reachable and that maintenance won’t violate availability expectations.
Deploy and configure vSAN Data Protection (and Recovery Plans)
Think in two layers:
- Deploy vSAN Data Protection: enable the capability and confirm prerequisites are met.
- Configure vSAN Data Protection: create what users actually rely on—schedules/targets and recovery expectations.
- Create/configure a vSAN Data Protection Recovery Plan: define “what to restore, in what order, with what checks,” so recovery is an engineered workflow rather than a panic event.
Configure vSAN services: File, iSCSI, Encryption, and Capacity Sharing
- vSAN File Services: deploy the service and then create file shares that teams can consume.
- vSAN iSCSI Target Service: offer block storage endpoints from the cluster; you must understand the networking and the target/LUN presentation logic.
- vSAN Encryption: implement data-at-rest protection (with a KMS dependency).
- Configure vSAN Cross-Cluster Capacity Sharing and vSAN Storage Clusters: share capacity so a compute cluster can consume storage capacity provided by another cluster—great for reducing “stranded capacity,” but requires disciplined role clarity (provider vs consumer).
Deploy a VCF Workload Domain cluster with supported (non-vSAN) Storage
Operationally, you’re proving:
- Every ESXi Host can access the datastore(s) consistently.
- The datastore(s) behave well under multipathing and failover conditions.
- You can build higher-level constructs like Datastore Clusters (Storage DRS grouping) when appropriate.
4. Common mistakes, risks, and troubleshooting hints
- Skipping “cluster-wide consistency” checks: a config that works on one ESXi Host is not good enough; exams love “only some hosts see the datastore.”
- Turning on features without confirming prerequisites:
- Encryption without validating KMS reachability and trust
- File Services without confirming the environment supports the service and networking is ready
- iSCSI target service without clear network separation expectations
- Misunderstanding provider/consumer roles in capacity sharing: treating the consumer cluster as if it “owns” the storage is a common conceptual error.
- Day 2 blind spots:
- vSAN: ignoring resync/repair windows, not monitoring policy compliance, or entering maintenance mode without understanding the consequences.
- Stretched: ignoring site fault domains/witness health; confusing “site issue” with “host issue.”
- Non-vSAN: forgetting Storage DRS behavior in a datastore cluster, or not validating multipathing after changes.
5. Exam relevance & study checkpoints
You should be able to do (at a high level, without memorizing UI clicks):
- Describe the deployment differences between:
- a standard vSAN cluster, a vSAN Stretched Cluster, and a vSAN 2-Node cluster
- Explain what it means to:
- create/configure a vSAN Storage policy (SPBM) and apply it to workloads
- enable and configure vSAN Data Protection and articulate what a recovery plan is for
- enable vSAN File Services and then create a file share
- enable vSAN iSCSI Target Service and explain what is being presented to consumers
- configure vSAN Encryption and state the added dependency (KMS)
- configure vSAN cross-cluster capacity sharing and clearly label provider vs consumer
- For non-vSAN storage, list the first three things you verify when a datastore is invisible on some hosts: access control (export/masking/zoning/CHAP), host configuration consistency, and pathing/multipathing.
6. Summary and suggested next steps
In VCF storage, “deploy and configure” is really about repeatable, verifiable outcomes:
- vSAN deployments differ by topology (standard vs stretched vs 2-node), but all depend on consistent host/network readiness and policy-driven intent.
- Storage services (File Services, iSCSI, Encryption, Data Protection, capacity sharing) introduce dependencies—each should be enabled with an explicit verification mindset.
- Non-vSAN storage is integration-heavy: consistency across all hosts and stable pathing matter more than any single setting.
Next, we’ll focus on how to monitor, troubleshoot, and optimize both vSAN and supported (non-vSAN) storage using the right tools and a structured troubleshooting approach.
Install, Configure, Administrate the VMware Solution (Additional Content)
Deployment verification playbooks (what to prove, fast)
Context & why it matters
In exam scenarios, you usually don’t need “click paths.” You need to know what must be true immediately after deployment so you can judge whether a design is valid, and how to isolate the first failing layer when it isn’t.
Advanced explanation
A) Deploy a vSAN Cluster within a VCF Workload Domain (minimum viable proof set)
- All ESXi Hosts: consistent vSAN networking and stable cluster membership.
- vSAN datastore: visible and usable; a simple placement test works as expected.
- SPBM: a baseline storage policy can be applied and the object reaches compliance (or you can clearly explain why it cannot).
B) Deploy a vSAN Stretched Cluster within a VCF Workload Domain (what changes)
- Two-site semantics must be correct: the cluster understands site fault domains (site A vs site B).
- Witness role must be reachable and stable; site impairment behavior is predictable (tie-breaker works).
- “Site problem” vs “host problem” must be distinguishable: if you can’t explain which one you’re seeing, you’re not truly verifying stretched correctness.
C) Deploy a vSAN 2-Node Cluster (the witness dependency reality)
- Two data nodes + witness-style tie-breaker behavior must be stable.
- Maintenance tolerance is the key risk: your verification must include “what happens if one node is in maintenance or offline.”
Troubleshooting & decision patterns
- If a scenario says “deployment succeeded but datastore is missing,” treat it as a foundational enablement/visibility failure (don’t jump to advanced services).
- If a stretched/2-node scenario says “intermittent availability during link issues,” suspect witness reachability and fault domain correctness before suspecting disk failures.
Exam relevance
A frequent trap is choosing an answer that “adds a service” (File Services, iSCSI, Data Protection) when the stem indicates the baseline cluster is not yet verifiably healthy.
SPBM and policy intent: making storage policies exam-proof
Context & why it matters
“Create/configure a vSAN Storage policy” is rarely about knowing every rule. The tested skill is mapping intent to consequence and interpreting “compliance” language correctly.
Advanced explanation
A policy answer is strongest when it states:
- Intent (availability / performance / site tolerance)
- Constraint (cluster must physically support it: capacity headroom, fault domains, healthy components)
- Observable outcome (compliant vs noncompliant vs reduced availability during maintenance)
High-value interpretations:
- Noncompliant often means “the cluster cannot currently satisfy the policy,” not “someone typed the wrong setting.”
- If the stem includes maintenance or recent host failures, interpret noncompliance as potentially transient (rebuild/resync underway) unless the scenario clearly indicates insufficient resources or invalid fault domain layout.
Troubleshooting & decision patterns
When a question is “what should you do next,” and you see noncompliance:
- verify cluster capability/headroom and current repair/resync state,
- validate failure domain correctness (especially for stretched),
- only then consider policy changes.
Exam relevance
The exam favors answers that preserve intent and fix root capability issues, instead of “lowering the policy” to make warnings disappear.
vSAN services enablement: dependency-first reasoning (Encryption, File, iSCSI, Data Protection)
Context & why it matters
These features are common distractors. The exam often provides incomplete prerequisite signals; you must infer whether the environment can safely support the dependency chain.
Advanced explanation
A) Configure vSAN Encryption (KMS trust is the real prerequisite)
- Encryption adds an external trust dependency: the cluster must reliably reach and trust key services.
- Operational consequence: encryption state becomes a Day 2 consideration (key availability and change control), not a one-time checkbox.
B) Configure the vSAN File Service + configure a File Share
- File Services is “a service layer” running on top of the vSAN-backed environment.
- Your verification mindset should include: service health is stable, share can be created, and clients can consistently access it according to the intended access model.
C) Configure the vSAN iSCSI Target Service
- Treat it as “presenting block storage endpoints from a cluster-owned capacity pool.”
- Key reasoning points: network readiness, initiator access alignment, and “can initiators discover targets and see LUNs reliably.”
D) Deploy vSAN Data Protection + configure vSAN Data Protection + create/configure a vSAN Data Protection Recovery Plan
- Split your thinking into:
- Enablement (feature is available and prerequisites are satisfied)
- Operationalization (protection jobs/schedules create restore points predictably)
- Recoverability (a recovery plan is a documented, ordered workflow with verification steps)
Troubleshooting & decision patterns
- “Encryption enablement failed” → most exam-correct first checks are dependency/trust and reachability, not “reinstall vCenter Server.”
- “Share exists but access denied” → think access configuration and service readiness; avoid overfitting to vSAN object health unless the stem indicates broader datastore issues.
- “Initiator can’t discover targets” → think discovery/access/network alignment first; if only some initiators fail, think identity consistency.
- “Protection jobs don’t run / restore points missing” → think prerequisites + scheduling intent + policy alignment; then verify feature health.
Exam relevance
When multiple answers mention “restart” or “recreate,” the exam often prefers “verify prerequisites and dependency chain” unless the stem clearly indicates corruption.
Cross-cluster capacity sharing and vSAN Storage Clusters: provider/consumer discipline
Context & why it matters
Cross-cluster capacity sharing is easy to misunderstand in exam stems because both clusters “see storage,” but only one is truly providing it.
Advanced explanation
A safe mental model:
- Provider cluster: owns and serves storage capacity.
- Consumer cluster: consumes that capacity to place workloads, but does not magically become the storage owner.
Your verification checklist should prove:
- role clarity (provider vs consumer),
- connectivity and permissions needed for consumption,
- placement behavior aligns with policy expectations (“where does data live?” remains consistent with the model).
Troubleshooting & decision patterns
- “Consumer can’t see capacity” → start with role assignment, connectivity, and permissions before changing policies.
- “Policy placement confusion” → clarify whether the policy expects local vSAN behavior vs consumed capacity behavior; mismatched expectations are a common cause of wrong answers.
Exam relevance
A common trap: choosing an answer that treats capacity sharing like “just add a datastore.” The exam often expects you to enforce role clarity and verify prerequisites.
Non-vSAN datastores and datastore clusters: protocol-first checks + Storage DRS realism
Context & why it matters
External storage questions often test whether you can (1) ensure cluster-wide visibility, and (2) understand what changes when you introduce a Datastore Cluster (Storage DRS).
Advanced explanation
A) Configure a Datastore (non-vSAN) in a VCF Workload Domain Cluster (verification ladder)
- First: prove every ESXi Host sees the storage consistently.
- Then: validate protocol-specific access control alignment:
- NFS: export permissions and stable mounts
- iSCSI: discovery + sessions + LUN visibility
- FC/NVMe-oF: zoning + masking + stable paths
- Finally: validate multipathing stability and behavior under a link/path event (predictable failover matters as much as steady-state success).
B) Configure a Datastore Cluster in a VCF Workload Domain Cluster (what changes)
- A Datastore Cluster adds a placement/automation layer. Storage DRS can influence where VMs land based on space and (depending on configuration) load balancing.
- Exam-relevant consequence: “I put it on datastore X” can become “Storage DRS recommends datastore Y,” and you must know how that changes your troubleshooting and expectations.
C) Day 2 administration tasks on non-VSAN Datastores and Datastore Clusters
- Post-change validation becomes critical: after lifecycle actions, confirm visibility and pathing across all ESXi Hosts.
- For Datastore Clusters: validate that Storage DRS behavior aligns with your operational expectations (or is configured conservatively when requirements demand predictability).
Troubleshooting & decision patterns
- “Datastore visible on some hosts only” → access control alignment and host configuration drift first; array failure later.
- “Unexpected datastore choice” → check whether a Datastore Cluster / Storage DRS is influencing placement before assuming a storage outage.
Exam relevance
Many wrong answers jump straight to “array is down.” The exam often wants the first verification step you can perform safely inside vSphere/VCF: visibility, access control alignment, and pathing consistency.
Day 2 operations: the “maintenance + recovery” exam core (vSAN and stretched)
Context & why it matters
Day 2 scenarios combine routine actions (maintenance, replacements) with hidden constraints (small clusters, stretched semantics, compliance requirements).
Advanced explanation
A) Day 2 administration tasks on a vSAN Cluster (high-yield patterns)
- Treat maintenance as a policy-impact event: understand what “safe” means for your resilience intent.
- Watch resync/repair pressure: it is the most common reason performance degrades after “normal” operations.
B) Day 2 administration tasks on a vSAN Stretched Cluster (stretched-specific patterns)
- Site maintenance sequencing matters: you can accidentally create a “single site carrying everything” state if you don’t reason about site fault domains.
- Witness health checks are not optional; they are part of the operational heartbeat for stretched semantics.
Troubleshooting & decision patterns
- “Everything slow after maintenance” → suspect resync/repair pressure and reduced redundancy first.
- “Intermittent issues during site events” → suspect fault domain drift or witness reachability first.
Exam relevance
The exam often rewards answers that prioritize “keep the system safe and compliant” (verify health/compliance, avoid risky maintenance states) over “make the warning disappear quickly.”