vcf architecture planning

Architecture & Planning

Consolidated architecture on a 3-node cluster — hardware BOM, DNS design, VLAN layout, and IP planning for VCF 9.0.2.

Introduction: The “Consolidated” Architecture

For this home lab build, we are deploying a Consolidated Architecture on a 3-Node Cluster.

In a Consolidated deployment, the Management Domain functions as a hybrid environment. It runs the VCF management stack (vCenter, NSX, SDDC Manager) alongside your user workloads (VMs, Kubernetes). This eliminates the need for a separate VI Workload Domain, significantly reducing hardware requirements.

Key VCF 9.0.2 Feature: Support for 1GbE NICs on the Management Network. You no longer need 10GbE for the management interface, though 10GbE is used here for Data/Storage.

Target End State:

  • Identity Management (Active Directory Authentication & Custom Certificates).
  • Offline Lifecycle Management (Offline Depot).
  • Full BGP dynamic routing.
  • vSAN ESA with Memory Tiering.
  • NFS Storage for Bulk Data (Additional Datastores).
  • VCF Operations for Networks (Network Insight).
  • Supervisor Cluster (vSphere with Tanzu/VKS).
  • VCF Automation configured for VPCs and Tenancy.

High Level Deployment Steps

The following is the recommended end-to-end deployment path for this lab. Each step links to the relevant guide section.

StepActionGuide
1Architecture planning, BOM, DNS, VLAN design1 - Architecture & Planning
2Physical host prep, networking, kickstart2 - Infrastructure Preparation
3VCF Bring-Up, SDDC Manager deployment3 - Deployment
4Day 2: Identity, certificates, NSX, Edge nodes, Supervisor Deployment4 - Post-Deployment
5Operations stack (VCF Ops, Ops for Networks)5 - Operations
6Logs deployment6 - Logs Deployment & Use
7Automation (VCF Automation)7 - Automation
8NSX network operations8 - NSX & Network Operations
9VKS services: Contour, Harbor, Argo CD9 - VKS & Supervisor Services
10Windows Active Directory10 - Windows AD Deployment
11Salt automation states11 - Salt
warning
Step 4 must be complete before Step 9Supervisor deployment (in Step 4) must be complete and healthy before configuring VKS workloads, Contour, Harbor, or any other Supervisor Services in Step 9.

Phase 1: Architecture & Planning (The “Paperwork”)

Do not skip this. 90% of VCF failures happen because of missing DNS records or VLAN mismatches here.

Bill of Materials (BOM)

  • Software:
    • VMware Cloud Foundation 9.0.2 Installer OVA.
    • ESXi 9.0 ISO (Ensure hosts are on the exact supported build).
    • VCF Bundle Transfer Utility (OBTU): Required for downloading bundles on the internet-connected workstation.
  • Hardware (Your Lab Spec):
    • VCF Cluster (Target): 3x Physical Hosts (Group B).
      • pgesxa1.pgnet.io (10.200.1.220)
      • pgesxa2.pgnet.io (10.200.1.222)
      • pgesxa3.pgnet.io (10.200.1.224)
    • Supporting Cluster: A separate, existing cluster.
      • Role: Hosts the “Day 0” and “Day 1” infrastructure services (DNS/NTP, Router/Firewall, VCF Installer).
      • Requirement: Must have network reachability to VLAN 201 (Mgmt).
    • Depot Workstation: A separate physical or virtual machine.
      • Role: Has Internet Access to Broadcom Support.
      • Purpose: Runs the Bundle Transfer Utility to populate the repository “Offline” for speed and control.
    • NFS NAS Appliance:
      • Hostname: pgnas.pgnet.io
      • Role: Provides additional datastores for bulk storage (ISO files, backup targets, heavy content libraries) to offload the vSAN ESA tier.
      • Connection: Connected via 10GbE to VLAN 204.
    • CPU: Minimum 16 Cores per VCF host.
    • RAM: 128GB minimum per VCF host (Augmented by Memory Tiering).
    • Storage: vSAN ESA (Express Storage Architecture).
      • Requirement: vSAN ESA compliant NVMe drives for both storage pool and system boot.
    • Network: Physical Uplinks configured for Trunking.
  • Lab Specific Workarounds:
    • vSAN ESA Mock VIB: Required to bypass strict Hardware Compatibility List (HCL) checks on consumer NVMe drives. (Reference: William Lam - vSAN ESA Hardware Mock VIB)
    • AMD Ryzen Configuration: Specific kernel parameters required to allow sensitive workloads (like NSX Edge Nodes and Memory Tiering) to function correctly on consumer AMD CPUs. (Reference: William Lam - AMD Ryzen Workarounds)

VLAN & Subnet Schedule

The following schedule matches your specific lab network (UniFi/PGGB-UDM).

Traffic TypeVLAN IDCIDRPurpose
VM Management20110.200.1.0/24ESXi Mgmt, vCenter, SDDC Mgr, NSX Mgrs
vMotion20210.200.2.0/24Host-to-Host live migration
vSAN20310.200.3.0/24Storage traffic (East-West)
NFS Storage20410.200.4.0/24NAS Appliance / Bulk Data Datastores
Host TEP20510.200.5.0/24NSX Host Overlay (Geneve) - No DNS required
Edge TEP20610.200.6.0/24NSX Edge Overlay - No DNS required
VM Net20710.200.7.0/24General VM Workload Traffic
VKS Mgmt20810.200.8.0/24Tanzu / K8s Management
K8 Workload20910.200.9.0/24Tanzu / K8s Workload
RouterNet 125010.200.250.0/24BGP Uplink 1 (Peer: 10.200.250.1)
RouterNet 225110.200.251.0/24BGP Uplink 2 (Peer: 10.200.251.1)

DNS & Identity Strategy

Your Infrastructure Server (10.200.1.240) hosted on the Supporting Cluster is the single source of truth for core services.

  • DNS Server: 10.200.1.240 (Running BIND 9).
    • Role: Authoritative for the pgnet.io infrastructure zone.
    • Forwarding: Forwards authentication requests (SRV records, Kerberos) to the Active Directory domain controllers (pggb.local).
  • NTP Server: 10.200.1.240 (Running NTPD/Chrony).
    • Role: Point all VCF hosts and appliances to 10.200.1.240 to ensure time sync matches the DNS/Identity environment.
  • Authentication Domain: pggb.local (Active Directory).
    • Note: While infrastructure lives in pgnet.io, all user authentication will occur against pggb.local.

DNS Validation Table

Hostname (FQDN)VCF RoleIP AddressPTR Status
sddc.pgnet.ioSDDC Manager10.200.1.27✅ Valid
fleet.pgnet.ioFleet Manager10.200.1.10✅ Valid
vc.pgnet.ioManagement vCenter10.200.1.11✅ Valid
nsx.pgnet.ioNSX Manager VIP10.200.1.15✅ Valid
nsxm1.pgnet.ioNSX Manager Node 110.200.1.24✅ Valid
installer.pgnet.ioVCF Installer Appliance10.200.1.30✅ Valid
pgesxa1.pgnet.ioESXi Host 1 (Target)10.200.1.220✅ Valid
pgesxa2.pgnet.ioESXi Host 2 (Target)10.200.1.222✅ Valid
pgesxa3.pgnet.ioESXi Host 3 (Target)10.200.1.224✅ Valid
pgen1.pgnet.ioNSX Edge Node 110.200.1.50✅ Valid
pgen2.pgnet.ioNSX Edge Node 210.200.1.51✅ Valid
ops.pgnet.ioVCF Operations10.200.1.12✅ Valid
opscol.pgnet.ioVCF Ops Collector10.200.1.13✅ Valid
opsnet.pgnet.ioVCF Ops for Networks (Platform)10.200.1.44✅ Valid
opsnetcol.pgnet.ioVCF Ops for Networks (Collector)10.200.1.45✅ Valid
auto.pgnet.ioVCF Automation10.200.1.16✅ Valid
log.pgnet.ioVCF Logs10.200.1.19✅ Valid
pgnas.pgnet.ioNFS NAS Appliance10.200.1.110✅ Valid
t0-gateway.pgnet.ioTier-0 Gateway VIPTBDCreate Record
router-uplink-1.pgnet.ioPhysical Router Peer 110.200.250.1✅ Valid
router-uplink-2.pgnet.ioPhysical Router Peer 210.200.251.1✅ Valid
api.pgnet.ioSupervisor API10.200.208.100