VCF 9.0 GA Mental Model Part 3: Day-0 to Day-2 Ownership Across Fleets, Instances, and Domains

TL;DR

If you want clean accountability in VCF 9.0, anchor your operating model to the official hierarchy:

VCF private cloud -> VCF fleet -> VCF instance -> VCF domain -> vSphere clusters.

This post translates that hierarchy into an operating model: who owns what, where day-0/day-1/day-2 work happens, and how topology (single site vs two sites vs multi-region) changes your posture.

Scope and code levels referenced in this article (VCF 9.0 GA component set):

  • SDDC Manager: 9.0.0.0 build 24703748
  • vCenter: 9.0.0.0 build 24755230
  • ESX: 9.0.0.0 build 24755229
  • NSX: 9.0.0.0 build 24733065
  • VCF Operations: 9.0.0.0 build 24695812
  • VCF Operations fleet management: 9.0.0.0 build 24695816
  • VCF Automation: 9.0.0.0 build 24701403
  • VCF Identity Broker: 9.0.0.0 build 24695128
  • VCF Installer: 9.0.1.0 build 24962180 (used to deploy the 9.0.0.0 component set)

Architecture Diagram

Legend:

  • Fleet-level management components centralize governance and shared services across a fleet.
  • Instances retain discrete infrastructure management stacks (they do not merge into one giant vCenter/NSX).
  • Domains are your lifecycle and isolation boundary for compute + storage + network resources.

Table of Contents

  • Scenario
  • Assumptions
  • Scope and Code Levels
  • The Ownership Model You Actually Need
  • Day-0, Day-1, Day-2 Map
  • Decision Criteria: Fleet vs Instance vs Domain vs Cluster
  • Topology Posture
  • Identity Boundaries and SSO Scope
  • Failure Domain Analysis
  • Operational Runbook Snapshot
  • Anti-patterns
  • Troubleshooting workflow
  • Conclusion

Scenario

You are aligning architects, operations, and leadership on what VCF 9.0 is actually managing, and how responsibilities split across:

  • Fleet-level services (VCF Operations, VCF Operations fleet management, VCF Automation, identity)
  • Instance-level foundations (SDDC Manager and the instance management domain)
  • Domain-level lifecycle and isolation (management domain and VI workload domains)
  • Consumption teams (VMs, Kubernetes, platforms, and automation templates)

The goal is predictable ownership and predictable blast radius.

Assumptions

  • You are deploying greenfield VCF 9.0.
  • You are building at least one VCF private cloud containing one or more VCF fleets.
  • You plan to deploy VCF Operations and VCF Automation early enough that the platform team uses them as the primary operational console for the environment.
  • You will support three physical site configurations over time:
    • Single site
    • Two sites in one region
    • Multi-region

Scope and Code Levels

This article is pinned to the VCF 9.0 GA component set (versions and builds listed in TL;DR). If you are on a later 9.0.x maintenance release, terminology remains consistent, but exact UI placement and lifecycle sequencing can shift.

Version Compatibility Matrix

LayerComponentVersionBuildWhy you care operationally
DeploymentVCF Installer9.0.1.024962180The day-1 bring-up entrypoint for fleets and instances
Instance foundationSDDC Manager9.0.0.024703748Drives instance lifecycle workflows and inventory control
Domain foundationvCenter9.0.0.024755230Domain-level management boundary and API surface
Host layerESX9.0.0.024755229Your cluster capacity and patching blast radius unit
Network layerNSX9.0.0.024733065Network segmentation, security policy, and edge services
Fleet servicesVCF Operations9.0.0.024695812Central ops, visibility, grouping, and platform workflows
Fleet servicesVCF Operations fleet management9.0.0.024695816Lifecycle for fleet services and related management components
Fleet servicesVCF Automation9.0.0.024701403Self-service, governance, and policy-driven provisioning
IdentityVCF Identity Broker9.0.0.024695128Enables VCF Single Sign-On models and SSO scope decisions

The Ownership Model You Actually Need

A clean VCF program usually stabilizes when you stop assigning ownership by product name and start assigning it by boundary:

  • Fleet boundary -> governance and shared services
  • Instance boundary -> discrete infrastructure footprint and operational control plane
  • Domain boundary -> lifecycle, isolation, and workload placement
  • Cluster boundary -> scale unit and maintenance blast radius

“Who owns what” chart

Use this as a starting point for your internal RACI.

Construct or capabilityPrimary ownerSecondary ownerDay-2 responsibilities that must be explicit
VCF private cloud (org boundary)Platform teamSecurity/GRCPortfolio decisions, fleet count, policy and compliance guardrails
VCF fleetPlatform teamArchitectureFleet service lifecycle, shared governance, change windows, identity posture
Fleet-level management components (VCF Operations, VCF Operations fleet management, VCF Automation)Platform teamSRE/OperationsBackups, upgrades, integrations, tenant and RBAC guardrails
VCF instancePlatform teamRegional opsCapacity lifecycle, adding domains, instance-level networking standards
Management domainPlatform teamVI admin“Keep the platform running” discipline: patching, certificates, backups
VI workload domainVI adminPlatform teamDay-2 LCM inside guardrails, cluster operations, domain health
Domain networking (NSX segments, T0/T1 patterns, edge capacity)Platform teamNetwork/securityNetwork design standards, firewall policy model, edge scaling ceilings
VM provisioning and templatesApp/platform teamsVI adminGolden image ownership, config drift control, tagging standards
Kubernetes platform on vSphereApp/platform teamsPlatform teamNamespace policy, cluster lifecycle, RBAC, platform SLOs
VCF Automation catalogs, projects, policiesPlatform teamApp/platform teamsSelf-service guardrails, approvals, quotas, blueprint governance
FinOps reporting and showbackPlatform teamFinanceTagging accuracy, allocation rules, cost anomaly response

Design-time vs day-2 operations

This split is where most teams get surprised.

Design-time decisions (day-0) are expensive to unwind:

  • Fleet count and boundaries
  • Instance placement (site and region alignment)
  • Domain topology (number of workload domains, shared vs dedicated services)
  • Identity model and SSO scope
  • Network consumption model (and how much change control you want to enforce)

Day-2 operations should be routine, repeatable, and low-toil:

  • Adding workload domains and clusters
  • Capacity rebalancing
  • Patch and upgrade sequencing
  • RBAC lifecycle and access review
  • Drift detection and remediation

Day-0, Day-1, Day-2 Map

Use this map to stop “platform work” from leaking into “workload work”, and vice versa.

PhaseWhat you doWhere it happensWhy it matters
Day-0Decide VCF private cloud -> fleets -> instances -> domains topologyArchitecture/designThis locks your governance and blast radius posture
Day-0Choose identity model and SSO scopeArchitecture/securityIdentity boundaries are hard to change later without operational pain
Day-0Define network consumption model and tenant isolation modelPlatform + network/securityNetwork decisions dictate scale ceilings and operational toil
Day-1Deploy first fleet + first instance management domainVCF Installer + first instance management domainThe first instance becomes the anchor location for fleet services
Day-1Stand up fleet-level management componentsFleet services (hosted in first instance management domain)This is your “platform services layer” for operations and governance
Day-1Deploy initial VI workload domain(s)Instance lifecycle workflowsWorkload domains become your default lifecycle and isolation unit
Day-2Add instances (new sites or regions)Fleet services + new instance management domainExpands footprint while keeping governance centralized
Day-2Add workload domains and clustersInstance workflows + domain operationsExpands capacity and isolates workloads cleanly
Day-2Operate identity, automation, and lifecycleFleet servicesCentralizes day-2 governance across attached instances

Decision Criteria: Fleet vs Instance vs Domain vs Cluster

Most “VCF design debates” are actually “where do I want the blast radius to stop?”

Quick decision table

If you need…Add a fleetAdd an instanceAdd a domainAdd a cluster
Separate governance plane and change windowsYesNoNoNo
Regulated isolation with hard separationOften yesSometimesSometimesNo
New site or region footprintSometimesYesNoNo
More lifecycle isolation for workloadsNoNoYesSometimes
Different SLA or patch cadence for a workload groupNoNoYesSometimes
More capacity in same workload boundaryNoNoNoYes
Separate SSO boundaryYes (cleanest)SometimesNoNo
Reduced shared service blast radiusYesSometimesSometimesNo

Architecture Tradeoff Matrix

OptionGovernance isolationOperational overheadScale ceilingTypical use
One private cloud, one fleetLowestLowestMedium to high, depending on identity modelStandard enterprise starting point
One private cloud, multiple fleetsHighHigherHigher overall, but duplicated servicesRegulated zones, different change windows
Multiple private cloudsHighestHighestHighestMergers, hard org separation, distinct GRC boundaries

Topology Posture

You can support all three topologies with the same mental model. What changes is how you set your fleet and instance boundaries.

Single site

This is the simplest operating posture:

  • One VCF private cloud
  • One fleet
  • One instance
  • Management domain + one or more workload domains

Operational posture:

  • One change window for platform services is acceptable for most orgs.
  • Use workload domains to isolate “platform workloads” from “business workloads”.
  • Treat the management domain as the platform control plane. Keep it boring.

Two sites in one region

Challenge:

You want higher availability and operational continuity, but you do not want to turn every incident into a “distributed systems lesson”.

Solutions:

A) One fleet, one instance, stretched where justified

  • Best when latency is low enough and your network supports the design.
  • Operational reality: stretched designs raise complexity. Treat it as an advanced pattern, not the default.

B) One fleet, two instances (one per site)

  • Clear failure domains.
  • You can keep lifecycle and capacity operations site-aligned.
  • Fleet-level services still centralize governance.

C) Two fleets (one per site)

  • You get maximum isolation at the cost of duplicated fleet services and duplicated operations.
  • Use this when you need separate change windows or regulated separation even within a region.

Multi-region

Challenge:

Regions are real failure domains. Latency and inter-region dependency will punish “single control plane” assumptions.

Solutions:

A) One private cloud, one fleet, multiple instances (region aligned)

  • You keep centralized governance while allowing region-specific instances.
  • Plan identity carefully. SSO scope can become too large too quickly.

B) One private cloud, multiple fleets (region aligned or regulation aligned)

  • You isolate fleet services, identity posture, and change windows.
  • You pay in duplicated management footprint and duplicated operational effort.

C) Multiple private clouds

  • Use when organizational or regulatory boundaries require hard separation.
  • Expect duplicated tooling and duplicated platform practices unless you standardize aggressively.

Identity Boundaries and SSO Scope

In VCF 9.0, identity is not a footnote. It is a design-time decision that changes:

  • Admin experience
  • Audit posture
  • Blast radius during identity incidents
  • How your teams move between instances and domains

VCF Single Sign-On models you should reason about

Treat these as scope control knobs.

ModelSSO scopeAvailability postureOperational overheadWhen it fits
Fleet-wide SSOLargeLower (single identity service per fleet)LowOne fleet, tight governance, smaller instance count
Cross-instance SSOBalancedBalancedMediumLarger fleets, want to limit identity blast radius
Single-instance SSOSmall (per instance)Higher per instanceHigherRegulated isolation or region autonomy

Scale note you should plan for:

  • A common planning guideline is to size a VCF Identity Broker deployment for a limited number of instances. If you intend to exceed that, plan multiple identity brokers or multiple fleets.

Separate IdP and separate SSO boundaries (do both)

You typically implement “separate identity boundaries” two ways.

A) Separate fleets, separate IdPs

  • Cleanest separation.
  • Strongest isolation for regulated tenants.
  • Duplicates fleet services footprint.

B) One fleet, multiple identity brokers (cross-instance model)

  • Keeps a single governance plane.
  • Reduces blast radius of identity events.
  • You must be disciplined about which instances authenticate through which broker.

Design-time warning:

  • Resetting VCF Single Sign-On is a non-trivial event. Treat identity changes like a change program, not a quick admin task.

Failure Domain Analysis

This is the mental model that reduces panic during incidents.

Practical blast radius map

FailureWhat breaks firstWhat usually keeps runningYour first triage question
Fleet services outage (VCF Operations, VCF Automation)Visibility, governance workflows, self-service provisioning, central policy operationsExisting workloads in domains, core hypervisor operationsIs this governance down or is core infrastructure down?
Identity broker outage (in-scope instances)Logins and SSO flows for in-scope componentsExisting workloads and dataplane continueWhat is the SSO scope for this identity broker?
Instance management domain incidentInstance lifecycle workflows, management vCenter/NSX for that instanceWorkloads can keep running, but operations become constrainedCan you still reach workload domain vCenter/NSX?
Workload domain incidentDomain-specific provisioning and lifecycleOther domains and instancesIs isolation working the way you intended?
Cluster-level capacity failurePlacement, HA behavior, performanceOther clusters/domainsDid you design cluster boundaries around maintenance and failure?

Operational Runbook Snapshot

This is the minimum you want documented before you call the platform “ready”.

Fleet services runbook (platform team)

  • Backups and restore procedures for:
    • VCF Operations
    • VCF Operations fleet management
    • VCF Automation
    • VCF Identity Broker (if used)
  • Certificate lifecycle and rotation plan
  • Upgrade sequencing plan:
    • Fleet-level management components first
    • Core instance components next (SDDC Manager, NSX, vCenter, ESX, vSAN)
  • Health checks:
    • Fleet service availability and telemetry ingestion
    • Identity broker health and readiness
    • Automation integration health

Real-world RTO/RPO examples you can start with

These are starting targets that many teams use to set expectations. Tune them to your recovery strategy and staffing model.

  • Fleet services (ops and automation):
    • RPO: 4 to 24 hours depending on your backup cadence and whether you treat it as “governance state” vs “mission critical”
    • RTO: 2 to 8 hours depending on appliance recovery automation and runbook maturity
  • Identity services:
    • RPO: 1 to 8 hours
    • RTO: 1 to 4 hours, because identity outages create broad administrative impact
  • Instance management domain:
    • RPO: 15 minutes to 4 hours depending on backup tools and datastore replication
    • RTO: 4 to 24 hours depending on whether you can rebuild vs restore

The key is consistency: define targets per boundary and test them.

Anti-patterns

These are the patterns that inflate toil and create “mystery outages”.

  • Treating “fleet” and “instance” as synonyms.
  • Putting regulated tenants in the same fleet without a clear identity and change-window strategy.
  • Running business workloads in the management domain.
  • Sprawling workload domains with no lifecycle boundary strategy.
  • Trying to retrofit complex identity changes without a reset and rollback plan.
  • Assuming multi-region behaves like a LAN.

Troubleshooting workflow

When something breaks, the fastest teams classify the problem by boundary first.

  • Identify the boundary:
    • Fleet service issue?
    • Identity issue?
    • Instance management domain issue?
    • Workload domain issue?
    • Cluster or host issue?
  • Confirm scope:
    • One domain, one instance, or the whole fleet?
  • Validate impact:
    • Provisioning impacted?
    • Visibility impacted?
    • Existing workloads impacted?
  • Choose the right console:
    • Fleet-level visibility and operations -> start with VCF Operations
    • Instance lifecycle workflows -> validate instance health and management domain state
    • Domain operational state -> validate domain vCenter and NSX health
  • Stabilize, then remediate:
    • Restore service first
    • Then fix drift, misconfiguration, or lifecycle backlog

Conclusion

If you want VCF 9.0 to feel operable at scale, you need an ownership model that matches the platform hierarchy:

  • VCF private cloud is your organizational boundary for platform outcomes.
  • Fleet is where you place shared governance services and shared operational responsibility.
  • Instance is your discrete infrastructure footprint, often aligned to a site or region.
  • Domain is the lifecycle and isolation boundary you use to protect workloads from each other.
  • Cluster is your capacity and maintenance blast radius unit.

When those boundaries map cleanly to “who owns what”, day-2 operations becomes repeatable instead of heroic.

Sources

VMware Cloud Foundation 9.0 Documentation (Tech Docs): https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0.html
VMware Cloud Foundation 9.0 Release Notes: https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/vmware-cloud-foundation-90-release-notes.html
Design (VMware Cloud Foundation 9.0): https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/design.html
Architectural Options in VMware Cloud Foundation: https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/design/vmware-cloud-foundation-concepts.html
Fleet Management (VCF Operations): https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/overview-of-vmware-cloud-foundation-9/what-is-vmware-cloud-foundation-and-vmware-vsphere-foundation/vcf-operations-overview/fleet-management.html
VCF Single Sign-On Architecture: https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/fleet-management/what-is/sso-architecture.html
Identity Providers and Protocols Supported for VCF Single Sign-On: https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/fleet-management/what-is/protocols-suported-for–sso.html
Linking vCenter instances in VCF Operations: https://techdocs.broadcom.com/us/en/vmware-cis/vcf/vcf-9-0-and-later/9-0/fleet-management/linking-vcenter-systems-in-vmware-cloud-foundation-operations.html

Leave a Reply

Discover more from Digital Thought Disruption

Subscribe now to keep reading and get access to the full archive.

Continue reading