VMware vSAN — Hyper-Converged Storage Architecture

VSAN

How VMware vSAN pools local NVMe, SSD, and HDD storage across ESXi hosts into a shared distributed datastore — covering the Original Storage Architecture versus the Express Storage Architecture introduced in vSAN 8, fault domains, storage policies, and stretched cluster design.

vmwarevsanhyper-convergedstoragevcpdcv

Overview

Traditional virtualised infrastructure separates compute and storage into distinct tiers. ESXi hosts provide CPU and memory; external SANs or NAS arrays provide the shared storage those hosts require. This model works but introduces cost, complexity, and a dedicated storage network as a potential bottleneck and single point of failure.

vSAN dissolves that boundary. It is a software-defined storage layer built directly into the ESXi kernel that aggregates local disks across all hosts in a cluster into a single shared datastore. From the perspective of virtual machines and vCenter, that datastore looks like any other — VMs can be created on it, vMotion works to and from it, and storage policies govern how their data is protected. The difference is that the storage physically lives inside the same servers running the workloads, eliminating the external storage array entirely.

vSAN is managed entirely through vCenter Server and requires either a standalone vSAN licence or the vSphere Enterprise Plus edition. A minimum of three hosts is required for a standard vSAN cluster, and each host must contribute at least one disk group’s worth of storage to the pool.

Original Storage Architecture (OSA) vs Express Storage Architecture (ESA)

vSAN 8 introduced a fundamentally different internal storage architecture alongside the existing one, making the choice of architecture a key design decision:

CharacteristicOSA (Original Storage Architecture)ESA (Express Storage Architecture)
IntroducedvSAN 1.0vSAN 8.0
Disk organisationDisk groups: one cache device + 1–7 capacity devicesNo disk groups; all devices contribute to a single pool
Cache tierDedicated SSD or NVMe device per disk groupNo separate cache tier; intelligent caching across all devices
Supported mediaAll-flash (NVMe/SSD) or hybrid (SSD cache + HDD capacity)NVMe only
Max disk groups per host5N/A
RAID-5/6 efficiencySupported but higher overheadMore efficient erasure coding
Write pathWrites land in cache tier first, then destaged to capacitySingle-tier write path, lower write amplification

OSA is the proven architecture that supports a wider range of hardware, including hybrid configurations where spinning HDD disks serve as capacity tier with a flash cache. ESA is the forward-looking architecture for environments where all storage is NVMe flash — it removes the disk group abstraction and delivers higher throughput and lower latency because the entire storage tier operates at flash speeds with no cache destaging overhead.

vSAN Requirements

Before enabling vSAN on a cluster, several prerequisites must be met:

Storage Policies — SPBM

vSAN does not protect data by default through a fixed RAID level. Instead, each virtual machine’s storage requirements are defined through a storage policy assigned at the VM or virtual disk level. This is the Storage Policy Based Management (SPBM) framework.

The most important policy settings are:

Policy SettingDescription
Failures To Tolerate (FTT)How many simultaneous host, disk, or network failures the object can survive (0, 1, 2, or 3)
Failure Tolerance MethodRAID-1 (mirroring) or RAID-5/6 (erasure coding)
Number of disk stripes per objectHow many capacity devices an object is striped across (default 1)
Object space reservationPercentage of object size reserved as thick-provisioned (0% = thin)
IOPS limit for objectMaximum IOPS per object; 0 means unlimited
Flash read cache reservationPercentage of logical size reserved in flash read cache (hybrid OSA only)

The interaction between FTT and Failure Tolerance Method determines both the protection level and the minimum cluster size required:

FTTMethodMinimum hostsSpace overhead
1RAID-132× (100% overhead)
1RAID-541.33× (33% overhead)
2RAID-153× (200% overhead)
2RAID-661.5× (50% overhead)

RAID-5 and RAID-6 erasure coding are more space-efficient than RAID-1 mirroring but require more hosts and are more CPU-intensive. For ESA clusters, erasure coding performs significantly better than on OSA due to the single-tier architecture.

Fault Domains

A fault domain is a named group of hosts that share a common physical failure boundary — typically a rack or chassis. Without fault domains, vSAN distributes data components across hosts but has no awareness of which hosts share the same power circuit, top-of-rack switch, or physical rack. A rack-level failure could affect multiple hosts and violate the FTT guarantee.

With fault domains configured, vSAN treats the entire fault domain as a single failure unit and ensures that no two components of the same data object land on hosts within the same fault domain. A minimum of three fault domains is required to provide FTT=1 protection with rack-awareness. Fault domains are created in vCenter under the vSAN cluster configuration and hosts are assigned to them — vSAN then enforces placement policy automatically.

Stretched Cluster

A vSAN stretched cluster spans two geographically separated sites and adds a witness host at a third location. This design provides site-level resilience: if an entire site goes offline, the VMs on that site restart on the surviving site.

Key design parameters:

A stretched cluster effectively implements FTT=1 across sites. Each data site constitutes one fault domain, and the witness provides the third domain for quorum purposes.

vSAN File Services

vSAN File Services extends the vSAN datastore to provide NFS v3 and NFS v4.1 shares. Applications and VMs that require shared file-level access — rather than block storage — can consume these shares directly. File Services runs as a set of file service agent VMs deployed automatically on the cluster and requires a vSAN Enterprise licence.

Summary

vSAN turns the local storage in every ESXi host into a participant in a cluster-wide shared pool, removing the need for external SAN or NAS infrastructure. OSA accommodates a broad hardware range including hybrid spinning-disk configurations; ESA raises the performance ceiling by requiring all-NVMe hardware and eliminating the disk group abstraction. Storage policies give per-VM control over protection level and space efficiency. Fault domains and stretched clusters extend that protection to rack-level and site-level failure scenarios respectively.