VPLEX
In computer data storage, the EMC VPLEX (VPLEX) is a virtual storage solution introduced by EMC in May 2010[1]. VPLEX implements a distributed "virtualization" layer within and across geographically disparate Fibre Channel storage area networks and Data Centers.[2][3][4].
Architecture[2]
VPLEX is deployed as cluster consisting of one or more engines. Each engine consists of two 1U redundant io directors and one 1U IO annex. Each engine has 32 Fibre Channel ports and is protected by two redundant stand-by power supplies. Each director is a bladed multi-core multi-processor x86 virtualization processing unit containing 4 hot-swappable io modules. The 1U IO annex is used for intra-cluster director communication. Each director runs a Linux kernel and a specialized Virtualization Storage Software environment called GeoSynchrony, that provides proprietary clustering capability. Each cluster has a service management station which provides all alerting and software management capabilities.
VPLEX is based on standard EMC building block hardware architecture components such as those used in its Symmetrix product line.
VPLEX uses an in-band architecture which means that data flowing between a host and a storage controller flows through one or more directors. On the front end, VPLEX presents an interface to a host which looks like a storage controller (like a target). On the VPLEX back end, it provides interface to a storage controller that looks like a host (like an initiator).
A VPLEX cluster consists of one or more pairs of directors (up to 4 pairs). Any director from any engine can failover to any other director in the cluster in the case of hardware or path failure.
Terminology[2]
- Director - a single 1U virtualization processor.
V-Plex models |
Type-model |
Cache [GB] |
FC speed [Gb/s] |
Engines |
FC Ports |
Announced |
VPLEX VS1 Single |
64 |
8 |
1 |
32 |
10 May 2010 |
VPLEX VS1 Dual |
128 |
8 |
2 |
64 |
10 May 2010 |
VPLEX VS1 Quad |
256 |
8 |
4 |
128 |
10 May 2010 |
- Cluster - a set of one or more pairs of directors, that are managed as a single entity.
- Cluster IP address - a single IP address of a cluster, that provides administrative interfaces (SSH and HTTPS).
- VPLEX Managemet Console - a management GUI for V-Plex. Installed on the System Management Server (SMS).
- Virtual Volume - a unit of storage presented to the host by VPLEX.
- Device - a logical unit constructed from one or more extents. Devices can be of type Raid-0, Raid-1, or Raid-C and be recursively constructed from other devices.
- Extent - an atomic unit of storage; an extent consists of some or all of a storage volume; a Device is formed from extents.
- Storage View - a logical container consisting of front end ports, registered host initiator ports, and virtual volumes. Storage Views determine host access to virtual volumes from VPLEX.
- VPLEX Local - a VPLEX cluster within a single data center.
- VPLEX Metro - two VPLEX clusters located within or across multiple data centers separated by up to 5ms of rtt latency.
- VPLEX Geo - two VPLEX clusters located within or across multiple data centers separated by up to 50 ms of rtt latency.
Logical Layout
Performance[5]
A VPLEX Quad achieves the following:
- IOPS
- Up to 935,000
- Throughput
- Up to 10.8 GB/s
Features
As of release 4.0.0.00.11, the base major features of VPLEX are[3]:
- Virtual Storage
- Servers access VPLEX as if it were a storage array. The SCSI LUNs they see represent virtual disks (virtual volumes) which are allocated in VPLEX from a pool of storage volumes provided by one or more back-end storage arrays. A storage volume is simply a storage LUN provided by one of the storage arrays that VPLEX is connected to.
- Data migration
- V-Plex can move data between different devices or between different extents, while maintaining I/O access to the data.
- Importing existing LUNs via a feature called Application Consistent mode.
- Application consistent mode virtual volumes are a one-to-one representations of an existing storage volumes; such volumes can be easily imported by a host after removing VPLEX from the data path. The ability to easily move from virtualized to non-virtualized disk storage is the main advantange to this approach. This approach limits the usable extent size to that of the underlying storage volume and imposes upper level limits on device layout and construction.
- Host LUN Mapping
- The set of presented virtual volumes can be configured independently for each server.
- Write-Through cache (Local and Metro)
- Writes from hosts are cached by VPLEX, but only acknowledged back to the host once they have been acknowledge by the back-end storage array. In the initial VS1 release, VPLEX cache is very beneficial in read skewed environments. Cache size is 32 GB per director.
- Write-Back cache (Geo only)
- Writes from hosts are cached by VPLEX, protected, and then acknowledged back to the host. For the VS1 hardware, cache size is 32 GB per director.
- Power and Space efficient[6]
- Virtual Volume Mirroring
- Provides the ability to make two copies of a LUN within and across heterogenous storage arrays.
- Distributed Devices
- Presentation of a logical device to hosts across geographically disparate (<100km | <5ms latency) clusters with full Read/Write host access provided by each VPLEX Cluster.
- Application layer manages (prevents) concurrent update from multiple hosts.
- AccessAnywhere ensures all hosts read the most recent updates, independent of source.
Base licensing includes up to 10 TB of attached back-end storage and then priced per TB per price tier beyond the base. There are some optional features (i.e. Metro), separately licensed.
See also
References