VPLEX

In computer data storage, the EMC VPLEX (VPLEX) is a virtual storage solution introduced by EMC in May 2010[1]. VPLEX implements a distributed "virtualization" layer within and across geographically disparate Fibre Channel storage area networks and Data Centers.[2][3][4].

Contents

Architecture[2]

VPLEX is deployed as cluster consisting of one or more engines. Each engine consists of two 1U redundant io directors and one 1U IO annex. Each engine has 32 Fibre Channel ports and is protected by two redundant stand-by power supplies. Each director is a bladed multi-core multi-processor x86 virtualization processing unit containing 4 hot-swappable io modules. The 1U IO annex is used for intra-cluster director communication. Each director runs a Linux kernel and a specialized Virtualization Storage Software environment called GeoSynchrony, that provides proprietary clustering capability. Each cluster has a service management station which provides all alerting and software management capabilities.

VPLEX is based on standard EMC building block hardware architecture components such as those used in its Symmetrix product line.

VPLEX uses an in-band architecture which means that data flowing between a host and a storage controller flows through one or more directors. On the front end, VPLEX presents an interface to a host which looks like a storage controller (like a target). On the VPLEX back end, it provides interface to a storage controller that looks like a host (like an initiator).

A VPLEX cluster consists of one or more pairs of directors (up to 4 pairs). Any director from any engine can failover to any other director in the cluster in the case of hardware or path failure.

Terminology[2]

V-Plex models
Type-model Cache [GB] FC speed [Gb/s] Engines FC Ports Announced
VPLEX VS1 Single 64 8 1 32 10 May 2010
VPLEX VS1 Dual 128 8 2 64 10 May 2010
VPLEX VS1 Quad 256 8 4 128 10 May 2010

Logical Layout

Performance[5]

A VPLEX Quad achieves the following:

IOPS
Up to 935,000
Throughput
Up to 10.8 GB/s

Features

As of release 4.0.0.00.11, the base major features of VPLEX are[3]:

Virtual Storage
Servers access VPLEX as if it were a storage array. The SCSI LUNs they see represent virtual disks (virtual volumes) which are allocated in VPLEX from a pool of storage volumes provided by one or more back-end storage arrays. A storage volume is simply a storage LUN provided by one of the storage arrays that VPLEX is connected to.
Data migration
V-Plex can move data between different devices or between different extents, while maintaining I/O access to the data.
Importing existing LUNs via a feature called Application Consistent mode.
Application consistent mode virtual volumes are a one-to-one representations of an existing storage volumes; such volumes can be easily imported by a host after removing VPLEX from the data path. The ability to easily move from virtualized to non-virtualized disk storage is the main advantange to this approach. This approach limits the usable extent size to that of the underlying storage volume and imposes upper level limits on device layout and construction.
Host LUN Mapping
The set of presented virtual volumes can be configured independently for each server.
Write-Through cache (Local and Metro)
Writes from hosts are cached by VPLEX, but only acknowledged back to the host once they have been acknowledge by the back-end storage array. In the initial VS1 release, VPLEX cache is very beneficial in read skewed environments. Cache size is 32 GB per director.
Write-Back cache (Geo only)
Writes from hosts are cached by VPLEX, protected, and then acknowledged back to the host. For the VS1 hardware, cache size is 32 GB per director.
Power and Space efficient[6]
Virtual Volume Mirroring
Provides the ability to make two copies of a LUN within and across heterogenous storage arrays.
Distributed Devices
Presentation of a logical device to hosts across geographically disparate (<100km | <5ms latency) clusters with full Read/Write host access provided by each VPLEX Cluster.
Application layer manages (prevents) concurrent update from multiple hosts.
AccessAnywhere ensures all hosts read the most recent updates, independent of source.

Base licensing includes up to 10 TB of attached back-end storage and then priced per TB per price tier beyond the base. There are some optional features (i.e. Metro), separately licensed.

See also

References