Red Hat Cluster Suite
From Wikipedia, the free encyclopedia
The Red Hat Cluster Suite includes software to create a high availability and load balancing cluster, it currently does not contain functionality for distributed computing (see below).
Contents |
[edit] Cluster types
The cluster suite itself contains two products. Both can be used on the same system although this use case is unlikely. Both products have been initiated in the community and repackaged by Red Hat. The Red Hat MRG (currently in beta) is an attempt to add high throughput computing, and, although this is clustering, it is not being sold as part of the standard Red Hat Cluster suite at this time.
[edit] High Availability Cluster
A Red Hat cluster suite when configured for high availability attempts to ensure service availability by monitoring other nodes of the cluster. All nodes of the cluster must agree on their configuration and shared services state before the cluster is considered Quorate and services are able to be started.
The primary form of communicating node status is via a network device (commonly Ethernet), although in the case of possible network failure, quorum can be decided through secondary methods such as shared storage or multicast.
Software services, filesystems and network status can be monitored and controlled by the cluster suite, services and resources can be failed over to other network nodes in case of failure.
The Cluster suite forcibly terminates a cluster node's access to services or resources to ensure the node and data is in a known state. The node is terminated by removing power or access to the shared storage.
Service locking and control is guaranteed through fencing and STONITH, more recent versions of Red Hat use a distributed lock manager (DLM), to allow fine grained locking and no single point of failure. Earlier versions of the cluster suite relied on GULM (Grand Unified Lock Manager) which could be clustered, but still presented a point of failure) if the nodes acting as gulm servers were to fail. GULM as a locking manager is available but deprecated in Red Hat Cluster Suite 5.
[edit] Technical Details
- Support for up to 128 nodes ( 16 on Red Hat Enterprise Linux 3 and 4)
- NFS (Unix) /CIFS (Windows)/GFS (Multiple Operating systems) File system failover support
- Service failover support
- Fully shared storage subsystem
- Comprehensive Data Integrity
- SCSI and Fibre Channel support
[edit] Load balancing Cluster
Red Hat adapted the Piranha load balancing software to allow for transparent load balancing and failover between servers. The application being balanced does not require special configuration to be balanced, instead a Red Hat Enterprise Linux server with the load balancer configured, intercepts and routes traffic based on metrics/rules set on the load balancer.
[edit] Support and product life cycle
Red Hat Cluster suite support is tied to a matching version of Red Hat Enterprise Linux and follows the same maintenance policy. The product has no activation, time limit or remote kill switch, it will remain working after the support life cycle has ended. It is not supported running under VMWare Virtual Machine.
[edit] History
The cluster suite is available in Red Hat Enterprise Linux 2.1, 3, 4 and 5. Supported Global File System as a filesystem in 3 and above. The load balancing software was a fork of the open source Piranha load balancing software.
[edit] See also
|
[edit] References
http://www.redhat.com/docs/manuals/csgfs/browse/rh-cs-en/s1-config-fence-devices.html
http://www.redhat.com/archives/linux-cluster/2006-June/msg00204.html