Saturday, August 27, 2011




Storage area network switches overview










Director switches
Enterprise data centers often incorporate large director-class switches that offer massive scalability for the storage area network, yet can be managed as a single platform. In addition, director switches feature fault tolerance, nondisruptive software upgrades, hot-swappable components, redundant power and cooling systems, and advanced security features. Director switches can frequently support multiple network protocols and can even be logically organized into multiple virtual switches. Heterogeneous server and storage system support is critical for a director switch. The two principle issues with director switches are cost and management: Directors are significantly more expensive to acquire than other switch classes, and the general lack of friendly wizard-driven interfaces usually means that management tasks can be more cumbersome and time consuming.
You'll typically find products from Cisco Systems and Brocade Communications Systems in the director space. For example, Cisco's MDS 9513 is one of the largest director switches available, sporting as many as 528 Fibre Channel ports in a single chassis, each port capable of 1 Gbps, 2 Gbps, 4 Gbps or 10 Gbps. The MDS 9513 includes support for VSANs and VSAN routing, along with advanced security features. Brocade's DCX Backbone director supports up to three hundred and eight four 4 Gbps or 8 Gbps ports in one chassis, and connects to other Brocade directors, such as the 48000 and Mi10K.
QLogic offers a SANbox 9000 director, which at 128 ports is smaller and less expensive than the Brocade and Cisco directors, but it lacks the mainframe connectivity and redundancy of its rivals' products.


Fabric switches
Along with large modular, redundant and chassis-based directors, organizations also use smaller fixed-port fabric switches in departments and outside the data center. The Brocade 5000 is 32-port switch, Cisco's MDS 9124 is a 24-port switch and the QLogic SANbox 5802 is a 24-port switch that supports 8 Gbps connectivity. Besides being smaller than directors, Fibre Channel switches lack features, such as automatic failover and hot swappable components that would allow them to continue to work in event of a failure. Fabric switches are also known as edge switches.

Intelligent switches

In terms of SAN switches, "intelligence" suggests the presence of advanced features, such as storage virtualization, remote mirroring, data sharing, protocol conversion, quality of service (QoS) management and strong authentication/security features. As storage networks become more sophisticated and support a growing range of storage products, adding intelligence to the switch offers compelling benefits -- most notable is the centralization of network-based services. Rather than managing network services handled by a variety of applications running through individual servers in different locations around the network, managing an intelligent switchpotentially lowers labor costs and software licensing costs (e.g., volume managers) and improves resource utilization by reducing overprovisioning.

However, intelligence has nothing to do with the size of a switch. Thus, director-class and intelligent switches are not the same thing. A director switch may not have many intelligent features, if any, and often must be configured with service modules to support intelligent functions. Conversely, intelligent features are appearing more frequently in blade switches and even in some lower end products, like edge/workgroup switches.
Brocade, Cisco and QLogic sell intelligent switches.


Blade switches
Blade switches are basically modular switches designed to fit in blade chassis produced by major manufacturers like Dell, IBM, and Hewlett-Packard. Although there are no features or capabilities unique to blade switches, the blade form factor offers a substantial advantage to blade server users. By mounting a blade switch, other blade servers in the chassis can be connected through the blade switch to the rest of the network. This is cleaner and more convenient than cabling each blade server to an external switch.
However, it's important to remember that blade switches generally offer limited scalability, and they are more expensive to deploy because blade chassis are not yet mainstream. Blade servers may also experience issues booting from the network through a blade switch.

Sunday, July 31, 2011


The SAN components interact as follows:

1. When a host wants to access a storage device on the SAN, it sends out a
block‐based access request for the storage device.

2 .SCSI commands are encapsulated into FC packets. The request is accepted by the
HBA for that host and is converted from its binary data form to the optical form
required for transmission on the fiber optic cable.

3. At the same time, the request is packaged according to the rules of the FC protocol.

4 .The HBA transmits the request to the SAN.

5 .Depending on which port is used by the HBA to connect to the fabric, one of the
SAN switches receives the request and sends it to the storage processor, which
sends it on to the storage device.
The remaining sections of this white paper provide additional information about the



SAN Components:

The components of an FC SAN can be grouped as follows and are discussed below:

1.Host Components
2.Fabric Components
3.Storage Components


Host Components

The host components of a SAN consist of the servers themselves and the components
that enable the servers to be physically connected to the SAN.
HBAs are located in the servers, along with a component that performs
digital‐to‐optical signal conversion. Each host connects to the fabric ports through
its HBAs.
HBA drivers running on the servers enable the servers’ operating systems to
communicate with the HBA.

Fabric Components

All hosts connect to the storage devices on the SAN through the SAN fabric. The
network portion of the SAN consists of the following fabric components:
SAN Switches — SAN switches can connect to servers, storage devices, and other
switches, and thus provide the connection points for the SAN fabric. The type of
SAN switch, its design features, and its port capacity all contribute to its overall
capacity, performance, and fault tolerance. The number of switches, types of
switches, and manner in which the switches are interconnected define the fabric
topology.

for smaller SANs, the standard SAN switches (called modular switches) can
typically support 16 or 24 ports (though some 32‐port modular switches are
becoming available). Sometimes modular switches are interconnected to
create a fault‐tolerant fabric.

 For larger SAN fabrics, director‐class switches provide a larger port capacity
(64 to 128 ports per switch) and built‐in fault tolerance.
  Data Routers — Data routers are intelligent bridges between SCSI devices and FC
devices in the SAN. Servers in the SAN can access SCSI disk or tape devices in the
SAN through the data routers in the fabric layer.
  Cables — SAN cables are usually special fiber optic cables that are used to connect
all of the fabric components. The type of SAN cable and the fiber optic signal
determine the maximum distances between SAN components and contribute to
the total bandwidth rating of the SAN.
! Communications Protocol — Fabric components communicate using the FC
communications protocol. FC is the storage interface protocol used for most of
today’s SANs. FC was developed as a protocol for transferring data between two
ports on a serial I/O bus cable at high speeds. FC supports point‐to‐point,
arbitrated loop, and switched fabric topologies. Switched fabric topology is the
basis for most current SANs.
Storage Components
The storage components of a SAN are the storage arrays. Storage arrays include storage
processors (SPs). The SPs are the front end of the storage array. SPs communicate with
the disk array (which includes all the disks in the storage array) and provide the
RAID/LUN functionality.
Storage Processors
SPs provide front‐side host attachments to the storage devices from the servers, either
directly or through a switch. The server HBAs must conform to the protocol supported
by the storage processor. In most cases, this is the FC protocol.



Storage Devices
  
Data is stored on disk arrays or tape devices (or both).
Disk arrays are groups of multiple disk devices and are the typical SAN disk storage
device. They can vary greatly in design, capacity, performance, and other features.
Storage arrays rarely provide hosts direct access to individual drives. Instead, the
storage array uses RAID (Redundant Array of Independent Drives) technology to
group a set of drives. RAID uses independent drives to provide capacity, performance,
and redundancy. Using specialized algorithms, several drives are grouped to provide
common pooled storage. These RAID algorithms, commonly known as RAID levels,
define the characteristics of the particular grouping.
In simple systems that provide RAID capability, a RAID group is equivalent to a single

LUN. A LUN is a single unit of storage. Depending on the host system environment, a
LUN is also known as a volume or a logical drive. From a VI Client, a LUN looks like many other storage unit available for access.
In advanced storage arrays, RAID groups can have one or more LUNs created for
access by one or more servers. The ability to create more than one LUN from a single
RAID group provides fine granularity to the storage creation process. You are not
limited to the total capacity of the entire RAID group for a single LUN.
Most storage arrays provide additional data protection and replication features such as
snapshots, internal copies, and remote mirroring.
  •   A snapshot is a point‐in‐time copy of a LUN. Snapshots are used as backup sources for the overall backup procedures defined for the storage array.

  •   Internal copies allow data movement from one LUN to another for an additional copy for testing.

  •  Remote mirroring provides constant synchronization between LUNs on one

storage array and a second, independent (usually remote) storage array for
disaster recovery.


Tape Storage Devices
  
Tape storage devices are part of the SAN backup capabilities and processes.

  • Smaller SANs might use high‐capacity tape drives. These tape drives vary in their transfer rates and storage capacities.
  •  A high‐capacity tape drive might exist as astandalone drive, or it might be part of a tape library.

  • Typically, a large SAN, or a SAN with critical backup requirements, is configured with one or more tape libraries. A tape library consolidates one or more tape drives into a single enclosure. Tapes can be inserted and removed from the tape drives in the library automatically with a robotic arm. Many tape libraries offer large storage.



SAN Ports and Port Naming

A port is the connection from a device into the
SAN. Each node in the SAN  each host, storage device, and fabric component
(router or switch)—has one or more ports that connect it to the SAN.

 Ports can be identified in a number of ways:

 1.WWPN — World Wide Port Name. A globally unique identifier for a port which
allows certain applications to access the port. The FC switches discover the WWPN of a device or host and assign a port address to the device.

2. Port_ID (or port address) — Within the SAN, each port has a unique port ID that
serves as the FC address for the port. This enables routing of data through the SAN
to that port. 
The FC switches assign the port ID when the device logs into the fabric. The port ID is valid only while the device is logged on.



Indepth information on SAN ports can be found at http://www.snia.org, the Web site
of the Storage Networking Industry Association.

Multipathing and Path Failover

An FC path describes a route:

From a specific HBA port in the host,Through the switches in the fabric, and Into a specific storage port on the storage array.

A given host might be able to access a LUN on a storage array through more than one
path. Having more than one path from a host to a LUN is called multipathing.

By default, VMware ESX Server systems use only one path from the host to a given LUN at any time.
 If the path actively being used by the VMware ESX Server system fails, the server selects another of the available paths. 
The process of detecting a failed path and switching to another is called path failover
A path fails if any of the components along the path.
HBA, cable, switch port, or storage processor. fails.

Active/Active and Active/Passive Disk Arrays

It is useful to distinguish between active/active and active/passive disk arrays:
 An active/active disk array: allows access to the LUNs simultaneously through all the storage processors that are available without significant performance degradation.
 All the paths are active at all times (unless a path fails).

 In an active/passive disk array, one SP is actively servicing a given LUN. The other SP acts as backup for the LUN and may be actively servicing other LUN I/O. 

I/O can be sent only to an active processor. If the primary storage processor fails, one of the secondary storage processors becomes active, either automatically or
through administrator intervention.

Zoning

Zoning provides access control in the SAN topology; it defines which HBAs can connect to which SPs. 
You can have multiple ports to the same SP in different zones to reduce the number of presented paths.
When a SAN is configured using zoning, the devices outside a zone are not visible to the devices inside the zone. 
In addition, SAN traffic within each zone is isolated from
the other zones.
Within a complex SAN environment, SAN switches provide zoning.
 Zoning defines and configures the necessary security and access rights for the entire SAN.
Typically, zones are created for each group of servers that access a shared group of storage devices and LUNs. You can use zoning in several ways. Here are some examples:

 1.Zoning for security and isolation . You can manage zones defined for testing
independently within the SAN so they don.t interfere with the activity going on in
the production zones. Similarly, you could set up different zones for different departments.

 Zoning for shared services . Another use of zones is to allow common server access for backups. 
SAN designs often have a backup server with tape services that require SANwide access to host servers individually for backup and recovery processes. 
These backup servers need to be able to access the servers they back up.
A SAN zone might be defined for the backup server to access a particular host to perform a backup or recovery process. The zone is then redefined for access to another host when the backup server is ready to perform backup or recovery
processes on that host.

3. Multiple storage arrays . Zones are also useful when there are multiple storage arrays. 
Through the use of separate zones, each storage array is managed separately from the others, with no concern for access conflicts between servers.

LUN Masking

LUN masking is commonly used for permission management. LUN masking is also referred to as selective storage presentation, access control, and partitioning, depending on the vendor.
LUN masking is performed at the SP or server level; it makes a LUN invisible when a target is scanned.
 The administrator configures the disk array so each server or group SAN Conceptual and Design Basics
VMware, 













Inc. 10 of servers can see only certain LUNs. Masking capabilities for each disk array are vendor specific, as are the tools for managing LUN masking.

SAN Installation Considerations

Installing a SAN requires careful attention to details and an overall plan that addresses all the hardware, software, storage, and applications issues and their interactions as all
the pieces are integrated.

Requirements

To integrate all co mponents of the SAN, you must meet the vendor.s hardware and software compatibility requirements, including the following:

 HBA (firmware version, driver version, and patch list) Switch (firmware) Storage (firmware, host personality firmware, and patch list)


zoning: done at switch level used to segment the fabric        LUNs
LUN masking:done at SP or server level;
makes a LUN invisible when a target is scanned
storage array.
SP
0 11 12
FC switch
HBA HBA

Mask LUN 11
21:00:00:E0:8B:19:AB:31 21:00:00:E0:8B:19:B2:33
50:05:01:60:10:20:AD:87
Mask LUN 12

WWN
(world-wide name)
unique, 64-bit address assigned to Fibre Channel node
SAN Conceptual and Design Basics.

SAN Setup

When you are ready to set up the SAN, complete these tasks.

To prepare the SAN

1 Assemble and cable together all hardware components and install the
corresponding software.
a Check the versions.
b Set up the HBA.
c Set up the storage array.

2 .Change any configuration settings that might be required.

3. Test the integration.
During integration testing, test all the operational processes for the SAN environment.
 These include normal production processing, failure mode testing, backup functions, and so forth.
4 .Establish a baseline of performance for each component and for the entire SAN.
Each baseline provides a measurement metric for future changes and tuning. See
ESX Server SAN Configuration Guide for additional information.
5 .Document the SAN installation and all operational procedures.

SAN Design Basics
When designing a SAN for multiple applications and servers, you must balance the performance, reliability, and capacity attributes of the SAN.
 Each application demands resources and access to storage provided by the SAN. 
The SANs switches and storage must provide timely and reliable access for all competing applications.
This section discusses some SAN design basics. It does not focus on SAN design for ESX Server hosts  .