Monday, October 22, 2012

Information storage and Management - EMCISA v2

Module 1 – Introduction to information storage

Data -> Information -> Knowledge -> Wisdom

Data is a collection of raw facts from which conclusions may be drawn
Data can be classified as Structured and Unstructured.
Majority of the data being created is unstructured that is 80% to 90% and in future it will increase further.

Structured data is typically stored using a database management system (DBMS).

Data Center : Core elements of data center are
Application
DBMS
Host or Compute
Network
Storage

All these components work together in Data Centre.

Storage Centric Architecture – Storage is managed centrally and independent of servers

Module 2 – Data Centre Environment

hypervisor software VMware ESXi.
The hypervisor abstracts CPU, memory, and storage resources to run multiple virtual machines concurrently on the same physical server.
VMware ESXi is a hypervisor that installs on x86 hardware to enable server virtualization.s
All the files that make up a VM are typically stored in a single directory on a cluster file system called Virtual Machine File System (VMFS).
The physical machine that houses ESXi is called ESXi host.

ESXi has two key components: VMkernel and Virtual Machine Monitor.
VMkernelprovides functionality similar to that found in other operating systems, such as process creation, file system management, and process scheduling
The virtual machine monitor is responsible for executing commands on the CPUs and performing Binary Translation (BT).

Module 3 – Data Protection - RAID
RAID techniques –striping, mirroring, and parity –form the basis for defining various RAID levels.

Striping is a technique of spreading data across multiple drives (more than one) in order to use the drives in parallel. Striped RAID does not provide any data protection unless parity or mirroring is used

Mirroring is a technique whereby the same data is stored on two different disk drives, yielding two copies of the data.
Mirroring involves duplication of data—the amount of storage capacity needed is twice the amount of data being stored. Therefore, mirroring is considered expensive and is preferred for mission-critical applications that cannot afford the risk of any data loss. Mirroring improves read performance because read requests can be serviced by both disks. However, write performance is slightly lower than that in a single disk because each write request manifests as two writes on the disk drives. Mirroring does not deliver the same levels of write performance as a striped RAID.

Parity is a method to protect striped data from disk drive failure without the cost of mirroring. An additional disk drive is added to hold parity, a mathematical construct that allows re-creation of the missing data. Parity is a redundancy technique that ensures protection of data without maintaining a full set of duplicate data. Parity calculation is a bitwise XOR operation.

However, there are some disadvantages of using parity. Parity information is generated from data on the data disk. Therefore, parity is recalculated every time there is a change in data. This recalculation is time-consuming and affects the performance of the RAID array.

RAID 0 configuration uses data striping techniques, where data is striped across all the disks within a RAID set. RAID 0 is a good option for applications that need high I/O throughput.
RAID 0 does not provide data protection and availability.

RAID 1 is based on the mirroring technique. In this RAID configuration, data is mirrored to provide fault tolerance. A RAID 1 set consists of two disk drives and every write is written to both disks.
RAID 1 is suitable for applications that require high availability and cost is no constraint.

RAID 1+0 combines the performance benefits of RAID 0 with the redundancy benefits of RAID 1. It uses mirroring and striping techniques and combine their benefits. This RAID type requires an even number of disks, the minimum being four. RAID 1+0 is also known as RAID 10 (Ten) or RAID 1/0. RAID 1+0 is also called striped mirror. The basic element of RAID 1+0 is a mirrored pair, which means that data is first mirrored and then both copies of the data are striped across multiple disk drive pairs in a RAID set.

RAID 3 stripes data for performance and uses parity for fault tolerance. Parity information is stored on a dedicated drive so that the data can be reconstructed if a drive fails in a RAID set. RAID 3alwaysreads and writes complete stripes of data across all disks because the drives operate in parallel. There are no partial writes that update one out of many strips in a stripe.

RAID 4, Similar to RAID 3 stripes data for high performance and uses parity for improved fault tolerance. Data is striped across all disks except the parity disk in the array. Parity information is stored on a dedicated disk. Unlike RAID 3, data disks in RAID 4 can be accessed independently so that specific data elements can be read or written on a single disk without reading or writing an entire stripe. RAID 4 provides good read throughput and reasonable write throughput. RAID 4 is rarely used.

RAID 5 is a versatile RAID implementation. It is similar to RAID 4 because it uses striping. The drives (strips) are also independently accessible. The difference between RAID 4 and RAID 5 is the parity location. In RAID 4, parity is written to a dedicated drive, creating a write bottleneck for the parity disk. In RAID 5, parity is distributed across all disks to overcome the write bottleneck of a dedicated parity disk.

RAID 6 works the same way as RAID 5, except that RAID 6 includes a second parity element to enable survival if two disk failures occur in a RAID set. Therefore, a RAID 6 implementation requires at least four disks. RAID 6 distributes the parity across all the disks. The write penalty (explained later in this module) in RAID 6 is more than that in RAID 5; therefore, RAID 5 writes perform better than RAID 6. The rebuild operation in RAID 6 may take longer than that in RAID 5 due to the presence of two parity sets.

RAID impact on performance: In both mirrored and parity RAID configurations, every write operation translates into more I/O overhead for the disks, which is referred to as a write penalty. It is 4 (2 Read + 2 Write) for RAID5 and 6 (3 Read + 3 Write) for RAID 6

A hot spare refers to a spare drive in a RAID array that temporarily replaces a failed disk drive by taking the identity of the failed disk drive. A hot spare can be configured as automatic or user initiated, which specifies how it will be used in the event of disk failure.

MODULE 4

Intelligent Storage Systems: Intelligent storage systems are feature-rich RAID arrays that provide highly optimized I/O processing capabilities. These storage systems are configured with a large  amount of memory (called cache) and multiple I/O paths and use sophisticated algorithms to meet the requirements of performance-sensitive applications.
Support for flash drives and other modern-day technologies, such as virtual storage provisioning and automated storage tiering,
An intelligent storage system consists of four key components: front end, cache, back end, and physical disks.
In modern intelligent storage systems, front end, cache, and back end are typically integrated on a single board ( referred as a storage processor or storage controller).

Each front-end controller has processing logic that executes the appropriate transport protocol, such as Fibre Channel, iSCSI, FICON, or FCoE for storage connections. Front-end controllersroute data to and from cache via the internal data bus. When the cache receives the write data, the controller sends an acknowledgment message back to the host.

Cache is semiconductor memory where data is placed temporarily to reduce the time required to service I/O requests from the host.

Cache can be implemented as either dedicated cache or global cache. With dedicated cache, separate sets of memory locations are reserved for reads and writes. In global cache, both reads and writes can use any of the available memory addresses.
Cache management is more efficient in a global cache implementation because only one global set of addresses has to be managed.
Cache Management Algorithms : Least Recently Used (LRU): discards the data that has not been accessed for long time
Most Recently Used (MRU): This algorithm is the opposite of LRU, where the pages that have been accessed most recently are freed up or marked for reuse.

Cache Management- Watermarking : As cache fills, the storage system must take action to flush dirty pages (data written into the cache but not yet written to the disk) to manage space availability. Flushing is the process that commits data from cache to the disk. On the basis of the I/O access rate and pattern, high and low levels called watermarks are set in cache to manage the flushing process.
Low watermark (LWM) is the point at which the storage system stops flushing data to the disks.

Idle flushing: Occurs continuously, at a modest rate, when the cache utilization level is between the high and low watermark.
High watermark flushing: Activated when cache utilization hits the high watermark.
Forced flushing: Occurs in the event of a large I/O burst when cache reaches 100 percent of its capacity, which significantly affects the I/O response time.

Cache Data Protection: Cache is volatile memory, so a power failure or any kind of cache failure will cause loss of the data that is not yet committed to the disk. This risk of losing uncommitted data held in cache can be mitigated using cache mirroring and cache vaulting:

Cache mirroring: Each write to cache is held in two different memory locations on two independent memory cards. In cache mirroring approaches, the problem of maintaining cache coherency is introduced. Cache coherency means that data in two different cache locations must be identical at all times.

Cache vaulting: use a set of physical disks to dump the contents of cache during power failure. This is called cache vaulting and the disks are called vault drives. When power is restored, data from these disks is written back to write cache and then written to the intended disks.

Server Flash Caching Technology: Server flash-caching technology uses intelligent caching software and PCI Express (PCIe) flash card on the host. This dramatically improves application performance by reducing latency and accelerates throughput. Server flash-caching technology works in both physical and virtual environments and provides performance acceleration for read-intensive workloads. This technology uses minimal CPU and memory resources from the server by offloading flash management onto the PCIe card.

Storage Provisioning : provisioning is the process of assigning storage resources to hosts based on capacity, availability, and performance requirements of applications running on the hosts. Storage provisioning can be performed in two ways: traditional and virtual.

It is highly recommend that the RAID set be created from drives of the same type, speed, and capacity to ensure maximum usable capacity, reliability, and consistency in performance. For example, if drives of different capacities are mixed in a RAID set, the capacity of the smallest drive is used from each disk in the set to make up the RAID set’s overall capacity.

Logical units are created from the RAID sets by partitioning (seen as slices of the RAID set) the available capacity into smaller units. Logical units are spread across all the physical disks that belong to that set. Each logical unit created from the RAID set is assigned a unique ID, called a logical unit number(LUN). LUNs hide the organization and composition of the RAID set from the hosts. LUNs created by traditional storage provisioning methods are also referred to as thick LUNs to distinguish them from the LUNs created by virtual provisioning methods.
When a LUN is configured and assigned to a non-virtualized host, a bus scan is required to identify the LUN. This LUN appears as a raw disk to the operating system. To make this disk usable, it is formatted with a file system and then the file system is mounted.

In a virtualized host environment, the LUN is assigned to the hypervisor, which recognizes it as a raw disk. This disk is configured with the hypervisor file system, and then virtual disks are created on it. Virtual disks are files on the hypervisor file system. The virtual disks are then assigned to virtual machines and appear as raw disks to them. To make the virtual disk usable to the virtual machine, similar steps are followed as in a non-virtualized environment. Here, the LUN space may be shared and accessed simultaneously by multiple virtual machines.
Virtual machines can also access a LUN directly on the storage system. In this method the entire LUN is allocated to a single virtual machine. Storing data in this way is recommended when the applications running on the virtual machine are response-time sensitive, and sharing storage with other virtual machines may impact their response time.

LUN Expansion – MetaLUN: MetaLUN is a method to expand LUNs that require additional capacity or performance. A metaLUN can be created by combining two or more LUNs. A metaLUN consists of a base LUN and one or more component LUNs. MetaLUNs can be either concatenated or striped.

All LUNs in both concatenated and striped expansion must reside on the same disk-drive type: either all FibreChannel or all ATA.

Virtual provisioning enables creating and presenting a LUN with more capacity than is physically allocated to it on the storage array. The LUN created using virtual provisioning is called a thin LUN to distinguish it from the traditional LUN.
Thin LUNs do not require physical storage to be completely allocated to them at the time they are created and presented to a host. Physical storage is allocated to the host “on-demand” from a shared pool of physical capacity. Shared pools can be homogeneous (containing a single drive type) or heterogeneous (containing mixed drive types, such as flash, FC, SAS, and SATA drives).
Virtual provisioning improves storage capacity utilization and simplifies storage management.

Both traditional and thin LUNs can coexist in the same storage array. Based on the requirement, an administrator may migrate data between thin and traditional LUNs.

LUN Masking: is a process that provides data access control by defining which LUNs a host can access. The LUN masking function is implemented on the storage array. This ensures that volume access by hosts is controlled appropriately, preventing unauthorized or accidental use in a shared environment.

Intelligent storage systems generally fall into one of the following two categories: high-end storage systems, and midrange storage systems. High-end storage systems, referred to as active-active arrays, are generally aimed at large enterprise applications. These systems are designed with a large number of controllers and cache memory. An active-active array implies that the host can perform I/Os to its LUNs through any of the available controllers.


Midrange storage systems are also referred to as active-passive arrays and are best suited for small-and medium-sized enterprise applications.
In an active-passive array, a host can perform I/Os to a LUN only through the controller that owns the LUN. The host can perform reads or writes to the LUN only through the path to controller A because controller A is the owner of that LUN. The path to controller B remains passive and no I/O activity is performed through this path.
Midrange storage systems are typically designed with two controllers, each of which contains host interfaces, cache, RAID controllers, and interface to disk drives.

Practical examples of Intelligent storage systems are EMC VNX, EMC Symmetrix VMAX
The EMC VNX storage array is EMC’s midrange storage offering. EMC VNX is a unified storage platform that offers storage for block, file, and object-based data within the same array.

EMC Symmetrix is EMC’s high-end  storage offering. The EMC Symmetrix offering includes Symmetrix Virtual Matrix (VMAX) series.

•Incrementally scalable to 2,400 disks
•Supports up to 8 VMAX engines (Each VMAX engine contains a pair of directors)
•Supports flash drives, fully automated storage tiering(FAST), virtual provisioning, and Cloud computing
•Supports up to 1 TB of global cache memory
•Supports FC, iSCSI, GigE, and FICON for host connectivity
•Supports RAID levels 1, 1+0, 5, and 6
•Supports storage-based replication through EMC TimeFinder and EMC SRDF
•Highly fault-tolerant design that allows non-disruptive upgrades and full component-level redundancy with hot-swappable replacements


Module 5 - Fibre Channel Storage Area Network (FC SAN)

Overview of FC SAN
Direct-attached storage (DAS) is often referred to as a stovepipe storage environment. Hosts “own” the storage, and it is difficult to manage and share resources on these isolated storage devices. Efforts to organize this dispersed data led to the emergence of the storage area network (SAN).

What is SAN
SAN is a high-speed, dedicated network of servers and shared storage devices. It enables storage consolidation and enables storage to be shared across multiple servers. This improves the utilization of storage resources compared to direct-attached storage architecture. SAN also enables organizations to connect geographically dispersed servers and storage.

Common SAN deployments are Fibre Channel (FC) SAN and IP SAN. Fibre Channel SAN uses Fibre Channel protocol for the transport of data, commands, and status information between servers (or hosts) and storage devices. IP SAN uses IP-based protocols for communication.

Understanding Fibre Channel:  Fibre Channel is a high-speed network technology that runs on high-speed optical fiber cables and serial copper cables.

High data transmission speed is an important feature of the FC networking technology. In comparison with Ultra-SCSI that is commonly used in DAS environments, FC is a significant leap in storage networking technology. The latest FC implementations of 16 GFC (Fibre Channel) offers a throughput of 3200 MB/s (raw bit rates of 16 Gb/s), whereas Ultra640 SCSI is available with a throughput of 640 MB/s. Credit-based flow control mechanism in FC delivers data as fast as the destination buffer is able to receive it, without dropping frames. Also FC has very little transmission overhead. The FC architecture is highly scalable, and theoretically, a single FC network can accommodate approximately 15 million devices.

Note: FibRE refers to the protocol, whereas fibber refers to a media.
Components of FC SAN : Servers and storage are the end points or devices in the SAN (called ‘nodes’). FC SAN infrastructure consists of node ports, cables, connectors, interconnecting devices (such as FC switches or hubs), along with SAN management software.


Node Ports:  Each node requires one or more ports to provide a physical interface for communicating with other nodes. These ports are integral components of host adapters, such as HBA, and storage front-end controllers or adapters. In an FC environment a port operates in full-duplex data transmission mode with a transmit (Tx) link and a receive (Rx) link.

Cables: SAN implementations use optical fiber cabling. Copper can be used for shorter distances for back-end connectivity because it provides acceptable signal-to-noise ratio for distances up to 30 meters.

There are two types of optical cables: multimode and single-mode. Multimode fiber (MMF) cable carries multiple beams of light projected at different angles simultaneously onto the core of the cable. Based on the bandwidth, multimode fibers are classified as OM1 (62.5μm core), OM2 (50μm core), and laser-optimized OM3 (50μm core). An MMF cable is typically used for short distances because of signal degradation (attenuation) due to modal dispersion.

Single-mode fiber (SMF) carries a single ray of light projected at the center of the core. These cables are available in core diameters of 7 to 11 microns; the most common size is 9 microns. single-mode provides minimum signal attenuation over maximum distance (up to 10 km).
MMFs are generally used within data centers for shorter distance runs, whereas SMFs are used for longer distances.

Connectors : A connector is attached at the end of a cable to enable swift connection and disconnection of the cable to and from a port. A Standard connector (SC) and a Lucent connector (LC) are two commonly used connectors for fiber optic cables. Straight Tip (ST) is another fiber-optic connector, which is often used with fiber patch panels.

Interconnecting Devices : FC hubs, switches, and directors are the interconnect devices commonly used in FC SAN. Hubs provide limited connectivity and scalability. Hubs are used as communication devices in FC-AL implementations. Because of the availability of low-cost and high-performance switches, hubs are no longer used in FC SANs.
Switches are more intelligent than hubs and directly route data from one physical port to another. Therefore, nodes do not share the data path. Instead, each node has a dedicated communication path.
Directors are high-end switches with a higher port count and better fault-tolerance capabilities.
Switches are available with a fixed port count or with modular design. In a modular switch, the port count is increased by installing additional port cards to open slots. The architecture of a director is always modular.

SAN Management Software: SAN management software manages the interfaces between hosts, interconnect devices, and storage arrays. The software provides a view of the SAN environment and enables management of various resources from one central console.

FC Interconnectivity options: The FC architecture supports three basic interconnectivity options: point-to-point, fibre channel arbitrated loop (FC-AL), and fibre channel switched fabric (FC-SW).

Point-to-point connectivity is the simplest FC configuration—two devices are connected directly to each other.
The point-to-point configuration offers limited connectivity, because only two devices can communicate with each other at a given time. Standard DAS uses point-to-point connectivity

In the FC-AL connectivity configuration, devices are attached to a shared loop. FC-AL has the characteristics of a token ring topology and a physical star topology. In FC-AL, each device contends with other devices to perform I/O operations. Devices on the loop must “arbitrate” to gain control of the loop. At any given time, only one device can perform I/O operations on the loop.


FC-AL uses only 8-bits of 24-bit Fibre Channel addressing (the remaining 16-bits are masked) and enables the assignment of 127 valid addresses to the ports. Hence, it can support up to 127 devices on a loop. One address is reserved for optionally connecting the loop to an FC switch port. Therefore, up to 126 nodes can be connected to the loop.

FC-SW connectivity is also referred to as fabric connect. A fabric is a logical space in which all nodes communicate with one another in a network. This virtual space can be created with a switch or a network of switches. Each port in a fabric has a unique 24-bit Fibre Channel address for communication.
In a switched fabric, the link between any two switches is called an interswitch link (ISL).
Unlike a loop configuration, a FC-SW network provides dedicated path and scalability. The addition or removal of a device in a switched fabric is minimally disruptive; it does not affect the ongoing traffic between other devices.

Ports types in Switched Fabric :  Ports in a switched fabric can be one of the following types:


N_Port:An end point in the fabric. This port is also known as the node port. Typically, it is a host port (HBA) or a storage array port that is connected to a switch in a switched fabric.


E_Port:A port that forms the connection between two FC switches. This port is also known as the expansion port. The E_Porton an FC switch connects to the E_Portof another FC switch in the fabric ISLs.
F_Port:A port on a switch that connects an N_Port. It is also known as a fabric port.

G_Port:A generic port on a switch that can operate as an E_Portor an F_Portand determines its functionality automatically during initialization.

Fibre Channel (FC) Architecture : Traditionally, host computer operating systems have communicated with peripheral devices over channel connections, such as ESCON and SCSI.
The FC architecture represents true channel/network integration and captures some of the benefits of both channel and network technology. FC SAN uses the Fibre Channel Protocol (FCP) that provides both channel speed for data transfer with low protocol overhead and scalability of network technology.
Fibre Channel provides a serial data transfer interface that operates over copper wire and optical fiber. FCP is the implementation of SCSI over an FC network. In FCP architecture, all external and remote storage devices attached to the SAN appear as local devices to the host operating system.

FC Protocol Stack : FCP defines the communication protocol in five layers: FC-0 through FC-4 (except FC-3 layer, which is not implemented).



FC Addressing in Switched Fabric : An FC address is dynamically assigned to Node when a node port logs on to the fabric. The FC address has a distinct format, as shown

The first field of the FC address contains the domain ID of the switch. A Domain ID is a unique number provided to each switch in the fabric. Although this is an 8-bit field, there are only 239 available addresses for domain ID because some addresses are deemed special and reserved for fabric management services. For example, FFFFFC is reserved for the name server, and FFFFFE is reserved for the fabric login service.


The area ID is used to identify a group of switch ports used for connecting nodes. An example of a group of ports with common area ID is a port card on the switch. The last field, the port ID, identifies the port within the group.
Therefore, the maximum possible number of node ports in a switched fabric is calculated as:
239 domains X 256 areas X 256 ports = 15,663,104

World Wide Name (WWN) : Each device in the FC environment is assigned a 64-bit unique identifier called the World Wide Name (WWN). The Fibre Channel environment uses two types of WWNs: World Wide Node Name (WWNN) and World Wide Port Name (WWPN). Unlike an FC address, which is assigned dynamically, a WWN is a static name for each node on an FC network. WWNs are similar to the Media Access Control (MAC) addresses used in IP networking. WWNs are burned into the hardware or assigned through software.
Several configuration definitions in a SAN use WWN for identifying storage devices and HBAs. The name server in an FC environment keeps the association of WWNs to the dynamically created FC addresses for nodes.
Structure and Organization of FC Data : In an FC network, data transport is analogous to a conversation between two people, whereby a frame represents a word, a sequence represents a sentence, and an exchange represents a conversation.

Exchange : An exchange operation enables two node ports to identify and manage a set of information units. The structure of these information units is defined in the FC-4 layer. This unit maps to a sequence. An exchange is composed of one or more sequences.

Sequence:A sequence refers to a contiguous set of frames that are sent from one port to another. A sequence corresponds to an information unit, as defined by the ULP.

Frame:A frame is the fundamental unit of data transfer at Layer 2. An FC frame consists of five parts: start of frame (SOF), frame header, data field, cyclic redundancy check (CRC), and end of frame (EOF).

Fabric Services :  All FC switches, regardless of the manufacturer, provide a common set of services as defined in the FibreChannel standards. These services are available at certain predefined addresses.

Login Types in Switched Fabric : Fabric services define three login types:

Fabric login (FLOGI): Performed between an N_Portand an F_Port. To log on to the fabric, a node sends a FLOGI frame with the WWNN and WWPN parameters to the login service at the predefined FC address FFFFFE (Fabric Login Server).

Port login (PLOGI): Performed between two N_Portsto establish a session. The initiator N_Portsends a PLOGI request frame to the target N_Port, which accepts it.

Process login (PRLI): Also performed between two N_Ports. This login relates to the FC-4 ULPs, such as SCSI.

FC SAN Topologies and Zoning :

Mesh Topology : A mesh topology may be one of the two types: full mesh or partial mesh. In a full mesh, every switch is connected to every other switch in the topology. A full mesh topology may be appropriate when the number of switches involved is small. In a full mesh topology, a maximum of one ISL or hop is required for host-to-storage traffic. However, with the increase in the number of switches, the number of switch ports used for ISL also increases. This reduces the available switch ports for node connectivity.

Core-Edge Topology : The core-edge fabric topology has two types of switch tiers. The edge tier is usually composed of switches and offers an inexpensive approach to adding more hosts in a fabric. Each switch at the edge tier is attached to a switch at the core tier through ISLs.
The core tier is usually composed of directors that ensure high fabric availability. In addition, typically all traffic must either traverse this tier or terminate at this tier. In this configuration, all storage devices are connected to the core tier, enabling host-to-storage traffic to traverse only one ISL.
In core-edge topology, the edge-tier switches are not connected to each other.

The core-edge topology provides higher scalability than mesh topology and provides one-hop storage access to all servers in the environment
  
Zoning : Zoning is an FC switch function that enables node ports within the fabric to be logically segmented into groups and communicate with each other within the group.
Each Zone comprises Zone members (HBA and array ports)
Whenever a change takes place in the name server database, the fabric controller sends a Registered State Change Notification (RSCN) to all the nodes impacted by the change. If zoning is not configured, the fabric controller sends an RSCN to all the nodes in the fabric.
In the presence of zoning, a fabric sends the RSCN to only those nodes in a zone where the change has occurred.

Zoning also provides access control, along with other access control mechanisms, such as LUN masking. Zoning provides control by allowing only the members in the same zone to establish communication with each other.
Multiple zone sets may be defined in a fabric, but only one zone set can be active at a time.

Types of Zoning : Zoning can be categorized into three types:
Port zoning: Uses the thephysical address of switch ports to define zones. In port zoning, access to node is determined by the physical switch port to which a node is connected.

WWN zoning: Uses World Wide Names to define zones. The zone members are the unique WWN addresses of the HBA and its targets (storage devices).
Mixed zoning: Combines the qualities of both WWN zoning and port zoning.

Virtualization in SAN :
Block-level storage virtualization:  aggregates block storage devices (LUNs) and enables provisioning of virtual storage volumes, independent of the underlying physical storage. A virtualization layer, which exists at the SAN, abstracts the identity of physical storage devices and creates a storage pool from heterogeneous storage devices.
Virtual volumes are created from the storage pool and assigned to the hosts. Instead of being directed to the LUNs on the individual storage arrays, the hosts are directed to the virtual volumes provided by the virtualization layer.
Typically, the virtualization layer is managed via a dedicated virtualization appliance to which the hosts and the storage arrays are connected.

Block-level storage virtualization also provides the advantage of nondisruptive data migration. In a traditional SAN environment, LUN migration from one array to another is an offline event because the hosts needed to be updated to reflect the new array configuration.

Previously, block-level storage virtualization provided nondisruptivedata migration only within a data center. The new generation of block-level storage virtualization enables nondisruptivedata migration both within and between data centers. It provides the capability to connect the virtualization layers at multiple data centers.

Virtual SAN (also called virtual fabric) : is a logical fabric on an FC SAN, which enables communication among a group of nodes regardless of their physical location in the fabric.
Multiple VSANs may be created on a single physical SAN.
Each VSAN acts as an independent fabric with its own set of fabric services, such as name server, and zoning.
VSANs improve SAN security, scalability, availability, and manageability.  

Practical examples are EMC Connectrix and VPLEX.

The EMC Connectrix family represents the industry’s most extensive selection of networked storage connectivity products.
Connectrix integrates high-speed FC connectivity, highly resilient switching technology, options for intelligent IP storage networking, and I/O consolidation with products that support Fibre Channel over Ethernet (FCoE). The connectivity products offered under the Connectrix brand are: Enterprise directors, departmental switches, and multi-purpose  switches.

EMC VPLEX : EMC VPLEX is the next-generation solution for block-level virtualization and data mobility both within and across datacenters. The VPLEX appliance resides between the servers and heterogeneous storage devices. It forms a pool of distributed block storage resources and enables creating virtual storage volumes from the pool. These virtual volumes are then allocated to the servers. The virtual-to-physical-storage mapping remains hidden to the servers.

The VPLEX family consists of three products: VPLEX Local, VPLEX Metro, and VPLEX Geo.


Module 6 - IP SAN and FCoE
IP SAN : Two primary protocols that leverage IP as the transport mechanism are Internet SCSI (iSCSI) and Fibre Channel over IP (FCIP).

Traditional SAN enables the transfer of block I/O over FibreChannel and provides high performance and scalability. These advantages of FC SAN come with the additional cost of buying FC components, such as FC HBA and switches

Technology of transporting block I/Os over an IP is referred to as IP SAN.
Advantages of IP SAN are existing n/w infrastructure can be leveraged. Reduced cost compared to investing in new FC SAN h/w are s/w. Many long-distance disaster recover solutions already leverage IP-based network. Many robust and mature security options are available for IP n/w.

IP SAN Protocol: iSCSI :  iSCSI is encapsulation of SCSI I/O over IP. iSCSI is an IP based protocol that establishes and manages connections between host and storage over IP.

Components of iSCSI : An initiator (host) eg iSCSI HBA, target (storage or iSCSI gateway), and an IP-based network are the key iSCSI components . In an implementation that uses an existing FC array for iSCSI communication, an iSCSI gateway is used. If an iSCSI-capable storage array is deployed, then a host with the iSCSI initiator can directly communicate with the storage array over an IP network.

iSCSI Host Connectivity Options : A standard NIC with software iSCSI initiator, a TCP offload engine (TOE) NIC with software iSCSI initiator, and an iSCSI HBA are the three iSCSI host connectivity options.
A standard NIC with a software iSCSI initiator is the simplest and least expensive connectivity option, but places additional overhead on the host CPU.

A TOE NIC offloads TCP management functions from the host and leaves only the iSCSI functionality to the host processor. Although this solution improves performance, the iSCSI functionality is still handled by a software initiator that requires host CPU cycles

An iSCSI HBA is capable of providing performance benefits because it offloads the entire iSCSI and TCP/IP processing from the host processor. The use of an iSCSI HBA is also the simplest way to boot hosts from a SAN environment via iSCSI (NIC needs to obtain an IP address before the operating system loads)

iSCSI Topologies : Native iSCSI :
Two topologies of iSCSI implementations are native and bridged. Native topology does not have FC components. The initiators may be either directly attached to targets or connected through the IP network.
FC components are not required for iSCSI connectivity if an iSCSI-enabled array is deployed.

Bridged iSCSI : Bridged topology enables the coexistence of FC with IP by providing iSCSI-to-FC bridging functionality. An external device, called a gateway or a multiprotocol router, must be used to facilitate the communication between the iSCSI host and FC storage (does not have iSCSI ports). The gateway converts IP packets to FC frames and vice versa.

Combining FC and Native iSCSI Connectivity: The most common topology is a combination of FC and native iSCSI. Typically, a storage array comes with both FC and iSCSI ports that enable iSCSI and FC connectivity in the same environment. No bridge device needed.

iSCSI Protocol Stack : SCSI is the command protocol that works at the application layer of the Open System Interconnection (OSI) model. iSCSI is the session-layer protocol that initiates a reliable session between devices that recognize SCSI commands and TCP/IP.

iSCSI Discovery : An initiator must discover the location of its targets on the network and the names of the targets available to it before it can establish a session. This discovery can take place in two ways: SendTargets discovery or internet Storage Name Service (iSNS).

In SendTargets discovery, the initiator is manually configured with the target’s network portal to establish a discovery session
iSNS enables automatic discovery of iSCSI devices on an IP network. The initiators and targets can be configured to automatically register themselves with the iSNS server.

iSCSI Name : A unique worldwide iSCSI identifier, known as an iSCSI name, is used to identify the initiators and targets within an iSCSI network to facilitate communication.

There are two types of iSCSI names commonly used :

iSCSI Qualified Name (IQN):An organization must own a registered domain name to generate iSCSI Qualified Names. A date is included in the name to avoid potential conflicts caused by the transfer of domain names. An example of an IQN is iqn.2008-02.com.example:optional_string.

Extended Unique Identifier (EUI): An EUI is a globally unique identifier based on the IEEE EUI-64 naming standard. An EUI is composed of the euiprefix followed by a 16-character hexadecimal name, such as eui.0300732A32598D26.


IP SAN Protocol: FCIP: FCIP is a tunneling protocol that enables distributed FC SAN islands to be interconnected over the existing IP-based networks. In FCIP FC frames are encapsulated onto the IP payload. An FCIP implementation is capable to merge interconnected fabrics into a single fabric.

Majority of FCIP implementations today use switch-specific features such as IVR (Inter-VSAN Routing) or FCRS (Fibre Channel Routing Services) to create a tunnel. In this manner, traffic may be routed between specific nodes without actually merging the fabrics.

FCIP is extensively used in disaster recovery implementations in which data is duplicated to the storage located at a remote site.
In an FCIP environment, an FCIP gateway is connected to each fabric via a standard FC connection. The fabric treats these gateways as layer 2 fabric switches. An IP address is assigned to the port on the gateway, which is connected to an IP network.

FCIP Protocol Stack:


Fibre Channel over Ethernet ( FCoE ) : Data centers typically have multiple networks to handle various types of I/O traffic—for example, an Ethernet network for TCP/IP communication and an FC network for FC communication.
The need for two different kinds of physical network infrastructure increases the overall cost and complexity of data center operation.
FibreChannel over Ethernet (FCoE) protocol provides consolidation of LAN and SAN traffic over a single physical interface infrastructure. FCoE uses the Converged Enhanced Ethernet (CEE) link (10 Gigabit Ethernet) to send FC frames over Ethernet.

Components of an FCoE Network: CNA : Converged Network Adapters (CNAs). A CNA replaces both HBAs and NICs in the server and consolidates both the IP and FC traffic.
CNA offloads the FCoE protocol processing task from the server, thereby freeing the server CPU resources for application processing. A CNA contains separate modules for 10 Gigabit Ethernet, Fibre Channel, and FCoE Application Specific Integrated Circuits (ASICs).

Cable : Currently two options are available for FCoE cabling: Copper based Twinax and standard fiber optical cables. A Twinax cable is composed of two pairs of copper cables covered with a shielded casing. The Twinax cable can transmit data at the speed of 10 Gbps over shorter distances up to 10 meters. Twinax cables require less power and are less expensive than fiber optic cables.
The Small Form Factor Pluggable Plus (SFP+) connector is the primary connector used for FCoE links and can be used with both optical and copper cables.

FCoE Switch: An FCoE switch has both Ethernet switch and Fibre Channel switch functionalities. The FCoE switch has a Fibre Channel Forwarder (FCF), Ethernet Bridge, and set of Ethernet ports and optional FC ports.

FCoE Frame mapping: The FC stack consists of five layers: FC-0 through FC-4. Ethernet is typically considered as a set of protocols that operates at the physical and data link layers in the seven-layer OSI stack. The FCoE protocol specification replaces the FC-0 and FC-1 layers of the FC stack with Ethernet. This provides the capability to carry the FC-2 to the FC-4 layer over the Ethernet layer.

To maintain good performance, FCoE must use jumbo frames to prevent a Fibre Channel frame from being split into two Ethernet frames.

Converged Enhanced Ethernet : Converged Enhanced Ethernet (CEE) or lossless Ethernet provides a new specification to the existing Ethernet standard that eliminates the lossy nature of Ethernet.
Lossless Ethernet requires following functionalities

•Priority-based flow control (PFC): created 8 virtual links on single physical link, uses PAUSE capability of Ethernet for reach virtual link for lossless
•Enhanced transmission selection (ETS) : Allocates  bandwidth to different traffic classes, such as LAN, SAN, and Inter Process Communication (IPC). When a particular class of traffic does not use its allocated bandwidth, ETS enables other traffic classes to use the available bandwidth.
•Congestion notification (CN) : Congestion notification provides end-to-end congestion management for protocols, such as FCoE.
•Data center bridging exchange protocol (DCBX)

Module 7 - Network-Attached Storage (NAS).

File Sharing Environment : File sharing, enables users to share files with other users . A user who creates the file (the creator or owner of a file) determines the type of access (such as read, write, execute, append, delete) to be given to other users. When multiple users try to access a shared file at the same time, a locking scheme is required to maintain data integrity and, at the same time, make this sharing possible.

Some examples of file-sharing methods are; File Transfer Protocol (FTP), Distributed File System (DFS), client-server models that use file-sharing protocols such as Network File System NFS and Common Internet File System CIFS, and the peer-to-peer (P2P) model.

NAS is a dedicated, high-performance file sharing and storage device. NAS enables its clients to share files over an IP network. NAS uses network and file-sharing protocols to provide access to the file data. These protocols include TCP/IP for data transfer, and Common Internet File System (CIFS) and Network File System (NFS) for network file service. NAS enables both UNIX and Microsoft Windows users to share the same data seamlessly.
A NAS device uses its own operating system and integrated hardware and software components to meet specific file-service needs. Its operating system is optimized for file I/O and, therefore, performs file I/O better than a general-purpose server

Components of NAS : A NAS device has two key components: NAS head and storage. In some NAS implementations, the storage could be external to the NAS device and shared with other hosts.

Common Internet File System (CIFS) : is a client-server application protocol that enables client programs to make requests for files and services on remote computers over TCP/IP. It is a public, or open, variation of Server Message Block (SMB) protocol.
It uses file and record locking to prevent users from overwriting the work of another user on a file or a record. It is Stateful protocol CIFS server maintains connection information regarding every connected client.
Users refer to remote file systems with an easy-to-use file-naming scheme:
\\server\share or \\servername.domain.suffix\share.

Network File System (NFS) is a client-server protocol for file sharing that is commonly used on UNIX systems. It uses Remote Procedure Call (RPC) as a method of inter-process communication between two computers.
NFS (NFSv3 and NFSv2) is a stateless protocol and uses UDP as transport layer protocol (or TCP for NSFv3). NFS version 4 (NFSv4): Uses TCP and is based on a statefulprotocol design. It offers enhanced security.

NAS I/O Operation: The NAS operating system keeps track of the location of files on the disk volume and converts client file I/O into block-level I/O to retrieve data and converts back to file I/O for client.

NAS Implementation – Unified NAS: common NAS implementations are unified, gateway, and scale-out.
The unified NAS consolidates NAS-based (file-level) and SAN-based (block-level) data access within a unified storage platform. It supports both CIFS and NFS protocols for file access and iSCSIand FC protocols for block level access. A unified NAS contains one or more NAS heads and storage in a single system. The storage may consist of different drive types, such as SAS, ATA, FC, and flash drives, to meet different workload requirements.

Gateway NAS : device consists of one or more NAS heads and uses external and independently managed storage. Management functions in this type of solution are more complex than those in a unified NAS environment because there are separate administrative tasks for the NAS head and the storage. A gateway solution can use the FC infrastructure, such as switches and directors for accessing SAN-attached storage arrays or direct-attached storage arrays.

The gateway NAS is more scalable compared to unified NAS because NAS heads and storage arrays can be independently scaled up when required. NAS gateway and the storage system in a gateway solution is achieved through a traditional FC SAN.

Scale-out NAS implementation:  Pools multiple nodes together in a cluster. A node may consist of either the NAS head or storage or both. The cluster performs the NAS operation as a single entity. A scale-out NAS provides the capability to scale its resources by simply adding nodes to a clustered NAS architecture. The cluster works as a single NAS device and is managed centrally.

Scale-out NAS creates a single file system that runs on all nodes in the cluster. All information is shared among nodes, so the entire file system is accessible by clients connecting to any node in the cluster. Scale-out NAS stripes data across all nodes in a cluster along with mirror or parity protection. Scale-out NAS clusters use separate internal and external networks for back-end and front-end connectivity, respectively. An internal network (uses high-speed networking technology, such as InfiniBand or Gigabit Ethernet.) provides connections for intracluster communication, and an external network connection enables clients to access and share file data.

File-level Virtualization : File-level virtualization eliminates the dependencies between the data accessed at the file level and the location where the files are physically stored. It creates a logical pool of storage, enabling users to use a logical path, rather than a physical path, to access files. A global namespace is used to map the logical path of a file to the physical path names. File-level virtualization enables the movement of files across NAS devices, even if the files are being accessed.

Concepts in Practice:

EMC Isilon : Scale-out NAS solution. Isilon has a specialized operating system called OneFS that enables the scale-out NAS architecture.

EMC VNX Gateway: Gateway NAS Solution: The VNX Series Gateway contains one or more NAS heads, called X-Blades, that access external storage arrays, such as Symmetrix and block-based VNX via SAN. The VNX Gateway supports both pNFS and EMC patented Multi-Path File System (MPFS) protocols.

Module 8 - Object-based and the Unified Storage

Object-based Storage: more than 90 percent of data generated is unstructured. Traditional Solutions NAS is inefficient to handle growth as there is high overhead on NAS to managing large number of permission and nested directories.

NAS also manages large amounts of metadata generated by hosts, storage systems, and individual applications. Typically this metadata is stored as part of the file and distributed throughout the environment. This adds to the complexity and latency in searching and retrieving files.
These changes demanded a smarter approach Object-based storage is a way to store file data in the form of objects on flat address space based on its content and other attributes rather than the name and location.


OSD uses flat address space to store data. Therefore, there is no hierarchy of directories and files; as a result, a large number of objects can be stored in an OSD system.

Each object stored in the system is identified by a unique ID called the object ID. The object ID is generated using specialized algorithms such as hash function on the data and guarantees that every object is uniquely identified.

Components of Object-based Storage device: The OSD system is typically composed of three key components: nodes, private network, and storage.

The OSD system is composed of one or more nodes. A node is a server that runs the OSD operating environment and provides services to store, retrieve, and manage data in the system. The OSD node divides the file into two parts: user data and metadata. The OSD node has two key services: metadata service and storage service.
The metadata service is responsible for generating the object ID from the contents (may also include other attributes of data) of a file. It also maintains the mapping of the object IDs and the file system namespace.
The storage service manages a set of disks on which the user data is stored. OSD typically uses low-cost and high-density disk drives to store the objects.

Traditional storage solutions, such as SAN, and NAS, do not offer all these benefits as a single solution. Object-based storage combines benefits of both the worlds. It provides platform and location independence, and at the same time, provides scalability, security and data-sharing capabilities.

These capabilities make OSD a strong option for cloud-based storage. Cloud service providers can leverage OSD to offer storage-as-a-service. OSD supports web service access via representational state transfer (REST) and simple object access protocol (SOAP). REST and SOAP APIs can be easily integrated with business applications that access OSD over the web.


Content Addressed Storage (CAS): A data archival solution is a promising use case for OSD.
Content Addressed storage (CAS) is a special type of object-based storage device purposely built for storing fixed content.
The stored object is assigned a globally unique address, known as a content address (CA). This address is derived from the object’s binary representation. Data access in CAS differs from other OSD devices. In CAS, the application server can access the CAS device only via the CAS API running on the application server.

Unified Storage : Deploying disparate storage solutions ( NAS, SAN, and OSD) adds management complexity, cost and environmental overheads. An ideal solution would be to have an integrated storage solution that supports block, file, and object access.
It supports multiple protocols for data access and can be managed using a single management interface.
Components of Unified Storage: A unified storage system consists of following key components: storage controller, NAS head, OSD node, and storage.
The storage controller provides block-level access to application servers through iSCSI, FC, or FCoE protocols. It contains iSCSI, FC, and FCoE front-end ports for direct block access.

In a unified storage system, block, file, and object requests to the storage travel through different I/O paths. Block I/O request (FC, iSCSI, or FCoE) File I/O request (NFS or CIFS ) Object I/O request (REST or SOAP)

Concepts in Practice : EMC Atmos, EMC VNX, and EMC Centera.
EMC Atmos : object-based storage for unstructured data. Atmos can be deployed in two ways: as a purpose-built hardware appliance or as software in VMware environments.

EMC Centera : EMC Centera is a simple, affordable, and secure repository for information archiving. EMC Centera is designed and optimized specifically to deal with the storage and retrieval of fixed content by meeting performance, compliance, and regulatory requirements.



EMC VNX : EMC VNX is a unified storage platform that consolidates block, file, and object access in one solution. It implements a modular architecture that integrates hardware components for block, file, and object access. EMC VNX delivers file access (NAS) functionality via X-Blades (Data Movers) and block access functionality via storage processors. Optionally it offers object access to the storage using EMC Atmos Virtual Edition (Atmos VE).

Module 9- Introduction to Business Continuity

Information is an organization’s most important asset. Cost of unavailability of information to an organization is greater than ever.

Business Continuity: Is a process that prepares for, responds to, and recovers from a system outage that can adversely affects business operations. In a virtualized environment, BC solutions need to protect both physical and virtualized resources.

It involves proactive measures, such as business impact analysis, risk assessments, BC technology solutions deployment (backup and replication), and reactive measures, such as disaster recovery and restart, to be invoked in the event of a failure.

Information Availability: It is the ability of an IT infrastructure to function according to business expectations, during its specific time of operation.
Information Availability can be defined with help of

·        Accessibility: Info should be accessible to right user when required.
·        Reliability : Information should be reliable and correct in all aspects
·        Timeliness: Defines the time window during which information must be accessible.

Causes of Information Unavailability: Various planned and unplanned incidents result in information unavailability

·        Unplanned Outages (20%) – Failure like Database corruption, component failure, Human error
·        Planned Outages (80%) – Competing workloads- backup, reporting, include installation, maintenance of new hardware, software upgrades or patches, and refresh/migration
·        Disaster (<1% of occurrences) – Natural or man made – Flood, Fire, Earthquake

Impact of Downtime: Information unavailability or downtime results in loss of productivity, loss of revenue, poor financial performance, and damages to reputation.

The business impact of downtime is the sum of all losses sustained as a result of a given disruption. An important metric, average cost of downtime per hour, provides a key estimate

Average cost of downtime per hour = average productivity loss per hour + average revenue loss per hour
Where:
Productivity loss per hour = (total salaries and benefits of all employees per week) / (average number of working hours per week)
Average revenue loss per hour = (total revenue of an organization per week) / (average number of hours per week that an organization is open for business)

Measuring Information Availability : MTBF and MTTR
Mean Time Between Failure (MTBF): It is the average time available for a system or component to perform its normal operations between failures. MTBF is calculated as: Total uptime/Number of failures

Mean Time To Repair (MTTR):It is the average time required to repair a failed component. MTTR is calculated as: Total downtime/Number of failures

IA = MTBF/ (MTBF+MTTR) or IA= uptime/ (uptime + downtime)
Availability Measurement – Levels of ‘9s’ availability. For example, a service that is said to be “five 9s available” is available for 99.999 percent of the scheduled time in a year (24 ×365).


BC terminologies and BC planning :
Disaster recovery: This is the coordinated process of restoring systems, data, and the infrastructure required to support ongoing business operations after a disaster occurs. It is the process of restoring a previous copy of the data and applying logs or other necessary processes to that copy to bring it to a known point of consistency. Generally implies use of Backup technologies

Disaster restart: This is the process of restarting business operations with mirrored consistent copies of data and applications. Generally implies use of replication technologies.

Recovery-Point Objective (RPO): This is the point-in-time to which systems and data must be recovered after an outage. It defines the amount of data loss that a business can endure.
Based on the RPO, organizations plan for the frequency with which a backup or replica must be made. Example

RPO of 24 hours:Backups are created at an offsite tape library every midnight. The corresponding recovery strategy is to restore data from the set of last backup tapes.


RPO of 1 hour: Shipping database logs to the remote site every hour. The corresponding recovery strategy is to recover the database to the point of the last log shipment.

RPO in the order of minutes: Mirroring data asynchronously to a remote site.

RPO of zero: Mirroring data synchronously to a remote site.

Recovery-Time Objective (RTO): The time within which systems and applications must be recovered after an outage. It defines the amount of downtime that a business can endure and survive. Some examples of RTOs

RTO of 72 hours:Restore from tapes available at a cold site.
RTO of 12 hours:Restore from tapes available at a hot site.
RTO of few hours:Use disk-based backup technology, which gives faster restore than a tape backup.
RTO of a few seconds:Cluster production servers with bidirectional mirroring, enabling the applications to run at both sites simultaneously.

BC Planning Lifecycles: The BC planning lifecycle includes five stages:
1. Establishing objectives 2. Analyzing 3. Designing and developing 4. Implementing
 5. Training, testing, assessing, and maintaining

Business Impact Analysis: A business impact analysis (BIA) identifies which business units, operations, and processes are essential to the survival of the business. It evaluates the financial, operational, and service impacts of a disruption to essential business processes

Based on the potential impacts associated with downtime, businesses can prioritize and implement countermeasures to mitigate the likelihood of such disruptions.

BC Technology Solutions : Solutions that enable BC are

•Resolving single points of failure
•Multipathing software
•Backup and replication

Single points of failure :refers to the failure of a component that can terminate the availability of the entire system or IT service. To mitigate single points of failure, systems are designed with redundancy, such that the system fails only if all the components in the redundancy group fail.

e.g. Configuration of NIC teaming at a server allows protection against single physical NIC failure.

Multipathing software: Configuration of multiple paths increases the data availability through path failover. Multiple paths to data also improve I/O performance through load balancing among the paths. Multipathing software intelligently recognizes and manages the paths to a device by sending I/O down the optimal path based on the load balancing and failover policy setting for the device. In a virtual environment, multipathing is enabled either by using the hypervisor’s built-in capability or by running a third-party software module, added to the hypervisor.

Concept in Practice : EMC PowerPath : EMC PowerPath is host-based multipathing software. Every I/O from the host to the array must pass through the PowerPath software

Module 10 - Backup and Archive

A backup is an additional copy of production data, created and retained for the sole purpose of recovering lost or corrupted data and also for regulatory requirements compliance.
Backups are performed to serve three purposes: disaster recovery, operational recovery, and archival.

Note: Backup window is the period during which a source is available for performing a data backup
Backup Granularity : Backup granularity depends on business needs and the required RTO/RPO.
backups can be categorized as full, incremental, and cumulative (or differential).

Incremental backup copies the data that has changed since the last full or incremental backup, whichever has occurred more recently. This is much faster than a full backup.
but takes longer to restore.
Cumulative backup copies the data that has changed since the last full backup. This method takes longer than an incremental backup but is faster to restore

Another way to implement full backup is synthetic (or constructed) backup. It is usually created from the most recent full backup and all the incremental backups performed after that full backup. This backup is called synthetic because the backup is not created directly from production data.It enables a full backup copy to be created offline without disrupting the I/O operation on the production volume.

Restore from Incremental backup: The process of restoration from an incremental backup requires the last full backup and all the incremental backups available until the point of restoration. It takes less time to backup , less files and storage and longer restore time.

Restore from cumulative backup: More files to be backed up, taks more time to backup and more storage space and faster recovery b’coz only last full backup and last cumulative back up must be applied

Backup Architecture : A backup system commonly uses the client-server architecture with a backup server and multiple backup clients.

The backup server manages the backup operations and maintains the backup catalog, which contains information about the backup configuration and backup metadata.
The role of a backup client is to gather the data that is to be backed up and send it to the storage node.
The storage node (media server) is responsible for writing the data, to the backup device. In a backup environment, a storage node is a host that controls backup devices. In many cases, the storage node is integrated with the backup server, and both are hosted on the same physical platform

When a backup operation is initiated, significant network communication takes place between the different components of a backup infrastructure. The backup operation is typically initiated by a server, but it can also be initiated by a client.

After the data is backed up, it can be restored when required. A restore process must be manually initiated from the client

Backup Methods: Hot backup or online (application up and running, with data being accessed, open file agent can be used to backup open files) and
cold backup or Offline ( application is shutdown)are the two methods deployed for backup. They are based on the state of the application when the backup is performed.

In a disaster recovery environment, bare-metal recovery (BMR) refers to a backup in which OS, hardware, and application configurations are appropriately backed up for a full system recovery.
Some BMR technologies—for example server configuration backup (SCB)—can recover a server even onto dissimilar hardware.

Server configuration backup (SCB) : The process of system recovery involves reinstalling the operating system, applications, and server settings and then recovering the data. During a normal data backup operation, server configurations required for the system restore are not backed up. Server configuration backup (SCB) creates and backs up server configuration profiles based on user-defined schedules.
In a server configuration backup, the process of taking a snapshot of the application server’s configuration (both system and application configurations) is known as profiling.

Two types of profiles used
Base profile: contains the key elements of the operating system required to recover the server
Extended Profile : typically larger than the base profile and contains all the necessary information to rebuild the application environment.

Backup topologies - such as Direct-attached, LAN-based, SAN-based and mixed backup.

<not complete ...>
Module 13- Cloud Computing


Cloud Computing: A model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g. servers, network, storage, application and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction.

According to NIST, the cloud infrastructure should have five essential characteristics:

·        On-Demand Self-Service – like provision computing capabilities, view catalogue via web interface to request for a service

·        Broad Network Access - Capabilities are available over the network and accessible from thin or thick clients like mobile, laptop, tablet etc.

·        Resource Pooling – to serve multiple customers using multitenant model. Virtualization enables resource pooling and multitenancy in the cloud.

·        Rapid Elasticity  - Capabilities can be elastically provisioned and released, in some cases automatically, to scale rapidly outward and inward commensurate with demand.

·        Measured Service - Resource usage can be monitored, controlled, and reported, providing transparency for both the provider and consumer of the utilized service.

Cloud Enabling Technologies – Grid Computing, Utility Computing, Virtualization, Service Oriented Architecture (SOA)

Cloud Service and Deployment Models : three primary cloud service models –Software-as-a-Service (SaaS), Platform-as-a-Service (PaaS), and Infrastructure-as-a-Service (IaaS). Thislesson also covers cloud deployment models –Public, Private, Community, and Hybrid.

·        Cloud Service : Infrastructure-as-a-Service (IaaS) : The capability provided to the consumer is to provision processing, storage, networks, and other fundamental computing resources where the consumer is able to deploy and run arbitrary software, which can include operating systems and applications. Amazon Elastic Compute Cloud (Amazon EC2) is an example of IaaS

·        Platform-as-a-Service (PaaS)  : The capability provided to the consumer is to deploy onto the cloud infrastructure consumer-created or acquired applications created using programming languages, libraries, services, and tools supported by the provide.
·        Google App Engine and Microsoft Windows Azure Platform are examples of PaaS.

·        Software-as-a-Service (SaaS): The capability provided to the consumer is to use the provider’s applications running on a cloud infrastructure.
·        Example Salesforce.com is a provider of SaaS-based CRM applications.

Deployment Models: cloud computing is classified into four deployment models—public, private, community, and hybrid.

·        Public: In a public cloud model, the cloud infrastructure is provisioned for open use by the general public. It may be owned, managed, and operated by a business, academic, or government organization, or some combination of them. It exists on the premises of the cloud provider.

·        Private :  In a private cloud model, the cloud infrastructure is provisioned for exclusive use by a single organization. It may be owned, managed, and operated by the organization (on-premise), a third party (Externally hosted private cloud), or some combination of them, and it may exist on or off premises.

·        Community:  In a community cloud model, the cloud infrastructure is provisioned for exclusive use by a specific community of consumers from organizations that have shared concerns. It may be owned, managed, and operated by one or more of the organizations in the community, a third party, or some combination of them, and it may exist on or off premises.

·        Hybrid : the cloud infrastructure is a composition of two or more distinct cloud infrastructures (private, community, or public) that remain unique entities, but are bound together by standardized or proprietary technology that enables data and application portability (example, cloud bursting for load balancing between clouds).

Cloud computing infrastructure, challenges of cloud computing: Cloud computing infrastructure usually consists of the following layers:

·  Physical infrastructure: The physical infrastructure consists of physical computing resources, which include physical servers, storage systems, and networks.
·  Virtual infrastructure: Virtualization enables fulfilling some of the cloud characteristics, such as resource pooling and rapid elasticity.
·   Applications and platform software: - SaaS , PaaS

Cloud management and service creation tools:  The cloud management and service creation tools layer includes three types of software:

·  Physical and virtual infrastructure management software
·   Unified management software - The key function of the unified management software is to   automate the creation of cloud services.
·   User-access management software
Cloud Optimized Storage: Cloud-optimized storage typically leverages object-based storage technology. Key characteristics of cloud-optimized storage solution are:

• Massively scalable infrastructure that supports large number of objects across a globally distributed infrastructure

• Unified namespace that eliminates capacity, location, and other file system limitations
 
• Metadata and policy-based information management capabilities that optimizes data protection, availability and cost, based on service levels

• Secure multitenancy that enables multiple applications to be securely served from the same infrastructure. Each application is securely partitioned and data is neither co-mingled nor accessible by other tenants

• Provide access through REST and SOAP web service APIs and file-based access using variety of client devices

Cloud Challenges – Consumer’s Perspective: Security and regulation, Network Latency, Supportability, Vendor lock-in

Provider’s Perspective: Service warranty and service cost, Complexity in deploying vendor software in the cloud, No standard cloud access interface (API).

Financial Aspect: 


Concept in Practice: Vblock:
Vblock is completely integrated cloud infrastructure offering that includes compute, storage, network, and virtualization products. These products are provided by EMC, VMware, and Cisco, who have formed a coalition to deliver Vblocks.

Vblocks enables organizations to build virtualized data centers and cloud infrastructures. Vblocks are pre-architected, preconfigured, pretested and have defined performance and availability attributes.
EMC Unified Infrastructure Manager (UIM) is the unified management solution for Vblocks.

Reference:
http://www.snia.org/education/certification/practice/index.htm
FCIA - www.fibrechannel.com
www.t11.org - Technical committee
Infiband Assoc - www.infibandta.org
http://www.iscsistorage.com/

No comments: