Storage Concepts Student Guide
User Manual:
Open the PDF directly: View PDF .
Page Count: 262
Download | |
Open PDF In Browser | View PDF |
Storage Concepts Introduction Welcome and Introductions Student Introductions • Name • Position • Experience • Your expectations Course Description This 2 day instructor-led course provides a comprehensive introduction to storage technology concepts, terminology and technologies of today’s storage industry. The course examines the need for storage solutions to manage and optimize an IT infrastructure to meet business requirements. The course also examines the major components of a storage system, common storage architectures and the various means of connecting storage elements. It compares network attached storage (NAS) and storage area network (SAN) implementations and data protection issues. It provides detail on industry-defined tiered storage, virtualization, and storage management strategies. Prerequisites Understanding of basic computer concepts Experience working with PCs or servers (Windows or UNIX) Course Objectives By completing the course, you will gain an understanding of: • Storage industry concepts and technologies • Industry-defined tiered storage, virtualization, and storage management strategies Course Topics Modules 1. Introduction to Data Management and Storage Systems 1a.Overview of Storage Concepts 2. Storage Components and technologies 3. Business Continuity and Replication 4. Virtualization of Storage Systems 5. Archiving and File and Content Management 6. Storage System Administration 7. Business Challenges 8. Storage Networking and Security (Optional) Activities Learning activities appear throughout the course. Storage Concepts Introduction to Data Management and Storage Systems Module Objectives Upon completion of this module, you should be able to: • Explain the different types of data • Explain what Cloud Computing is • Explain a storage systems’ view of data • Distinguish between the physical and logical levels of data processing • Understand and explain the basic concepts of data consistency and data integrity Module Topics Introduction to Data Management – Data Types Structured and Unstructured Data Data versus Information Data Processing Levels A Storage Systems View of Data Data Consistency and Data Integrity Concepts and Principles Introduction to Data Management Common Users: Data Examples Photos Movies Documents Email Personal web pages Data backed up to online storage such as Microsoft Sky Drive Increasingly popular cloud systems and on-line applications Movies Business Sector Data Accounting, invoices, and financial records Databases that contain data about clients Email communication Digitalization of printed documents Archiving State Institutions Data Databases Audio and video records Hospital records Confidential and classified data Archiving and digitalization Medical Records Data Lifecycle Data Lifecycle applies to all data that comes into existence • Data Retention Period applies to certain types of data and is governed by law Offline Structured and Unstructured Data Structured Data • Databases Unstructured Data • Medical Images – MRI scans • Photographs • Digital documents – check images • Satellite images • Biotechnology • Digital video • Email The Changing Forms of Data New data types and business models compound the problem of exploding storage. Consider the demands brought on by these new business models as a result of the ubiquity of the Internet. Total Digital Archive Capacity, by Content Type – Worldwide (TB) 30,000,000 25,000,000 20,000,000 15,000,000 10,000,000 5,000,000 0 2005 2006 ESG Research Report: Digital Archiving: EndUser Survey & Market Forecast 2006-2010 2007 2008 Unstructured Data Database Email 2009 2010 Question: What’s The Difference Between Data and Information? Data Information Data Versus Information Data Information • A physical and written representation of information and knowledge • Succession of written characters, which can be represented by numbers, letters, or symbols • Meaningful interpretation of data • Does not have to be written in characters Data in binary code… Data is stored to preserve and pass information on New Trends in Data Management – Cloud Computing Video – What Is Cloud Computing? YouTube video http://www.youtube.com/watch?v=ae_DKNwK_ms&feature=yout u.be Three Key Cloud Characteristics “Cloud is a way of using technology, not a technology in itself – it's a self-service, on-demand pay-per-use model. Consolidation, virtualization and automation strategies will be the catalysts behind cloud adoption.” – The 451 Group The 3 key characteristics of a cloud are: Self-service Pay-per-use Dynamic scale up and down Levels of Data Processing – Logical And Physical Hard Drive Partitioning (Physical Level) For example, a laptop running Microsoft Windows software: • Installed Hard Drives — CD / DVD / memory card drives • Each drive assigned a letter = valid address within the Operating System For example, hard drives have C: & D: addresses, DVD drive has E: address Exercise: Can You Think of Some Reasons for Partitioning a Hard Drive? A user may want to encrypt a partition that will contain critical data Other reasons: You can also send your answers using the CHAT Logical Level – Volumes / File Systems Volume – logical interface used by an operating system to access data stored on a particular media while using a single instance of a file system File System – the way in which files are named, organized, and stored on the hard drive Summary – File System Stores information about where data is physically located Stores metadata, containing additional information Maintains data integrity and allows users to set access restrictions and permissions Installed on a homogenous storage area called a volume Applications access a file system using an application programming interface (API) Sets up the way files are named and organized How the NTFS Works Logical Units on a Storage System A storage system is partitions of logical units striped across a large number of hard drives Windows UNIX Logical Devices Logical Unit Number (LUN) Concept A LUN is a logical device mapped to a storage port Logical Devices Windows UNIX LUNS are mapped to servers Some Definitions Microcode and Firmware Microcode: built-in software that works on the lowest layer of instructions, directly controlling hardware equipment • Most basic software – no graphical user interface (GUI) • Types: Microcode stored on individual hard drives Microcode on a storage system also containing an integrated interface, either a command line interface (CLI) or a GUI – firmware Firmware: contains microcode and some kind of user interface (menus, Icons) Data Consistency Data Consistency: means you have valid, usable and readable data • Point-in-time consistency: data is consistently as it was at any single instant in time For example, synchronous (continuous) data replication • Transaction consistency: preventing “lost” transactions • Application consistency Data Integrity Data Integrity: describes accuracy, reliability and correctness in terms of security and authorized access to a file • Policies: containing rules that govern access, preventing possible data alterations • Permissions and restrictions of access tools Module Summary Upon completion of this module, you should have learned to: • Explain the different types of data • Explain what Cloud Computing is • Explain a storage systems’ view of data • Distinguish between the physical and logical levels of data processing • Understand and explain the basic concepts of data consistency and data integrity Storage Concepts Differentiate between common basic storage architectures Module Objectives Upon completion of this module, you should be able to: • Differentiate between common basic storage architectures Storage Architecture Connectivity DAS Application NAS Application File System Parallel SCSI or Serial SCSI (FC) Application File System Local Area Network Storage SAN File System Storage Area Network Storage Storage Direct Attach Storage Network Attach Storage Storage Area Network Direct Attached Storage Storage is directly attached to the server No other device on the network can access the stored data Example of DAS – A PC with an attached external disk drive, another; A Mainframe direct connect to SAN storage array DAS Application File System Best used for accessing personal data or high speed non-shared access Storage Direct Attach Storage Network Attached Storage File oriented data access NAS Optimized for file serving Easy Installation and Monitoring Application No server intervention (or layer) required for data access Low Total Cost of Ownership Use existing Network/cabling Multiple protocol support (file sharing) using NFS, SMB, CIFS, HTTP, etc. NAS Heads NAS Blades (New HDS-G-Series 400-800) NAS Filers (HDS-HNAS 3000/4000 Series) Local Area Network File System Storage Network Attach Storage Network Attached Storage Targets midsize customers and remote, branch offices of large organizations Ideal for customers with a collaborative environment that requires sharing of files such as project management teams, law offices, and design firms – – – – – – – File serving Software development CAD/CAM Rich media Publishing and broadcast Archiving Near-line Data Storage to meet regulatory requirements NAS Server Local Area Network Storage Introduction to SAN Storage Area Network • Is a separate network that includes computer servers, disks, and other storage devices • Allows networking concepts to be applied to a server/storage model • Has its own connections rather than using a fixed backbone network • Has connections that utilize Fibre Channel equipment • Allows very fast access among servers and storage resources • Enables many servers to share many storage devices • Designed for very high speed shared data storage up to 16 million nodes Storage Area Network Designed to attach computer storage devices such as disk array controllers and tape libraries to servers Primary purpose is the transfer of data between computer systems and storage elements (switches/ directors [large switches]) Consists of a communication infrastructure and a management layer so that data transfer is secure and robust SAN Server Storage Area Network SAN and NAS can co-exist on the same network infrastructure Complicated (relatively) and expensive to implement Storage Storage Area Network Best suited for complex data centers that require high availability, scalability, reliability, and performance • Storage hosting providers SAN Server • International organizations that have multiple data centers • Businesses that implement Service Level Agreements (SLAs) • Disaster Recovery and, or Business Continuity requirements Storage Area Network Storage Summary Direct Attached Storage (DAS) – Storage is directly attached to the application or file server • Only one computer can access the storage • A PC with an externally-attached hard disk drive Network Attached Storage (NAS) – Multiple computers can access and share the storage devices • Accessed over an IP network • Ideal for collaborative file sharing Storage Area Networks (SAN) – Designed to attach computer storage devices to servers • Complex communication infrastructure organizes the connections, storage elements, and computer systems so that data transfer is secure and robust • Best suited for: Storage hosting providers International organizations that have multiple data centers Anyone with need for high speed shared data access Storage Concepts Storage Components and Technologies Module Objectives Upon completion of this module, you should be able to: • Identify and describe the major components of a storage system • Explain the different RAID levels and configurations • Describe the midrange storage system architecture and its components • Explain the factors directly influencing the performance of a storage system • Understand and explain the differences between a midrange and an enterprise storage system Module Topics How a hard drive works What means of hardware redundancy we have How to describe a midrange storage system architecture and components Module Flow Hard Drive and how it works Hard drive connectivity Factors influencing performance Other storage system components Overview of Disk Array Components Disk Drive Components and Connectivity A hard disk drive (HDD) is a nonvolatile storage device that stores data on a magnetic disk. Key components of a disk drive: Disk Drive Components and Connectivity Spindle – A spindle holds one or more platters. It is connected to a motor that spins the platters at constant revolutions per minute (RPM) Platter – A platter is the disk that stores the magnetic patterns. It is made from a nonmagnetic material, usually glass, aluminum, or ceramic, and has a thin coating of magnetic material on both sides. A platter can spin at a speed of 7,200 to 18,000 RPM. The cost of an HDD increases for a higher speed. Disk Drive Components and Connectivity Head – The read-write head of an HDD reads data from and writes data to the platters. It detects (when reading) and modifies (when writing) the magnetization of the material immediately underneath it. Information is written to the platter as it rotates at high speed past the selected head. There is one head for each magnetic platter surface on the spindle; these are mounted on a common actuator arm. Actuator – An actuator arm moves the heads in an arc across the spinning platters, allowing each head to access the entire data area, similar to the action of the pick-up arm of a record player. Disk Drive Components and Connectivity The performance of an HDD is measured using the following parameters: • Capacity – The number of bytes an HDD can store. The current maximum capacity of an HDD is 4TB. • Data transfer rate – The amount of digital data that can be moved to or from the disk within a given time. It is dependent on the performance of the HDD assembly and the bandwidth of the data path. The average data transfer rate ranges between 50-300 MB per second. • Seek time – The time the HDD takes to locate a particular piece of data. The average seek time ranges from 3 to 9 milliseconds. Disk Drive Components and Connectivity Transfer Rates – Performance HDD Bus HDD Connectivity Interfaces Disk Drive Components and Connectivity Bus Cables connect the storage central processing unit (CPU) to the HDD interface. Interface is a device that enables the connection of electrical circuits together. Interfaces use the following standards: • Parallel advanced technology attachment (PATA) • Serial advanced technology attachment (SATA) • Small computer systems interface (SCSI) • Serial attached SCSI (SAS) Disk Drive Components and Connectivity PATA PATA is a standard used to connect HDDs to computers, based on parallel signaling technology. PATA cables are bulky and can be a maximum of 18 inches long, so they can be used only in internal drives. Disk Drive Components and Connectivity SATA SATA evolved from PATA. It uses serial signaling technology. SATA is a standard used to control and transfer data from a server or storage appliance to a client application. Compared to PATA, SATA has the following advantages: • Greater bandwidth • Faster data transfer rates – up to 600GB/sec • Easy to set up and route in smaller computers • Low power consumption • Hot-swap support SATA does not perform as well as SAS Disk Drive Components and Connectivity SCSI A parallel interface standard used to transfer data between devices on both internal and external computer buses. SCSI advantages over PATA and SATA: • Faster data speeds • Multiple devices can connect to a single port • Device independence; can be used with most SCSI compatible hardware SCSI has the following disadvantages: • SCSI interfaces do not always conform to industry standards. • SCSI is more expensive than PATA and SATA. Disk Drive Components and Connectivity Serial Attached SCSI (SAS) Serial Attached SCSI has evolved from the previous SCSI standards as it uses serial signaling technology. SAS is a standard used to control and transfer data with SCSI commands from a server or storage appliance to a client application. SAS advantages over SCSI: • Greater bandwidth • Faster data transfer rates • Easy set up and routing in smaller computers • Low power SAS cables are similar to SATA cables. Redundant Array of Independent Drives RAID Redundant Array of Inexpensive/Independent Drives (RAID) – A method of storing data on multiple disks by combining various physical disks into a single logical unit. A logical disk is a combination of physical disks RAID provides the following advantages: Data consistency and integrity (security, protection from corruption) Fault tolerance Capacity Reliability Better Speed Different types of RAID can be implemented, according to application requirements. RAID The different types of RAID implementation are known as RAID levels. Common RAID levels: • RAID-0 • RAID-1 • RAID-1+ 0 • RAID-5 • RAID-6 RAID-0 (Data Striping) RAID-0 • RAID-0 implements striping; data is spread evenly across two or more disks. RAID-0 Benefits: • Easy to implement • Increased performance in terms of data access (more disks equals more heads, which enables parallel access to more data records) RAID-0 Disadvantage: • This RAID level has no redundancy and no fault tolerance. If any disk fails, data on the remaining disks cannot be retrieved, which is the major disadvantage. RAID-0 uses only data striping. RAID-1 (Data Mirroring) RAID-1 • Implements mirroring to create exact copies of the data on two or more disks. RAID-1 Advantages: • Reduces the overhead of managing multiple disks and tracking the data. • Read time is fast because the system can read from either disk. • If a disk fails, RAID-1 ensures there is an exact copy of the data on the second disk. RAID-1 Disadvantages: • The storage capacity is only half of the actual capacity as data is written twice. • RAID-1 is expensive; doubles the storage. RAID-1+ RAID-1+ • RAID-1+0 is an example of multiple or nested RAID levels. • A nested RAID level combines the features of multiple RAID levels. The sequence in which they are implemented determines the naming of the nested RAID level. • For example, if RAID-0 is implemented before RAID-1, the RAID level is called RAID-0+1. RAID-1+0 combines the features of RAID-0 and RAID-1 by mirroring a striped array. RAID-1+ has the following advantages: • Easy to implement • Fast read/write speed • Data protection RAID-1+ has the disadvantage of high cost to implement. RAID-1+ RAID-5 RAID-5 • RAID-5 consists of a minimum of three disks (two data and one parity) • RAID-5 distributes parity information across all disks to minimize potential bottlenecks if one disk fails, in which case parity data from the other disks is used to recreate the missing information. RAID-5 has the following advantages: • The most common and secure RAID level • Fast read speed • Ensures data recovery if a disk failure occurs RAID-5 has the following disadvantages: • Extra overhead required to calculate and track parity data • Slower writes because it has to calculate parity before writing data RAID-5 RAID-6 RAID-6 • RAID-6 is similar to RAID-5 with an additional parity disk • In RAID-5, if a second disk fails before the first failed disk has been rebuilt, data can be irretrievably lost. The additional parity drive in RAID-6 provides a solution to this problem. RAID-6 is designed for large environments and offers the following benefits: • Provides protection against double-disk failure • Has a fast read speed RAID-6 has the following disadvantages: • Similar to RAID-5 plus the cost of the extra parity disk • Slight performance overhead RAID-6 RAID Type Configuration and Usage Spare Disks — Sparing Correction Copy – Occurs when a drive in a RAID group fails and a compatible spare drive exists. Data is then reconstructed on the spare drive. Spare Disks — Sparing Dynamic Sparing – occurs if the online verification process (built-in diagnostic) determines that the number of errors has exceeded the specified threshold of a disk in a RAID group. Data is then moved to the spare disk, which is a much faster process than data reconstruction Correction Copy Parameters Copy Back No Copy Back Exercise: RAID Configuration Options Considering the following data storage needs, which RAID options would provide the best performance? 1. Online transactional database (banking, stock market) where performance and reliability is key 2. Search engine, or catalog system for a library 3. Overnight billing and inventory system If vILT class, write your answers on blank lines Building a Midrange Storage System – Components Expansion Unit / Disk Enclosure A typical expansion unit or disk enclosure (based on SAS architecture), consists of following components: • Individual hard drives (SSD, SAS, SATA) • Expander (buses and wires for connecting drives together) • Power supplies • Cooling systems (fans) • Chassis Expansion Unit Connectivity Each expansion unit is connected to both controllers Back-End Architecture – Midrange Storage System A midrange storage system contains main controller boards that are equipped with components that can be put into three categories: • Front end (connection to hosts or other storage systems) • Cache (works as a buffer, has major influence on performance) • Back end (connections to hard drive enclosures, RAID operations) Back-End SAS Architecture Example Cache Cache Cache is a temporary storage area for frequently used data. A system can access the cached copy instead of the data in the original location. This reduces the time taken to access data. Cache Operations The storage system’s microcode contains algorithms that should anticipate what data is advantageous to keep (for faster read access) and what data should be erased. There are two basic algorithms that affect the way cache is freed up: LRU Queue MRU Least Recently Used (LRU) — Data that is stored in cache and has not been accessed for a given period of time (i.e., data that is not accessed frequently enough) is erased. Most Recently Used (MRU) — Data accessed most recently is erased. This is based on the assumption that recently used data may not be requested for a while. LRU Cache Data Protection Cache Mirroring Controller 0 Controller 1 x GB Cache x GB Cache x GB Controller 0 x GB Controller 0 Read/Write Cache Mirrored Write Cache x GB Controller 1 Mirrored Write Cache x GB Controller 1 Read/Write Cache Front-End Architecture Interface Board – Ports connecting storage system to servers Protocols • Fibre Channel (FC) – Defines a multi-layered architecture for moving data; allows data transmission over twisted pair and over fiber optic cables • Fibre Channel over Ethernet (FCoE) – Encapsulation of Fibre Channel frames over Ethernet networks • iSCSI – Interface connecting storage system to the LAN; allows organizations to utilize their existing TCP/IP network infrastructure without investing in expensive Fibre Channel switches Front-End Architecture – Other Components In the frontend, we have QE8 FC port controllers that are part of the interface board. These controllers are mainly responsible for conversion of FC transfer protocol into PCIe bus used for internal interconnection of all components.. Notice the CPU and local RAM memory (not cache). Active-Passive Architecture In the event of a path failure, LUN ownership does not change. Data is transferred via the backup path to CTL1 and then internally to CTL0, bypassing CTL0 front end ports. This diminishes performance because of internal communication. Symmetric Active-Active Controllers Multiple paths to a single LUN are possible. LUN ownership automatically changes in the event of path failure. Unlike active-passive architecture, active-active architecture offers equal access to the particular LUN via both paths. This means the performance is not influenced by what path is currently used. Controller Load Balancing – Active-active Architecture No need to configure LUN ownership manually. Communication goes either via CTL0 or CTL1. In the event of path failure, no bypassing is necessary; therefore there is no communication overhead. Fibre Channel Ports and their Configuration Host connected to a Target port Storage connected to an External port for virtualization Storage systems connected via an Initiator port for replication iSCSI Interface Production Test/Development Mixed SAN showing Fibre Channel SAN connection for production servers and an iSCSI LAN for the Test/Development servers Module Summary Upon completion of this module, you should have learned to: • Identify and describe the major components of a storage system • Explain the different RAID levels and configurations • Describe the midrange storage system architecture and its components • Explain the factors directly influencing the performance of a storage system • Understand and explain the differences between a midrange and an enterprise storage system Storage Concepts Business Continuity and Replication Module Objectives Upon completion of this module, you should be able to: • Describe basic concepts of business continuity and replication • Understand business impact analysis and risk assessment • Describe basic concepts of disaster recovery Module Topics Business Continuity concepts Business Impact Analysis and Risk Assessment Back-up strategies and their implementation How Business Continuity and Disaster Recovery connect to IT Replication options Concepts of clusters and geoclusters Business Continuity Concepts Exercise: Data Center Disaster Tornado destroys ABC Corporation’s data center • What needs to be prepared ahead of time? Set up a Business Continuity team • Impacts on the business? Write your answers on blank lines or send your answers via the CHAT Exercise: Data Center Disaster Tornado destroys ABC Corporation’s data center • What needs to be prepared ahead of time? Set up a Business Continuity team A Disaster Recovery Plan Another data center location Suppliers ready to replace damaged equipment Employees trained in responding to these situations Staff in place at the alternate site to take over Failback to main data center • Impacts on the business? Financial loss Damaged reputation and company image Business Continuity Management: Regulatory Requirements Identifies critical business processes and establishes rules and procedures • Implementation governed by regulations and standards Sarbanes Oxley – Corporate reporting and financial results; tells CEO and CFO that they must be able to defend accuracy of their books Basel II – International banking regulation that deals with amortizing cost of risk into financial markets British Standard For Business Continuity Management (BS25999) – guidance for determining business processes and their importance North American Business Continuity Standard – (NFPA1600) Business Continuity Planning The implementation of business continuity concepts with respect to the particular organization. The planning normally includes: • Business Impact Analysis: Identifies which business processes, users and applications are critical to the survival of the business • Risk Assessment : Determines probability of threats to an organization • Policies: Aligns BC policies with the company’s business strategy • Business Recovery Plan: Defines procedures to be taken when a particular situation occurs • Disaster Recovery Plan: Describes how to get the critical applications working again after an incident takes place • Testing and Training Schedules: Provides a timeline for business continuity plan testing Business Impact Analysis – RPO and RTO Recovery Point Objective (RPO) – Worst case time between last backup and interruption time • Represents how much data must be recovered • How much can you afford to lose? Recovery Time Objective (RTO) – How long is the customer willing to live with downed systems? • Represents outage duration Business Continuity versus Disaster Recovery Disaster Recovery – Part of business continuity that focuses only on IT Infrastructure Disaster Recovery Plan contains: Basic information – purpose, area of application, requirements, log of DR plan modifications, members of DR team, their roles and responsibilities Notification/activation phase – notification procedure, call tree, damage assessment, activation criteria, plan activation Recovery procedures – succession of recovery procedures according to their importance, logging, escalation Standard operation resumption – checking whether all systems work properly, termination of DR plan Amendments – call book, vendor SLA, RTO of processes Data Backup and Data Replication Concepts Exercise: Data Backup and Data Replication Is data backup different from data replication? Backup = Replication = Write your answers on blank lines or send your answers via the CHAT Example: Data Backup Configuration Back-up over LAN is the simplest solution. The back-up server pulls data from production servers and then it sends data to a tape library or NAS device. Data Backup Models Full Backup: Data is stored in exact copies Incremental Backup: Only the data that was changed or added since last time backup is recorded Differential Backup: Only the data that differs from the initial full backup is recorded Reverse Delta Backup: At every scheduled backup, the initial full backup image is synchronized so that it mirrors the current state of data on servers Backup Requirements To back-up the data from production servers you need: A back-up device – in enterprise environment it will most probably be a tape library, but it can also be a storage system with a LUN dedicated for back-up. A back-up server – in most cases you need a back-up server that is communicating directly with the back-up device and that controls back-up from all the servers Back-up software – software that runs on a back-up server and that allows to make configuration according to your needs Back-up agents – these agents are small applications installed on all your production servers. They are part of back-up software and they allow communication between a back-up server and production servers. Backup Optimization Techniques used to achieve better backup utilization: Compression – the output archive file is smaller than total of the original files Deduplication – the technique that eliminates duplicities in data Multiplexing – the ability of software and equipment to back-up data from several sources simultaneously Staging – back up to a disk first and then transfer the data from this disk to tape – known as Disk-to-disk-to-tape (D2D2T) Data Replication Overview A volume with source data is called a Primary Volume (P-VOL), and a volume to which the data is copied is a Secondary Volume (S-VOL). In-system – all operations with logical units (LUs) within the same storage system Remote – all operations with LUs across different storage systems P-VOL S-VOL Data Replication Overview Data Replication Within and between array heterogeneous replication Data replication (or protection) provides operational and disaster recovery • Replicates data within or between storage systems without disruption • Creates multiple protected copies from each source volume • Can run independent of host OS, database, file system • Mirrors image of data • Offers quick restart and recovery in disaster situations Once created, copies can be used for: • Data warehousing or data mining applications • Backup and recovery • Application development Exercise: Data Backup Or Data Replication? In the event of a major disaster, would you use your backup solution or data replication solution to recover your business critical online applications? • Break up into teams, discuss, and present your reasons for choosing one of the solutions over the other. If a vILT class, present your answers over the phone Data Replication – Copy Operations Data copy operations: • Initial Copy All data is copied from P-VOL to S-VOL Copies everything including empty blocks • Update Copy Only differentials are copied Requirements for All Replication Products Any volumes involved in replication operations (source and destination) should be: • Same size (in blocks) • Must be mapped to a port Source can be online and in use. Destination must not be in use/mounted. Intermix of RAID levels and drive type is supported. Licensing • License is capacity independent. Data Replication Operations – Establish Pairs Primary Server Backup Server Create Pair Status change: SIMPLEX to COPY to PAIR Updates P-VOLs S-VOLs Pairs S-VOLs are inaccessible after the paircreate command is issued. Paircreate Data Replication Operations – Split Pairs Primary Server Backup Server Pairsplit Status change: PAIR to PSUS Updates are marked in differential bitmaps Backup S-VOL backup data to tape P-VOL S-VOL S-VOL now accessible. Updates are marked in S-VOL differential bitmaps Data Replication Operations – Resynchronizing Pairs Primary Server Backup Server Online Resynchronize Pair Updates P-VOL S-VOL Caution: Any changes applied to S-VOL while the pair is split are discarded during NORMAL resync. PAIRRESYNC Normal Resync Reverse Resync Status: PSUS to PAIR Data Replication Operations – Quick Restore Primary Server Backup Server Quick Restore Pair Swaps LDEV Mapping Status: Split to PAIR with copy direction reversed Updates S-VOL P-VOL DUPLEX Quick Restore In-System Replication In-System Replication In-system hardware-based copy facility that provides: • Full volume point-in-time copies • Host Independent Storage System • Nondisruptive replication • Clone images are RAID protected copies Production Data LU # 1 (P-VOL) Clone Image (S-VOL) LU # 2 In-System Replication: Copy-On-Write Snapshots Snapshots • Create a point-in-time (PiT) copy or “snapshot” of the data • Uses less space than full copies or clones Frequent, cost effective, point-in-time copies • Multiple copies of a primary volume • Immediate read/write access to virtual copy • Fast restore from any virtual copy Copy on write snapshot. Notice that both P-VOL and V-VOL are accessible for I/O operations. Snapshots can be created instantly In-System Replication: Copy-On-Write Snapshots Copy-on Write virtual volume (V-VOL) maintains a view of the primary volume (P-VOL) at a particular point in time V-VOL is a composite of original data in the P-VOL and change data in the pool V-VOL presents as a full volume copy to any secondary host Since V-VOL does not copy all data, it can be created or deleted almost instantly Remote Replication Remote Replication Scheme Remote replication scheme; DWDM, ATM or IP connections to remote site are possible. Primary Host DWDM / ATM / IP Any Distance Fibre Channel Extender P-VOL Extender Fibre Channel S-VOL Synchronous Remote Replication The remote I/O is not posted “complete” to the application until it is written to a remote system The remote copy is always a “mirror” image Provides fast recovery with no data loss Limited distance – response-time impact 4 1 2 P-VOL 3 S-VOL Synchronous Replication Provides a remote “mirror” of any customer data • The remote copy is always identical to the local copy • Allows very fast restart/recovery with no data loss • No dependence on host operating system, database, or file system • Distance limit is variable, but typically less than 100 kilometers • Impacts application response time • Distance depends on application read/write activity, network bandwidth, response-time tolerance and other factors Asynchronous Remote Replication • The local I/O is disconnected from the remote I/O • Very little impact to response time over any distance • Data integrity and update sequence maintained over any distance • Fast restart/recovery 2 1 Fibre Channel Extender P-VOL Extender 3 Fibre Channel S-VOL Dynamic Replication Appliance Dynamic Replication Appliance A possible implementation of a Replication Appliance. Data is collected from servers over LAN and then it is send to a storage system. Each server is running an agent that splits the data. Remote Replication and Geoclusters Geocluster interconnection scheme. Both sites (local and remote) are equipped with the same nodes. Data from the Disk Array A are synchronously replicated to Disk Array B over iFCP, FCIP or Dark Fiber technology, both SANs are interconnected. Servers in both locations are also interconnected, usually using TCP/IP protocol. . Three Data Center Multi-target Replication Three data center multi-target replication. The maximum possible data protection is ensured by using two remote sites for data replication. Diversity in Data Protection Requirements Solution Area of Cost, Performance and Distance Exercise: Replication Scenario Scenario: • Financial services business with two data centers 300 miles apart Your task: • Describe a disaster recovery strategy for this business using data replication. If a vILT class, use your drawing tools to show your configuration on the slide Module Summary Upon completion of this module, you should have learned to: • Describe basic concepts of business continuity and replication • Understand business impact analysis and risk assessment • Describe basic concepts of disaster recovery Storage Concepts Virtualization of Storage Systems Module Objectives Upon completion of this module, you should be able to: • Understand and explain virtualization concepts and its benefits • Explain the difference between fat and thin provisioning • Describe SAN virtualization concepts Module Topics Virtualization concepts, features, and benefits Different types of virtualization “Fat” and “Thin” provisioning concepts and features Virtualization Concepts What Is Virtualization? Definition (source: www.wikipedia.org) • Virtualization is the abstraction of computer resources. • Hides the physical characteristics of computing resources from the way in which other systems, applications, or end users interact with those resources • Single physical resources appear to function as multiple logical resources, or multiple physical resources appear as a single logical resource. Virtualization has moved “out of the box” and into the infrastructure, or cloud, and virtualization solutions are available at these layers: • Virtualization of applications • Virtualization of computers • Virtualization of networks • Virtualization of storage Server Virtualization The traditional architecture model requires one physical server per operating system and application. A virtualized server is able to run several virtual machines that all share the physical hardware. Virtualization Elements Elements of Virtualization Virtualization Areas Server Virtualization Server virtualization is increasingly used; it provides better utilization of server resources One-to-many virtualization – Makes one physical server look like many servers; allows multiple operating systems on one physical server Layers of Virtualization Users can access the application from s a virtual desktop. Applications can run on several virtual machines The server can be virtualized The host can access a virtualized volume VMware® Based Server Virtualization Hyper-V™ Based Server Virtualization The hypervisor virtualization layer is thin and optimized for direct access to hardware resources. Virtual machines in child partitions access the virtualization layer through the VMBus interface. Device drivers for virtual machines are loaded from the parent partition with the original instance of Windows 2008 Server. Virtual machine configuration and management are also done in the parent partition operating system. Virtualization with Blade Servers Blade servers are installed in a blade server chassis. The chassis is then placed in a standard rack. These blade servers offer logical partitioning, which is a highly sophisticated form of server virtualization. Storage System Virtualization Every storage system offers RAID functionality but additionally may offer other virtualization capabilities such as: Cache Partitions Virtual Ports Storage virtualization (of other storage systems) Thin provisioning and automatic tiering Virtualization Benefits Migration – VMs and LUNs can be easily transferred from one physical device to another. Backup – encapsulation simplifies backup of the VM Hardware platform independence – physical servers can be of different configurations Enhanced utilization – allows effective use of resources Lower power consumption Lower RPO and RTO Physical resources can be added without disruption Migration Virtualized Server Cluster Migration Virtualization – Thin Provisioning and Automated Tiering Comparison of Fat and Thin Provisioning To avoid future service interruptions, today it is common to overallocate storage by approximately 50% - 75%. Purchased, Allocated BUT UNUSED Actual DATA Thin Provisioning Parity groups are added to a thin provisioning pool. Virtual volumes are mapped to servers. Virtual volumes do not contain any actual data. Data is stored in the storage pool. Virtual volumes contain pointers that point to the location of data in the pool. Thin Provisioning Benefits • Increased physical disk striping for better performance • Reduces the need for performance expertise • Simplifies storage capacity planning and administration • Increases storage utilization • Eliminates downtime for application storage capacity expansion • Improves application uptime and SLAs Zero Page Reclaim The storage system scans the storage pool for used data blocks that contain only zeros. These blocks are then erased and freed automatically. Thin Provisioning An example of how thin provisioning can help you save the cost of buying all the capacity in advance. Exercise: Virtualization 1. Definition of Virtualization • Virtualization is the ____________ of computer resources. • Hides the __________________ of computing resources from the way in which other systems, applications, or end users interact with those resources. • Single physical resources appear to function as ____________________ . • Or multiple physical resources appear as _____________________. 2. Identify some of the common objects that can be virtualized: • ___________________________ • ___________________________ If a vILT class, write your answers on blank lines Virtualization in Storage System Controller Virtualization in Storage System Controller Controller based virtualization of external storage. Hitachi Virtual Storage Platform (VSP) is an example of an enterprise-level storage system that supports virtualization of external storage. Before virtualization the data center consists of various heterogeneous storage systems in a SAN. Virtualization in Storage System Controller After virtualization the VSP storage system provides access to virtualized volumes from the other storage systems Virtualization – Logical Partitioning Storage Virtualization – Logical Partitioning In this figure we see one storage system (VSP) with three external storage systems that create a virtualized storage pool. The VSP is then virtualized to Partition 1 provide two logical partitions — private virtual storage machines. Hosts are then able to access and use only the resources (cache, ports and disks) assigned to the respective partition. FC/IP SAN Exercise: Benefits From your point of view, what do you think are the benefits of partitioning storage? If a vILT class, send your answers via the CHAT or over the phone Dynamic (Automated) Tiering All the benefits of Dynamic Provisioning Dynamic Tiering Dynamic • Provisioning • • Further simplified management Further reduced OPEX Better Performance Dynamic Tiering in Hitachi Virtual Storage Platform The Future of Virtualization Storage systems are heading towards fully virtualized solutions. Market Adoption Cycles Direct Attached Storage Network Storage Cloud or Virtualized Storage Cloud Storage “Cloud is a way of using technology, not a technology in itself – it is a self-service, on-demand pay-per-use model. Consolidation, virtualization and automation strategies will be the catalysts behind cloud adoption.” – The 451 Group Key characteristics of the cloud are: The ability to scale and provision dynamically in a cost efficient way The ability to make the most of new and existing infrastructure without having to manage the complexity of the underlying technology Cloud architecture can be: • Private: Hosted within an organization’s firewall • Public: Hosted on the internet • Hybrid: A combination of private and public Module Summary Upon completion of this module, you should have learned to : • Understand and explain virtualization concepts and its benefits • Explain the difference between fat and thin provisioning • Describe SAN virtualization concepts Storage Concepts Archiving and File and Content Management Module Objectives Upon completion of this module, you should be able to: • Explain the basic concepts and features of archiving • Describe what is meant by Fixed Content • Understand the differences between archiving and backup Module Topics Fixed Content and its characteristics Components of a digital archive Content management Introduction to Archiving Fixed Content What is Fixed Content? • Data objects that have a long-term value, do not change over time, and are easily accessible and secure Legal Records Satellite Images Biotechnology Email Digital Video Medical Records Motivation for Data Archiving There are several reasons for implementing data archiving policies and technical solutions. Some of the most prominent reasons are: • • • • • Effective utilization of high performance tiers Cheaper storage for fixed content Data retention regulation Simplified content management Indexing and searching capabilities of a digital archive Archive Solutions? Seeking an archiving solution. The storage systems we have discussed up until now are not very suitable for fixed content storage and archiving. Legal Requirements of Data Retention Periods Many organizations have increasing regulations, especially in the pharmaceutical industry, the food processing industry, healthcare, financial services and auditing. The Sarbanes-Oxley Act very strictly regulates the length of retention of financial records and accounting in companies. The Need for a Better Archiving Solution An example of a decentralized and fragmented archiving solution. Disparate storage systems do not provide a common search engine, and they are not very scalable. A digital archive can solve this problem. A Digital Archive A digital archive works on the object level. Each object contains fixed content data, metadata and description of policies. Block Level Storage Compared to Object Level Storage A traditional block level storage system compared to an object level storage system. The object level storage system consists of powerful proprietary servers and management software. These servers are connected to a RAID array. Internal Object Representation A data object and its components in detail. This example illustrates how objects are handled by a object storage’s digital archive. Digital Archive Features Active functions of a digital archive are: • Content verification – Ensures authenticity and integrity of each data object • Protection service – Ensures stability of the digital archive • Compression service – Achieves better utilization of storage space assigned to the digital archive • Deduplication service – Detects and removes duplicities • Replication service – Ensures redundancy of archived data • Search capabilities – Allows users to search documents Digital Archive Accessibility In this example, a digital archive can be accessed using multiple independent standard protocols. WebDAV is an extension to HTTP protocol that allows remote management of files stored on web server Digital Archive Compliance Features The most important compliance features of a digital archive are: Write once read many (WORM) Retention period definition Data shredding Data encryption Exercise: Fixed Content Which 2 statements are true about the definition of fixed content? (Choose 2) Fixed content is … a. Content that cannot be archived and restored b. Content that can only be changed by the system administrator c. Static data that is in a final state d. Content that will not / cannot change Answers= and If vILT class, write your answers on blank lines, Exercise: Backup versus Archive Explain the differences between what is meant by Backup and Archive. If vILT class, Send your answers to the instructor via the WebEx CHAT tool or phone. Exercise: Object Representation Fixed-content data (Data) 1. Describe what each is. System metadata 2. What does an object contain? Send your answers using CHAT Custom metadata If vILT class, write your answers on blank lines Module Summary Upon completion of this module, you should have learned to: • Explain the basic concepts and features of archiving • Describe what is meant by Fixed Content • Understand the differences between archiving and backup Storage Concepts Storage System Administration Module Objectives Upon completion of this module, you should be able to: • Describe everyday storage administrator tasks • Explain how to configure and monitor storage systems • Describe tools used by the storage administrator in managing storage Module Topics Storage system administrator tasks and common functions Storage system management software Storage system implementer tasks Storage Administrator Who is a Storage System Administrator? In charge of maintaining a storage system infrastructure • Tasks based on Service-level Agreements (SLAs) Storage Administrator Tasks Capacity Management Amount of data to be stored; size and performance of LUNs; hard drive performance; I/O performance and R/W operations Availability Management Replication, backup, and archive strategies; protection against component failures Continuity Management Part of business continuity planning and disaster recovery procedures Financial Management Budget preparation; cost calculation and invoicing and TCO Storage Administrator Tasks – Other Common Operations Configuration of RAID groups and volumes Implementation of changes in volume configuration Data replication optimization Configuration of cache Cache partitioning Backup of storage system configuration Integration of a New Storage System When purchasing a new storage system, the storage administrator must think through the whole implementation process, including the following items: Storage system model Switch model Cabling Rack usage and floor space Power requirements Air conditioning LAN infrastructure Tasks of a Storage System Implementer Installation and initial configuration of the storage system Basic training to familiarize the customer with the new device. Conduct all hardware replacement and upgrade procedures Monitor the storage system remotely Help with performance tuning Microcode updates Module Summary Upon completion of this module, you should have learned to: • Describe everyday storage administrator tasks • Explain how to configure and monitor storage systems • Describe tools used by the storage administrator in managing storage Storage Concepts Business Challenges Module Objectives Upon completion of this module, you should be able to: • Identify business challenges driving the need for storage • Understand why storage systems are important for business • Explain the energy and green issues faced by today’s businesses Module Topics Business challenges companies face Advanced data classification and tiered storage Data center operations environmental concerns Business Challenges Business Challenges Business challenges can be classified as follows: Accelerating storage growth — need more and more capacity Increasing requirements on high availability — cannot afford any disruptions because our data has become too important for our business Fast and effective response to business growth — need to be able to react quickly to new conditions Heterogeneous infrastructure — result of fast infrastructure growth, which was not properly planned, causing TCO to increase rapidly Compliance and security challenges — need to process, protect and retain data according to legal requirements and regulations Power and cooling challenges — pay too much for electricity and cooling and may be running out of space in the server room Data center challenges — specific needs for those whose business is focused primarily on cloud type provision of storage and computing capacity Data Growth Forecast An overview of the storage requirements in the past years provide you with the necessary information to forecast data growth. Data Growth Forecast – Tier View Data growth forecast in relation to performance tiers Structured, Unstructured and Replicated Data Growth Advanced Data Classification and Tiering Tiered storage infrastructure Power Requirements and Cooling Greatest issues organizations and companies face What is the greatest facility problem with your primary data center? 29% Power and cooling exceeds server Spending Spending (US$B) $80 29% $70 21% 13% 6% Installed base (M units) 18 New server spending Power and cooling 12 $50 10 8 $30 6 $20 Excessive Heat Insufficient Power Insufficient Raised Floor None of the above Poor Location N = 112 Gartner, Best Practices in Data Center Facilities Excess Facility Cost 14 $60 $40 3% 16 4 $10 2 $0 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 Power Challenges Facing Data Centers Some of the challenges facing data centers with respect to electricity, cooling and environmental requirements include: Running out of power, cooling and space Growing energy costs Increasing regulatory compliance issues Data center expansion without consideration for future power and cooling requirements Data storage configured without adequate consideration to heat distribution (equipment racks should be installed with cold rows and hot rows Difficulty relieving data center hot spots without disrupting applications Other Challenges Facing Data Centers In addition to power consumption metrics in kW, there are other metrics that should be considered: Total five-year power and space costs Heat loading (kW/sq ft) Space requirements (sq ft) Floor loading (lbs/sq ft) Controller-based virtualization and thin provisioning can also yield substantial environmental advantages because they reduce the need for storage capacity. HDD and Fan Power Savings Features that work on the HDD and fan level and lead to significant power savings: Spin down drives in selected RAID groups SATA drives will park heads when idle for more than 2 hours Adjust fan speeds to maintain correct temperatures Keep data in cache as long as possible Green Data Center Hot and Cold Rows Arrange racks in alternating rows with cold air intakes facing one way and hot air exhausts facing the other Module Summary Upon completion of this module, you should have learned to: • Identify business challenges driving the need for storage • Understand why storage systems are important for business • Explain the energy and green issues faced by today’s businesses Storage Concepts Storage Networking and Security Module Objectives Upon completion of this module, you should be able to: • Describe basic networking concepts • Explain how common network devices operate • Explain how devices communicate in a network • Explain storage area network security Module Topics • Basic networking concepts • Operations of common network devices • Possibilities we have in storage system networking • How devices communicate with each other through the network • How to secure Storage Area Networks Introduction to Networks – Components Twisted Pair Cable Twisted Pair Cable structure, RJ45 connectors Fiber Optic Cable Fiber Optic Cables – Fiber optics is a technology that uses glass or plastic fibers to transmit data as light impulses. A fiber optic cable consists of a bundle of fibers and each fiber can transmit millions of messages modulated onto light waves. Fiber Optic Connectors Fiber Optic Cable LC connector and SFP transceiver Storage Network Components – Nodes Network nodes are all the devices connected in the network. We distinguish between endpoint communication nodes and data redistribution nodes. Storage Network Components – Ports Ports On a storage network, a port enables a node to communicate with another node over a Fibre Channel connection. • A node can contain multiple ports. On a storage network, a port enables the following connections: • Server to switch • Switch to switch • Switch to storage Storage Network Components – HBAs Host Bus Adapter (HBA) – In a storage system, an HBA is a Fibre Channel interface card installed in a server. It connects a computer and storage devices on a network. Each HBA has a unique WWN. The two types of WWNs on an HBA are these: • Node WWN: Shared by all ports on an HBA • Port WWN: Unique to each port on the HBA Storage Network Components – WWNs WWN address, World Wide Name, is a unique label, which identifies a particular device in Fiber Channel network WWN example - 5 0 0 6 0 E 8 0 1 0 4 5 3 0 3 0 1 6 The Open Systems Interconnection (OSI) Model The Open Systems Interconnection (OSI) model is a conceptual model that characterizes and standardizes the internal functions of a communications system by partitioning it into abstraction layers. Storage Network Components – HUB Hub – a simple device that allows interconnection and communication among nodes. Storage Network Components – Fibre Channel Switches Switch – also allows interconnection and communication among nodes within one network Storage Network Components – Routers Router – provides an interface between two different networks Storage Network Components – Directors Directors A director is a large and complex switch. It is: • Highly available, reliable, scalable, and manageable • Fault tolerant with the ability to recover from a non-fatal error • Designed with redundant hardware components • Capable of supporting Fibre Channel and fiber connectivity (FICON) • Potentially expensive and complex • Designed for enterprises with large data centers Large networks often use Fibre Channel switches and directors in the same implementation. Exercise: Storage Network Components Match the storage network component with the appropriate description: 1. Connects and transmits signals between nodes a. Node 2. Transmits or receives data over a network b. Port 3. Fibre Channel interface card c. WWN d. HBA e. Cable 4. Enables a node to communicate with another node 5. Unique number used to identify elements on a FC storage network If vILT class, write your answers on blank lines Storage Networking Topologies SAN Topologies Point-to-Point (FC-P2P) – A point-to-point (P-P) topology is considered the simplest topology, in which two devices are directly connected using Fibre Channel. • It has fixed bandwidth; data is transmitted serially over a single cable. • It can be used with DAS. SAN Topologies Arbitrated Loop (AL) – An FC topology where all devices are part of a loop and only one device can communicate with another device at a time In AL, devices use an access request mechanism called arbitrate (ARB), which circles the loop. • A device can use ARB depending on its priority and access rights. • The device with the highest priority gets first access. SAN Topologies Switched Fabric (FC-SW) – A Fibre Channel topology that connects multiple devices by using Fibre Channel switches In a switched fabric topology, bandwidth is not shared between devices, enabling devices to transmit and receive data at full speed at all times. Direct Attached Storage Direct attached storage infrastructure. Server is directly connected to a storage system. Storage system can be accessed only through the server. Server can be accessed from Local Area Network (LAN). Storage Area Network (SAN) Storage Area Network Storage Area Network (SAN) is a high-speed network of shared storage devices. Servers attached to a SAN can access any SAN attached storage devices. The only components in a SAN are storage devices and switches. SAN is designed to connect computer storage devices, such as disk array controllers and tape libraries, to multiple servers, or hosts. SAN Components Storage Area Networks are using Fiber Channel infrastructure, which includes Host Bus Adaptors installed in servers, Fiber Channel cables and switches, Fiber Channel ports installed in the storage system front end and proprietary network protocols. Host Bus Adaptor (HBA) and an example of WWN number. SAN over iSCSI Interface Internet Small Computer Systems Interface (iSCSI) – SAN can be implemented by using a network protocol standard called iSCSI which uses the SCSI protocol to transmit data over TCP/IP networks. iSCSI allows organizations to use their existing TCP/IP network infrastructure without investing in expensive Fibre Channel switches. iSCSI HBA Network Attached Storage Network Attached Storage Network Attached Storage (NAS) is represented by the server that functions as the NAS Head and common storage system. There are solutions that integrate both these functionalities in one package (NAS Appliances). NAS devices work relatively independently; they do not require servers with applications. All clients, application and other servers can access files stored in a NAS device. File Access Protocols supported include: • CIFS • NFS, • HTTP • HTTPS • FTP Network Attached Storage In NAS, the storage device is directly connected to the LAN and there is no server between the data and other network devices. Data is presented to the server at file level. LAN Network Attached Storage Advantages of NAS • Offers storage to different open-systems operating systems over LAN • Data is presented to servers at file level, reducing server overhead • Dedicated file server, optimized for sharing files between many users • Minimizes overhead by centrally managing storage • Facilitates easy and inexpensive implementation Disadvantages of NAS • Relies on the client-server model for communication and data transport which creates network overhead • Lower performance than a SAN NAS Implementations Methods of NAS Implementation – An organization can implement NAS architecture by following methods: • NAS appliance or filer • NAS blade • NAS gateway NAS Appliance NAS Appliance – Combines a front-end file server and back-end storage system in a single unit. This approach is called a closed-box approach. NAS appliance has the following advantages: • Combines a file server with the storage array • Provides efficient performance • Has high reliability • Enables easy installation, management, and use • Provides the least expensive NAS implementation NAS appliance has the following disadvantages: • Is not scalable • No pool storage, which makes it hard to achieve high utilization. NAS Blade NAS Blade – Allows multi-protocol data storage in a large disk array NAS blade has following advantages in addition to a NAS appliance: • Is scalable • Provides backup of storage data • Supports multiple NAS blades NAS Gateway NAS Gateway – All devices communicate directly with the file system. A NAS gateway overcomes the limitations of a NAS appliance. It has the following advantages: • Separates file server from storage device • Is less expensive than a NAS appliance • Supports multiple NAS gateways • Has better utilization rates • Combines NAS with SAN capacity to meet growing storage requirements • Provides NAS functionalities to SAN storage NAS gateway controller uses FC protocol to connect to external storage. Converged Solution – SAN and NAS Together Converged solution – SAN and NAS together • NAS head with storage over the SAN • NAS scales to the limits of the SAN Limited by NAS file system’s capacity Co-exists with application servers Users Centrally managed LAN NAS Gateway Hosts SAN Raid Storage Storage Networking Architectures…Side by Side DAS NAS SAN Application Application IP Network SAN File System Direct Connected File System Exercise: Storage Networking Concepts 1. A DAS device is not shared, so no other network device can access the data without first accessing the server. True or False? 2. Select the best description for the Fibre Channel topology known as FC-AL. a) Two devices, data transmitted serially over a single cable b) Multiple devices connected in a loop, highest priority device gets first access c) Multiple devices connected using Fibre Channel switch, devices transmit and receive data at full speed at all times Network Protocols Protocol Protocol – A set of rules that govern communication between computers on a network. It regulates the following characteristics of a network: • • • • Access method Physical topologies allowed in the network Types of cable that can be used in the network Speed of data transfer Protocols The different types of protocols that can be used in a network are: • Ethernet • Fibre Channel protocol (FCP) • Fiber connectivity (FICON) • Internet protocol (IP) • Internet small computer system interface (iSCSI) • Fibre Channel over IP (FCIP) • Internet Fibre Channel protocol (iFCP) • Fibre Channel over Ethernet (FCoE) Protocols Ethernet Uses an access method called carrier sense multiple access/collision detection (CSM/CD). Before transmitting, a node checks whether any other node is using the network. If clear, the node begins to transmit. Ethernet allows data transmission over twisted pair or fiber optic cables and is mainly used in LANs. There are various versions of Ethernet with various speed specifications. FCP Defines a multi-layered architecture for moving data. FCP packages SCSI commands into Fibre Channel frames ready for transmission. FCP also allows data transmission over twisted pair and over fiber optic cables. It is mainly used in large data centers for applications requiring high availability, such as transaction processing and databases. FICON Connects a mainframe to its peripheral devices and disk array. Ficon is based on FCP and has evolved from the older ESCON protocol. Protocols IP/TCP IP is used to transfer data across a network. Each device on the network has a unique IP address that identifies it. IP works in conjunction with the TCP, iSCSI and FCIP protocols. When you transfer messages over a network by using IP, IP breaks the message into smaller units called packets (third layer in OSI model). Each packet is treated as an individual unit. IP delivers the packets to the destination. TCP Is the protocol that combines the packets into the correct order to reform the message that was sent from the source. iSCSI Establishes and manages connection between IP-based storage devices, and it hosts and enables deployment of IP-based storage area networks. It facilitates data transfers over intranets, manages storage over long distances and is cost-effective, robust and reliable. iSCSI is best-suited for web server, email and departmental business applications in small to medium sized businesses. Protocols FCIP Fibre Channel over IP is a TCP/IP based tunnelling protocol that connects geographically distributed Fibre Channel SANs. FCIP encapsulates Fibre Channel frames into frames that comply with TCP/IP standards. It can be useful when connecting two SAN networks over the Internet tunnel, in a similar fashion to virtual private networks (VPNs) allowing connection to a distant LAN over the Internet. iFCP iFCP is again TCP/IP based. It is basically an adaptation of FCIP using routing instead of tunneling. It interconnects Fibre Channel storage devices or SANs by using an IP infrastructure. iFCP moves Fibre Channel data over IP networks by using iSCSI protocols. Both FCIP and iFCP provide means to extend Fibre Channel networks over distance. Both these protocols are highly reliable and scalable. They are best suited for connecting two data centers for centralized data management or disaster recovery. Protocols FCoE Fibre Channel over Ethernet is an encapsulation of Fibre Channel frames over Ethernet networks. This allows Fibre Channel to use 10Gb Ethernet networks while preserving the Fibre Channel protocol. FCoE provides these advantages: Network (IP) and storage (SAN) data traffic can be consolidated using a single network switch. It reduces the number of network interface cards required to connect disparate storage and IP networks. Reduces the number of cables and switches. Reduces power and cooling costs. Thus, you can build your SAN using Ethernet cables (mostly twisted pair). You can use one switch for your IP-based network traffic (LAN) and for creating SAN infrastructure. Even though the switch and cabling are the same, LAN will run on TCP/IP while SAN runs on FCP. Exercise: Storage Networking Match the following list of components with the appropriate definition: a. Client-server 1. Network connecting devices in a small geographic area b. Protocol 2. Relationship between two computers – one sends requests; one responds with data c. LAN d. WAN 3. Set of rules governing communication among computers on a network 4. Network connecting devices across larger geographical areas If vILT class, write your answers on blank lines Storage Area Network Security LUN Mapping A LUN is a logical device mapped to a storage port. Logical Devices Windows UNIX LUNS are mapped To servers Zoning Zones – Defined to establish rules governing communication of network devices • WWN Zoning (Soft Zoning) • Port Based Zoning (Hard Zoning) • Mixed Zoning An example of WWN based zoning Module Summary Upon completion of this module, you should have learned to: • Describe basic networking concepts • Explain how common network devices operate • Explain how devices communicate in a network • Explain storage area network security
Source Exif Data:
File Type : PDF File Type Extension : pdf MIME Type : application/pdf PDF Version : 1.5 Linearized : No Page Count : 262 Language : en-US Tagged PDF : Yes Title : Storage Concepts Author : Brenda Bateman Create Date : 2017:01:04 06:53:41+01:00 Modify Date : 2017:01:04 06:53:41+01:00 Producer : Microsoft PowerPoint 2010 Trial Creator : Microsoft PowerPoint 2010 TrialEXIF Metadata provided by EXIF.tools