You are on page 1of 78

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Foundations

2006 EMC Corporation. All rights reserved.

Welcome to NAS Foundations. The AUDIO portion of this course is supplemental to the material and is not a replacement for the student notes accompanying this course. EMC recommends downloading the Student Resource Guide from the Supporting Materials tab, and reading the notes in their entirety.
These materials may not be copied without EMC's written consent. EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED AS IS. EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Use, copying, and distribution of any EMC software described in this publication requires an applicable software license. EMC is a registered trademark of EMC Corporation. All other trademarks used herein are the property of their respective owners.

NAS Foundations - 1

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Foundations
Upon completion of this course, you will be able to: Identify the concepts and value of Network Attached Storage List Environmental Aspects of NAS Identify EMC NAS Platforms and their differences Identify and describe Celerra Software Features Identify and describe Celerra Management Software offerings Identify and describe Windows Specific Options with respect to EMC NAS environments Identify and describe NAS Business Continuity Options with respect to the various EMC NAS platforms

2006 EMC Corporation. All rights reserved.

NAS Foundations - 2

The objectives for this course are shown here. Please take a moment to read them.

NAS Foundations - 2

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Network Attached Storage


Identify what constitutes a NAS environment

2006 EMC Corporation. All rights reserved.

NAS Foundations - 3

NAS environment components are reviewed in this section.

NAS Foundations - 3

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

What Is Network-Attached Storage


Built on the concept of shared storage on a Local Area Network Leverages the benefits of a network file server and network storage Utilizes industrystandard network and file sharing protocols
File Server + Network-Attached Storage = NAS
2006 EMC Corporation. All rights reserved. NAS Foundations - 4

Application

Application

Application

Unix Client

Windows Client

Unix Client

Network

The benefit of NAS is that it now brings the advantages of networked storage to the desktop through file-level sharing of data via a dedicated device. NAS is network-centric and typically used for client storage consolidation on a variety of network topologies such as LANs (Local Area Network), MANs (Metropolitan Area Network), WANs (Wide Area Network), etc. NAS is a preferred storage capacity solution for enabling clients with unregulated access to files quickly and directly via purpose built data sharing equipment. This eliminates several bottlenecks users often encounter when accessing files from a general-purpose servers. In addition, NAS can serve UNIX and Microsoft Windows users seamlessly, sharing the same data between the different architectures. NAS provides security and performs all file and storage services through standard network protocols: TCP/IP for data transfer Ethernet and Gigabit Ethernet for media access CIFS, http, ftp, and NFS for remote file service

NAS Foundations - 4

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Why NAS?
Highest availability Scales for growth Avoids file replication
S1

Firewall

Internal Network

Internet
S2
. . . .

NAS

Increases flexibility Reduces complexity Improves security Reduces Costs


2006 EMC Corporation. All rights reserved.

Sn Web Servers

Data Center

NAS Foundations - 5

Through the advent of NAS applications that use file system level access, the data can now be shared to large numbers of users, that may be geographically dispersed, simultaneously. Therefore many users can now take advantage of the availability and scalability of networked storage. Centralizing file storage can reduce system complexity and system administration costs, along with simplifying backup, restore, and disaster recovery solutions. Although NAS trades some performance for manageability and simplicity, it is by no means a lazy technology. Gigabit Ethernet allows NAS to scale to high performance and low latency, making it possible to support a myriad of clients through a single interface. Many NAS devices support multiple interfaces and can support multiple networks at the same time.

NAS Foundations - 5

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Operations
Traditional IO operations use file level IO protocols File system is mounted remotely using a network file access protocol, such as:
Network File System (NFS) for Unix Common Internet File System(CIFS) for Microsoft Windows
NAS
Application

IP Network

IO is redirected to remote system Utilizes mature data transport (e.g., TCP/IP) and media access protocols NAS device assumes responsibility for organizing data (R/W) on disk and managing cache
2006 EMC Corporation. All rights reserved.

NAS Device Direct Attach OR

SAN

Disk

NAS Foundations - 6

One of the key differences of a NAS disk device, compared to DAS or other networked storage solutions such as SAN, is that all traditional I/O operations use file level I/O protocols. File I/O is a high level type of request that, in essence, specifies only the file to be accessed, but does not directly address the storage device. The client file I/O is converted into block level I/O by the NAS device operating system to retrieve the actual data. Once the data has been retrieved it is once again converted back to file level I/O for return to the client. A file I/O specifies the file. It also indicates an offset into the file. For instance, the I/O may specify Go to byte 1000 in the file (as if the file were a set of contiguous bytes), and read the next 256 bytes beginning at that position. Unlike block I/O, there is no awareness of a disk volume or disk sector in a file I/O request. Inside the NAS appliance, the operating system keeps tracks of where files are located on disk. The OS issues a block I/O request to the disks to fulfill the file I/O read and write requests it receives. The disk resources can be directly attached to the NAS device or using a SAN, referred to as a gateway configuration. Block level IO support by NAS devices is discussed later in this module.

NAS Foundations - 6

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Architecture
To Storage Application Remote I/O request Operating System I/O Redirector NFS/CIFS

NFS and CIFS handle file requests to remote file system I/O is encapsulated by TCP/IP Stack to move over the network NAS device converts requests to block IO and reads or writes data to NAS disk storage

Storage Network Protocol (Fibre Channel) Drive Protocol (SCSI) NAS Operating System Network File Protocol Handler TCP/IP Stack

TCP/IP Stack Network Interface

Network Interface File I/O to NAS

2006 EMC Corporation. All rights reserved.

NAS Foundations - 7

The Network File System (NFS) protocol and Common Internet File System (CIFS) protocol handle file I/O requests to the remote file system, which is located in the NAS device storage. I/O requests are packaged by the initiator into the TCP/IP protocols to move across the IP network. The remote NAS file system converts the request to block I/O and reads or writes the data to the NAS disk storage. To return data to the requesting client application, the NAS appliance software re-packages the data to move it back across the network. Here we see an example of an IO being directed to the remote NAS device and the different protocols that play a part in moving the request back and forth to the remote file system located on the NAS server.

NAS Foundations - 7

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Device
Single-purpose machine or component, serves as a dedicated, high-performance, high-speed communication of data using both file level and block level IO Is sometimes called a filer or a network appliance Uses one or more Network Interface Cards (NICs) to connect to the customer network Uses proprietary optimized operating system; DART, Data Access in Real Time Uses industry standard storage protocols to connect to storage resources
2006 EMC Corporation. All rights reserved.

Client Application

IP Network

NAS Device
Network Drivers and Protocols NFS CIFS

NAS Device OS (DART) Storage Drivers and Protocols

Disk Storage

NAS Foundations - 8

A NAS server is not a general-purpose compute. NAS devices use a significantly streamlined/tuned OS in comparison to general purpose computer. It is sometimes called a filer because it focuses all of its processing power solely on file service and file storage. The NAS device is sometimes called a network appliance, referring to the plug and play design of many NAS devices. Common network interface cards (NICs) include gigabit Ethernet (1000 Mb/s) or Fast Ethernet (10Mb/s), ATM, and FIDDI. Most NAS devices also support NDMP (Network Data Management Protocol) for backup, Novell Netware, FTP and HTTP protocols. The NAS operating system for Network Appliance products is called Data ONTAP. The NAS operating system for EMC Celerra is DART - Data Access in Real Time. These operating systems are tuned to perform file operations including open, close, read, write, etc. The NAS device generally uses a standard drive protocol, some form of SCSI, to manage data to and from the disk resources.

NAS Foundations - 8

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Applications
CAD/CAM environments, where widely dispersed engineers have to share and modify design drawings Serving Web pages to thousands of workstations at the same time Easily sharing company-wide information among employees Database application
Low transaction rate Low data volatility Smaller in size Not performance constrained

2006 EMC Corporation. All rights reserved.

NAS Foundations - 9

Database applications have traditionally been implemented in a SAN architecture. The primary reason is the conclusive performance of a SAN. This characteristic is especially applicable for very large, online transactional applications with high transaction rates and high data volatility. However, NAS might be appropriate where the database transaction rate is low and performance is not constrained. Extensive application profiling should be done in order to understand the specific database application requirement and, if in fact, a NAS solution would be appropriate. When considering a NAS solution, the databases should: be sequentially accessed, non-indexed or have a flat file structure have a low transaction rate have low data volatility be relatively small not have performance / timing constraints require multiple dynamic path access to application servers

NAS Foundations - 9

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Environment
Identify components in a common networking environment

2006 EMC Corporation. All rights reserved.

NAS Foundations - 10

Key components of NAS and networking infrastructure are reviewed in this section.

NAS Foundations - 10

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Terminology
Ethernet: Local network protocol that uses coaxial or twisted pair cables Network Topology: Geometric arrangement of nodes and cable links in a LAN; used in two general configurations: bus and star Protocol: Defines how computers identify one another on a network, the form that the data should take in transit, and how this information is processed once it reaches its final destination IP Address: Unique number that identifies a computer to all other computers connected to the network
2006 EMC Corporation. All rights reserved. NAS Foundations - 11

Ethernet is a local-area network protocol that uses coaxial or twisted pair cables as a means for communication. Ethernet is popular because it strikes a good balance between speed, cost, and ease of installation. These benefits, combined with wide acceptance in the computer marketplace and the ability to support virtually all popular network protocols, make Ethernet an ideal networking technology for most computer users today. A network topology is the geometric arrangement of nodes and cable links in a LAN, and is used in two general configurations: bus and star. A protocol defines how computers identify one another on a network, the form that the data should take in transit, and how this information is processed once it reaches its final destination. TCP/IP is a common protocol used in sending information via the Internet. Protocols also define procedures for handling lost or damaged transmissions, or "packets. An Internet Protocol (IP) address is a four octet number in the commonly used IP version 4, for example 155.10.20.11, that uniquely identifies a computer to all other computers connected to the network.

NAS Foundations - 11

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

What is a Network?
LAN Physical Media WAN MAN Site 1

Site 2
2006 EMC Corporation. All rights reserved. NAS Foundations - 12

LAN A network is any collection of independent computers that communicate with one another over a shared network medium. LANs are networks usually confined to a geographic area, such as a single building or a college campus. LANs can be small, linking as few as three computers, but often linking hundreds of computers used by thousands of people. Physical Media An important part of designing and installing a network is selecting the appropriate medium. There are several types in use today: Ethernet, Fiber Distributed Data Interface (FDDI), Asynchronous Transfer Mode (ATM), and Token Ring. WAN Wide area networking combines multiple LANs that are geographically separate. Services such as dedicated leased phone lines, dial-up phone lines (both synchronous and asynchronous), satellite links, and data packet carrier services connect the different LANs. Wide area networking can be as simple as a modem and remote access server for employees to dial into, or it can be as complex as hundreds of branch offices globally linked using special routing protocols and filters to minimize the expense of sending data over vast distances. MAN Metropolitan area networking is a networking infrastructure size that falls in-between a LAN and a WAN. They are generally used to consolidate networking infrastructures in a campus sized, generally between five (5) and fifty (50) kilometers in diameter, area to provide sharing of localized resources. They typically use wireless or optical interconnections between localized sites within the MAN. The IEEE 802.6 standard specifies the unique way that the MAN can communicate between sites to minimize latency and congestion. This is known as a distributed queue dual-bus network, (DQDB), which utilizes a dual bus, distributed queuing.

NAS Foundations - 12

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Physical Components
Network Interface Card (NIC)
Switch NIC

Switches Routers
NIC Router

155.10.10.XX

NIC 155.10.20.XX Switch NIC


2006 EMC Corporation. All rights reserved. NAS Foundations - 13

Network Interface Card Network interface cards, commonly referred to as NICs, are used to connect a Host, Server, Workstation, PC, etc. to a network. The NIC provides a physical connection between the networking cable and the computer's internal bus. The rate at which data passes back and forth can be different.

Switches LAN switches can link multiple network connections together. Todays switches accept and analyze the entire packet of data to catch certain packet errors and keep them from propagating through the network before forwarding it to its destination. Each of the segments attached to an Ethernet switch has the full bandwidth of the switch 10Mb/100Mb/1Gigabit.

Routers Routers pass traffic between networks. Routers also divide networks logically instead of physically. An IP router can divide a network into various subnets so that only traffic destined for particular IP addresses can pass between segments.

NAS Foundations - 13

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Network Protocols

Network transport Protocols


Universal Datagram Protocol (UDP) for non-connection oriented networks Transmission Control Protocol (TCP) for connection oriented networks

Network filesystem Protocols


NFS to manage files in a networked Unix environment CIFS to manage files in a networked Windows environment

2006 EMC Corporation. All rights reserved.

NAS Foundations - 14

Network transport protocols are standards that allow computers to communicate. They are used to manage the movement of data packets to devices communicating across the network. UDP and TCP are examples of transport protocol. In a non-connection oriented communication model, the data is sent out to a recipient using a best effort approach with no acknowledgement of the receipt of the data being sent back to the originator by the recipient. Error correction and resend must be controlled by a higher layer application to ensure data integrity. In a connection oriented model, all data packets sent by an originator are acknowledged by the recipient and transmission errors / lost data packets are managed at the protocol layer. TCP/IP (for UNIX, Windows NT, Windows 95 and other platforms), IPX (for Novell NetWare), DECnet (for networking Digital Equipment Corp. computers), AppleTalk (for Macintosh computers), and NetBIOS/NetBEUI (for LAN Manager and Windows NT networks) are examples of network transport protocols in use today. Network filesystem protocols are used to manage how data requests are processed once it reaches its final destination. Both NFS and CIFS support UDP and TCP transport protocols. Network block level protocols are discussed later in this presentation.

NAS Foundations - 14

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Network Addressing
155.10.10. 14 Host name Mary 155.10.10.13 Host Name Peter

IP Addressing DHCP DNS


DNS Server

DHCP Server 155.10.10.XX

155.10.10.11

155.10.10.12 Router 155.10.20.XX

Switch

Host Name = Account1


2006 EMC Corporation. All rights reserved.

155.10.20.11

NAS Foundations - 15

Several things must happen in order for computers to be able to communicate data across the network. First, the computer must have a unique network address, referred to as the IP Address. An address can be assigned in one of two ways; dynamically or statically. A static address requires entering the IP address that the computer uses in a local file. However, if two computers on the same subnet are assigned the same IP address, they would not be able to communicate. Another approach is to set up a computer on the network to dynamically assign an IP address to a host when it joins the network. This is called the Dynamic Host Configuration Protocol (DHCP Server). In our example, the host Mary is assigned an IP address 155.10.10.14, and the host Peter is assigned an IP address 155.10.10.13 by the DHCP server. The NAS device, Account1, is a File server. Servers normally have a statically assigned IP address. In this example, it has the IP address 155.10.20.11. A second requirement for communications is to know the address of the recipient of the communication. The more common approach is to communicate by name, for example, the name you place on a letter. However, the network uses numerical addresses. A more efficient solution is the Domain Name Service (DNS). The DNS is a hierarchical database, which resolves host names to IP addresses. In our example, if someone on host Mary wants to talk to host Peter, it is the DNS server that resolves Peter to 155.10.20.13.

NAS Foundations - 15

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Volume and Files


155.10.10. 14 Host name Mary 155.10.10.13 Host Name Peter

Create Volumes Volume Create Network Filesystem

DNS Server

DHCP Server

155.10.10.11

155.10.10.12 Router NAS Account1

155.10.20.11 File System


2006 EMC Corporation. All rights reserved.

Array

/Acct_ Rep
NAS Foundations - 16

Create Array Volume The first step in a network attached storage environment is to create logical volumes on the array and assign it a LUN Identifier. The LUN is then presented to the NAS device. Create NAS Volume The NAS device performs a discovery operation when it first starts or when directed. In the discovery operation, the NAS device sees the array LUN as a physical drive. The next task is to create logical volumes at the NAS device level. The Celerra creates meta volumes using the volume resources presented by the array. Create Network File When the logical volumes are created on the Celerra, it can use them to create a file system. In this example, we have created a file system /Acct_Rep on the NAS server Account1. Mount File System Once the file system has been created, it must be mounted. With the file system mounted, we can then move to the next step, which is publishing the file system on the network.

NAS Foundations - 16

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Publish
155.10.10. 14 Host name Mary User Mary MS Windows Share

Group Name = SALES 155.10.10. 13 Host name Peter User Peter Unix Export DHCP Server

Export Share

DNS Server

155.10.10.11

155.10.10.12 Router

Group Name = Accounting NAS

155.10.20.11 ACCOUNT1 /Acct_ Rep


2006 EMC Corporation. All rights reserved.

Array

NAS Foundations - 17

Now that a network file system has been created and mounted, there are two ways it can be accessed using the network. The first method is through the UNIX environment using NFS. This is accomplished by performing an Export. The Export publishes to those UNIX clients who can mount (access) the remote file system. Access permissions are assigned when the export is published. The second method is through the Windows environment using CIFS. This is accomplished by publishing a share. The share publishes to those Windows clients who map a drive to access the remote file system. Access permission are assigned when the share is published. In our example, we may only allow Mary and Peter, who are in the Sales organization, share or export access. At this level, NFS and CIFS are performing the same function but are used in different environments. All members of the Group SALES, which include the users Mary and Peter, are granted access to /Acct_Rep.

NAS Foundations - 17

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Client Access
155.10.10. 14 Host name Mary User Mary MS Windows MAP

Group Name = SALES 155.10.10. 13 Host name Peter User Peter Unix DNS Server DHCP Server nfsmount

Mount MAP

155.10.10.11

155.10.10.12 Router

Group Name = Accounting NAS

155.10.20.11 ACCOUNT1 /Acct_ Rep


2006 EMC Corporation. All rights reserved.

Array

NAS Foundations - 18

To access the network file system, the client must mount a directory or map a drive pointing to the remote file system. Mount is a UNIX command performed by a UNIX client to set a local directory pointer to the remote file system. The mount command uses NFS protocol to mount the export locally. For a UNIX client to perform this task, it executes the nfsmount command. The format for the command is: nfsmount /name of the NAS server:name of the remote file system/name of the local directory

For example: nfsmount/Account1:Acct_Rep /localAcct_Rep.

For a Windows client to perform this task, it executes a map network drive. The sequence is my computer> tools>map network drive. Select the drive letter and provide the server name and share name in the Folder field. For example: G: \\Account1\Acct_Rep If you make a comparison, the same information is provided: the local drive (Windows) or the local directory and the name of the NAS server and the name of the export or the share.

NAS Foundations - 18

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

File Permissions
155.10.10. 14 Host name Mary User Mary MS Windows

Group Name = SALES 155.10.10. 13 Host name Peter User Peter Unix

Create File File Request

DNS Server

DHCP Server

155.10.10.11

155.10.10.12 Router

Group Name = Accounting NAS

Account1 /Acct_ Rep Files


2006 EMC Corporation. All rights reserved.

155.10.20.11 PRPT2

Array

MRPT1

NAS Foundations - 19

Create file Once access is gained by the client, files can be created on the remote file system. When a file is created by a client, normal permission is assigned. The Client can also modify the original permissions assigned to a file. File permission is changed in UNIX using the chmod command. File permission in Windows is changed through right clicking on the selected file, then selecting Properties> Security add or remove group add or remove permissions. It should be noted that in order to modify the file permissions, one must have the permission to make the change. File request If a request for a file is received by the NAS server, the NAS server first authenticates the user locally or over the network. If the user identity is confirmed, then the user is allowed to perform operations contained in the file permissions of the users Group. In our example, user Mary on host Mary creates a file MRPT1 on the NAS server Accout1. She assigns herself the normal permission for this file, which allows her to read and write to this file. She also limits file permissions to other members of the Group Sales to read only. User Peter on host Peter is a member of the Group SALES. Peter has access to the export / Acct_Rep. If user Peter attempts to write to file MRPT1, he would be denied the permission to write to the file.

NAS Foundations - 19

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC NAS Platforms


Identify products from the EMC NAS range of equipment

2006 EMC Corporation. All rights reserved.

NAS Foundations - 20

EMC NAS products are reviewed in this section.

NAS Foundations - 20

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Celerra Platforms


Broadest Range of NAS Products
SIMPLE WEB-BASED MANAGEMENT

NS350 NS40* NS500 NS700


High availability One or two Data Movers * X-Blade technology Upgradeable to gateway Integrated CLARiiON

NS704

NS40G* NS500G NS700G


High availability One or two Data Movers * X-Blade Technology NAS gateway to SAN

NS704G NS80G*

NSX

Advanced clustering

Advanced clustering Four Data Movers * X-Blade Technology

Advanced clustering Four to eight X-Blades

Four Data Movers

Upgradeable to gateway Integrated CLARiiON

NAS gateway to SAN

NAS gateway to SAN

CLARiiON, Symmetrix

CLARiiON, Symmetrix

CLARiiON, Symmetrix

2006 EMC Corporation. All rights reserved.

NAS Foundations - 21

An important decision you must make is, What is the right information platform that meets my business requirements? EMC makes it easy by offering the broadest range of NAS platforms in the industry. Rate your requirements and choose your solution. The range of EMC NAS all use DART, Data Access in Real Time, operating system, which is specially developed to provide efficient data transfer between the front end network connections and the backend disk interfaces. There are at present two configurations available, Gateway and Integrated. The Gateway models provide a NAS interface to SAN/Fabric attached storage arrays, while the Integrated have their storage arrays contained within the same frames as the NAS heads, Data Movers, which are solely dedicated to NAS functionality, (no shared host access to disks). The Celerra NS Gateway (can be configured with up four Data Movers) and the Celerra NS GS (configured with a single Data Mover) connects to CLARiiON CX arrays and/or Symmetrix DMX arrays through a fibre channel switch or directly connected (in the case of CLARiiON). The NSX gateway model can be configured with between four and eight Data Movers.

NAS Foundations - 21

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra NAS - SAN Scalability


Consolidated storage infrastructure for all applications NAS front end scales independently of SAN back end Allocate storage to Celerra and servers as needed Centralized management for SAN and NAS iSCSI gateway to SAN
2006 EMC Corporation. All rights reserved.

Celerra NS G Family

CLARiiON CX Family

Celerra NSX Connectrix SAN

Windows UNIX

Symmetrix DMX Family

NAS Foundations - 22

One of the reasons that EMC NAS scales impressively is due to the gateway architecture that separates the NAS front end (Data Movers) from the SAN back end (Symmetrix or CLARiiON). This allows the front end and back end to grow independently. Customers can merely add Data Movers to the EMC NAS to scale the front-end performance to handle more clients. As the amount of data increases, you can add more disks, or the EMC NAS can access multiple Symmetrix or CLARiiON. This flexibility leads to improved disk utilization. EMC NAS supports simultaneous SAN and NAS access to the CLARiiON and Symmetrix. and can be added to an existing SAN, with general purpose servers now able to access non-NAS back-end capacity. This extends the improved utilization, centralized management, and TCO benefits of SAN plus NAS consolidation to EMC NAS, Symmetrix, and CLARiiON. The configuration can also be reconfigured via software. Since all Data Movers can see the entire file space, it is easy to reassign filesystems to balance the load. In addition, filesystems can be extended online as they fill. Even though the architecture splits the front end among multiple Data Movers and a separate SAN back end, the entire NAS solution can be managed as a single entity.

NAS Foundations - 22

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Hardware


Describe and identify common EMC NAS components

2006 EMC Corporation. All rights reserved.

NAS Foundations - 23

Due to the diversity of the range of EMC NAS systems we now briefly review some of the major hardware components, to differentiate between the various options available.

NAS Foundations - 23

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra NS Family Control Station Hardware


The control station provides an interface to the control; to manage and configure the NAS solution

2006 EMC Corporation. All rights reserved.

NAS Foundations - 24

Control Station provides the controlling subsystem of the Celerra, as well as the management interface to all file server components. The Control Station provides a secure user interface as a single point of administration and management for the whole Celerra solution. Control Station administrative functions are accessible via the local console, Telnet (not recommended), or a Web Browser. The Control station is single Intel processor based, with high memory capacity. Dependent on the model, the Control Stations may have internal storage. The local LAN switch provides the internal communications network for the Data Movers and the Control Station and should NOT be integrated into a client networking infrastructure. Within the NSX model there are no serial interconnections between the Control Station and the Data Movers and the internal switch has been built into the Control Station functionality.

NAS Foundations - 24

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NSX Next Generation Control Station


Celerra NSX Front view

Reset Switch NMI Switch Power Switch

ID Switch LEDs and Switches

Power Boot Sequence LED Status LED HDD Act LED HDD Fault LED
2006 EMC Corporation. All rights reserved.

Serial Port COM2 USB Connectors 2 and 3 ID LED Gb # 1 and Gb # 2 LED


NAS Foundations - 25

The Control Station is a dedicated management Intel processor-based computer that monitors and sends commands to the blades. The private network connects the two Control Stations (always shipped on NSX systems) to the blades through the system management switch modules. Like previous versions it provides software installation and upgrade services, and high-availability features such as fault monitoring, fault recovery, fault reporting (CallHome), and remote diagnosing. Two Control Stations can be connected to a public or private network for remote administration. Each Control Station has a serial port that connects to an external modem so that the Control Station can call home to EMC or a service provider if a problem should arise.

NAS Foundations - 25

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NSX Next Generation Control Station


Celerra NSX Rear view

eth3 - Public LAN Port COM1 - To serial modem (for Call-Home) eth0 Internal Network (To Mgmt. Switch-A in Enclosure 0) Video Port Gb2 Internal Network (To Mgmt. Switch-B in Enclosure 0) Gb1 IPMI (To eth1 of the other Control Station)

2006 EMC Corporation. All rights reserved.

NAS Foundations - 26

This slide displays the rear view of the Next Generation Control Station. Note the lack of a 25-pin quad serial port and spider cable.

NAS Foundations - 26

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Data Mover Hardware


Single or Dual Intel Processors PCI or PCI-X based High memory capacity Multi-port Network cards Fibre Channel connectivity to storage arrays No internal storage devices Redundancy mechanism
NS40 Data Mover Fibre I/O module GbE I/O module NSX Data Mover/ NS80G Data Mover

2006 EMC Corporation. All rights reserved.

NAS Foundations - 27

Each Data Mover is an independent, autonomous file server that transfers requested files to clients and are managed as a single entity. Data Movers are hot pluggable and can be configured with standbys to implement N to 1 unaffected should a problem arise with another Data Mover. The multiple Data Movers (up to 8 in the NSX and 4 in the NS range availability. A Data Mover (DM) connects to a LAN through FastEthernet and/or Gigabit Ethernet. The default name for a Data Mover is server n, where n was its original slot location in the first NAS frames. This has been continued into the new frames and the naming convention remains slot related. For example, in the Golden Eagle/ Eagle frame, a Data Mover can be in slot location 2 through 15 (i.e. server_2 - server_15), therefore the first Data Mover is any frame remains server_2, the second server_3, etc. There is no remote login capability on the DM, nor do they run any binaries (very secure) and all access to the Data Mover for management and configuration must be performed via the Control Station. Data Mover redundancy is the mechanism by which the Celerra family reduces the network data outage in the event of a Data Mover failure. The ability to failover the Data Movers is achieved by the creation of a Data Mover configuration database on the Control Station system volumes and is managed via the Control Station. No Data Mover failover occurs if the Control Station is not available for some reason.

NAS Foundations - 27

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Data Mover Hardware (Cont.)


Standby Data Mover Configuration Options
Each standby Data Mover, as a standby for a single primary Data Mover Each standby Data Mover, as a standby for a group of primary Data Movers

Failover Operational Modes


Automatic Retry Manual

2006 EMC Corporation. All rights reserved.

NAS Foundations - 28

These Standby Data Movers are powered and ready to assume the personality of their associated Primary Data Movers in the event of a failure. If a Primary Data Mover fails, the Control Station detects the failure and initiates the failover process. The failover procedure, in an Automatic configuration, is as follows. The Control Station: 1. Removes power from the failed Data Mover. 2. Sets the location for the Standby Data Mover to assume its new personality in the configuration database. 3. Controls the personality take over and allows the Standby Data Mover to assume the primary role, thereby enabling clients to re-access their data transparently via the standby. Once the failed Data Mover is repaired, the failback mechanism is always manually administrator initiated. This process is the reverse of the failover process and restores the primary functionality to the repaired Primary Data Mover and returns the Standby Data Mover into its standby state in preparation for any future outage. There are three operational modes of operation for Failover: Automatic, Retry, and Manual. 1. Automatic Mode: Control Station detects the failure of a Data Mover. The failover process occurs without trying any recovery process first. 2. Retry Mode: Control Station detects the failure, an attempt to reboot the failed Data Mover is tried first before the failover procedure is initiated. 3. Manual Mode: Control Station detects the failure and removes power from the failed Data Mover. However, no further Data Mover recovery action is taken until administrative intervention. Recovery after a Data Mover failover is always a manual process.

NAS Foundations - 28

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NAS Reference Documentation

NAS Interoperability Matrix


Data Movers Control Stations Software supported features

Website
www.emc.com/horizontal/interoperability

2006 EMC Corporation. All rights reserved.

NAS Foundations - 29

The NAS interoperability Guide provides support information on the Data Movers and Control Station models, NAS software version, supported features, storage models, and microcode. This interoperability reference can be found at: http://www.emc.com/horizontal/interoperability.

NAS Foundations - 29

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Software


Describe operating systems used by EMC NAS

2006 EMC Corporation. All rights reserved.

NAS Foundations - 30

Having briefly reviewed some of the major hardware components, the software environment of the high-end EMC NAS offering is covered next.

NAS Foundations - 30

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Software Operating Systems


EMC Linux
This is an industry hardened and EMC modified Operating System loaded on the Control Station to provide:
Secure NAS management environment Growing in popularity and corporate acceptance

DART Data Access in Real Time


This is a highly specialized Operating System designed to optimize network traffic Input/Output throughput and is loaded on the Data Movers Is multi-threaded to optimize load balancing capabilities of the multi-processor Data Movers Advanced volume management - UxFS
Large file size and filesystem support Ability to extend filesystems online Metadata logging for fast recovery Striped volume support

Feature rich to support the varied specialized capabilities of the Celerra range
Data Mover Failover Networking functionality Port Aggregation, FailSafe Network device, multi-protocol support Point in time Filesystem copies Windows environmental specialties

2006 EMC Corporation. All rights reserved.

NAS Foundations - 31

EMC Linux OS is installed on the Control Station. Control Station OS software is used to install, manage, and configure the Data Movers, monitor the environmental conditions and performance of all components, and implement the Call Home and dial-in support feature. Typical Administration functions include the volume and filesystem management, configuration of network interfaces, creation of filesystems, exporting filesystems to clients, performing filesystem consistency checks, and extending filesystems. The OS that the Data Movers run is EMCs Data Access in Real Time (DART) embedded system software, which is optimized for file I/O, to move data from the EMC storage array to the network. DART supports standard network and file access protocols: NFS, CIFS, and FTP.

NAS Foundations - 31

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Software Management


Describe user interfaces available for EMC NAS management

2006 EMC Corporation. All rights reserved.

NAS Foundations - 32

The two user interfaces available for EMC NAS management are reviewed in this section.

NAS Foundations - 32

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Management Command Line


The command line can be accessed on the Control Station via
An ssh interface tool (i.e. PuTTy) Telnet

Its primary function is for scripting of common repetitive tasks that may run on a predetermined schedule to ease administrative burden It has approximately 80 UNIX command-like commands

2006 EMC Corporation. All rights reserved.

NAS Foundations - 33

Telnet access is disabled, by default, on the Control Station due to the possibility of unauthorized access if the Control Station is placed on a publicly accessible network. If this is the case, it is strongly recommended that this service is not enabled. The preferred mechanism of accessing the Control Station is the SSH (Secure Shell) daemon via an SSH client such as PuTTy.

NAS Foundations - 33

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Manager Management

2006 EMC Corporation. All rights reserved.

NAS Foundations - 34

GUI management has become consolidated into one product with two options; Celerra Native Manager Basic Edition and Celerra Management Advanced Edition. The Basic Edition is installed, along with the DART OS, and provides a complete set of common management functionality for a single Celerra at a time. The Advanced Edition adds multiple Celerra support, along with some advanced feature GUI management, and is licensed separately from the DART code.

NAS Foundations - 34

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Manager Wizards

2006 EMC Corporation. All rights reserved.

NAS Foundations - 35

Celerra Manager offers a number of configuration Wizards for various tasks to assist with new administrator ease of implementation.

NAS Foundations - 35

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Manager Tools

2006 EMC Corporation. All rights reserved.

NAS Foundations - 36

Celerra Manager offers a set of tools to integrate Celerra monitoring functionality and launch Navisphere Manager. With the addition of the Navisphere Manager Launch capability, the SAN/NAS administrator has a more consolidated management environment.

NAS Foundations - 36

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC ControlCenter V5.x.x NAS Support


Discovery and Monitoring
Data Movers Devices and volumes Network adapters and IP interfaces Mount points Exports Filesystems (including snapshots and checkpoints)

2006 EMC Corporation. All rights reserved.

NAS Foundations - 37

The EMC flagship management product, EMC ControlCenter, has the capability of an assisted discovery of both EMC NAS and third party NAS products, namely NetApps filers Currently, management of the EMC NAS family is deferred to the specific product management products due to the highly specialized nature of the NAS environment. Therefore, this product functionality (shown on this slide) is focused mainly around discovery, monitoring, and product management software launch capability ControlCenter V5.x.x has enhanced device management support for the Celerra family. The ControlCenter Celerra Agent runs on Windows and has enhanced discovery and monitoring capabilities. You can now view properties information on Celerra Data Movers, devices, network adapters and interfaces, mount points, exports, filesystems (including snapshots and checkpoints), and volumes from the ControlCenter Console. You can also view alerting information for the Celerra family as well.

NAS Foundations - 37

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Software Management


Describe the implementation of VLANs (Virtual Local Area Networks) for environmental management with EMC NAS

2006 EMC Corporation. All rights reserved.

NAS Foundations - 38

Next, an overview of the virtual local area networking environment, or VLANs, is reviewed.

NAS Foundations - 38

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

VLAN Support
Create logical LAN segment
Divide a single LAN into logical segments Join multiple separate segments into one logical LAN
Broadcast Domain LAN Bridge or Switch
Hub Hub Hub

VLAN A

Collision Domain LAN Segment Collision Domain LAN Segment


Hub

VLAN B

VLAN Tagging
802.1q

Collision Domain LAN Segment Bridge or Switch Workstation VLAN B


Hub Hub

Simplified Management
No network reconfiguration required for member relocation
2006 EMC Corporation. All rights reserved.

VLAN A

Router

Broadcast Domain LAN

NAS Foundations - 39

Network domains are categorized into Collision, a LAN segment within which data collisions are contained, or Broadcast, the portions of the network through which broadcast and multicast traffic is propagated. Collision domains are determined by hardware components and how they are connected together. The components are usually client computers, hubs, and repeaters. A network switch or a router that generally does not forward broadcast traffic separates a Collision domain from a Broadcast domain. VLANs allow multiple, distinct, possibly geographically separate network segments to be connected in to one logical segment. This can be done by subnetting or using VLAN tags (802.1q.), which is an address added to network packets to identify the VLANs to which the packet belongs. This could allow servers that were connected to physically separate networks to communicate more efficiently and it could prevent servers that were attached to the same physical network from impeding one another. By using VLANs to logically segment the Broadcast Domains, the equipment contained within this logical environment need not be physically located together. This now means that if a mobile client moves location, an administrator need not do any physical network or software configuration for the relocation as bridging technology would now be used, and a router would only be needed to communicate between VLANS. There are two commonly practiced ways of implementing this technology: 1. IP Address subnetting or 2. VLAN Ethernet packet tagging When using the IP address subnetting methodology, the administrator configures the broadcast domains to include the whole network area for specific groups of computers by using BridgeRouter technology. When using the VLAN tagging methodology, the members of a specific group have an identification tag embedded into all of their Ethernet packet traffic. VLAN Tagging allows a single Gigabit Data Mover port to service multiple logical LANs (Virtual LANs). This allows data network nodes to be configured (added and moved as well as other changes) quickly and conveniently from the management console, rather than in the wiring closet. VLAN also allows a customer to limit traffic to specific elements of a corporate network and protect against broadcasts (such as denial of service) affecting whole networks. Standard router based security mechanisms can be used with VLANs to restrict access and improve security.

NAS Foundations - 39

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

VLAN Benefits
Performance Reduced Overhead Reduced Costs Security

VLAN-A
2006 EMC Corporation. All rights reserved.

VLAN S

VLAN E

NAS Foundations - 40

The benefits of VLAN support include: Performance: In all networks, there is a large amount of broadcast and multicast traffic and VLANS can reduce the amount of traffic being received by all clients. Virtual Collaborative Work Divisions: by placing widely dispersed collaborative users into a VLAN, broadcast and multicast traffic between these users are kept from affecting other network clients and reduce the amount of routing overhead placed on their traffic. Simplified Administration: with the large amount of mobile computing today, physical user relocation generates a lot of administrative user reconfiguration (adding, moving and changing). If the user has not changed company functionality, but has only relocated, VLANs can achieve undisrupted job functionality. Reduced Cost by using VLANS: expensive routers and billable traffic routing costs can be reduced. Security, by placing users into a tagged VLAN environment, external access to sensitive broadcast data traffic can be reduced. VLAN support enables a single Data Mover with Gigabit Ethernet port(s) to be the standby for multiple primary Data Movers with Gigabit Ethernet port(s). Each primary Data Mover's Gigabit Ethernet port(s) can be connected to different switches. Each of these switches can be in a different subnet and different VLAN. The standby Data Mover's Gigabit Ethernet port is connected to a switch which is connected to all the other switches.

NAS Foundations - 40

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Filesystem Management


Describe filesystem Quotas implementation on EMC NAS

2006 EMC Corporation. All rights reserved.

NAS Foundations - 41

Next, file system controls supported by Celerra Management software are reviewed.

NAS Foundations - 41

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Filesystem Controls - User Quota Restrictions


There are three main types of quotas used in data space control:
Soft Quota
Amount of data space or number of files used under normal working conditions

Hard Quota
Total space or number of files a user/group can use or create on a file system

Tree Quota
Total space or number of files that a user/group can use or create on a data directory tree. They are used as a logical mechanism to segment large file systems into smaller administrative portions that do not affect each others operation

2006 EMC Corporation. All rights reserved.

NAS Foundations - 42

One of the most common concerns in a distributed data environment is that users tend to save many copies of the same information. When working in a collaborative distributed environment, the amount of data space required by each user expands rapidly and, in some cases, uncontrollably. To minimize data space outages, the user space can be controlled by imposing Quotas on users, or groups of users, to limit the number of blocks of disk space they can use or the number of files they can create. The Soft Quota is a logical limit placed on a user that can be exceeded without the need for any administrative intervention. Once the soft quota limit has been exceeded, the user has a grace period to use the extra space defined by the hard quota limit. However, the user/group cannot exceed the hard limit The grace period is a time limit during which the user, or group, can continue to increase the amount of disk space used or number of files created. If the grace period expires, the user/group must reduce the amount of space used or the number of files to below the soft limit before any new space or files can be created. The Celerra family supports all of these Quota methodologies, thereby assisting administrators used to these management tools, with a seamless transition into an EMC NAS environment.

NAS Foundations - 42

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Management Software


Describe some Windows-specific options for environmental management using EMC NAS including:
Usermapper Virtual Data Movers Microsoft Management Console Snap-ins

2006 EMC Corporation. All rights reserved.

NAS Foundations - 43

Celerra family high availability features are reviewed in this section.

NAS Foundations - 43

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Usermapper Windows and UNIX integration


Usermapper is the methodology by which Windows SIDs (Security Identifiers) are equated with UNIX UIDs (User/Group Identifiers) on EMC NAS devices There are two configurable environments to achieve these mappings
Internal
Part of the Data Mover's software. It does not require a separate installation or additional configuration procedures for a new Celerra Network Server

External
Runs as a daemon on a Celerra Control Station. Requires a separate installation as well as additional configuration and management procedures

2006 EMC Corporation. All rights reserved.

NAS Foundations - 44

EMC NAS device Data Mover operating system, DART, utilizes a very specialized UNIX like file system and thus has the same security structures. To support disparate clients, NFS and CIFS, the various environmental security structures need to be equated to the Data Mover structures. In the NFS environment no translation needs to be performed, however in the Microsoft environment the Security Identifiers, (SID), need to be equated to the security structures on the filesystem. Usermapper is the mechanism that is used in an EMC NAS device to achieve this mapping.

NAS Foundations - 44

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Usermapper Windows and UNIX integration (Cont.)


Other considerations about choosing the Usermapper methodology is determined by the client environment
Windows only
Internal Usermapper in Windows-only environments is recommended. Celerra Network Server installations after version 5.2 use this by default

Mixed protocol UNIX and Windows


In multiprotocol environments, file systems can be accessed by UNIX and Windows users. Some of the methodologies that enable this to be achieved are:
Active Directory (using Microsoft Management Console snap-ins) A Data Movers local user and group files Network Information Service (NIS)

2006 EMC Corporation. All rights reserved.

NAS Foundations - 45

ACL= Access Control Lists ACE= Access Control Entry In multiprotocol environments, file systems can be accessed by UNIX and Windows users. File access is determined by the permissions on the file or directory, the UNIX permissions, Windows access control lists (ACLs), or both permissions and ACLs. Therefore, if a user has a UNIX and Windows user accounts, you should choose a mapping method that allows you to indicate that the two accounts represent the same user. Some of the methodologies that enable this to be achieved are: Active Directory (using Microsoft Management Console snap-ins) A Data Movers local user and group files Network Information Service (NIS) If a user in a multiprotocol environment only uses a single logon ( through Windows or UNIX), then it is acceptable to use Usermapper. If a user has only one account, mapping to an equivalent identity in the other environment is not necessary.

NAS Foundations - 45

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Usermapper - Pre & Post DART v5.2


Primary Server

(4) Adds a new entry from specified

Example: New User Requests Resource


(New mapping required)
Data Mover

UID/GID range

(5) Notifies all other Secondary Servers that


they should initiate a cache update request

Usermapper DB

DART v5.2 (3)

Usermapper DB
Pre - DART v5.2 (6)
New Entry

Usermapper Resolver

New Mapping Request

(1) Resolver queries


First server configured

Secondary Server

(2) Mapping is not in DB (7) Updates cache with new mapping

(8) Replies to the Data Mover's request with the UID/GID mapping Configuration/Installation
2006 EMC Corporation. All rights reserved.

Usermapper DB
NAS Foundations - 46

This slide illustrates the steps to grant access. Step 1: A client request is received at a Data Mover, with the resolver stub running, without a valid UID/ID. The resolver then contacts the first usermapper server configured in the configuration file with a request for a UID/GID Step 2 and 3: A secondary server is contacted due to its configuration priority over the primary server. If this secondary server does not have a listing for the particular user making the request in its cache, a request is made to the primary server for a UID / GID new mapping. Step 4 and 5: When the primary server receives a request for a new mapping, an entry from the specified UID/GID range is added to the database and a notification is issued to all secondary servers that their cache entries must be updated. Step 6 and 7: The secondary server making the request for the new mapping updates its cache with new information from the primary server upon the receipt of the update notification. Step 8: The secondary server that received the initial request now responds back to the requesting Data Mover with the new mapping information and the user is granted access (or denied access) to the requested resource. Note: DART v 5.2 introduces a fundamental upgrade to the usermapper process. Each Data Mover now maintains its own usermapper data base of user mappings. This upgrade assists with Data Mover failover connectivity continuance and access is unaffected by possible Control Station failure.

NAS Foundations - 46

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Virtual Data Movers


Virtual Data Movers on Single Physical Data Movers
Ability to create multiple virtual CIFS servers on each logical Data Mover Consolidation of multiple server file serving functionality onto single Data Movers as each virtual Data Mover can maintain isolated CIFS servers with their own root filesystem environment Allows whole Virtual Data Mover environments to be loaded, unloaded, or even replicated between physical Data Movers for ease in Windows environmental management

2006 EMC Corporation. All rights reserved.

NAS Foundations - 47

Currently, in pre DART v5.2, a Data Mover supported one NFS server and multiple CIFS servers, where each server has the same view of all the resources. The CIFS servers are not logically isolated and although they are very useful in consolidating multiple servers into one data mover, they do not provide the isolation between servers as needed in some environments such as data from disjoint departments hosted on the same data mover. Now, VDMs support separate isolated CIFS servers, allowing you to place one or multiple CIFS servers into a VDM, along with their file systems. The servers residing in a VDM store their dynamic configuration information (such as local groups, shares, security credentials, and audit logs, etc.) in a configuration file system. A VDM can then be loaded and unloaded, moved from Data Mover to Data Mover, or even replicated to a remote Data Mover as an autonomous unit. The servers, their file systems, and all of the configuration data that allows clients to access the file systems are available in one virtual container. VDMs provide virtual partitioning of the physical resources and independently contain all the information necessary to support the contained CIFS servers. Having the file systems and the configuration information contained in a VDM does the following: 1. Enables administrators to separate CIFS servers and give them access to specified shares; 2. Allows replication of the CIFS environment from primary to secondary without impacting server access, 3. Enables administrators to easily move CIFS servers from one physical Data Mover to another. A VDM can contain one or more CIFS servers. The only requirement is that you have at least one interface available for each CIFS server you create. The CIFS servers in each VDM have access only to the file systems mounted to that VDM, and therefore can only create shares on those file systems mounted to the VDM. This allows a user to administratively partition or group their file systems and CIFS servers.

NAS Foundations - 47

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Additional Tools: MMC Snap-ins


UNIX User Management
Active Directory migration tool MMC plug-in extension for Active Directory uses and computers Celerra Management tool snap-in (MMC Console)

Virus Checker Management


Celerra Management tool: (MMC Console)

Home Directory snap-in


Allows multiple points of entry to a single share

Data Mover security snap-in


Manage user rights and auditing
2006 EMC Corporation. All rights reserved. NAS Foundations - 48

Celerra offers a number of Windows management tools with the Windows look and feel. For example, Celerra shares and quotas can be managed by the standard Microsoft Management Console (MMC). The tools include: The Celerra Management Tool (MMC Console): Snap-in extension for Dart Virus Checker Management which manages parameters for the DART Virus Checker. The Active Directory (AD) Migration tool: Migrates the Windows/UNIX user and group mappings to Active Directory. The matching users/groups are displayed in a property page with a separate sheet for users and groups. The administrator selects the users/groups that should be migrated and de-selects those that should not be migrated or should be removed from Active Directory. The Microsoft Management Console (MMC): Snap-in extension for AD users and computers. This adds a property page to the users property sheet to specify UID (user ID) /GID (group ID)/Comment and adds a property page to the group property sheet to specify GID/Comment. You can only manage users and a group of the local tree. The Celerra Management Tool (MMC Console): Snap-in extension for Dart UNIX User Management displays Windows users/groups which are mapped to UNIX attributes. It also displays all domains that are known to the local domain (Local Tree, Trusted domains). The Home Directories capability in the Celerra allows a customer to set up multiple points of entry to a single Share/Export so as to avoid sharing out many hundreds of points of entry to a filesystem for each individual user for storing their Home Directories. The MMC Snap-in provides a simple and familiar management interface for Windows administrators for this capability. The Data Mover Security Settings Snap-in provides a standard Windows interface for managing user rights assignments, as well as the settings for which statistics Celerra should audit, based on the NT V4 style auditing policies.

NAS Foundations - 48

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Software


Describe some network high availability features incorporated into the EMC NAS solution

2006 EMC Corporation. All rights reserved.

NAS Foundations - 49

Celerra family high availability features are reviewed in this section.

NAS Foundations - 49

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NS Series Networking
Network interfaces
Ethernet Gigabit Ethernet

Network protocols
TCP/IP, UDP/IP CIFS, NFS V2, V3 and V4 FTP, TFTP, and SNMP NDMP V2, V3, and V4 NTP, SNTP iSCSI target Link aggregation FailSafe Networking Ethernet Trunking Virtual LAN
FSN Ethernet

SNMP TCP iSCSI

Gigabit Ethernet

NFS

Feature support

VLAN

Trunking NDMP

CIFS

FTP

2006 EMC Corporation. All rights reserved.

NAS Foundations - 50

The NS Series implements industry-standard networking protocols: The network ports supported by the NS700, NS704, NS700G, and NS704G consist of 10/100/1000 Ethernet (Copper) and Optical Gigabit Ethernet. All other NS Series platforms support Copper 10/100/1000 Ethernet only. Network protocols supported include Transmission Control Protocol over Internet Protocol (TCP/IP) and User Datagram Protocol over IP (UDP/IP). File-sharing protocols are CIFS (Common Internet File System), used by Windows; and NFS (Network File System) V2, V3, and V4, used by UNIX and Linux. File transfers are supported with the FTP and TFTP protocols. NDMP V2, V3, and V4 are supported for LAN-free backups. Network management can be accomplished with Simple Network Management Protocol (SNMP). NTP and SNTP protocols allow Data Movers to synchronize with a known time source. SNTP is more appropriate for LAN environments. The NS Series supports iSCSI Target for block access. VLAN Tagging allows a single Gigabit port to service multiple logical LANs (virtual LANs). FailSafe Networking extends the failover functionality to networking ports.

NAS Foundations - 50

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Network FailSafe Device


Network outages, due to environmental failure, are more common than Data Mover failures Network FailSafe Device
DART OS mechanism to minimize data access disruption due to these failures Logical device is created using physical ports or other logical ports combined together to create redundant groups of ports Logically grouped Data Mover network ports monitor network traffic on the ports Active FailSafe Device port senses traffic disruption Standby (non-active) port assumes the IP Address and Media Access Control address in a very short space of time, thus reducing data access disruption
2006 EMC Corporation. All rights reserved. NAS Foundations - 51

Having discussed the maintenance of data access via redundant Data Movers, we now discuss the same concept utilizing network port mechanisms. First lets look at the Network Failsafe device. Network outages due to environmental failures are more common than Data Mover failures. To minimize data access disruption due to these failures, the DART OS has a mechanism that is environment agnostic, the Network FailSafe Device. This is a mechanism by which the Network ports of a Data Mover may be logically grouped together into a partnership that monitor network traffic on the ports. If the currently active port senses a disruption of traffic, the standby (non-active) port assumes the active role in a very short space of time, thus reducing data access disruption. The way this works is a logical device is created, using physical ports or other logical ports, combined together to create redundant groups of ports. In normal operation, the active port carries all network traffic. The standby (nonactive port) remains passive until a failure is detected. Once a failure has been detected by the FailSafe Device, this port assumes the network identity of the active port, including IP Address and Media Access Control address. Having assumed the failed port identity, the standby port now continues the network traffic. Network disruption due to this change over is minimal and may only be noticed in a high transaction oriented NAS implementation or in CIFS environments due to the connection-oriented nature of the protocol. There are several benefits achieved by configuring the network FailSafe device: 1. Configuration is handled transparently to client access; 2. the ports that make up the FailSafe device need not be of the same type; 3. Rapid recovery from a detected failure; 4. can be combined with logical Aggregated Port devices to provide even higher levels of redundancy. Although the ports that make up the FailSafe device need not be of the same type, care must be taken to ensure that once failover has occurred, that client expected response times remain relatively the same and data access paths are maintained.

NAS Foundations - 51

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Link Aggregation - High Availability


Link aggregation
Combining of two or more data channels into a single data channel for high availability Two Methods: IEEE 802.3ad LACP & CISCO FastEtherChannel

IEEE 802.3ad LACP


Combining links for improved availability If one port fails, other ports take over Industry standard IEEE 802.3ad Combines 212 Ethernet ports into a single virtual link Deterministic behavior Does not increase single client throughput
2006 EMC Corporation. All rights reserved.

LINK Industry Standard Switch

Celerra

NAS Foundations - 52

Having discussed the network FailSafe device, the next methodologies we look at are the two Link Aggregation methodologies. Link aggregation is the combining of two or more data channels into a single data channel. There are two methodologies that are supported by EMC NAS devices. They are IEEE 802.3ad Link Aggregation Control Protocol and CISCO FastEtherChannel using Port Aggregation Protocol (PAgP). The purpose for combining data channels in the EMC implementation is to achieve redundancy and fault tolerance of network connectivity. It is commonly assumed that link aggregation provides a single client with a data channel bandwidth equal to the sum of the bandwidth of individual member channels. This is not, in fact, the case due to the methodology of channel utilization and, it may only be achieved with very special considerations to the client environment. The overall channel bandwidth is increased, but the client only receives, under normal working conditions, the bandwidth equal to one of the component channels. To implement Link Aggregation, the network switches must support the IEEE 802.3ad standard. It is a technique for combining several links together to enhance availability of network access and applies to a single Data Mover and not across Data Movers. The current implementation focuses on availability, therefore check the NAS support matrix. Only full duplex operation is currently supported. Always check the NAS Interoperability Matrix for supported features at the following: http://www.emc.com/horizontal/interoperability

NAS Foundations - 52

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Link Aggregation - High Availability (Cont.)

CISCO EtherChannel
Port grouping for improved availability Combines 2,4, or 8 Ethernet ports into a single virtual device Inter-operates with trunking-capable switches High availability: if one port fails, other ports take over Does not increase single client throughput
2006 EMC Corporation. All rights reserved.

Channel CISCO Switch Celerra

NAS Foundations - 53

Ethernet Trunking (Ether Channel) increases availability. It provides statistical load sharing by connecting different clients to different ports. It does not increase single-client throughput. Different clients get allocated to different ports. With only one client, the client accesses Celerra via the same port for every access. This DART OS feature interoperates EtherChannel capable Cisco switches. EtherChannel is Cisco proprietary.

NAS Foundations - 53

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Network Redundancy - High Availability

An example of FSN and Port aggregation co-operation

2006 EMC Corporation. All rights reserved.

NAS Foundations - 54

This example shows a fail-safe network device that consists of a FastEtherChannel comprising the four ports of an Ethernet NIC and one Gigabit Ethernet port. The FastEtherChannel could be the primary device but, per recommended practices, the ports of the Fail Safe Network (FSN) would not be marked primary or secondary. FSN provides the ability to configure a standby network port for a primary port, and the two or more ports can be connected to different switches. The secondary port remains passive until the primary port link status is broken, then the secondary port takes over operation. An FSN device is a virtual device that combines 2 virtual ports. A virtual port can consist of a single physical link or an aggregation of links (EtherChannel, LACP). The port types or number need not be the same when creating a failsafe device group. For example, a quad Ethernet card can be first trunked and then coupled with a single Gigabit Ethernet port. In this case, all four ports in the trunk would need to fail before FSN would implement failover to the Gigabit port. Thus, Celerra could tolerate four network failures before losing the connection. Note: An active primary port/active standby port configuration on the Data Mover is not recommended practice.

NAS Foundations - 54

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Family Business Continuity


Describe EMS NAS disk based replication and recovery solutions

2006 EMC Corporation. All rights reserved.

NAS Foundations - 55

Having integrated the Celerra into the environment, data replication and recovery solutions that augment the environment are reviewed next.

NAS Foundations - 55

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Disk-Based Replication and Recovery Solutions


NS & NSX / Symmetrix

Synchronous Disaster Recovery


SRDF FUNCTIONALITY

File-based FileReplication
NS / CLARiiON
TimeFinder/FS Celerra Replicator EMC OnCourse

File Restoration
Celerra SnapSure
Celerra Celerra/ FC4700 NS

RECOVERY TIME

Hours
2006 EMC Corporation. All rights reserved.

Minutes

Seconds
NAS Foundations - 56

High-end environments require non-stop access to the information pool. From a practical perspective, not all data carries the same value. The following illustrates that EMC Celerra provides a range of disk-based replication tools for each recovery time requirement. File restoration: This is the information archived to disk and typically saved to tape. Here we measure recovery in hours. Celerra SnapSure enables local point-in-time replication for file undeletes and backups. File-based replication: This information is recoverable in time frames measured in minutes. Information is mirrored to disk by TimeFinder, and the copy is made accessible with TimeFinder/FS. The Celerra Replicator creates replicas of production filesystems locally or at a remote site. Recovery time from the secondary site depends on the bandwidth of the IP connection between the two sites. EMC OnCourse provides secure, policy-based file transfers. The Replicator feature supports data recovery for CIFS and NFS by allowing the secondary filesystem (SFS) to be manually switched to read/write mode after the Replicator session has been stopped, manually or due to a destructive event. Note: There is no re-synch or failback capability. Synchronous disaster recovery: This is the information requiring disaster recovery with no loss of transactions. This strategy allows customers to have data recovery in seconds. SRDF, in synchronous mode, facilitates real-time remote mirroring in campus environments (up to 60 km). File restoration and file-based replication (Celerra Replicator, EMC OnCourse) are available with Celerra /CLARiiON. The entire suite of file restoration, file-based replication, and synchronous disaster recovery are available with Celerra /Symmetrix.

NAS Foundations - 56

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Disaster Recovery
Describe EMC NAS disaster recovery methodology using Celerra SRDF (Symmetrix Remote Data Facility)

2006 EMC Corporation. All rights reserved.

NAS Foundations - 57

Celerra disaster recovery, when integrated with the Symmetrix, utilizes a very synergistic combination of Celerra and the Symmetrix functionality.

NAS Foundations - 57

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra SRDF Disaster Recovery


Network Campus (60 km) distance Uni or bi-directional

Celerra

Celerra

Increases data availability by combining the high availability of the Celerra family with the Symmetrix Remote Data Facility Celerra synchronous disaster recovery solution
Allows an administrator to configure remote standby Data Movers waiting to assume primary roles in the event of a disaster occurring at the primary data site SRDF allows administrator to achieve a remote synchronous copy of production filesystems at a remote location Real-time, logically synchronized and consistent copies of selected volumes Uni-directional and bi-directional support Resilient against drive, link, and server failures No lost I/Os in the event of a disaster Independent of CPU, operating system, application, or database Simplifies disaster recovery switchover and back
2006 EMC Corporation. All rights reserved.

In the NAS environment, data availability is one of the key aspects for implementation determination. By combining the high availability of the Celerra family with the Symmetrix Remote Data Facility, data available increases exponentially. What the SRDF feature allows an administrator to achieve is a remote synchronous copy of production filesystems at a remote location. However, as this entails the creation of Symmetrix specific R1 and R2 data volumes, this functionality is currently restricted to Celerra / Symmetrix implementations only. This feature allows an administrator to configure remote standby Data Movers waiting to assume primary roles in the event of a disaster occurring at the primary data site. Due to data latency issues, this solution is restricted to a campus distance of separation between the two data sites (60 network km). The SRDF solution for Celerra can leverage an existing SRDF transport infrastructure to support the full range of supported SAN (storage area network) and DAS (direct-attached storage) connected general purpose server platforms. The Celerra disaster recovery solution maintains continuously available filesystems, even with an unavailable or non-functioning Celerra. Symmetrix technology connects a local and remote Celerra over a distance of up to 40 miles (66 km) via an ESCON or Fiber Channel SRDF connection. After establishing the connection and properly configuring the Celerra, users gain continued access to filesystems in the event that the local Celerra and/or the Symmetrix becomes unavailable. The Celerra systems communicate over the network to ensure the primary and secondary Data Movers are synchronized with respect to meta data, while the physical data is transported over the SRDF link. In order to ensure an up to date and consistent copy of the filesystems on the remote Celerra, the synchronous mode of SRDF operation is currently the only supported SRDF operational mode. Implementation of Celerra disaster recovery software requires modification of the standard Celerra configuration. SRDF has two modes of operation: active-passive and active-active. Active-passive (Unidirectional) SRDF support means that one Celerra provides active Data Mover access while a second (remote) Celerra provides all Data Movers as failover. Active-active (Bi-directional) SRDF support means that one Celerra can serve local needs while reserving some of its Data Movers for recovery of a remote Celerra, which reserve some of its Data Movers for recovery of the first Celerra . In addition, local failover Data Movers can be associated with Data Movers in the primary Symmetrix to ensure that local failover capability is initiated in the unlikely event there is a hardware related issue with a specific Data Mover. The mode of operation with SRDF/S is Active-Active. With active-active (SRDF/S only) support, one NS Series/NSX Gateway can serve local needs while reserving some of its Data Movers for recovery of a remote NS Series/NSX Gateway, which reserve some of its Data Movers for recovery of the first NS Series/Gateway.

NAS Foundations - 58

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Data Replication
Describe various data replication technologies provided for in EMC NAS, including:
SnapSure, TimeFinder/FS Celerra SRDF/A Celerra Replicator FileMover and DiskXtender

2006 EMC Corporation. All rights reserved.

NAS Foundations - 59

Besides SRDF, Celerra has several other mechanisms for data replication and recovery. We briefly review these in the next few slides.

NAS Foundations - 59

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra SnapSure Data Replication


Enables speedy recovery
Low volume activity, read-only applications Simple file undelete Incremental backup

Celerra

CLARiiON or Symmetrix

Logical point-in-time view of Celerra data


Works for all Celerra Implementations Saves disk space Maintains pointers to track changes to the primary filesystem Not a mirror; creation of specialized volumes (R1/R2, BCVs) not required Production Multiple Checkpoints for recovery of filesystem different point-in-time images
GUI Checkpoint schedule manipulation Checkpoint out of order delete Automatic mounting - NAS v5.2
2006 EMC Corporation. All rights reserved. NAS Foundations - 60

Checkpoint

Due to the business demands for high data availability and speedy recovery, there are many methodologies utilized to facilitate this requirement. The first methodology discussed is the SnapSure feature of the Celerra family. This methodology uses a logical point-intime view of a Production Filesystem to facilitate Incremental backup views of a Production File System, PFS, individual file recovery, and roll back of an entire filesystem to a previous point-in-time image. SnapSure maintains pointers to changes to the primary filesystem and reads data from the primary filesystem or a copy area. The copy area is defined as a meta-volume (SavVol). One of the obvious benefits of this solution is that it is storage array agnostic, i.e. works for all NAS DART implementations. This also means that there are no specialized volumes that need to be configured for this feature to function. Some other replication methodologies, such as SRDF and TimeFinder/FS, are dependent on the creation of Symmetrix Remote Data Facility and Business Continuity Volumes in the Symmetrix. SnapSure does not require any specialized volume creation and therefore works with any back-end storage array, CLARiiON, or Symmetrix. Multiple Checkpoints can be done on the Production Filesystem and thereby facilitate the ability to recover different pointin-time images of files or filesystems. Without using any other similar replication methodologies, i.e. Celerra Replicator, the currently supported maximum of Checkpoints per filesystem is 32. For ease of management, Checkpoints can be manipulated with the GUI management interfaces, along with the ability to schedule the frequency of the Checkpoints. Most Checkpoint technology is chronologically linked; however, the DARTv 5.2 solution supports out of order deletion of checkpoints, while maintaining SnapSure integrity. SnapSure Enhancements allow customers to delete a Checkpoint out of order. This feature allows customers to delete any Checkpoint instead of being constrained to having to delete Checkpoints from the oldest to maintain integrity. A customer may also delete an individual scheduled checkpoint instead of the entire schedule, and may refresh any checkpoint instead of the only the oldest. Checkpoints created in DART v5.2 are automatically mounted upon creation and maintenance of a hidden checkpoint directory in any subdirectory. This new hidden directory now also allows changing the default name (yyy_dd_hh_mm_ss_GMT) into something more administratively friendly.
NAS Foundations - 60

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra TimeFinder/FS Data Replication


Point-in-time copy of file system Provides an independent mirror copy of Celerra data for out-of-band business processes and support functions Provides read and write functionality independent of the original Requires Symmetrix storage Celerra controlled features
Point-in-time copies Dynamic mirroring Multiple BCVs Spans volumes Entire filesystem Backup and restore Data warehouses Live test data Batch jobs

Symmetrix

Celerra

Point-in-time copy FSA PFS PFS Copy

BCV = Business Continuance Volume

Applications

2006 EMC Corporation. All rights reserved.

A second Celerra data replication method that provides high availability and rapid recovery is the TimeFinder /FS feature. TimeFinder is a well established and mature product. In the Symmetrix Storage array, it is utilized to produce a standalone copy of a production volume. It uses a specially defined volume, called a Business Continuance Volume (BCV), to facilitate this functionality. As the Symmetrix Array is the only array currently able to define a BCV, TimeFinder/FS on the Celerra Family is currently restricted to implementations with Symmetrix only. The TimeFinder/FS implementation is different from a standard TimeFinder implementation. It allows users to copy filesystems into Business Continuance Volumes (BCVs) for a wide variety of purposes. How it happens is outlined below. TimeFinder/FS creates a point-in-time copy, or a dynamic mirror, of a filesystem. Integrated into the Celerra Control Station. The TimeFinder/FS option allows users to create filesystem copies (with only a brief suspension of access to the original filesystem). These copies permit independent read/write copies of data, useful for non-disruptive file backups, live copy test beds for new applications, and mirror copies of files for redundancy and business continuity. It facilitates backup and restore of older versions of a specific file, directory, (by mounting the snapshot filesystem and manually recovering the file or directory) or complete file system. It can also function in mirroring and continuous updates mode for an active filesystem. Filesystem copies require that the configuration of the Symmetrix system attached to the Celerra include BCVs. A BCV, which attaches to a standard volume on which a filesystem resides, provides the foundation for the filesystem copy. Filesystems can share BCVs, although the BCV remains dedicated to a volume. After issuing the single TimeFinder split command, the Celerra initiates the synchronization of the primary and BCV filesystem volumes. When the synchronization is almost complete (typically within two seconds), access to the primary filesystem is halted and Data Mover memory is flushed to ensure a consistent copy of the data is split off. When consistently synchronized, the split command is issued and the BCV is available for use. The BCV is mounted to a second Data Mover for backup or for use in secondary business processes.

NAS Foundations - 61

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

TimeFinder/FS Near Copy


Synchronous disk-based disaster recovery and data replication solution
Requires Symmetrix storage R1 / R2 data is synchronized for disaster recovery Read-only R2 data accessible via BCV for backup and testing purposes Synchronous SRDF as base technology
60 km distance
R1

Windows

UNIX

Data Network

SRDF
R2 BCV Celerra Symmetrix Symmetrix Celerra

ESCON / Fibre Channel


NAS Foundations - 62

2006 EMC Corporation. All rights reserved.

Combining the TimeFinder/FS product with another mature Symmetrix product, Symmetrix Remote Data Facility (SRDF), has enabled the TimeFinder/FS concept to be utilized for Disaster Recovery, as the stand alone copy of the data can now be synchronously updated at a remote site. It allows this solution to be utilized as a Data Replication and Disaster Recovery solution. This is known as TimeFinder/FS Near Copy, as the supported SRDF network distance between the two sites is 60 km (campus), due to the synchronous nature of the R2 volume updates. The Remote TimeFinder/FS Near Copy solution applies to environments that have a requirement for real-time, synchronous, disk-based recovery. Synchronous SRDF is used to maintain the R1/R2 pair. TimeFinder BCVs can be generated from the R2 and made available (read-only) to independent Data Movers in the remote Celerra. The Celerra at the remote site can make the content available for secondary business processes such as testing or backup. This solution works for environments with SRDF active-active mode, where R1s and R2s exist in both sites, as well as active-passive, where all the R1s are located in one site, with SRDF to a passive R2 only Symmetrix. Synchronous SRDF operates over ESCON/Fibre Channel and is limited to 60 km distances. The BCV at the R2 site is readonly, and restore must be done manually.

NAS Foundations - 62

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

TimeFinder/FS with SRDF Far Copy


Asynchronous data replication solution
Replicated point-in time copy of primary site filesystem Data replication time does not impact production file system performance Requires Symmetrix storage Uses SRDF Adaptive Copy as base technology Sites can be geographically distant
2006 EMC Corporation. All rights reserved. NAS Foundations - 63

Windows UNIX

Windows UNIX

R1 Site Network

R2 Site Network

STD R1/ BCV Celerra Symmetrix

SRDF
Adaptive Copy

STD

R2/ BCV Symmetrix Celerra

All remote copies of data are not designated for Disaster Recovery, but could be used for data replication, such as Web Site data replication, Inventory Replication, Off-site Backup facility, or Employee directories. To facilitate these kinds of solutions, where time taken to replicate the data does not impact the performance of the Production Filesystem, TimeFinder/FS Far Copy can be utilized. The Remote TimeFinder/FS Far Copy solution applies to environments that have a requirement for remote pointin-time copies of the filesystems beyond the typical distances associated with synchronous SRDFthat is, greater than 60 km. Adaptive SRDF is used to replicate the information over geographical distances. The readonly copy at the remote site can be made available for secondary business processes such as testing or backup. Implementation of this solution allows data to be replicated asynchronously over a very wide area to where it is needed. It does not affect the Production file system, PFS, as would the Celerra SRDF solution if the distances were over 60 kilometers, because a BCV copy of the PFS is first made and then the BCV is copied to the remote location, while the Production Filesystem continues serving data to the clients uninterruptedly. As this solution is dependent on the TimeFinder/FS, it is only supported with the Celerra/Symmetrix configuration. The process for performing this action is: 1. Create a R1/BCV of STD 2. Sync R1/BCV with R2/BCV over SRDF link 3. Restore R2/BCV to Local STD (read-only if the relationship between the R2BCV needs to be maintained) 4. Import File System on R2 Celerra

NAS Foundations - 63

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra SRDF/A
SRDF/A: Asynchronous Disaster Recovery
Delta Set-synchronized copies of volumes Extended distances

Network

SRDF/A

NS Series/NSX Gateway

NS Series/NSX Gateway

2006 EMC Corporation. All rights reserved.

NAS Foundations - 64

Celerra also integrates with the Symmetrix SRDF/A, asynchronous, product for extended distance support between the data centers. However data replication in this case is asynchronous and therefore is considered a Data Replication solution as opposed to Disaster Recovery. SRDF/A: an asynchronous mechanism for extending the distance with manual failover and failback. Asynchronous disaster recovery: Symmetrix SRDF/A technology connects a local NS Series/NSX Gateway and a remote NS Series/NSX Gateway to replicate the information over geographical distances. After a manual failover and recovery process, users gain access to file systems in the event that the local NS Series/NSX Gateway and/or the Symmetrix becomes unavailable. The mode of operation supported with this solution is Active-Passive With active-passive (SRDF/S or SRDF/A) support, one NS Series/NSX Gateway provides active Data Mover access, while a second (remote) NS Series/NSX Gateway provides all Data Movers for failover.

NAS Foundations - 64

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Replicator Data Replication


Point-in-time read-only filesystem copy over IP Windows Production filesystem available during replication Only sends changed data over wire (after initial synchronization) One-way link for single target content distribution Asynchronous Data Recovery Data recovery for CIFS data Requires Symmetrix storage
R1 Site Network

UNIX Windows

UNIX

R2 Site Network

Production Filesystem Primary SavVol Log

IP Network

Secondary Filesystem Secondary SavVol

Celerra/Symmetrix

Celerra NS

Replication Process

2006 EMC Corporation. All rights reserved.

NAS Foundations - 65

Celerra Replicator is an IP-based replication solution. Replication between a primary and a secondary filesystem can be on the same or a remote Celerra system. The data flow is described in the following steps. Celerra Replicator Events: Step 1: Manually synchronize production file system and secondary file system - Initial Copy: Before starting Celerra Replicator, a full copy of the Production File System, PFS, has to be made on the Secondary File System, SFS. Step 2: Any changes to the production filesystems are recorded in the log - Log Changes: On the primary, where PFS is mounted read/write, all CIFS block addresses of changed blocks are logged in a main memory log. Periodically, copy changes to SavVol: Transfer is triggered by special events. The trigger can be controlled by a user-defined policy or via an explicit request on the Control Station. Step 3: Remote replication copies log changes to primary SavVol, and begins movement to secondary SavVol - The transfer step comprises creating a local copy of all of the changes made to the log file and copying them to a local delta set called SavVol. The copy process is designed to copy a Checkpoint of the blocks, allowing the PFS to be modified during the copy. IP Transfer: A transfer is set up with the remote replica for the newly updated set of SavVol blocks. While the transfer is in process, read/write activity on the PFS is not halted. Instead, a new log area is set up to track subsequent changes. Concurrently with the copy process, the newly created delta set is transferred to the secondary over IP. In local replication (when the secondary Data Mover is in the same cabinet as the primary), no transfer is required. The delta set is accessible as a shared volume. Step 4: Remote SavVol changes are incorporated into secondary filesystems. Playback on Secondary: The delta set is made available to the secondary in its SavVol. When the delta set arrives at the SavVol of the secondary and has been flagged as valid, the secondary starts to replay the blocks from its local SavVol (a local copy of the replicated delta set) and apply the delta set to the SFS. This operation occurs transparently and with almost no interruption to SFS access. The Replicator feature is able to support data recovery for both CIFS and NFS by allowing the secondary filesystem (SFS) to be manually switched to read/write mode after the Replicator session has been stopped, manually or due to a destructive event. Note: Through version Dart 5.1, there is no automated re-synch or failback capability. This capability was added in Dart 5.2.
NAS Foundations - 65

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Replicator for iSCSI


Provides an application-consistent point-in-time replication solution for iSCSI
Replicates application-consistent iSCSI snaps (via VSS) Sends only changed data over the wire Managed by EMC Replication Manager/SE
Windows + Replication Manager/SE

Primary network

Secondary network

Production LUN available during replication; remote copy available for backup or repurposing Cost-effective, long-distance data recovery
In the event of failure, the LUN copy can be made available as read/write Changes to the LUN copy can be reapplied to the production LUN on failback
2006 EMC Corporation. All rights reserved.

Production LUN

LUN Copy

IP network
iSCSI Snaps

4
iSCSI Snaps

NSX

NS700
NAS Foundations - 66

As iSCSI is deployed within the infrastructure, protection of files that are accessed over iSCSI is critical. EMC Celerra Replicator now supports application-consistent iSCSI replication. Managed by EMC Replication Manager/SE, only the changed data is sent to the replica, improving network efficiencies. The copy can be made read/write in the event of a failure of the primary. Celerra Replicator for iSCSI support is currently limited to Windows, and the failback process is manual.

NAS Foundations - 66

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra FileMover and DiskXtender ILM Solution FileMover is an API, when used in conjunction with a policy engine, migration software and a secondary storage infrastructure Data supports concept of ILM when combined with DiskXtender software

2006 EMC Corporation. All rights reserved.

NAS Foundations - 67

ILM = Information Lifecycle Management FileMover is an API, when used in conjunction with a policy engine, migration software and a secondary storage infrastructure, e.g. FSMNAS (File System Manager for NAS), allows the right data to be placed in the right place at the right time When combined with DiskXtender software the data can be further migrated to a Centera, which supports the concept of ILM, Information Lifecycle Management, and provides a complete data management infrastructure using EMC network attached storage solutions

NAS Foundations - 67

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Backup Options


Describe backup options for the EMC NAS family

2006 EMC Corporation. All rights reserved.

NAS Foundations - 68

Celerra backup options are reviewed in this section.

NAS Foundations - 68

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Tape-Based Backup and Restore Options


Most backup utilities NFS / CIFS mounts over the client network or separate sub-network
Celerra Network DATA NetWorker Tape

Network Backup

NDMP Backup
(LAN-less (LANlocal backup)

VERITAS NetBackup Legato Networker CommVault Galaxy HP OmniBack Atempo Time Navigator

Celerra Data Network NetWorker Symmetrix or CLARiiON Tape

NDMP2D Backup

NDMP enabled backup utilities Backup local Celerra data to specially configured disks within the array
NetWorker

Network Celerra

2006 EMC Corporation. All rights reserved.

NAS is a fast growing market. Many NAS implementations have mission-critical data availability requirements and this is what Celerra does best. Fast and efficient backup and restore is an absolute requirement. There are a number of options as far as backup and restore is concerned. Network backups entail simply mounting the filesystems across the network and backing up to the backup server. EMC Data Manager, and most backup utilities, support this option. EDM also preserves the bi-lingual (CIFS and NFS) information with network backups. NDMP backups only use the LAN for control information (LAN-less) and the data is transferred to the local backup device. EDM, VERITAS NetBackup, Legato Networker, CommVault Galaxy, HP OmniBack, and Atempo Time Navigator support Celerra NDMP backups. NDMP backups preserve the bi-lingual file information.

NAS Foundations - 69

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra Network Backup


Third-party backup products
CIFS access through shares NFS access through mounts
Tape Library Celerra CLARiiON or Symmetrix Network

Data Flow Control Information Flow

2006 EMC Corporation. All rights reserved.

NAS Foundations - 70

Network backups utilize a backup server that mounts the remote filesystems and transfers them to tape. The data path includes the Celerra Data Mover, the network, and the backup server. Most third-party backup products support CIFS network backups through remote file shares and NFS network backups using remote file mounts. Third-party backups should not be used for multi-protocol files. For example, if they are backed up over a CIFS share, they lose any NFS attributes and permissions upon restoration.

NAS Foundations - 70

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Celerra NDMP Backup


Third-party NDMP VERITAS NetBackup Legato NetWorker CommVault Galaxy HP OmniBack Atempo Time Navigator Celerra backs up data to directly attached tape library unit (TLU) Backup is performed by client running NDMP-compliant ISV software No LAN performance impact: only control data goes via the LAN Multi-protocol support: both CIFS and NFS filesystem attributes
2006 EMC Corporation. All rights reserved.

Production Network Client with NDMP backup software Data Mover NDMP Server

Tape Library Unit

Celerra

CLARiiON or Symmetrix

Data Flow Control Information Flow

NAS Foundations - 71

NDMP is an industry standard LAN-less backup for NAS devices. NDMP backups only use the LAN for control information and the data is transferred to the local backup device. VERITAS NetBackup, Legato Networker, Commvault Galaxy, HP OmniBack, and Atempo Time Navigator also support Celerra NDMP backups. Backup activity can be localized to a single backup Data Mover, thus requiring only one Data Mover be physically attached to the TLU (tape library unit). This option is implemented through TimeFinder/FS. Filesystems are split off and mounted to the backup Data Mover and backed up with no impact to the primary filesystem. Tape library units are connected to a Data Mover via a SCSI interface. Backup traffic is offloaded from the network and allows for dual accessed filesystems to be backed up, preserving both permission structures on the filesystem.

NAS Foundations - 71

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

NDMP2D: Network Data Management Protocol to Disk


A software-based tape target that resides on a Data Mover that emulates SCSI Tape Library Units (TLU) and tape drives Emulation allows NDMP to work with disk storage, rather than tape storage. NDMP2D can be configured and operated from NDMP Client S/W as if it were a physical TLU attached to a Data Mover

NDMP2D

Network

PFS
NetWorker NDMP Control Meta Data Celerra

2006 EMC Corporation. All rights reserved.

NAS Foundations - 72

This feature is supported in DART v5.4 and later versions. It allows for the creation of a tape target on a Celerra using Fibre Channel or ATA drives. (PFS = Production File System)

NAS Foundations - 72

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Definition of NDMP2D
A software-based tape target that resides on a Data Mover
Emulates a physical tape library unit (TLU)

Allows compatibility between NDMP & disk-based storage


Backup software responsible for managing TLUs cannot distinguish between NDMP2D and a physical TLU.

2006 EMC Corporation. All rights reserved.

NAS Foundations - 73

NDMP2D is a filesystem-based storage sub-system that emulates SCSI Tape Library Units (TLUs) and tape drives. Applications communicating with NDMP2D get the impression of communicating with a standard commercial TLU. NDMP2D maximizes the attractiveness of NDMP-based PAX, VBB, VLC (aka tape silvering) and other EMC backup products. NDMP2D could eliminate the need for physical tape drives and provide a single, centralized data repository.

NAS Foundations - 73

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Celerra Database Solutions


Describe Celerra databases and business application requirements and advantages

2006 EMC Corporation. All rights reserved.

NAS Foundations - 74

Database solutions for the Celerra Family are reviewed in this section.

NAS Foundations - 74

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Celerra NAS Database Solutions


Database applications have traditionally been implemented in a SAN architecture
Primary reason was the deterministic performance of a SAN NAS was considered appropriate only for applications where the transaction rate was low with sequential I/O, flat file like access and or the performance was not constrained

Things have changed now


NAS is well suited for different types of I/O patterns random and sequential, for both types of applications OLTP and DSS Size of the database is not a determinant for storing DB on NAS Databases on NAS is an established market NAS is expected to grow at a faster rate than fiber channel Oracle Corporation, world leader in databases, has endorsed NAS for databases
2006 EMC Corporation. All rights reserved. NAS Foundations - 75

Databases on EMC NAS offer the following advantages: Lower TCO: Lower acquisition, administration and maintenance costs than equivalent DAS or SAN Greater Manageability: Easier Implementation, provisioning and volume management High Availability: NAS devices that implement a clustering architecture provide very high levels of data availability Increased Flexibility: Easy to re-deploy data with other servers. Simplicity in making databases, or copies of database, available ( via remounts) to other servers. Improved Protection: Integrated backup and availability High Performance: Superior price/performance for databases Benefits of EMC ILM: Ability to employ Tiered Storage.

NAS Foundations - 75

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Celerra NAS Database Solutions


Changes that made NAS Feasible for Databases Network Enhancement:
Adaptation of Gig Ethernet Network Data Separation

Increases in Host Processing Capabilities:


Increases in Mhz and Processing Power Decreases in cost of host RAM

Industry Knowledge of Managing Database Objects:


Technology Knowledge Sharing: Web Training, Metalink, etc. DBAs are getting better at managing all database objects

2006 EMC Corporation. All rights reserved.

NAS Foundations - 76

When it comes to backing up the database, backup/recovery processes are designed to run on the primary site during normal operations. EMC Snapsure technology is used in the backup process of data files and log files. The reason in using Snapsure is to create a read-only image of the entire production file system , which also includes data and log files. The idea here is to balance the performance and availability impact of backup processing with application processing.

NAS Foundations - 76

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Celerra NAS Database Solutions


Metrics that determine if NAS is a fit: Transaction Rates or IOPS Bandwidth or Throughput
2 GB/s DSS Workloads

BW GB/s

NFS/iSCSI Capability

Commercial Apps

OLTP Workloads

IOPS
2006 EMC Corporation. All rights reserved.

30k
NAS Foundations - 77

OLTP (Online Transaction Processing) workloads have lower throughput and higher IOPS (IOs per second) need. They have random reads/writes. DSS (Decision Support Systems) or Datawarehouse workloads have higher throughput and lower IOPS needs. They have sequential IOs. NAS (NFS+iSCSI) IOPS and throughput can meet most of the commercial application needs. Just as Fibre Channel (FC) pulls significantly ahead in the performance race with 4Gb/sec speed, IP/Ethernet is about to leapfrog FC with 10Gb/sec Ethernet over copper wire. Analysts see two uses for the greater storage networking speeds: disk-to-disk backup and disk archiving, and storage port consolidation.

NAS Foundations - 77

Copyright 2006 EMC Corporation. Do not Copy - All Rights Reserved.

Course Summary
Key points covered in this course: Concepts and value of Network Attached Storage Environmental components of NAS EMC NAS Platforms and their differences Celerra Software Features Celerra Management Software offerings Windows Specific Options with respect to EMC NAS environments NAS Business Continuity Options with respect to the various EMC NAS platforms Celerra NAS Database Solutions

2006 EMC Corporation. All rights reserved.

NAS Foundations - 78

These are the key points covered in this training. Please take a moment to review them.

NAS Foundations - 78

You might also like