You are on page 1of 226

Table of Contents

Volume 1
Course Introduction
Overview
Prerequisites
Learner Skills and Knowledge
Course Goal and Objectives
Course Flow
Additional References
Cisco Glossary of Terms

Cisco Wide Area Application Services

1
1
1
2
3
4
5
5

1-1

Overview
Module Objectives

1-1
1-1

Cisco WAAS Overview

1-3

Overview
Objectives
Application Delivery Initiatives
Cisco WAAS Introduction
Cisco WAAS Optimizations
Cisco WAE Platforms and Software Licensing
Summary

WAN Optimization Technical Overview


Overview
Objectives
Application Performance Barriers
Introduction to TCP
Transport Flow Optimization
Advanced Compression
Summary

Application Acceleration Technical Overview


Overview
Objectives
The Need for Application Acceleration
CIFS Acceleration
Connectivity Directive
Integrated Print Services
Summary
Module Summary
Module Self-Check
Module Self-Check Answer Key

Designing Cisco WAAS Solutions


Overview
Module Objectives

Network Design, Interception, and Interoperability


Overview
Objectives
Physical Inline Deployment
Off-Path Network Deployment
Interception Using WCCPv2
IOS Routing Platforms
Switching Platforms
Security Platforms
WCCPv2
Interception Using Policy-Based Routing
Data Center Deployment Using ACE

1-3
1-3
1-4
1-10
1-25
1-29
1-36

1-37
1-37
1-37
1-38
1-42
1-50
1-63
1-80

1-81
1-81
1-81
1-82
1-85
1-87
1-99
1-103
1-105
1-107
1-111

2-1
2-1
2-1

2-3
2-3
2-3
2-4
2-16
2-29
2-29
2-30
2-30
2-32
2-38
2-43

Automatic Discovery
Asymmetric Routing
Network Transparency
Summary

Performance, Scalability, and Capacity Sizing


Overview
Objectives
Cisco WAAS Design Fundamentals
Cisco WAE Device Positioning
WCCPv2 Design Considerations
Summary
Module Summary
Module Self-Check
Module Self-Check Answer Key

ii

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2-47
2-58
2-66
2-71

2-73
2-73
2-73
2-74
2-91
2-94
2-99
2-101
2-103
2-105

2007 Cisco Systems, Inc.

WAAS

Course Introduction
Overview
Cisco Wide Area Application Services Technical Training (WAAS) is a three-day course that
introduces Cisco WAAS to pre-sales and post-sales engineers, as well as server, storage,
application and network IT managers. This course teaches the business value of WAN
optimization and application acceleration technologies. You will learn how to design a basic
WAAS deployment and configure Cisco WAN Application Engine (WAE) devices, including
WAN optimization and application acceleration components.
In the lab, you will install and configure WAE devices, configure and test application traffic
policies, and configure acceleration for CIFS.

Prerequisites
You should have a fundamental knowledge of data networking and Microsoft Windows
networking technologies. This course includes appendices with key supplemental information
about Microsoft Windows Networking, the Common Internet File System (CIFS) protocol, and
other relevant course information.

Learner Skills and Knowledge


Basic understanding of data networking concepts and
technologies
Basic understanding of Microsoft Windows networking concepts
and technologies

2006 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.7Intro-3

Learner Skills and Knowledge


This subtopic lists the skills and knowledge that learners must possess to benefit fully from the
course.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Course Goal and Objectives


This topic describes the course goal and objectives.

Course Goal:
Design and deploy a solution using Cisco WAAS that
improves application performance over the WAN while
enabling infrastructure consolidation. Understand Cisco
WAAS optimization technology, design considerations,
network integration, and system troubleshooting aspects.

Cisco Wide Area Application Services

2006 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.7Intro-5

After you complete this course, you will be able to design and deploy a WAAS configuration
that includes Transport Flow Optimizations, application traffic policies, compression, and file
and print services.
Upon completing this course, you will be able to meet these objectives:

Explain the business value of WAN optimization and application acceleration technologies
and understand the technologies employed by Cisco Wide Area Application Services to
enable consolidation while improving application performance over the WAN

Design Cisco WAAS solutions

Describe Cisco WAAS implementation, integration, and management

Troubleshoot Cisco WAAS installations, including platform and network connectivity


issues, network interception issues, WAN optimization issues, and application acceleration
issues

Explain Microsoft Windows Networking concepts, and describe the training lab and
challenge topologies

2007 Cisco Systems, Inc.

Course Introduction

Course Flow
This topic presents the suggested flow of the course materials.

Course Flow
Day 1

A
M

Day 2

Day 3

Implementation and
Integration

Lab: Configuring
Application
Acceleration

Course Introduction
Introduction to Cisco
WAAS

Lunch
P
M

Designing Cisco WAAS


Solutions

Lab: Initial Cisco


WAAS Configuration

Lab: Designing
WAAS Solutions

Lab: Configuring WAN


Optimization

2006 Cisco Systems, Inc. All rights reserved.

Troubleshooting
Cisco WAAS

WAAS v4.0.7Intro-7

The schedule reflects the recommended structure for this course. This structure allows enough
time for the instructor to present the course information and for you to work through the lab
activities. The exact timing of the subject materials and labs depends on the pace of your
specific class.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Additional References
This topic presents the Cisco icons and symbols that are used in this course, as well as
information on where to find additional technical references.

Cisco Icons and Symbols

NAS

Workstation

Cisco WAE

Application Server

IP Router

NAS Filer

Ethernet Switch

Disk storage
subsystem

Multilayer Switches

Tape storage
subsystem

Firewall

2006 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.7Intro-9

Cisco Glossary of Terms


These are the icons and symbols you will see throughout this course.
For additional information on Cisco terminology, refer to the Cisco Internetworking Terms and
Acronyms glossary of terms at
http://www.cisco.com/univercd/cc/td/doc/cisintwk/ita/index.htm.

2007 Cisco Systems, Inc.

Course Introduction

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module 1

Cisco Wide Area Application


Services
Overview
IT organizations are tasked with two conflicting challenges. The first challenge is to provide
acceptable service levels for applications, content, and data for all users throughout the globally
distributed enterprise. The second challenge is to simplify and consolidate the infrastructure to
minimize costs associated with procurement, management, and data protection. Cisco Wide
Area Application Services (WAAS) helps to bridge the gap between centralized and
decentralized infrastructures by providing the tools necessary to ensure high performance
access to a centralized pool of resources. This module provides an introduction to Cisco WAAS
and explains how the technologies within the Cisco WAAS solution help to enable
consolidation of distributed infrastructure while maintaining performance expectations and
improving performance for already-centralized applications and services.

Module Objectives
Upon completing this module, you will be able to explain the business value of WAN
optimization and application acceleration technologies and understand the technologies
employed by Cisco Wide Area Application Services to enable consolidation while improving
application performance over the WAN. This includes being able to meet these objectives:

Describe the business value of WAN optimization and application acceleration


technologies

Explain Cisco WAAS WAN optimization features

Explain how Cisco WAAS application-specific acceleration improves performance for file
and print protocols

1-2

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Lesson 1

Cisco WAAS Overview


Overview
IT organizations are tasked with two conflicting challenges. The first challenge is to provide
acceptable service levels for applications, content, and data for all users throughout the globally
distributed enterprise. The second challenge is to simplify the infrastructure to minimize costs
associated with procurement, management, and data protection. Cisco Wide Area Application
Services (WAAS) helps to bridge the gap between centralized and decentralized infrastructure
by providing IT with the tools necessary to ensure high performance access to a centralized
pool of resources. This lesson provides an introduction to the Cisco WAAS solution.

Objectives
Upon completing this lesson, you will be able to describe the business value of WAN
optimization and application acceleration technologies. This includes being able to meet these
objectives:

Describe the infrastructure and application challenges faced by IT organizations today

Define how Cisco WAAS technologies help to enable application delivery and
infrastructure consolidation

Describe the optimization technologies provided by Cisco WAAS

Describe the Cisco WAE family of appliances and the router-integrated network module,
along with the software licenses that are necessary to enable WAAS functionality

Application Delivery Initiatives


This topic explains the key drivers and application challenges that customers face when
consolidating infrastructure.

The Application Delivery Problem


Increasingly distributed workforce
drives need for distribution of IT
resources to remote locations:

Data protection, availability,


compliance, and management
drives need for consolidation

Enable productivity

Fewer devices to manage

Drive revenue and profits

Fewer points to protect

Remote Offices

Distribution of
Resources
Data Center

Regional Offices
Data center
consolidation
Home Offices

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-4

IT faces two opposing forces. One force is the need to provide high performance access to files,
content, and applications to a globally distributed workforce, which is necessary to ensure
productivity and job satisfaction. This force causes IT to push resources, such as servers,
storage, and applications, out to the enterprise edge. The other force is driving consolidation,
because maintaining an IT infrastructure in a distributed deployment model is an expensive
task.
One force drives the movement of resources out to the enterprise edge to support the remote
users, while the other force drives the movement of resources inward toward the data center in
an effort to control costs, simplify management, protect data, and improve compliance with
regulations.
One factor that is commonly overlooked by both forces is the over-worked WAN. In most
cases, the WAN is running beyond capacity with voice traffic and other business critical
applications. The WAN presents a variety of challenges in meeting the needs of remote users
and in fulfilling the requirements of the IT organization when consolidating the remote office
infrastructure.
Explain the business value of WAN optimization and application acceleration technologies and
understand the technologies employed by Cisco Wide Area Application Services (WAAS) to
enable consolidation while improving application performance over the WAN
Another challenge for most global organizations is the rise in the use of web-based applications
happened just as there was a major push toward data center and management consolidation.
The web is well-suited for centralized deployment and it allows organizations to manage large
scale operations from a single site. The downside is that it quickly puts significant strain on the
network, especially the WAN. Many users might work in branch or remote offices, but it is just
1-4

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

as likely that they might work from home, on the road, or that you need business partners to
interact with your applications, as well. The vast majority of enterprise-class applications were
not built operate in this diffuse environment. Simply put, application interactions that work fine
on a campus LAN are wildly inefficient across the WAN. It is not that these are poorly written
applications; its just that they are being asked to serve in an environment for which they were
never meant. And where there are well established solutions to help distribute website content
and cache data, applications are filled with dynamic and changing information that requires a
new approach.
The implications of poor performance are profound. Users refuse to access corporate portals
and continue to use outdated information. Expectations are set that the application is slow,
causing employee stress over the inability to maintain productivity levels, leading to eroding
job satisfaction. Major application rollouts are stalled because managers refuse to give up on
outdated systems that at least work. Complaints to the help desk escalate, and IT becomes the
problem rather than the solution. Who suffers when new systems arent adopted, arent used,
and arent productive?

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-5

Typical Distributed Enterprise


Expensive distributed IT
infrastructure:
File and print servers
Email servers
Tape backup

Application delivery woes:


Congested WAN
Bandwidth and latency
Poor productivity

WAN

Data protection risks:


Failing backups
Costly off-site vaulting
Compliance
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-5

Many organizations have infrastructure silos in each of their remote, branch, and regional
offices. These silos are typically carbon-copies of the infrastructure in the data center; including
file servers, print servers, backup servers, application servers, e-mail servers, web servers,
storage infrastructure, and more. In any location where storage capacity is deployed with active
data, that data must be protected with disk drives, tapes drives, tape libraries, backup software,
service with an off-site vaulting company, and perhaps even replication. The remote office
infrastructure is costly to maintain.
The goal of the typical distributed enterprise is to consolidate as much of this infrastructure as
possible into the data center, without overloading the WAN, and without compromising the
performance expectations of remote office users who are accustomed to working with local
resources.

1-6

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

The WAN Is A Barrier To Consolidation


Applications are
designed for LAN
environments:

Round Trip Time (RTT) ~ 0mS


Client

Switch

Server

High bandwidth
Low latency
Reliability

WAN characteristics
hinder consolidation:
Already congested

Round Trip Time (RTT) ~ many milliseconds


Client

Switch

Routed Network

Switch Server

Low bandwidth
Latency
Packet Loss
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-6

Applications often do not perform well in WAN environments that are built in a utopian
environment like that shown at the top of the figure. Application developers and operating
system vendors tend to put servers adjacent to the clients on the same LAN when developing
their products. LANs provide a great environment for application development because the
challenges associated with a WAN are not encountered. LANs tend to have high bandwidth,
high reliability, low latency, low congestion, and low packet loss. There are few performance
barriers on the LAN.
In contrast, the WAN is generally low in bandwidth, high in latency, unreliable, and high in
packet loss. Separating the user from the server with a low performance long-distance
unreliable network can wreak havoc with application performance. In comparison to the
utopian LAN environment, WAN applications tend to fall apart for a variety of reasons,
including insufficient bandwidth, unreliable connections or lack of connection stability, packet
loss or congestion, retransmission, transport latency, and application latency. Some applications
require many hundreds of roundtrips to complete a trivial operation, to perform a Common
Internet File System (CIFS) file open for example. Web applications are also affected, requiring
the exchange of hundreds of operations. With each message that is exchanged, the user
application pays a roundtrip penalty of the WAN latency per operation. With 1,000 operations
that must ping-pong in a 40 millisecond WAN, this equates to about 40 seconds of response
time.
Unsatisfactory WAN application performance is the primary reason that most organizations do
not consolidate costly remote office infrastructures.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-7

Addressing the WAN Challenge


Source
Latency

Need

Application acceleration

Protocol optimization

Improved application response time on


congested links by reducing the amount of
data sent across the WAN

Object caching

Compression, suppression

Improved network throughput (total amount


of data) by improving transport behavior

TCP optimization

Adaptive congestion mgmt

Physical integration into existing platforms

Router modules, linecards

Compliance with network functions

Feature interoperability

Replacement for services that branch office


servers provide

Centrally managed remote


services interface

Align network resources with business


priority and application requirements

Quality of Service

Advanced Routing

NetFlow, RMON, monitoring tools

IP Service Level Agreement

Bandwidth Utilization

Transport Throughput

Network Integration

Administrative Traffic

Network Control

Network Visibility

Technology

Reduced number of network roundtrips


caused by chatty application protocols

Understand network utilization and


application performance metrics

React and respond to changes in network


conditions

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-7

The most silent yet largest detractor of application performance over the WAN is latency.
Latency is problematic because of the volume of message traffic that must be sent and received.
Some messages are very small, yet even with substantial compression and flow optimizations,
these messages must be exchanged between the client and the server to maintain protocol
correctness, data integrity, and so on. The best way to mitigate latency is to deploy intelligent
protocol optimizations, otherwise known as application acceleration, in the remote office. This
is done on a device that understands the application protocol well enough to make decisions on
how best to handle application traffic as it occurs and can closely mimic the performance of a
local server in many cases. On a per-message basis, this application accelerator could examine
messages to determine they can be suppressed, or handled locally. If the request is for data, the
application accelerator could determine if the data is best served from cache (if the object is
valid, the user is authenticated, and the appropriate state is applied against the object on the
origin server), or if a message must be sent to the origin server to maintain proper protocol
semantics.
Bandwidth utilization is another application performance killer. Transferring a file multiple
times can consume significant WAN bandwidth. If a validated copy of a file or other object is
stored locally in an application cache, it can be served to the user without using the WAN.
Application caching is typically tied to an application accelerator and is specific to that
application, but there are compression techniques that can be applied at the transport layer that
are application agnostic. One of these techniques is standards-based compression. Another
technique is called Data Redundancy Elimination (DRE), which is an advanced form of
suppressing the transmission of redundant network byte streams. Compression and application
caching provide another way to improve application performance by minimizing the amount of
data that must traverse the network. Minimizing the amount of data on the network improves
response time and leads to better application performance, while also freeing up network
resources for other applications.
The next application performance barrier in a WAN environment is transport throughput.
Application protocols run on top of a transport mechanism that provides connection-oriented or
non-connection-oriented delivery of data. In many cases, enterprise applications use TCP for its
1-8

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

inherent reliability. Although it is reliable, TCP presents performance barriers of its own. If
TCP could be optimized to perform better in WAN environments, then application throughput,
response time, and the user experience can all show improvement, due to better utilization of
existing network capacity and better response to network conditions.
Two factors should be considered with all consolidation-enabling solutions. The first factor is
network integration. Consolidation solutions should not disrupt the operation of existing
network features such as QOS, access-lists, NetFlow, firewall policies, and others. By
integrating with the network in a logical manner, that is maintaining service transparency
(preserving information in packets that the network needs to make intelligent feature-oriented
decisions), fundamental network-layer optimizations can continue to operate in the face of
application acceleration or WAN optimization. Physical integration allows such technology to
be directly integrated into existing network devices, thereby providing a far more effective
Total Cost of Ownership (TCO) and Return On Investment (ROI) model.
When possible, administrative services such as print services should be centrally managed but
locally deployed when possible in remote sites. This keeps such administrative traffic from
needing to traverse the WAN.
The network should be aligned with business priority and application requirements to ensure
the appropriate handling of traffic. Quality of Service, or QoS, for instance, allows network
administrators to configure network behavior in specific ways for specific applications. As all
applications are not created equal, the network must be prepared to handle traffic in different
ways based on how the application needs to be handled. This involves classification of data
(seeing what application it is and who is talking to who, among other metrics), pre-queuing
operations (immediate actions, such as marking, dropping, or policing), queuing and scheduling
(ensuring the appropriate level of service and capacity are assigned to the flow), and postqueuing optimizations (link fragmentation and interleaving, packet header compression). This
set of four functions is known as the QoS Behavioral Model, which relies on visibility (service
transparency), should acceleration technology be deployed to fully function. Also, the network
should be able to make on-the-fly path routing decisions (advanced routing) to ensure that the
right path is taken for the right flows. This includes policy-based routing (PBR), optimized
edge routing (OER), and more.
Finally, the network should be visible. That is, administrators need to know how the network is
performing, being used, and when network characteristics are performing as expected.
Technologies such as NetFlow and collection or analysis tools allow administrators to see how
the network is being utilized, top talkers, and more. Functions such as IP Service Level
Agreements (IP SLAs) allow the network to alert administrators when conditions exceed
thresholds, and furthermore allow the network to even react when such events occur.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-9

Cisco WAAS Introduction


This topic defines the market positioning for Cisco WAAS technologies, and how the
technologies within Cisco WAAS help IT organizations to consolidate infrastructure and
improve application performance for centralized applications.

Cisco Provides a Comprehensive Solution


Application Acceleration

Reduced WAN
Expenses
WAN
Optimization

Session-based
Compression

Protocol
Optimization

Data Redundancy
Elimination

Object Caching

TCP Flow
Optimization

Easily
Manage WAN

Consolidated
Branch
Wide-Area
File Services

Local Services

NetFlow
Performance
Visibility
Monitoring
Monitor and
IP SLAs
Provision

Cisco WAAS
Integrated with
Cisco IOS

Dynamic
Auto-Discovery

Queuing
Shaping
Policing
OER

Network Transparency
Compliance

QoS and
Control

Applications
Meet Goals

Preserve Network Services

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-9

Cisco WAAS provides the cornerstone for a holistic solution that provides an optimized
framework for application delivery and infrastructure consolidation. While Cisco WAAS is one
piece of this framework, the other piece is the network infrastructure itself. When combining
powerful technologies from both Cisco WAAS and the network together, IT organizations find
themselves well poised to better leverage existing network resources, improve performance,
consolidate infrastructure, simplify, and control costs. Each of the items shown in the solution
diagram above directly correlates to a problem area mentioned in the previous slide relative to
application performance challenges over the WAN.
Network technologies that help address performance barriers created by the WAN (not an
exhaustive list):

1-10

NetFlow: NetFlow gathers data about flows on the network from specific points within the
network. This data can then be sent to a collector and analyzed. NetFlow provides the
foundation for visibility into how the network is performing and how the network is being
used.

IP Service Level Agreements: IP SLAs are measurements taken from within the network
and actions taken when metrics are violated. For instance, IP SLAs can be used to monitor
the latency from point-to-point and trigger a reaction should the latency exceed a specific
threshold.

Optimized Edge Routing: OER is a function employed at the edge of the network
whereby a specific network path can be selected based on metrics such as latency,
bandwidth, and loss. OER can be combined with other optimized routing capabilities, such
as PBR, to ensure that high priority flows utilize the best available network path, and can

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

also be combined with high availability technologies such as Hot Standby Router Protocol
(HSRP), Virtual Router Redundancy Protocol (VRRP), and Gateway Load Balancing
Protocol (GLBP).

Quality of Service (QoS): QoS is an architecture rather than a feature. QoS provides a
series of components that allow the network to identify incoming flows and understand
what they are and how to respond (classification); and to perform immediate actions
against flows before network resources are consumed, such as dropping packets, policing,
marking packets, estimating bandwidth requirements (pre-queuing operations); and to
queue traffic according to application requirements and business priority, scheduling queue
service, and shaping (queuing); and to apply optimization to packets being serviced to
ensure performance (post-queuing operations).

Cisco WAAS technologies that help address performance barriers created by the WAN:

WAN optimization: Cisco WAAS overcomes performance limitations caused by WAN


bandwidth, latency, packet loss, and TCP through a series of optimizations that includes
Data Redundancy Elimination (DRE), Persistent LZ Compression, and Transport Flow
Optimization (TFO). WAN optimization is generic in nature and benefits a broad range of
applications concurrently.

Application Acceleration: Cisco WAAS not only overcomes unruly WAN conditions, but
also overcomes unruly conditions that exist in applications and application protocols
themselves as well. This capability of latency reduction, bandwidth utilization reduction,
and response time improvement is called application acceleration. Whereas WAN
optimization is application-agnostic and can benefit almost any application, application
acceleration is specific to a given application protocol.

Wide Area File Services (WAFS): WAFS is in many ways a component of application
acceleration in that it helps to improve user performance when accessing remote file
servers over the WAN, however, WAFS also contains other distinctive characteristics that
fall outside of the realm of application acceleration. These include disconnected mode of
operation, which helps to ensure some level of productivity during network outage
scenarios, as well as local services such as print services, which helps to keep unruly
administrative traffic off of the WAN.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-11

Cisco WAAS Overcomes the WAN


Cisco WAAS is a solution that leverages the hardware
footprint of the WAE in the remote office and in the
data center to overcome application performance
problems in WAN environments.
Remote Office
Data Center
Op
tim
ize
WAN

Remote Office

ed
Optimiz

dC

onn
ect

Connect

ion
s

ions

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-10

Cisco WAAS is a powerful new solution that overcomes the challenges presented by the WAN.
Cisco WAAS is a software package that runs on the Wide Area Application Engine (WAE) that
transparently integrates with the network to optimize applications without client, server, or
network feature changes.
A WAE is deployed in each remote office, regional office, and data center of the enterprise.
With Cisco WAAS, flows to be optimized are transparently redirected to the Cisco WAE,
which overcomes the restrictions presented by the WAN, including bandwidth disparity, packet
loss, congestion, and latency. Cisco WAAS enables application flows to overcome restrictive
WAN characteristics to enable the consolidation of distributed servers, save precious WAN
bandwidth, and improve the performance of applications that are already centralized.
As shown in the diagram, Cisco WAAS can be deployed using appliances that attach to the
network as nodes, or can be deployed within the Cisco Integrated Services Router.
Alternatively, Cisco WAAS can be deployed on devices that are deployed physically in-path.

1-12

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

The Problem with Tunneling


Other WAN optimization solutions use GRE tunnels
to encapsulate optimized WAN connections.
Tunneling hides Layer 3-7 headers.
This breaks many traffic monitoring and optimization
applications, like IDS and QoS.
IDS, QoS, Shaping, Policing, NBAR
NetFlow, NAM, PVM, IP SLA, OER

WAN

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-11

Competing WAN optimization solutions use GRE tunnels to encapsulate optimized WAN
connections. The problem with this approach is that tunneling prevents the network from
having visibility into Layer 37 data. Without visibility into Layer 37 data, functions like
Quality of Service (QoS), Network-Based Application Recognition (NBAR), NetFlow, and
Intrusion Detection Systems (IDS) cannot function. This means that the existing traffic
monitoring and optimization infrastructure is broken.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-13

WAAS Network Transparency


Cisco WAAS uses a TCP proxy architecture instead of
GRE tunnels.
WAAS integrates transparently with IOS features for
network control, optimization, visibility, and monitoring,
providing a holistic solution.
IDS, QoS, Shaping, Policing, NBAR
NetFlow, NAM, PVM, IP SLA, OER

WAN

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-12

Cisco WAAS provides a transparent framework for accelerating applications. Cisco WAAS
uses a TCP proxy architecture that preserves information critical to network feature operation,
such as Layer 3-7 packet header information. With its transparent optimization architecture,
Cisco WAAS is uniquely positioned to offer feature compatibility with functions that are
already used within the network, including:

Quality of Service: QoS policies provide traffic classification, prioritization, scheduling,


queuing, policing, and shaping.

Network-Based Application Recognition (NBAR): NBAR provides protocol and


application discovery.

Access Control Lists (ACLs): ACLs permit or deny traffic based on identification.

NetFlow: NetFlow provides statistics on nodes communicating on the network, and the
applications being used.

Firewall policies: Firewall policies are used to prevent malicious traffic from entering or
exiting a network.

By providing a transparent architecture, Cisco WAAS is able to provide interoperability with


end-to-end performance analysis systems such as Cisco Network Application Performance
Analysis (NAPA), and Performance Visibility Manager (PVM), as well as other 3rd party
products.

1-14

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Enables Consolidation


Cisco WAAS features include:

Transparent integration
Robust optimizations
Auto discovery
Policy-based configuration

Consolidation benefits include:


Remove costly branch servers
Centralize data protection
Save WAN resources

WAN

Improvements include:
Application acceleration
WAN optimization
Local infrastructure services

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-13

Cisco WAAS is designed to help consolidate infrastructure from remote offices into the data
center. Cisco WAAS is characterized by its ability to:

Integrate transparently into the existing infrastructure

Understand application protocols and how to optimize those applications

Provide compression and flow optimizations to improve delivery of data that must traverse
the WAN

Simplify consolidation by providing policy-based configuration and automatic discovery

Aside from cost savings, the primary goal of infrastructure consolidation is to give users the
same level of access that is available with a local infrastructure.
Maintaining performance while enabling consolidation entails a number of services:

Application-specific acceleration (file and print services)

WAN optimizations such as Transport Flow Optimization (TFO), DRE, and Persistent
Lempel-Ziv (LZ) compression

With Cisco WAAS, WAEs automatically discover each other to minimize the administrative
burden.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-15

WAAS Accelerates Broad Range of


Applications
Application

Protocol

Typical Improvement

Windows (CIFS)

2X-400X

UNIX (NFS)

2X-10X

Exchange (MAPI)

2X-10X

Notes

2X-10X

SMTP/POP3, IMAP

2X-50X

Internet and intranet

HTTP, HTTPS, WebDAV

2X-50X

Data transfer

FTP

2X-50X

SMS

Altiris

2X-400X

SQL

Oracle

2X-10X

Backup Applications

Replication Applications

2X-10X

Citrix ICA

Microsoft Terminal Services, RDP

2X-5X

Any TCP-based Application

2X-10X

File sharing

Email

Software distribution
Database applications
Data protection
Terminal
Other

* Performance improvement varies based on user workload, compressibility of data, and WAN
characteristics and utilization. Actual numbers are case-specific and results might vary.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-14

The table lists typical improvements that can be expected with specific applications and related
protocols. Note that the typical improvements shown are subjective and can be relative to a
variety of factors including, but not limited to:

Response time improvement how long does an operation or series of operations take
without Cisco WAAS as opposed to with Cisco WAAS

Bandwidth savings how much bandwidth capacity is consumed on the WAN without
Cisco WAAS as opposed to with Cisco WAAS

Examples of application performance improvements are shown in the next series of slides.

1-16

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Performance: File Services


Operations over T1 (1.544Mbps), 80mS RTT
20 Seconds

40 Seconds

60 Seconds

80 Seconds

Opening 5MB
PowerPoint
Saving 5MB
PowerPoint
Drag and Drop
of 5MB
PowerPoint

Legend
Operation Over Native WAN
First Operation with WAAS
Future Operation with WAAS
20 Seconds

40 Seconds

60 Seconds

Download of
8MB Package
Microsoft SMS

80 Seconds
Legend

Operation over native WAN


First operation with WAAS, no preposition
First operation with WAAS, with preposition
Future operation with WAAS

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-15

WAAS acceleration for file services protocols leads the industry. By coupling latencyreduction and bandwidth savings techniques such as read ahead, message and operation
batching, prediction, data caching, metadata caching, and local protocol handling, to a powerful
WAN optimization architecture (consisting of DRE, LZ, and TFO), Cisco WAAS can provide
up to a 400X performance improvement for CIFS file sharing over the WAN. The top graph in
the figure shows performance improvements when working with a 5MB PowerPoint file. The
bottom graph shows performance improvements when working with Microsoft Service
Management Solution (SMS) for package download using CIFS. With Cisco WAAS, file and
software distribution servers can be consolidated without compromising end-user performance
expectations.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-17

Cisco WAAS Performance: Exchange


Sending and Receiving of
E-mail with 1MB
Attachment over
T1 (1.544 Mbps) Line
with 80-ms Latency
Microsoft Exchange
No Cached Mode

16 Seconds

32 Seconds

48 Seconds

Sending and Receiving of


E-mail with 5MB
Attachment over
T1 (1.544 Mbps) Line
with 80-ms Latency
Microsoft Exchange
No Cached Mode

40 Seconds

80 Seconds

120 Seconds

Sending and Receiving


E-Mail with 1MB
Attachment over
T1 (1.544 Mbps) Line with
80-ms Latency
Microsoft Exchange
Cached Mode

12 Seconds

24 Seconds

36 Seconds

Sending and Receiving


E-Mail with 5MB
Attachment over
T1 (1.544 Mbps) Line with
80-ms Latency
Microsoft Exchange
Cached Mode

40 Seconds

80 Seconds

120 Seconds

Legend
Send and Receive over WAN
First Send and Receive with Cisco WAAS
Future Send and Receive with Cisco WAAS

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-16

Email is another application that can be accelerated with Cisco WAAS. Microsoft Exchange
and Outlook (5.5, 2000, 2003, with or without cached mode), Lotus Notes, or other mail servers
using TCP-based protocols such as IMAP, POP3, or Simple Management Transport Protocol
(SMTP) can all be accelerated. As shown in the figure, Cisco WAAS can dramatically reduce
bandwidth consumption and improve response time for email users. The example in the figure
uses a send and receive function, as the user sends an email with a large attachment to himself.

1-18

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Performance: SharePoint


Operations over 256kbps line with 120mS RTT and .5 percent packet loss
20 Seconds

40 Seconds

60 Seconds

80 Seconds

File Open

200KB
Word Document

File Save
500KB
Word Document

File Open

File Save
1MB
Word Document

File Open
Legend
Open Operation Over Native WAN
File Save

First Open Operation with Cisco WAAS


Subsequent Open Operation with WAAS
(Save opened first via WAAS)

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-17

SharePoint and other collaboration portal applications can be optimized by Cisco WAAS. The
results in this figure show that response time is significantly improved with Cisco WAAS.
Bandwidth savings are also significant, thanks to Cisco WAAS DRE and persistent LZ
compression.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-19

Cisco WAAS Performance: Data Protection


16 Minutes
First Replication
Operation of 1GB
over T3 (45Mbps) Line
with 80-ms Latency

32 Minutes

48 Minutes
51 Minutes

25 Minutes

Legend

4 Minutes
Replication over Native WAN

First Replication Operation with Cisco WAAS


Future Replication Operation with Cisco WAAS

7 Minutes
Backup Operation
of 83MB over WAN
T1 (1.544Mbps) Line
80-ms Latency
Restore Operation
of 83MB over WAN
T1 (1.544Mbps) Line
80-ms Latency

14 Minutes

21 Minutes
22 Minutes

8 Minutes
4 Minutes
23 Minutes
12 Minutes
Legend

2 Minutes
Native Operation over WAN
First Operation with Cisco WAAS
Subsequent Operation with Cisco WAAS

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-18

For replication applications such as Network Appliance SnapMirror, EMC Symmetrix Remote
Data Facility (SRDF) native over IP or over Fibre Channel over IP (FCIP, a fabric bridging
protocol that leverages TCP/IP as an interconnect between Storage Area Networks (SAN)
fabrics), EMC SANCopy (over FCIP), EMC Celerra IP Replicator, or any other replication
applications that use FCIP or native TCP/IP as a transport. Cisco WAAS can dramatically
improve performance and minimize bandwidth for these applications.
Cisco WAAS also accelerates backup and restore operations performed over the WAN. The
example in the figure shows how Cisco WAAS can improve the performance of Microsoft
NTBackup. With Cisco WAAS compression history through DRE, a previously-seen backup
provides huge performance increases for a file or system restore.

1-20

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Performance: Citrix

384Kbps

768Kbps

1.1Mbps

Animated Flash
with audio
Legend
Microsoft Word

Uncompressed ICA
Citrix Compression

Microsoft Excel

2007 Cisco Systems, Inc. All rights reserved.

Cisco WAAS

WAAS v4.0.71-19

Cisco WAAS can also provide acceleration for remote desktop applications such as Citrix
Presentation Server or Microsoft Terminal Services. With Cisco WAAS, client connections use
less bandwidth, experience greater connection stability, and improved application
responsiveness. For optimum optimization of Citrix and other remote desktop applications,
such as Terminal Services, it is recommended that native compression be disabled and
encryption be enabled only for login traffic.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-21

Cisco WAAS Deployment Architecture


Regional Office

Remote Office

WAE
Appliance

Branch Office

WAE
Appliance

ISR with
WAE Network Module

WAN

Data Center

WAAS Central Manager


Primary/Standby

WAE
Appliance

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-20

WAEs are deployed at network entry and exit points of WAN connections. If multiple entry
and exit points exist, you can deploy a single WAE that optimizes both connections by sharing
the interception configuration across those entry and exit routers. To provide and support
optimizations, WAAS requires that devices be deployed in two or more sites. To support
redundancy, more than one WAE is typically deployed in the data center. WAEs must also be
deployed to host the Central Manager application, which can be made highly available by using
two WAEs. To provide transparent optimizations, WAAS requires two devices in the path of
the connection to be optimized.
As shown in the figure, Cisco WAE devices can either be standalone appliances or network
modules that integrate physically into the Integrated Services Router (ISR).

1-22

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Seamless, Transparent Integration


The WAE integrates into the
network fabric with high
availability, load-balancing, and
failover regardless of interception
mechanism:

Physical inline
WCCPv2
Policy Based Routing
CSM and ACE modules

Src Mac AAA


Dst Mac BBB

Src IP 1.1.1.10 Src TCP 15131


Dst IP 2.2.2.10
Dst TCP 80

APP DATA

WAEs provide compliance with


network value-added features:
Preserves packet headers
Supports QoS, Network-Based
Application Recognition (NBAR),
queuing, policing, shaping
classifications
Supports firewall policies and Access
Control Lists (ACLs)
Supports NetFlow, monitoring, and
reporting
2007 Cisco Systems, Inc. All rights reserved.

Src Mac BBB


Dst Mac AAA

Src IP 1.1.1.10 Src TCP 15131


Dst IP 2.2.2.10
Dst TCP 80

optimized

WAAS v4.0.71-21

The WAE relies on network interception and redirection to receive packets to be optimized.
Cisco WAAS can leverage a variety of options:

Physical inline: The WAE can be deployed physically in-path within the network when
configured with the inline card. This allows the WAE to see all traffic traversing a network
path.

WCCPv2: The WCCPv2 protocol (Web Cache Communication Protocol version 2) allows
the WAE devices to be deployed virtually in-path but physically off-path.

PBR: Policy Based Routing allows the network to treat a Cisco WAE as a next-hop router
to automatically route traffic through it for optimization. Like WCCPv2, the WAEs are
virtually in-path but physically off-path.

CSM/ACE: The Content Services Module (CSM) or Application Control Engine (ACE)
modules for the Catalyst 6500 series switch can be used for enterprise data center
integration and scalability. Like WCCPv2 and PBR, the WAEs are virtually in-path, but
physically off-path.

After the packets are redirected to the WAE, the WAE applies the appropriate optimization as
determined by the application policy. For all traffic that is optimized (with some exceptions),
Cisco WAAS retains the packet header information to ensure that upstream network features
are not impaired. The source IP, destination IP, source port, and destination TCP port are fully
maintained. This is called service transparency and helps to ensure compatibility with existing
network features.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-23

Scaleable, Secure Central Management


Comprehensive management:
Central configuration
Device grouping
Monitoring, statistics
Alerts, reporting

Easy-to-use interface:
GUI, wizards
IOS CLI
Roles-based administration

Proven scalability and


security:
Up to 2500 managed nodes
Redundancy and recovery
SSL-encrypted
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-22

WAAS Central Manager is a secure, robust, and scaleable management platform with many
years of lineage from the Application and Content Networking System (ACNS). WAAS
Central Manager provides all of the features necessary to provide enterprise-wide system
management, including inter-device secure communications, secure access, roles-based
administration, global and local policy configuration, group configuration, and more. With
roles-based administration, a user can be configured to see only specific features or specific
devices or groups, thus facilitating a common management framework for all users working
with services that are consolidated by WAAS.
Central Manager scales to support up to 2500 nodes and can be configured for high availability
using active-standby clustering.
The Cisco WAE provides a command-line interface (CLI) that is similar to the IOS CLI.

1-24

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Optimizations


This section describes the application-specific and application-agnostic optimizations that are
provided by Cisco WAAS.

Application Acceleration
Application and protocol awareness:

Application
specific
acceleration

Eliminates unnecessary chatter and transfer


State managed by origin server only
Enables disconnected operations

Most workloads
handled
locally

Intelligent protocol proxy:

Improves application response time


Read-ahead, message prediction
Safe object caching with validation
Provides origin server offload

WAN

WAN
optimization
DRE/TFO/LZ

WAASv4 application adaptors:


CIFS (Windows file services)
Windows printing

2007 Cisco Systems, Inc. All rights reserved.

Origin
Server
100 percent
of capacity

WAAS v4.0.71-24

Cisco WAAS provides application-specific acceleration to improve performance for


applications that are chatty and inefficient. With application acceleration, the majority of
messages are kept off of the WAN altogether to mitigate latency and improve user
performance. Currently, Cisco WAAS includes application-specific acceleration for the CIFS
protocol for Windows file sharing, and also Windows print services. The CIFS acceleration can
be deployed transparently or non-transparently. Transparent mode allows the user to map
directly to an origin server, instead of mapping directly to the WAE. This method is useful in
situations where the server is logically moved, for example, to a virtual server on a Network
Attached Storage (NAS) system, or physically moved to another location such as a data center.
Transparent mode relies on network interception to re-route traffic from the network to the
WAE for local handling and optimization. Non-transparent mode allows the WAE to take on
the personality of a file server, thereby allowing users to map directly to a name being
published by the WAE.
The following are acceleration components provided by Cisco WAAS which help to improve
application performance over the WAN:

Message suppression: Message suppression handles protocol workload locally and


suppress unnecessary messages, respond to messages locally when determined safe.

Safe data caching: Safe data caching caches copies of objects accessed (partial and full) to
provide local data response upon subsequent request from an authorized, authenticated user
after appropriate state has been applied against the origin server (lock requests). Should a
cached object change on the origin server, it is evicted from the edge cache and the updated
content is fetched.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-25

Read-ahead optimization: Multiple forms of read-ahead are employed within Cisco


WAAS to overcome performance challenges. Read-ahead is used when accessing a file that
is not fully cached or can not be cached.

Write-behind optimization: Write-behind optimization examines blocks being written to


suppress redundant write operations. Only changed blocks are propagated to the server.

Message prediction: Message prediction relies on a programmatic understanding of the


way applications leverage the protocols, and dynamic learning of how protocols are used.
This knowledge can then be used to perform operations ahead of time on behalf of the user,
thereby mitigating latency.

Integration with WAN optimization: Cisco WAAS application acceleration capabilities


can also take advantage of the WAN optimization capabilities provided by WAAS. For
instance, DRE, TFO, and persistent LZ compression can also be employed to minimize
bandwidth consumption and better leverage available network capacity.

The acceleration capabilities provided by Cisco WAAS not only help to improve performance,
but also offload the origin file server to provide better scalability using existing hardware.
Application-specific acceleration offers other tangible benefits:

Disconnected modes of operation: This includes disconnected and guest printing, for
example, and read-only CIFS file server access.

Cache prepopulation: Cache prepopulation schedules the movement of data that is


frequently used, or where first access must be very high performance. Note that the DRE
cache is also populated.

WAN and origin server offload

Cisco WAAS fully encompasses the features provided by the original Cisco WAFS product
family. Procedures to migrate from a WAFS v3.0 installation to a WAAS v4.0 solution are
available. For more information, please contact your Cisco account team or product specialist
team.

1-26

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

DRE and LZ Manage Bandwidth Utilization


DRE provides advanced compression to eliminate redundancy
from network flows regardless of application. DRE generally
provides 2:1 to 100:1 compression
Persistent LZ compression provides session-based compression
for all traffic, even traffic with redundancy removed. Persistent LZ
compression generally provides 2:1 to 10:1 compression

WAN

FILEDOC

FILEDOC
DRE CACHE

DRE CACHE

LZ

2007 Cisco Systems, Inc. All rights reserved.

LZ

WAAS v4.0.71-25

DRE allows the WAE to maintain a local database of TCP segments that have already been
seen by the device. Those same segments can be safely suppressed if they occur again. When
redundant segments are identified, the WAE sends a small instruction set to the other WAE on
how to rebuild the message with zero loss and 100% coherency. As traffic comes into the
WAE, it is compared against the DRE database context, which is a partition of the DRE
database that is reserved for the peer WAE. If the segments are identified as new, they are
added to the context. If segments are identified as redundant, they are removed and replaced
with a lightweight signature (an instruction set) that instructs the other device on which context
entries to reference in its DRE database context to accurately rebuild the message.
Along with DRE, WAAS also uses persistent LZ compression. DRE signatures (instruction
sets) are heavily compressible, meaning an additional 2:1, 3:1, or 4:1 compression can be
applied in addition to the 2:1 to 100:1 compression applied by DRE. Persistent LZ compression
provides strong compression for DRE instruction sets, as well as strong compression for data
that has not been identified by DRE before.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-27

TFO Improves Efficiency and Utilization


TFO overcomes TCP and WAN bottlenecks, and shields
nodes connections from WAN conditions:
Clients experience fast acknowledgement
Minimizes perceived packet loss
Eliminates need to use inefficient congestion handling
WAN

LAN TCP
behavior

Window Scaling
Large Initial Windows
Congestion Mgmt
Improved Retransmit

2007 Cisco Systems, Inc. All rights reserved.

LAN TCP
behavior

WAAS v4.0.71-26

TFO is a mechanism that shields communicating nodes from WAN conditions. Running a TCP
proxy and optimized TCP stacks between devices prevents many of the problems that occur in
a WAN from propagating back to the communicating nodes. With this approach,
communicating nodes experience LAN-like TCP response times because the WAE is
terminating TCP locally. From a client or server perspective, packet loss is rare. When packet
loss is encountered on the WAN, the recovery is performed at LAN speed. As packets are lost
or congestion is encountered, the WAEs create a buffer or a boundary at the border of the
WAN to keep problematic situations from bleeding over and impacting the TCP stacks of the
communicating nodes.
The TCP proxy and TFO are not applied until after device auto-discovery, which occurs during
the TCP connection establishment. In the figure, the first round-trip is completed natively with
no optimization.
In addition to buffering the WAN condition, Cisco WAAS also provides performance
optimizations to improve the throughput and responsiveness of TCP connections that traverse
the WAN. These optimizations include:

1-28

Window scaling: Window scaling allows communicating nodes to better use the available
WAN capacity.

Large initial windows: Large initial windows helps TCP connections to exit the slow-start
phase and progress more quickly to receiving ample throughput.

Improved retransmission: Improved retransmission minimizes the amount of data that


must be retransmitted during packet loss or congestion scenarios.

Advanced congestion management: Advanced congestion management algorithms to


safely maximize throughput in lousy scenarios, thereby providing bandwidth scalability
and fairness with other application flows attempting to leverage the network.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAE Platforms and Software Licensing


This section introduces Cisco WAE hardware appliances and describes the router-integrated
network module. Performance and scalability characteristics of these components are also
briefly discussed, and discussed in more detail in the design module. This section also
introduces the Cisco WAAS software licenses, which are necessary to provide WAAS
functionality.

Cisco WAE Family Positioning


Enterprise
Data Center

Performance

ACE

WAE-7326
Regional
Office or Small
Data Center

WAE-612

Branch or Remote
Office
WAE-512

NME-WAE-502
NME-WAE-302

Scalability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-28

The Cisco WAAS WAE product family includes six models:

NME-WAE-302: The NME-WAE-302 provides WAN optimization capabilities (with no


application acceleration) in a router-integrated form factor for small-medium branch office
deployments. It is deployed physically within an ISR with an open module slot.

NME-WAE-502: The NME-WAE-502 provides WAN optimization and application


acceleration capabilities in a router-integrated form factor for small-medium branch office
deployments. It is deployed physically within an ISR with an open module slot.

WAE-512: The WAE-512 provides WAN optimization and application acceleration


capabilities in a 1RU appliance form-factor for medium-large branch office deployments,
regional office deployments, or small data center deployments.

WAE-612: The WAE-612 provides WAN optimization and application acceleration


capabilities in a 1RU appliance form-factor for enterprise branch office deployments, large
regional office deployments, or medium-large data center deployments.

WAE-7326: The WAE-7326 provides WAN optimization and application acceleration


capabilities in a 2RU appliance form-factor for the largest of branch office deployments,
regional office deployments, or enterprise data center deployments.

Application Control Engine: The ACE is a module for the Catalyst 6500 series switch
and provides data center integration for WAE devices in environments, where tremendous
levels of scalability and performance are required. The ACE module does not perform

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-29

WAN optimization or application acceleration; rather, it acts as a scalability mechanism for


a large number of WAE devices within a data center.

1-30

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

NME-WAE: Router Integrated Module


NME-WAE:

NME-WAE
Router-Integrated Network Module
for the Cisco Integrated Services Router

Provides the lowest CapEx and


OpEx; integrates within the ISR;
addresses the majority of remote
branch offices
Single processor system, can be
clustered with WCCPv2, PBR
Supported in ISR models 2811,
2821, 2851, 3825, and 3845

NME-WAE-302:
512MB of RAM, 80GB of disk
Supports up to 4Mbps WAN
connections and up to 250 optimized
TCP connections
Transport license only

NME-WAE-502:
Cisco Integrated Services
Router (ISR) Series

2007 Cisco Systems, Inc. All rights reserved.

1GB of RAM, 120GB of disk


Supports up to 4Mbps WAN
connections and up to 500 optimized
TCP connections
Enterprise license capable
WAAS v4.0.71-29

The WAE relies on network interception and redirection to receive packets to be optimized.
Cisco WAAS can leverage a variety of options:

Physical inline: The WAE can be deployed physically in-path within the network when
configured with the inline card. This allows the WAE to see all traffic traversing a network
path.

WCCPv2: The WCCPv2 protocol (Web Cache Communication Protocol version 2) allows
the WAE devices to be deployed virtually in-path but physically off-path.

PBR: Policy Based Routing allows the network to treat a Cisco WAE as a next-hop router
to automatically route traffic through it for optimization. Like WCCPv2, the WAEs are
virtually in-path but physically off-path.

CSM/ACE: The Content Services Module or Application Control Engine linecards for the
Catalyst 6500 series switch can be used for enterprise data center integration and
scalability. Like WCCPv2 and PBR, the WAEs are virtually in-path but physically offpath.

After the packets are redirected to the WAE, the WAE applies the appropriate optimization as
determined by the application policy. For all traffic that is optimized (with some exceptions),
Cisco WAAS retains the packet header information to ensure that upstream network features
are not impaired. The source IP, destination IP, source port, and destination TCP port are fully
maintained. This is called service transparency and helps to ensure compatibility with existing
network features.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-31

WAE Appliance Family


The WAE-512 appliance:

WAE-512
Remote Office Appliance

1RU, Single processor, 1 or 2GB RAM


Supports up to 20Mbps WAN
connections and up to 1500 optimized
TCP connections
Supports 250GB RAID-1 disk capacity

The WAE-612 appliance:

WAE-612
Regional Hub and Data Center Appliance

1RU, Dual-core processor, 2 or 4GB


RAM
Supports up to 155Mbps WAN
connections and up to 6000 optimized
TCP connections
300GB RAID-1 SAS disk capacity

The WAE-7326 appliance:

WAE-7326
Enterprise Data Center Appliance

2RU, Dual processor, 4GB RAM


Supports up to 310Mbps WAN
connections and up to 7500 optimized
TCP connections
900GB RAID-1 SCSI disk capacity

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-30

Cisco WAAS can be installed on a Cisco WAE appliance, or on a router-integrated network


module. Each platform offers different performance and scalability characteristics. The
numbers shown in the figure are recommended values. Actual results can vary, based on
application, workload, WAN capacity, and other relevant factors. System throughput
(bandwidth) is not limited in software; it is a device positioned to support up to 4Mbps WAN
links, and can actually support a larger WAN link. In many cases, the entire WAN link does not
need to be optimized, and as such, a smaller device can be sufficient to support a larger WAN
link.
WAE performance and scalability sizing is based on the number of concurrent TCP
connections and the amount of WAN bandwidth that can be driven by the device. The LANside application throughput is unrestricted.
The Cisco WAAS deployment architecture is centered around WCCP, which enables high
availability, scalability, and load-sharing for up to 32 nodes in a cluster. Any WAE network
module or appliance can be clustered using WCCP or PBR. Inline interception can be used with
the appliance family only (512, 612, 7326). The ACE module can only be used in conjunction
with appliance devices. In cases where the WAN link is larger than the capacity of a single
device, multiple devices can be used.
The WAE appliance family includes:

1-32

WAE-512 (a 1RU appliance): This single processor system succeeds the WAE-511 (and
FE-511, CE-511), and adds RoHS compliance. Supports one or two 250GB or 500GB
SATA-2 drives, running (RAID-1) or (JBOD) and can be configured with either the
Transport or Enterprise license. Supports WAN links up to 20Mbps and up to 1500
concurrently-optimized TCP connections.

WAE-612 (a 1RU appliance): This dual-core system succeeds the WAE-611 (and FE-611,
CE-566), and adds RoHS compliance. Supports one or two 300GB Serial Attached SCSI
(SAS) drives (running RAID-1 or JBOD) and can be configured with either the Transport
or Enterprise license. Supports WAN links up to 155Mbps and up to 6000 concurrentlyoptimized TCP connections.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WAE-7326 (a 2RU appliance): This dual-processor system succeeds the FE-7326 (and CE7325), and is RoHS compliant. Supports two to six 300GB SCSI drives (running RAID-1
or JBOD), and can be configured with either the Transport or Enterprise license. Supports
WAN links up to 310Mbps and up to 7500 concurrently-optimized TCP connections.

Note

2007 Cisco Systems, Inc.

Any Cisco WAE platform can be used as a branch office or data center platform, as long as
appropriate performance and sizing guidelines are followed per the design module.

Cisco Wide Area Application Services

1-33

Enterprise Data Center Integration


Cisco ACE:
Provides transparent integration,
interception, load-balancing, and
failover for:
Up to 16Gbps throughput
Cisco Application Control Engine
linecard for the Catalyst 6500 family

Up to 4M TCP connections
350K connections per second setup

ACE features and benefits include:


Integration into Catalyst 6500 series of
intelligent switches
Solution for scaling servers, appliances,
and network devices
Catalyst 6500 Series Multilayer
Intelligent Switching Platform

Virtual partitions, flexible resource


assignment, security, and control

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-31

Deploying Cisco WAAS using the Cisco ACE linecard for the Catalyst 6500 series of
intelligent switches provides a solution that meets the needs of the most demanding enterprise
data center environments in terms of performance, scalability, and availability. The ACE
module can scale to 4 million TCP connections, with a setup rate of 350 thousand TCP
connections per second, and up to 16Gbps of throughput. Additionally, ACE represents the
industrys most scaleable, high performance, secure, and feature-rich solution for server load
balancing, network device load balancing, virtualization, and application control. With physical
integration into the Catalyst 6500, operational costs are minimized through simplified
deployment and management. The ACE module supports a variety of features, including
virtualization and virtual partitions, contexts and domains, flexible resource assignment,
granular security, and control.

1-34

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Licensing


License
Transport

Description
Includes WAN optimization features only:
Data Redundancy Elimination (DRE)
Persistent Session-Based LZ Compression
Transport Flow Optimization (TFO)
Includes IOS-like Command Line Interface
and Cisco WAE Device GUI.

Enterprise

Includes all of the features of Transport license


and:
CIFS Acceleration (file and print)

Use
Used for deployments where applications
need to be optimized but protocol latency
does not need to be mitigated (nonserver-consolidation environments).
Provides optimizations for all TCP-based
applications, but no CIFS protocol
acceleration (latency mitigation/caching),
file server disconnected mode, or print.

Used for deployments where applications


need to be optimized and file servers are
being consolidated.
Provides optimizations for all TCP-based
applications AND protocol acceleration for
CIFS (file and print).

Central
Manager

Enables a WAE to act as Central Manager for


Cisco WAAS deployments:
Includes Central Manager GUI
Order two for active or standby deployments

2007 Cisco Systems, Inc. All rights reserved.

Required for each deployment of Cisco


WAAS. Deployments without Central
Manager are not supported under any
circumstance

WAAS v4.0.71-32

Cisco WAAS has three licensing options:

Transport license: A Transport license enables all WAN optimization capabilities of


Cisco WAAS, including DRE, persistent session-based LZ compression, and TFO. This
license does not include protocol acceleration for CIFS (safe caching, disconnected mode,
preposition, latency reduction). This license is useful for deployments where applications
needing acceleration are not CIFS-based, such as email, intranet, replication, and Citrix, or
where bandwidth savings are desired. This license does provide performance improvements
for CIFS in low-latency environments but is not useful for CIFS over the WAN. This
license does not include print services, but it does include the Cisco WAE Device GUI and
the CLI.

Enterprise license: An Enterprise license provides all of the features of the Transport
license plus full acceleration of the CIFS protocol, including safe data caching, read-ahead,
prediction, batching, and other latency mitigation and bandwidth utilization techniques.
This license also includes disconnected mode of operation and preposition capabilities.
This license is useful for environments that need to optimize applications, and consolidate
file and print servers.

Central Manager: The Central Manager license enables a Cisco WAE to be configured as
a Central Manager to provide central management of a Cisco WAAS deployment. One
license is required for each Cisco WAE configured as a Central Manager. Up to 2 Central
Managers can be configured per deployment; one as active, and one as standby. Cisco
WAAS must be configured with at least one WAE acting as a Central Manager.
Deployments without Central Manager are not supported.

Note

2007 Cisco Systems, Inc.

Cisco WAAS licensing is not enforced, and no license files or keys need to be added to the
WAEs running Cisco WAAS. This enforcement level might change in a future release of the
product.

Cisco Wide Area Application Services

1-35

Summary
This topic summarizes the key points that were discussed in this lesson.

Summary
IT organizations face significant challenges from two opposing
forces; one challenge requires consolidation to improve data
protection, and the other requires distribution to improve service
levels.
Cisco WAAS provides robust application-specific and networklayer optimizations to improve application delivery to remote
users while enabling consolidation.
Cisco WAAS optimizations include application latency mitigation,
application caching, DRE, and TFO.
The Cisco WAE appliance family and router-integrated network
module run the Cisco WAAS software, and are deployed at
network entry and exit points to enable application acceleration
and WAN optimization.

2007 Cisco Systems, Inc. All rights reserved.

1-36

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

WAAS v4.0.71-33

2007 Cisco Systems, Inc.

Lesson 2

WAN Optimization Technical


Overview
Overview
This lesson explains the challenges associated with application performance over the WAN and
provides a detailed technical overview of Cisco Wide Area Application Services (WAAS)
WAN optimization features. These features include Transport Flow Optimization (TFO), Data
Redundancy Elimination (DRE), and persistent session-based LZ compression.

Objectives
Upon completing this lesson, you will be able to explain Cisco WAAS WAN optimization
features. This includes being able to meet these objectives:

Identify the performance barriers created by the WAN

Describe the basic characteristics of TCP

Describe the functions provided by Cisco WAAS TFO

Identify the compression capabilities provided by Cisco WAAS

Application Performance Barriers


This topic examines the performance barriers created by the WAN, including bandwidth,
packet loss, throughput, and latency.

Bandwidth
Bandwidth constraints keep applications from performing well.
Too much data and too small of a pipe causes congestion, packet
loss, and backpressure.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-4

In low bandwidth environments such as the WAN, traffic encounters a point of bandwidth
disparity at the network element that connects the disparate networks. In most cases, this
network element is the router. The router has to negotiate the transmission of data from a high
speed network to a low speed network. Given the difference in bandwidth, data might need to
be queued for transmission for long periods of time before the next set of data can be sent. As
shown in the figure, data leaves the node on a high bandwidth network and enters the network
element managing the bandwidth disparity, and a fraction of the data is parsed across the long
distance, low bandwidth WAN.
The challenge of the low bandwidth environment is two-fold. First, only a small amount of data
can be sent. Second, when the flow reaches another high bandwidth network, it is not able to
fully leverage the capacity of that network. From the perspective of the server in this example,
the client is only sending small amounts of data. The same is true in the reverse direction,
where the server response is throttled at the WAN router, and the client sees the server as
responding slowly.
Other challenges arise with bandwidth constraints, including congestion and packet loss. As
packets are queued on each router in the network, and these queues become full, some packets
are lost. For applications using a reliable connection-oriented transport such as TCP, the data
must be retransmitted. Additionally, the amount of data outstanding on the wire is significantly
reduced when packet loss is encountered. These characteristics force the communicating nodes
to send less data, and application performance is derailed.

1-38

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Latency
Latency impairs application performance in three ways:
Network latency: The amount of time necessary for a message to
traverse the network
Transport latency: The amount of time necessary for the transport
mechanism (TCP) to acknowledge and retransmit data
Application latency: The chattiness of an application protocol that
causes messages to be exchanged across the network
Round Trip Time (RTT) ~ many milliseconds

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-5

Latency impairs application performance in three primary ways:

Network latency: Network latency is the amount of time it takes for data on the wire to go
from the sender to the recipient. Unfortunately, the speed of light is not infinite, so it does
take time for data to go from point-to-point. The longer the link, the higher the latency, and
the lower the resulting performance.

Transport latency: Transport latency is the latency associated with transport protocols.
With TCP, for example, data must be acknowledged to provide guaranteed delivery. This
acknowledgement requires that control traffic be exchanged periodically between two
nodes. Also, when packets are lost, the transport layer of TCP is responsible for
retransmitting that data. Thus, guaranteed delivery is achieved with a performance penalty
imposed by the overhead of transport management, which is directly impacted by latency.
Note that transport latency compounds the effects of network latency.

Application latency: To complicate matters, some applications use chatty application


protocols that require a significant amount of message exchange before the user can be
productive. Each application layer message leverages a transport protocol that must
traverse the physical network, meaning that application protocol chatter, transport protocol
management, and physical network latency, all add up to diminishing application
performance over the WAN. In most cases these protocols do not benefit from compression
or other types of optimization because application-induced latency is the real problem.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-39

Packet Loss, Congestion, and


Retransmission
Packet loss and congestion cause retransmission which hinders
application performance and throughput
Commonly caused by saturated device transmit queues in the
network path

Packet Loss
Congestion

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-6

When detected, packet loss and congestion cause transmitting nodes to retransmit data and
adjust transmission throughput. These events signal to the transmitting node that there is
contention in the network and that the available bandwidth capacity needs to be shared with
another node. As such, the transmitting node will slow down transmission rates to allow other
network nodes to consume bandwidth as well. In this way, when a packet is lost, the transmitter
needs to not only retransmit the lost data, but also slows down to accommodate a network that
is being shared.

1-40

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

The Impact of Latency and Packet Loss


R=
R

MSS 1 .2
RTT p 0 .5

: Average Throughput

MSS: Packet Size

Throughput

RTT: Round-Trip Time


Expected

: Packet Loss

1.544Mbps

Actual

500Kbps

80 ms
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-7

The formula in the top-right corner of this slide shows how latency (RTT) and packet loss (p)
negatively impact overall throughput. Given that both variables are in the lower half of the
equation, as they increase, the overall throughput will decrease. Packet loss and latency cause
an exponential drop in overall throughput, causing most long-distance networks to never reach
their utilization potential.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-41

Introduction to TCP
This topic describes the basic characteristics of TCP and examines its behavior in WAN
environments.

TCP Overview

Rx Data

Tx Data

TCP
Rx
Buffers

As the network is able to


handle transmission, TCP
drains data from the
application buffer and sends it
through the network layer.

Operating System

Tx
Buffers

TCP acts as an intermediary


between application data
buffers awaiting transmission
and the unreliable network
infrastructure.

IP

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-9

TCP provides reliable and guaranteed delivery of data from one application buffer to another.
After a TCP connection is established, TCP receives data from applications and the operating
system and places it into a transmit buffer. The TCP process then manages the transmission of
this data through the IP network by packetizing the data with control information, including
port numbers, TCP sequence (SEQ) numbers, and acknowledgement (ACK) numbers.
When the data is received by the distant node and drained from the receive buffer by the
application process, an ACK is sent to the sender, to tell the sender that the data was
successfully received and that it can be safely flushed from the send buffer. The data is then
removed from the recipient receive buffer, and after the ACK is received, the data is removed
from the senders transmit buffer.

1-42

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

TCP Connection Establishment

Attempt Connection
Src port, Dst port
Sequence Number
Window Size, Checksum
Options (MSS, SACK, etc.)

TCP SYN
Acknowledge Connection
Attempt Connection

TCP SYN, ACK

Src port, Dst port


Sequence Number
Acknowledgement Number
Window Size, Checksum
Options (MSS, SACK, etc.)

Acknowledge Connection
Sequence Number
Acknowledgement Number
Window Size, Checksum
Options (MSS, SACK, etc.)

GET HTTP/1.1

2007 Cisco Systems, Inc. All rights reserved.

TCP ACK

APPLICATION DATA

WAAS v4.0.71-10

Before transmitting TCP-based application, two communicating nodes must first establish a
connection through a process called a three-way-handshake. The establishment of the
connection determines the transmission and acknowledgement characteristics of the two
communicating nodes.
The synchronize/start (SYN) message is used to initiate the connection. This SYN message
includes information such as:

Source TCP port: This is a unique port number on the sender that logically maps to an
application process on the sender.

Destination TCP port: This is a unique port number on the recipient that logically maps to
an application process on the receiver.

SEQ number: SEQ numbers are a cumulative number of all data that has been received,
starting with a specific value at the beginning of the connection.

Window size: This figure is the advertised TCP window size supported by the client, that
is, the amount of data the client can safely hold in receipt.

Checksum: The checksum is a 16-bit summation of all data being transmitted. It is used to
verify data integrity at the receiver.

Options: Options are additional TCP settings, such as segment size, selective
acknowledgement, and window scaling.

For example, if a client is talking to a web server using HTTP, the source port would likely be a
random high-value port number (greater than 1024) and the destination port would be 80, the
well-known port for HTTP.
The SYN ACK packet is used to respond to the SYN packet to establish connectivity in the
reverse direction from the receiver to the sender. Characteristics of the SYN ACK packet
include the following:

The SEQ number is incremented based on the amount of data received

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-43

The ACK number acknowledges receipt of the SYN packet

After the SYN ACK packet is received by the original sender, an ACK packet is returned to
confirm the connection.
After the connection is established, the application processes on the two connected nodes can
exchange application data. For example, an HTTP/1.1 GET request can be issued for a web
server object.

1-44

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Maximum Window Size

The MWS is the maximum amount of a data a node can have


unacknowledged and outstanding in the network.
The node cannot continue transmission until previous
transmissions have been acknowledged.
A small MWS creates problems over Long Fat Networks.
A small MWS limits the ability to fully use the available network
resources.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-11

A window is used for managing the data that is in-flight and the data to be exchanged. A TCP
window is the amount of outstanding (unacknowledged by the recipient) data a sender can
transmit before it gets an acknowledgment back from the receiver saying that data is being
received.
For example, if a pair of hosts are exchanging data using a TCP connection that has a TCP
maximum window size (MWS) of 64 KB, the sender can only send up to 64 KB of data. It must
then stop and wait for an acknowledgment from the receiver saying that some or all of the data
has been received. If the receiver acknowledges that all the data has been received, then the
sender is free to send another 64 KB. If the sender receives an acknowledgment from the
receiver that it received the first 32 KB (which happens when the second 32 KB segment is still
in-transit or lost), the sender can only send another 32 KB without exceeding the maximum
limit of 64 KB of unacknowledged outstanding data.
The primary benefit of TCP windows is congestion control. The network connection consists of
hosts, routers, switches, and associated physical links, and usually has a bottleneck somewhere
that limits the speed of data passage. Bottlenecks cause transmissions to occur at a rate that the
network is not prepared to handle, often resulting in data that is lost in-transit. The TCP
window attempts to throttle the transmission speed down to a level where congestion and data
loss do not occur.
The challenge with the MWS is its relatively small size, which is commonly 64KB or 256KB.
On long fat networks (LFNs), or elephants, the small window size limits throughput because
TCP does not allow the communicating nodes to drive available network capacity to higher
levels of utilization.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-45

TCP Acknowledgements
4

ACK

Acknowledgements are sent when an entire TCP window of data


has been received, allowing additional data to be sent.
Upon encountering packet loss, the node must retransmit the
entire window of data:
Retransmissions cause problems over low-speed links.
Retransmissions cause problems with large windows.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-12

Acknowledgements are sent after an entire TCP window has been received and the receivers
application process has removed the data from the receive buffer. Standard TCP
implementations acknowledge entire windows, which can be smaller than the MWS. If any of
the data is lost in-transit, the entire window must be retransmitted, which leads to poor
performance in high packet loss environments or low-speed links.

1-46

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

TCP Window Management


Receive
buffer (4KB)
Application performs
2KB write

seq=1,ack=1,win=4096,2048B data

seq=1,ack=2048,win=2048
Application performs
2KB write

seq=2049,ack=2048,win=2048,2048B data
seq=2049,ack=4096,win=0
Application reads
first 2KB

Sender blocked!

seq=2049,ack=4096,win=2048
Application performs
1KB write

2007 Cisco Systems, Inc. All rights reserved.

seq=4097,ack=4096,win=2048,1024B data

WAAS v4.0.71-13

The figure shows how TCP buffers and the TCP window are used to throttle the amount of data
that can be exchanged between two application processes.
In this example, the receiver has a receive buffer MWS of 4KB. As the client application writes
2KB, starting at sequence 1, the 2KB of data are placed into the receive buffer, leaving 2KB of
empty space. The 2KB of data does not leave the receive buffer until requested by the
application process. After the data is placed safely in the 2KB receive buffer, an ACK is sent
with the value of 2048 (2KB of data received) and a window (WIN) value of 2048, specifying
that 2048 bytes remain empty in the receive buffer. At this point, the client can safely send 2KB
more of data.
The client sends another 2KB of data, this time with a SEQ of 2049, specifying that the new
2KB of data is appended to the end of the previous 2048 (SEQ numbers identify the
placeholder for data in-transit). When the receiver receives this 2KB of data, it is added to the
receive buffer, leaving no free space. An ACK is sent with a value of 4096 (4096 bytes have
been placed into the receive buffer) with a WIN value of 0 (no more data can be received). At
this point, the data has been safely placed in the receive buffer, but the client is blocked from
transmitting more data, because the receiver has no where to put it because the receive buffer is
full.
The receiver application then reads 2KB of data from the receive buffer. This 2KB of data is
then freed from the receive buffer, and an ACK is sent to the client with a value of 4096
(acknowledging all 4096 bytes) and a WIN value of 2048. The receiver can handle 2KB more
of data in the receive buffer.
The client then sends another 1KB of data, with a SEQ number of 4097. The incremented
sequence number specifies the placeholder for the new data, which is appended to the end of
the first 4096.
If data is lost and retransmitted, the SEQ number is used to identify the original location of the
data being transmitted.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-47

Bandwidth Delay Product


RTT 10 ms
Amount of data that can
be in-transit at any time:
155Mbps = 19.375MBps
19.375MBps * 10mS
BDP = 193KB

Bandwidth
155 Mbps
(OC-3)

RTT 200 ms

Bandwidth
155 Mbps
(OC-3)

2007 Cisco Systems, Inc. All rights reserved.

Amount of data that can


be in-transit at any time:
155Mbps = 19.375MBps
19.375MBps * 200mS
BDP = 3860KB

WAAS v4.0.71-14

Every network is able to store a certain amount of data in-transit. This data is constantly in
motion, and it takes time to travel from the entry point of the network to the exit point. This
storage capacity is called the Bandwidth Delay Product (BDP), which defines how much data
can be in-transit on a link at any given time. The BDP is calculated by converting the rate of
transmission to bytes (divide by eight) and then multiplying the resultant value by the latency
of the link. Multiplying by the one-way latency defines how much data can be sent by one node
to another node. Multiplying by the round-trip latency defines how much data can be
exchanged between the two nodes.
For example, an Optical Carrier-3 (155Mbps) that is 10mS long has a BDP of (155Mbps/8 =
19.375MBps * 10mS) = 193KB. This means that a maximum of 193KB of data can be in-flight
on the network link at any given time. Given that the link is only 10mS long, data exits the
network quickly, allowing higher levels of throughput that are closer to the maximum link
capacity. For larger and longer-distance links, as shown in the bottom figure, the BDP
continues to increase to a capacity that is difficult for a pair of communicating nodes to fully
leverage.
When the BDP of the network is higher than the MWS of the communicating nodes, a
percentage of the network capacity is left unused.

1-48

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Challenge
Standard TCP implementations on client and server
operating systems can bottleneck application
performance, resulting in:
An inability to use available bandwidth
Inefficient recovery from packet loss, requiring retransmission
Bandwidth starvation for short-lived connections

Cisco WAAS TFO uses industry-standard TCP


optimizations to remove these application performance
barriers.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-15

The challenge with standard TCP implementations is their inability to handle the network
connections that are available today. TCP/IP was developed almost 20 years ago and the
network landscape has changed significantly:

Longer distance links, including satellite, are more frequently used, with higher levels of
packet loss.

Larger levels of oversubscription and congestion are common in broadband environments.

High capacity links now span large geographic boundaries.

Cisco WAAS TFO uses standards-based TCP optimizations to remove TCP as a barrier to
application performance over the WAN.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-49

Transport Flow Optimization


This topic provides a high level overview of Cisco WAAS TFO, and describes the
optimizations provided and the process of implementation.

TCP Flow Optimizations


WAAS TFO enables applications that are TCP
throughput bound to achieve higher levels of
throughput and overall performance.
TFO uses a TCP proxy architecture and Layer 4 TCP
options markings on connection establishment
packets:
Optimizations are performed for each TCP connection
Used to auto-discover distant endpoints

After devices have discovered and defined


optimization configuration, optimizations can then be
applied to the TCP connection.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-17

WAAS TFO provides significant optimizations for TCP connections that traverse the WAN.
These optimizations enable applications to better use available network capacity and shield the
communicating nodes from problems that are encountered in the WAN such as packet loss and
congestion.
Starting a TCP Proxy after auto-discovery allows communicating nodes to experience LANlike TCP behavior through local acknowledgement and TCP handling. Reliable delivery is still
guaranteed through the Wide Area Application Engine (WAE). After the TCP proxy is started,
TFO optimizations can be applied, along with other optimizations such as DRE.
Note

1-50

Cisco WAAS uses TFO as the data path for optimized connections. The TCP proxy service
is the foundation for other optimizations that are applied by the system. The TCP proxy
service is automatically restarted when clearing the DRE cache, causing connections and
sessions to be broken. Clients and servers automatically regenerate these connections.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Without TCP Proxy


WAN

X
TIMEOUT! RESEND

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-18

If no TCP proxy infrastructure is in place, no LAN TCP features are available over a WAN. In
this example, each ACK sent has to traverse the WAN completely. For a 100mS WAN, after 10
ACKs, TCP has added up to 1 second of additional overhead to the application experience. If a
packet is lost, the entire window of data must be retransmitted after the timeout period has
passed.
Also note the slow ramp-up of throughput after the establishment of the connection.
Throughput is severely limited until after the connection enters congestion avoidance. This is
called the slow-start phase.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-51

TCP Proxy and TFO


Client receives LAN
TCP behavior.

Server receives LAN


TCP behavior.

WAN

Window Scaling
Large Initial Windows
Congestion Mgmt
Improved Retransmit

X
No retransmission
necessary: packet loss
is handled
by the WAE.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-19

A successful completion of automatic discovery initializes the TCP proxy service. TCP proxy
is a mechanism by which WAEs are inserted in the TCP connection to provide localized
handling of TCP buffering and control. With a TCP proxy in place, each WAE provides a TCP
termination and generation point. This allows the WAE to locally acknowledge TCP data to
keep the nodes communicating. By using larger buffers and WAN optimizations, the WAE can
more efficiently use the WAN and more effectively handle situations such as packet loss. Due
to the large windows, selective acknowledgements, and advanced congestion avoidance
functions of Cisco WAAS TFO, packet loss has minimal impact on overall throughput, and the
vast majority of WAN latency associated with TCP control handling is mitigated.
WAAS TFO is designed to overcome challenges common to standard TCP implementations:

1-52

Window scaling: Scaling allows you to capitalize on available bandwidth.

Selective acknowledgement: This mechanism provides efficient packet loss recovery and
retransmission mechanisms.

Large initial windows: A larger size maximizes transmission after connection


establishment.

Advanced congestion management: This function provides an adaptive return to


maximum throughput upon encountering congestion based on packet loss history, thereby
allowing bandwidth scalability and fairness.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Link Utilization and MWS, BDP


BDP

Bandwidth

Unusable network capacity

MWS
Link Utilization

Latency
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-20

The MWS determines the maximum amount of data that can be in-transit and unacknowledged
at any given time. The BDP (Bandwidth-delay product) defines the amount of data that can be
contained within a network at any given time:

If MWS > BDP, then the application might not be throughput bound and can fill the pipe.

If BDP > MWS, then the application can not fully use the network capacity, and can not fill
the pipe.

MWS does not account for application-layer (L7) latency such as that experienced with
protocol-specific messaging.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-53

Windows Scaling, MWS, and BDP

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-21

Without window scaling, high BDP networks can not be fully used by nodes that have a MWS
that is smaller than the BDP. Significant improvement can be achieved by virtually scaling the
TCP MWS, and handling TCP on the WAE. In this manner the communicating nodes can fully
use the available WAN capacity.

1-54

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Link Utilization After Window Scaling


BDP
Cisco WAAS TFO

Bandwidth

Able to fill the pipe

Original MWS

Latency
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-22

WAAS TFO window scaling is based on RFC 1323 and scales the TCP window to 2MB to
overcome problems when filling LFNs. Window scaling applies a binary shift to the decimal
value supplied in the specified window field. For instance, if an advertised window is 64KB,
and a binary shift of 2 (window scale factor of 2) were employed, this would indicate a 256KB
TCP window.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-55

Selective Acknowledgement
Standard TCP implementations acknowledge receipt of data by
acknowledging that the entire window has been received.
Loss of a packet causes retransmission of the entire TCP window,
causing performance degradation as the window becomes larger.

Receive

Transmit
3

ACK
Retransmit
3

1
ACK

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-23

With standard TCP implementations, ACKs are sent to acknowledge receipt of the entire
window of data. If any piece of that data is lost, the entire window must be retransmitted by the
sender. In cases of high-latency or low-bandwidth links, the propagation delay of the
acknowledgement could be very large, and the amount of data that needs to be retransmitted
could also be quite large. This combination of factors leads to degraded application
performance.
With standard TCP, the loss of a segment results in application throughput reductions of fifty
percent. Add in the need for increased management overhead and a high latency network, and
TCP becomes a barrier to application performance over the WAN.

1-56

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Selective Acknowledgement (Cont.)


Cisco WAAS uses Selective Acknowledgement and
extensions to improve acknowledgement of transmitted
data, improve delivery of missing segments, and
minimize unnecessary retransmission.

WAN
Transmit
3

Receive
3

Transmit
3

Receive

ACK

ACK
Retransmit
3
ACK

2007 Cisco Systems, Inc. All rights reserved.

Transmit

Receive

ACK

WAAS v4.0.71-24

Cisco WAAS uses selective acknowledgement (SACK) and extensions to minimize the amount
of data that must be retransmitted in the case of data loss. With selective acknowledgement, the
recipient is able to stream acknowledgements back to the sender for segments of data that have
been successfully received. These SACKs also free up capacity of the window to allow the
sender to continue transmission. If a segment is lost, only that segment is retransmitted. Add
Cisco WAEs to the path to handle the retransmission, and the communicating nodes never
know that a packet was lost.
SACK allows a receiver to identify blocks of data from within a window that has been
received. The sender is required to retransmit missing blocks only. SACK is defined in RFC
2018.
Forward acknowledgement (FACK) is an extension to SACK that is used to aggressively
request retransmission of blocks that are missing from within a window after later blocks have
been received.
Duplicate SACK (DSACK) is another extension to SACK. DSACK allows a receiver to notify
the sender that duplicate blocks of data have been received.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-57

Cisco WAAS Large Initial Windows


While 80% of network traffic (bytes) is typically associated with
long-lived connections, approximately 80% of network
connections are short-lived (mice).
Short-lived connections transmit smaller numbers of packets and
are torn down before leaving the slow-start phase of TCP.
Cisco WAAS Large Initial Windows, based on RFC3390,
increases initial window size to expedite entry into congestion
avoidance mode for high throughput.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-25

TCP slow-start is a mechanism that initially throttles a connection until it can determine the
optimum window size to allocate based on network conditions. With every successful
roundtrip, the congestion window is doubled, starting from a single segment size (1460 bytes).
At this rate, it can take a large number of successful roundtrips before a connection reaches the
maximum window size, or a packet is lost, and the connection is transitioned into congestion
avoidance mode. Congestion avoidance mode allows the connection to operate at high levels of
throughput as long as congestion or packet loss is not encountered.
Short-lived connections that typically complete in a small number of roundtrip exchanges are
typically starved of bandwidth because they do not live long enough to exit the slow-start
phase.

1-58

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS Large Initial Windows


(Cont.)
Packet Loss

Segments per Round Trip (cwnd)

TFO

Slow-Start
(discovery)

Congestion
Avoidance
(high-throughput)

TCP

Round Trips
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-26

Cisco WAAS Large Initial Windows increases the initial maximum segment size from 1460
bytes to 4380 bytes to help connections exit slow-start more quickly. With each successful
roundtrip, the congestion window (cwnd) is doubled, but the initial starting size is three times
the value of the unoptimized starting size. This feature allows connections to more quickly take
advantage of available WAN bandwidth during the congestion avoidance phase.
Note

2007 Cisco Systems, Inc.

The congestion window is the amount of data that can be outstanding and unacknowledged
in the network at any given time between two connected nodes. It is also referred to as the
number of segments that can be sent per roundtrip.

Cisco Wide Area Application Services

1-59

Standard TCP Congestion Avoidance


Return to maximum
throughput could take
a very long time!

Segments per Round Trip (Congestion Window)

Packet loss causes connection to enter into


linear congestion avoidance (+1 cwnd per ACK)
cwnd dropped by 50% on packet loss

loss

Linear Congestion
Avoidance
(+1 cwnd per ACK)

loss

Exponential
Slow Start
(2x pkts per RTT)
Low throughput
during this period
1

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

Round Trips
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-27

Standard TCP implementations employ an exponential slow-start to increase throughput to the


slow-start threshold. From the slow-start threshold, the congestion window is increased linearly
by one segment size per round-trip until packet loss is encountered. Upon encountering packet
loss, the congestion window is cut in half to return to a throughput level safe given the
congested environment. The net result is saw-tooth throughput, and a return to maximum
throughput can take hours for long-lived connections and LFNs.

1-60

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAAS TFO Congestion Avoidance


Adaptive Increase to cwnd
cwnd = cwnd + f(cwnd, history)

Packet loss

Packet loss

Cwnd decreased by 1/8 on


packet loss vs 1/2 with TCP

Packet loss

Packet loss

cw nd

Cisco
WAAS TFO

Slow start

Congestion avoidance

Time (RTT)
Standard
TCP

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-28

Cisco WAAS employs the Binary Increase Congestion (BIC) congestion avoidance system as
part of TFO to improve throughput and enable bandwidth scalability in environments that
experience higher levels of packet loss. WAAS TFO, and specifically the BIC congestion
avoidance system, maintains a history of packet loss encountered in the network to determine
the best rate at which to return to maximum throughput. TFO uses a binary search to adaptively
increase the size of the congestion window, resulting in a stable and timely return to higher
levels of throughput. Unlike standard TCP, TFO decreases the congestion window by only oneeighth (rather than one-half) when packet loss is encountered, to mitigate the majority of the
performance penalty.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-61

Comparing TCP and TFO


Cisco TFO provides significant throughput
improvements over standard TCP implementations

TFO

cw nd

TCP

Slow start

Congestion avoidance

2007 Cisco Systems, Inc. All rights reserved.

Time (RTT)

WAAS v4.0.71-29

Comparing TCP to TFO with advanced congestion management shows that TFO offers a more
timely return to maximum levels of throughput. The area shown in red in the figure is
potentially unused network capacity that TFO is able to fully use. With WAAS TFO and
advanced congestion management, the WAN link is used efficiently to its maximum potential.

1-62

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Advanced Compression
This topic examines the need for WAN compression and describes the benefits of Cisco WAAS
advanced compression, including DRE.

The Need for Compression


Advanced compression technologies allow customers to virtually
increase WAN capacity.
Advanced compression technologies allow customers to
leverage existing WAN capacity and mitigate the need for a
costly bandwidth upgrade.

WAN without
compression
WAN with compression

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-31

As applications grow to require additional network bandwidth resources, the network itself can
become a bottleneck, especially if insufficient bandwidth is available to support the
applications. Network compression functions compress the data in transit and then rebuild the
original messages at the other side of the link. This compression minimizes the amount of data
that must traverse the network while still maintaining message and data validity and integrity.
Some data sets are not good candidates for compression, unless adaptation is first performed:

Previously-compressed data: No additional compression can be provided by


computational compression. Previously-compressed data provides a good opportunity for
data suppression.

Previously-encrypted data: Minimal additional compression is provided by computation


compression. This data also provides a good opportunity for data suppression, as long as
session-based encryption is not being used.

Adaptation could include the local termination of encryption, the application of compression,
followed by a re-encryption of the transmission.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-63

Data Transfer Without Compression

WAN

WAN

Congestion!

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-32

When transmitting data across a low-bandwidth network, the transmission is throttled by


congestion and packet loss at the point of bandwidth disparity (generally the router), and the
distant network sees only a small amount of data. In this example, the server sees the client as
being slow to transmit.

1-64

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Data Transfer With Compression

WAN

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-33

Network compression not only minimizes the amount of data that must use the network, but
also allows for a larger degree of throughput as experienced by the recipients of the
transmission. As traffic enters the WAE and is DRE encoded and compressed, it takes up less
network capacity. When the encoded and compressed message reaches the other side of the
network, it is decompressed and decoded. This process allows traffic flows to take advantage of
the high bandwidth LAN on the other side of the link.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-65

Cisco WAAS Advanced Compression


Cisco WAAS employs two forms of advanced
compression:
DRE: data suppression
Persistent LZ compression: standards-based compression using
a long-lived session-oriented history
Original
Message

Original
Message

Compressed
Message

LZ

LZ

DRE

DRE
Synchronized
context

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-34

Cisco WAAS employs two distinct forms of compression, each addressing different bandwidth
requirements.
The first form of compression is DRE. DRE is a process that allows WAEs deployed on either
side of the WAN to maintain a history of previously-seen data segments from TCP messages,
based on the configured application policy. By keeping a history of previously-seen data
segments, the WAE can check to see if a piece of data being transmitted has been seen before
or not. Assuming the data is a repeated segment, instructions can be sent to the distant device,
defining how to rebuild the data segment, rather than sending the data segment itself. DRE
maintains data and message integrity without compromise by ensuring that the rebuilt message
maintains one-hundred percent coherency with the original message. DRE can dramatically
reduce network bandwidth consumption by ensuring that data only traverses the network once
within a history. With DRE, as much as 100:1 compression can be realized.
The second form of compression is persistent Lempel-Ziv (LZ) compression, which is a
standards-based compression method found in many applications, including zip technologies.
The Cisco WAAS implementation of LZ leverages a long-lived session-oriented history per
TCP connection to better improve compression ratios. Persistent LZ extrapolates patterns and
redundancy within the dataset being transferred, using a smaller connection-specific history
than DRE, and is useful in providing compression for application data that has never been seen
by DRE. With persistent LZ, as much as 10:1 compression can be realized, and persistent LZ
can even compress encoded messages that are generated by DRE.

1-66

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

DRE Overview

1
Branch WAE

Client A

Data Center WAE

5
LZ

LZ

3
Client B

2
DRE

2007 Cisco Systems, Inc. All rights reserved.

Server

DRE

WAAS v4.0.71-35

The figure shows the DRE compression process:


1. Client A requests a file from the server. This is the first time the file has been accessed.
2. The Core WAE chunks and fingerprints the file, and adds the signatures to its local DRE
database.
3. The Core WAE sends data chunks with signatures to the Edge WAE.
4. The Edge WAE adds the chunks and signatures to its local DRE database.
5. The Edge WAE forwards the file to Client A.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-67

DRE Overview (Cont.)

Server
Branch WAE

Client A

Data Center WAE


LZ

Client B

LZ

10
DRE

7
DRE

2007 Cisco Systems, Inc. All rights reserved.

1-68

WAAS v4.0.71-36

6.

Client B requests the same file. However, the file was changed on the server after the
previous transaction.

7.

The Core WAE chunks the file, and locates signatures in the local DRE database for the part
of the file that was not changed.

8.

The Core WAE sends signatures with no data for those chunks that are in the DRE database,
and sends data with new signatures for the remaining chunks.

9.

The Edge WAE receives the signatures and data. For signatures received without data, the
WAE retrieves the matching chunks from local DRE database. Chunks received with
signatures are added to the database.

10.

Edge WAE reassembles file and forwards to Client B.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

DRE Block Diagram


Signature Matching
Add New Entries

DRE
FIFO
Cache
Synchronization

DRE
FIFO
Cache
Synchronization

Fingerprint
Chunk Identification

Signature Matching
Add New Entries

Fingerprint
Chunk Identification
P-LZ

TCP proxy

2007 Cisco Systems, Inc. All rights reserved.

P-LZ

TCP proxy

WAAS v4.0.71-37

DRE and LZ compression are applied based on the configuration of the application traffic
policy that identifies the incoming flows. As data is buffered in the WAEs TCP proxy, it is
shifted into DRE, where the following processes are applied:

Chunk identification: This process breaks the data set into smaller, more manageable
chunks by using a fingerprinting function.

Signature generation: All identified chunks are assigned a signature.

Signature matching: Generated signatures are compared against the existing DRE context,
which is a shared database between two connected WAE peers. If redundant signatures are
identified, the chunk is removed and the signature remains, instructing the distant WAE to
replace the signature with the associated chunk from its context. If redundancy is not
identified, the chunk and signature are added to the local context, and remain in the
message to instruct the distant WAE to also update its context.

The output of this function is an encoded message which can then be compressed by LZ and
sent back to the TCP proxy for forwarding across the WAN.
When received by the other WAE, the message is first uncompressed (LZ) and then passed to
DRE to identify signatures and chunks. Signatures received with no data are replaced by data
segments from the DRE context, and chunks received with signatures are added to the local
DRE context. After the message is rebuilt and verified as accurate, the message is forwarded
toward the destination.
Following is a list of DRE terminology:

Signature: This is a 5-byte label that references a chunk of data, generated per identified
chunk.

Chunk: A chunk is a segment of data identified within a transmission. It is identified by a


fingerprint function.

Fingerprint: The fingerprint function is a sliding window operation used to identify chunk
boundaries within a transmission.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-69

FIFO: This term identifies a first-in-first-out database for chunks that have been identified
between communicating DRE peers (WAEs).

FIFO clock: This term identifies the timestamp associated with FIFO data. The FIFO clock
is used to synchronize communicating DRE peers (WAEs).

DRE and Persistent LZ compression are application policies that are configurable. Application
policy is negotiated between two WAEs upon establishment of the TCP proxy. Part of this
exchange includes synchronization of the DRE database, which is discussed later.

1-70

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Fingerprinting and Chunk Identification


Window
No boundary found
Window
No boundary found
Window
No boundary found
Window
No boundary found

DRE analyzes incoming data


streams using a sliding window
to identify chunks.
Each chunk is assigned a 5-byte
signature.
A single-pass is used to identify
chunks at multiple levels:
Basic chunks

Window
Boundary identified!

Chunk aggregation (nesting)

After chunks are identified, DRE


begins pattern matching:
Chunk1

Window

5-byte signature

2007 Cisco Systems, Inc. All rights reserved.

Looks for largest chunks first


Looks for smaller chunks if necessary

WAAS v4.0.71-38

The encode function is applied to traffic entering a WAE and is configured to leverage DRE.
TCP proxy temporarily buffers data to provide DRE with a large amount of data to analyze. Up
to 32KB of data can be analyzed at one time. The fingerprinting and chunk identification
process is accomplished in the following steps:

First, the encode function generates a 16-byte message validity signature in MessageDigest algorithm 5 (MD5) format to be used by the distant WAE to validate that the
message, after decoding, is an identical match to the encoded message. This step ensures
data and message integrity. If a hash collision is detected (that is, the same signature for
two different pieces of data), a synchronous instruction is sent to the peer to flush the
relevant entries in the DRE context. The calculation of this message validity signature also
includes device-specific key data as well to prevent against hash vulnerabilities.

Second, the encode function identifies chunks, using a sliding window to analyze data to
locate content-based break-points within the data. All content between break-points is
considered a chunk.

Finally, a 5-byte signature identifier is generated for each chunk identified.

DRE attempts to match chunks on four different levels in a process called chunk aggregation:

Level-0 chunk: This is a basic chunk. Level-0 chunks can be identified for data segments
that are as small as 32 bytes, but are commonly found approximately every 256 bytes.

Level-1 chunk: Level-1 chunks are commonly found approximately every 1024 bytes, and
reference multiple level-0 chunks.

Level-2 chunk: Level-2 chunks are commonly found approximately every 4096 bytes, and
reference multiple level-1 chunks.

Level-3 chunk: Level-3 chunks are the largest chunk. Level-3 chunks are commonly found
approximately every 16K bytes, and reference multiple level-2 chunks.

DRE pattern matching and chunk identification is done efficiently, and only one pass is made
against the entire data stream being analyzed.
2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-71

A Fully Chunked Message


Level-0 Chunk
Basic Chunk
~256 bytes

Level-1 Chunk
~1024 bytes

Level-2 Chunk
~4096 bytes

Level-3 Chunk
~16384 bytes

ORIGINAL DATA

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-39

A fully-chunked message is shown in the figure. Notice that for the original data, many level-0
chunks are identified. Multiple level-1 chunks are identified, and each is a representation of all
of the level-0 chunks that make up the level-1 chunk. Multiple level-2 chunks are identified,
and each is a representation of all of the level-1 chunks that make up the level-2 chunk.
Multiple level-3 chunks are identified, and each represents all of the level-2 chunks that make
up the level-3 chunk.
Each chunk identified receives a 5 byte signature. For redundant data that has been identified,
the chunk is replaced with the signature, to provide as much as 2500:1 compression for a chunk
of data. Note that pattern match on a level-3 chunk = 16KB / 5B = approximately 3000:1
compression. Overall, DRE can provide an overall compression ratio from 2:1 to as much as
100:1.

1-72

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Chunk Identification
Original Message
Match
5B Signature

5B Signature

Chunk1

Chunk2

Add

5B Signature
Chunk3

DRE Context

Add

Match
5B

5B

5B

Chunk3 Chunk4 Chunk5 Chunk6 3


Add

Add

5B 5B

5B 5B 5B 5B
3

4
Add

Match

Add

Add

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-40

After all of the level-0 through level-3 chunks are identified and signatures are assigned to
each, DRE begins to pattern match against the identified chunks. Given that the chunk
boundaries are found based on the content being transferred, the break-points are always
identified at the same location within the transmission. Changes to data are contained within a
chunk and do not affect neighboring chunks. If a change occurs on a chunk boundary, only the
two neighboring chunks are affected. Changes are isolated to the chunk or chunks where the
change was inserted:

Typical: A new chunk must be added to the local context and transmitted in full for each
chunk experiencing a change.

Worst case: A change inserted at the location of a chunk break-point only invalidates the
two adjacent chunks. The new chunk must be added to the local context and transmitted in
full.

After the message is fully chunked, DRE begins a top-down pattern match. It first searches for
matches on the largest chunks, and continues to smaller chunk sizes as necessary. Notice that as
each chunk is identified, the chunk of data is removed from the message and only the signature
for the data remains.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-73

Chunk Identification (Cont.)

A fully chunked message contains:


Signatures only for previously-seen patterns
Signatures and data for non-redundant patterns; used to update the
adjacent WAE
16-byte MD5 hash of original message to verify integrity after the
rebuild
Chunk aggregation (nesting)

The message is then passed to LZ compression, based on policy, and


to the TCP proxy to return to the network.
REDUNDANT DATA AND EXISTING SIGNATURES ONLY

MD5

NON-REDUNDANT DATA AND NEW SIGNATURES


2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-41

After the fully chunked message completes the pattern matching process, the resulting output is
an encoded message containing the following:

Signatures and chunks for non-redundant data to be added to the peer WAEs DRE context.

Signatures with no chunks for redundant data; the peer WAE inserts the data on behalf of
the chunk upon receipt.

The original 16 byte MD5 message validity signature, which is used by the peer WAE to
ensure that the rebuilt message is an identical representation of the original message.

The message is then passed to persistent LZ compression, based on the configured application
policy, for an additional layer of compression.
The decoding WAE first decompresses the message received using LZ compression, and then
begins a process of identifying:

Standalone signatures with no chunk attached

Signatures with attached chunks

Message validity signatures

After these components are identified, the following processes are performed:

Standalone signatures with no chunk attached are replaced by the corresponding chunk of
data contained in the local context.

Signatures with attached chunks are added to the local context and the signature is removed
from the message.

A new message validity signature is generated.

If a standalone signature is received and no chunk is stored in the local context, a negativeacknowledge character (NACK) is sent to the encoding WAE indicating that the signature and
chunk must be resent. This allows the WAE to update its context. If a standalone signature is

1-74

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

successfully replaced with a chunk, an asynchronous ACK is sent to the other WAE to notify
that the decoding for that chunk was successful.
After the message is fully rebuilt, a new message validity signature is generated and compared
against the message validity signature that was sent by the encoding WAE. If the two
signatures match, the message is considered completely valid and is forwarded toward the
intended destination. If the message validity signatures do not match, a synchronous NACK is
sent to the encoding WAE indicating that the message was rebuilt using a chunk that was
inaccurate. This message instructs the encoding WAE to resend all signatures and chunks
related to that message to the receiving WAE to update its local context, and to ensure the
message can be rebuilt safely.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-75

DRE Synchronization
DRE is a database that is provisioned into connection-based
contexts. Each context represents dedicated storage for each
connected pair of WAEs.
These logically separate contexts are dynamic in size and
adjusted based on connection activity:
More active connections are dynamically assigned additional capacity.
Fewer active connections result in capacity reduction.

FIFO clock information is exchanged to ensure the


synchronization of contexts.
Contexts are bidirectional and loosely synchronized.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-42

Upon connection establishment, DRE peers compare FIFO clock information from their
respective contexts. This includes the head and tail of the context, described as follows:

Head: The head is oldest entry contained in the context, and the first to be evicted if
additional capacity is needed.

Tail: The tail represents the newest entry contained in the context, and the last to be evicted
if additional capacity is needed.

FIFO clock timestamps are not related to actual system time. Instead, they are relative to the
connection time itself. Generally, FIFO tail and head times are nearly identical as context is
maintained per connection. However, there are circumstances where these timestamps might
not be closely identical, for example, when a Core context is partially or fully flushed to
support other, more active contexts; or when a disk failure is encountered.
FIFO clocks are synchronized as follows:

Upon connection establishment: An exchange of head and tail FIFO values occurs to
identify areas within the context that are still valid.

Upon eviction: Eviction causes synchronization to again exchange head and tail FIFO
values to identify areas within the context that are still valid.

Intermittently: Intermittent synchronization assumes that no eviction has taken place, and
DRE contexts continue to resynchronize every 5 minutes.

Using a bidirectional context, that is inherently cross-protocol, means that a single context is
used regardless of the direction of traffic flow. This approach is useful, for example, when the
download of a new file helps to provide significant compression when the file is returned in the
reverse direction of traffic flow, even if using a different application protocol.

1-76

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Successful Synchronization
DRE Database

DRE Database
Tail (newest)

Flush

Keep

Flush

S10 DATA Clock 155


S9 DATA Clock 154
S8 DATA Clock 153
S7 DATA Clock 152
S6 DATA Clock 151
S5 DATA Clock 150
S4 DATA Clock 149
S3 DATA Clock 148
S2 DATA Clock 147
S1 DATA Clock 146

S8 DATA Clock 153


S7 DATA Clock 152
S6 DATA Clock 151
S5 DATA Clock 150
S4 DATA Clock 149
S3 DATA Clock 148
S2 DATA Clock 147

Head (oldest)

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-43

In circumstances where the two communicating WAEs can successfully negotiate a range of
entries that are common on both WAEs, the context area containing common entries within the
least common denominator are maintained and used on both WAEs, as the signatures and data
appear on both WAEs. Entries that are outside of the negotiated range are flushed.
After the DRE contexts are synchronized, the two have identified areas from within the local
context that are still considered valid (data exists on both WAEs) and can be used for
transmissions involving this WAE pair. The areas within the context that can not be safely used
are immediately flushed from the context.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-77

Failed Synchronization
DRE Database

DRE Database
Tail (newest)

Flush

S10 DATA Clock 155


S9 DATA Clock 154
S8 DATA Clock 153
S7 DATA Clock 152
S6 DATA Clock 151

Flush

S5 DATA Clock 150


S4 DATA Clock 149
S3 DATA Clock 148
S2 DATA Clock 147
S1 DATA Clock 146

Head (oldest)

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-44

In circumstances where two communicating WAEs can not negotiate a range of entries to
maintain, the negotiated FIFO clock value is set to zero, and both contexts are flushed.
From here, the WAEs must rebuild the context with new data. No redundancy can be found
until chunks and signatures are populated in the context.

1-78

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Combining Cisco WAAS TFO and DRE

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-45

As shown in this slide, DRE can combine with TFO to provide exponential increases in overall
network throughput. TFO helps transmitting nodes to effectively fill-the-pipe. When coupled
with DRE, which is applying high levels of compression, TFO is then able to fill-the-pipe with
compressed data, resulting in even higher levels of throughput.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-79

Summary
This topic summarizes the key points that were discussed in this lesson.

Summary
Cisco WAAS TFO addresses the WAN performance challenges
of TCP, and shields communicating nodes from WAN conditions.
TCP provides a reliable, connection-oriented mechanism for
exchanging application data between communicating nodes, but
has serious performance challenges in WAN environments.
Cisco WAAS TFO provides a number of optimizations to improve
efficiency when communicating over WAN links, including
window scaling, SACK, large initial windows, and advanced
congestion management.
Cisco WAAS advanced compression includes data suppression
through DRE, and standards-based persistent LZ compression.

2007 Cisco Systems, Inc. All rights reserved.

1-80

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

WAAS v4.0.71-46

2007 Cisco Systems, Inc.

Lesson 3

Application Acceleration
Technical Overview
Overview
This lesson discusses the application-specific acceleration capabilities provided by Cisco Wide
Area Application Services (WAAS) for the Common Internet File System (CIFS) protocol to
enable server and storage consolidation while maintaining end-user performance expectations.

Objectives
Upon completing this lesson, you will be able to explain how Cisco WAAS application-specific
acceleration improves performance for file and print protocols. This includes being able to meet
these objectives:

Identify the need for application-specific acceleration

Identify the acceleration capabilities provided by Cisco WAAS for the CIFS file services
protocol

Discuss print services capabilities of Cisco WAAS

The Need for Application Acceleration


This topic examines the need for application-specific acceleration and explains why standard
compression, flow, and data suppression optimizations are insufficient to ensure highperformance access to centralized file servers over a WAN.

The Need for Application-Specific


Acceleration
The Common Internet File System is an example of a
protocol that requires protocol-specific acceleration.
CIFS makes a portion of a local file system network
accessible, and must maintain all of the semantics of
the local file system itself, including:

User (or process) authentication and authorization


Information security
Locating information, directory traversal
File access control and locking semantics
I/O operations, including open, close, read, write, seek

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-4

Many application protocols can not be adequately optimized through simple compression and
transport optimizations alone. Application protocols are commonly developed in utopian
environments where the client and the server are located on the same LAN or are positioned
close to each other. Application-induced or protocol-induced latency and unnecessary data
transfers hinder overall end-user performance.
CIFS is an example of protocols that require protocol-specific acceleration. By effectively
making a portion of a file system available on the network, the node then has to ensure all of
the same file system semantics are maintained. This includes:

1-82

Authentication: Authentication verifies that the requesting user, process, or node is a valid
entity. Authentication can be performed locally or through an authentication provider such
as Active Directory.

Authorization: Authorization ensures that the requesting user, process, or node has the
appropriate privileges to access the requested information.

Security: Security includes auditing and authentication, authorization and accounting


(AAA).

Access control and locking: These functions ensure that files or portions of files are
locked properly to enable collaboration, and maintain data integrity.

Input and output operations: These operations include open requests, create requests,
close requests, read operations, write operations, seek, and find.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

The Need for Application-Specific


Acceleration (Cont.)
The result is that hundreds upon thousands of messages
must traverse the network before any usable data is
served or the function is completed.
Protocol version selection
User authentication
User authorization
Meta data operations
Find file

File open, FID


Lock segment ranges
Read data

Write data

Close file

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-5

File system protocols are notoriously chatty and require a large number of ping-pong
operations. These operations are typically very small, highly uncompressible (commonly with a
zero-byte length), and must occur in sequence before the next process can begin.
Applying compression to communications between the client and server certainly minimizes
the amount of bandwidth consumed by each protocol message, and applying transport
optimizations to these communications improves the ability of each message to efficiently and
fully use available network capacity. However, many hundreds or thousands of messages must
still traverse the WAN in sequence.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-83

The Need for Application-Specific


Acceleration (Cont.)
In this example of
a 2MB Word
document open,
over 1000
messages are
exchanged.
With a 40mS RTT
WAN, this
equates to over
52 seconds of
wait time before
the document is
usable.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-6

This example shows the impact of accessing a file over a WAN. With a 40mS round trip time
(RTT) WAN, opening a 2MB Microsoft Word document requires the exchange of thousands of
messages, the vast majority occurring in sequence. This process directly impacts the response
time of the application that the user is accessing. Performance is hindered due to applicationinduced latency caused by the CIFS protocol.

1-84

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

CIFS Acceleration
This topic describes the protocol-specific acceleration capabilities applied by Cisco WAAS to
provide LAN-like performance to centralized file servers.

WAAS CIFS Acceleration Overview


Intelligent local handling and
optimization of protocol mitigates latency

Sessions are maintained end-to-end to


ensure no security reconfiguration

File caching removes the need for


unnecessary file transfer; validation
ensures stale data is never served

Auditing, access-control, and quotas are


fully preserved
Scheduled preposition to prepopulate
Data Redundancy Elimination (DRE)
and edge data cache

Transparent integration ensures no client


or server changes to apply optimization

WAN
Files

FILE.DOC

Cache

Disconnected mode of operation


allows R/O access to fully-cached
content when the server is unreachable

2007 Cisco Systems, Inc. All rights reserved.

Advanced WAN optimization layer


improves throughput and efficiency
DRE eliminates redundant network data
TCP optimizations to improve protocol
ability to fully use the network
WAAS v4.0.71-8

Cisco WAAS provides the most innovative and robust file services optimizations for industry:

Application protocol interface for CIFS to handle protocol message workload at the Edge
to mitigate the impact of latency through message suppression, local response handling,
protocol caching, operation batching, message prediction, read-ahead, and pre-fetch

Application data and meta data cache to serve usable content at the Edge to mitigate
unnecessary data transfers when safe; validate-on-open to verify that file data has not
changed; global locking to ensure coherency and enable global collaboration scenarios

Network compression through DRE and Lempel-Ziv (LZ) persistent compression to


minimize bandwidth usage during data transfer situations

Transport Flow Optimizations (TFO) to improve utilization of the available network


capacity

Download the Wide Area File Services (WAFS) Benchmark Tool from Cisco Connection
Online (CCO). This utility stages data to a file server and then executes a script that makes calls
against these files, including OPEN, READ, WRITE, SAVE, and CLOSE operations. The
amount of time taken to perform these tests can then be saved to a comma separated value
(CSV) file for viewing and graphing. The results shown in the figure represent the typical
performance improvement provided by Cisco WAAS in CIFS environments.
Cisco WAAS acceleration is safe and requires no coherency configuration. The level of
optimization applied is directly related to the type of file being opened, and the state of the
opportunistic lock that is granted to the user. For single-user situations, Cisco WAAS can
employ the breadth of its optimizations to dramatically improve performance. For multi-user

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-85

situations or no-oplock situations, Cisco WAAS can safely apply many optimizations to
improve performance.
For example; when a user is editing a single Microsoft Word file, WAAS employs all available
optimizations to improve performance, and also for Microsoft Access database files and other
collaborative data sets; when a single user is working with an object, Cisco WAAS employs the
full optimization suite. When multiple users are working with the same file, if necessary,
WAAS will downgrade its level of CIFS acceleration to accommodate multi-user scenarios in
order to preserve data integrity, coherency, and safety.
Cisco WAAS is effective for the most common CIFS applications including Microsoft Office
(Word, PowerPoint, Excel), MS Access (and other database applications that use CIFS),
computer-aided design/computer-aided manufacturing (CAD/CAM) applications, My
Documents storage, desktop backup and restore, and other applications such as imaging.

1-86

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WAAS CIFS Acceleration Services


Two services can be configured on a WAE to handle
different responsibilities:

IP
Network

NAS

Files

Edge
Service

Connectivity Directive
Optimized Connections

Core
Service
Core
Service
Core Cluster

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-9

Connectivity Directive
To perform acceleration for CIFS, a Wide Area Application Engine (WAE) must be configured
with the appropriate WAFS service:

WAFS Edge Service: The WAE with this service is deployed in close proximity to the
remote users that need high performance access to the centralized file server. This WAE
performs protocol optimizations, such as latency mitigation, read-ahead, operation
batching, message prediction, metadata caching, and safe file caching (with file coherency
validation). Note that this WAE always propagates messages that are critical to data
integrity and protocol correctness (authentication, authorization, file OPEN requests, file or
region LOCK requests, synchronous WRITE requests, flushes) to the origin server.

WAFS Core Service: The WAE with this service is deployed in close proximity to the file
servers that remote users need to access. The WAFS Core WAE performs termination of
protocol acceleration and provides aggregation and access to the centralized file servers. A
fan-out ratio of 50:1 is enforced on WAFS Core devices, meaning that a WAFS Core WAE
can support up to a maximum of 50 WAFS Edge devices.
A core cluster must be configured, even if only one WAFS Core WAE is deployed. Edge
WAEs are mapped to Core Clusters and not directly to the Core WAEs themselves. The
Core Cluster component enables high availability, load-sharing, and fail over between all
Core WAEs within the cluster.
A WAE configured with 2GB of memory or more can be configured to run both the WAFS
Edge and WAFS Core services concurrently. Running these services concurrently on a
WAE that has less than 2GB of memory is not supported and not recommended. It is also
recommended that any device participating as a WAFS Core have a minimum of 2GB of
memory.
Configuring a WAE with WAFS Edge and WAFS Core services concurrently should only
be used in scenarios where the WAE is:

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-87

1-88

Adjacent to file servers that remote users want to access; in this case, configure the Core
service

Adjacent to users that want to access remote file servers that have nearby WAEs; in this
case, configure the Edge service

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Data Caching and Integrity


Edge file segment caching and meta data caching:
Data is cached on-demand as files or directories are opened
Prepopulation of edge cache via CDN-like prepositioning

Coherency, concurrency, and ACL:


Cache validation guarantees no stale data is served
File locking and AAA are handled synchronously with server
IP
Network

NAS

OPEN
FILE.DOC

Files
AAA, OPEN, LOCK

FILE.DOC

Edge

2007 Cisco Systems, Inc. All rights reserved.

APPROVED, LOCKED, VALIDATED

Core

WAAS v4.0.71-10

The WAE running the Edge File Services service is able to cache previously-seen files and
metadata information so that a file can be served locally on the next request. Each time a file is
opened, the WAE validates the state of the file with the origin file server to make sure that
cached contents are identical to those stored on the file server. The file server is also
responsible for authentication, authorization, and file locking, and the WAE propagates related
control messages synchronously using the WAN optimization capabilities of WAAS.
Prepositioning can also be used to prepopulate an Edge WAE file cache to improve the
performance of first user access.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-89

Integration with WAN Optimization


File services adaptor leverages WAN optimization
capabilities provided by DRE, TFO, and LZ:
DRE and LZ improves open and save operation performance
through compression and data suppression.
TFO enables the protocol to more effectively and efficiently use
available WAN resources.

WAN
FILE.DOC
TRANSPORT FLOW OPTIMIZATION
DRE CACHE

FILE.DOC

DRE CACHE

LZ

LZ

Edge

2007 Cisco Systems, Inc. All rights reserved.

Core

WAAS v4.0.71-11

The file services optimizations provided in Cisco WAAS can leverage the WAN optimization
components of WAAS in any scenario where information must be transferred over the WAN.
Such examples include:

Cache miss: The file does not exist in cache; in this case DRE suppresses redundant data
segments during the transfer based on the data contained in the DRE history. Persistent
Lempel-Ziv (LZ) compression is used to compress any unsuppressed segments that must be
transferred, and TFO are employed to improve the behavior of TCP. For example, if the file
has been accessed previously but has been evicted from the file cache or otherwise
considered invalid (file has changed on the server), data from the file might be resident in
the DRE compression history. Otherwise, data from other files can be used if the data is
identical to data patterns found in the file being transferred.

Cache invalidation: In this situation, a user accesses a file that has changed on the origin
server after the file was cached. Assuming the segments that comprise the file are in the
DRE cache, the transfer of the data to rebuild the new version of the file in cache happens
very quickly and consumes very little bandwidth.

Write operations: The process of saving a file is accelerated due to the DRE cache having
cached contents from the file. Only the changed data must traverse the WAN, assuming the
segments that comprise the file are still contained in the DRE cache.

Other control messages: Other control messages such as authentication, authorization, and
locking are accelerated through DRE, TFO, and LZ.

Note

1-90

Clearing the DRE cache on a WAE also restarts the CIFS acceleration services. This
causes the sessions and connections that are established to be torn down. Clients
automatically regenerate these sessions and connections.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Intelligent File Prepositioning


Intelligent prepositioning capabilities with flexible configuration to
prepopulate cache with files before the first user request
Leverages DRE and LZ compression to improve transfer
performance and user save performance

IP
Network

NAS

Files
Distribute
FILE.DOC
at 3am

Edge

Fetch
FILE.DOC

Core

FILE.DOC

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-12

File prepositioning allows an administrator to schedule the distribution of a set of files and
directories to an Edge cache. This function transfers the contents of the file to the edge cache
and also populates the DRE cache. This method significantly improves performance for the
first user access, and is helpful in situations such as:

Software distribution environments: These environments require the installation of


service packs, hotfixes, and antivirus definition files.

Software development environments and CAD/CAM: These environments typically


involving large packages containing many small or large files. Objects that have not
changed can be served from the local cache; objects that have changed are delivered in an
accelerated manner through DRE and LZ. Only a handful of these objects tend to change
on a daily basis, and the write back of updated files is accelerated due to the DRE cache.

Imaging environments: It is helpful to prepopulate an Edge device with any images that
are relevant to the operations of the day or week, for example, patient medical images.
Transfer of new images is accelerated through DRE, LZ, and TFO.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-91

File Prepositioning Job Flow


The Core WAE scans the file server and filters the file set
according to directive criteria.
The results of the filtered scan are sent to the Edge WAE, and the
Edge WAE again filters the contents of the cache to determine
what is necessary to preposition.
The Edge WAE requests each required file or file segment from
the Core WAE, which leverages DRE and LZ compression for the
data transfer.

LIST

LIST
Send FILE123.DOC

NAS

Files

Core
WAE

FILE123
DOC

2007 Cisco Systems, Inc. All rights reserved.

Edge
WAE

WAAS v4.0.71-13

When the administrator defines a preposition directive in Central Manager, the following
processes are executed:
Step 1

The Edge WAE connects to the Core WAE and sends preposition parameters. These
include:

File server to connect to

Share to gather data from

Root path to search from

File pattern to attempt to match

Whether or not to search subdirectories from the root path

Time filters

File size filters

Step 2

The Core WAE performs the scan against the server based on the criteria provided
and returns a match list, representing the results of a filtered scan, to the Edge WAE.

Step 3

The Edge WAE compares the match list against the current state of the file cache
and creates a delta list. Any file that does not exist in the cache or has been changed
is added to the delta list.

Step 4

The Edge WAE then submits requests sequentially to the Core WAE based on the
files contained within the delta list.

Step 5

The Core WAE fetches the file and stores it in the preposition staging area. The
Core WAE then instructs the Edge WAE to download the file.

Steps 4 and 5 are repeated until the delta list has been exhausted or the limitation parameters of
the preposition directive have been met.

1-92

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Note

2007 Cisco Systems, Inc.

Prepositioning is only available for CIFS file servers. Prepositioning populates the DRE
cache on both WAEs involved in the transaction. This is useful when users access files that
have changed, as the rebuild of the cache is efficient and high-performance, assuming the
segments that made up the original transfer of the file still exist in the DRE context.
Prepositioning can also be used as a mechanism for warming the DRE context for other
applications, including web, email, video, database, and others.

Cisco Wide Area Application Services

1-93

Integration Example: Software


Distribution
Transparently optimize CIFS to a remote software distribution
server to provide LAN-like access to hotfixes, service-packs, and
other updates
Preposition files to prepopulate edge cache with large packages
that users will request to improve download and installation
performance

Data Center

Branch Office

Download
ServicePak.msi
from \\pluto

Router
NAS

WAN
Core

\\Pluto\SWUpdates

Distribute
ServicePak.msi
Edge

2007 Cisco Systems, Inc. All rights reserved.

Service
Pack

WAAS v4.0.71-14

Prepositioning is a useful feature in environments where large amounts of data must be readily
available in the remote office. Using prepositioning helps consolidate not only file servers, but
also software distribution servers.

1-94

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

File Blocking
Cisco WAAS can be configured to prevent specific types of files
from being stored on the data center file server or Network
Attached Storage (NAS) device.
This strategy prevents undesirable file types from consuming
valuable WAN resources, and improves productivity.
Save
SONG.MP3
IP
Network

NAS

Files

MP3

Edge

2007 Cisco Systems, Inc. All rights reserved.

Core

WAAS v4.0.71-15

The Edge WAE can be configured to block operations against specific types of files, for
example, MP3, and JPG files. This ability helps to control the types of files that can be stored
on the data center file server, and minimizes the transfer of unsanctioned data, to eliminate the
use of network resources. In many cases, this kind of functionality can also be employed on file
servers or network-attached storage (NAS) devices also.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-95

File Services Flexible Integration


Options
Non-transparent integration using published names:
Data Center
NAS

Branch1
WAN

Core

\\Pluto\Demo

Windows
Client

Edge

Core WAE
Name: Core1

Edge WAE
Name: BR1Cache

\\BR1-Pluto\Demo

Transparent integration using WCCPv2, PBR, ACE, or Physical Inline:


Data Center

Branch1
Router

NAS

WAN

\\Pluto\Demo

Core

\\Pluto\Demo

Windows
Client

Core WAE
Name: Core1

Edge
WAE

2007 Cisco Systems, Inc. All rights reserved.

Edge
WAAS v4.0.71-16

The WAE file services optimizations can be integrated into the network in either a transparent
or non-transparent fashion.
With non-transparent mode, the WAE takes on the personality of a file server and appears as a
node on the remote office LAN. In non-transparent mode, client computers map network drives
to shares that appear to be located on the network name that the WAE is publishing on the
network. In non-transparent mode, the WAE appears as a local file server. This published name
can also be added to Dynamic Frequency Selection (DFS) as a link target, allowing users that
have already deployed DFS for global namespace capabilities to continue to leverage that
investment. In the example shown in the figure, the Pluto server in the data center has a share
called demo. When a windows client starts browsing the network, the client on the right sees a
BR1-Pluto server (the WAE) with the same share when the local WAE is configured to publish
names. All the same security, authorization, and auditing settings are applied, because WAAS
simply passes through messages critical to security and data integrity. The exported file server
shares are available in the clients network neighborhood just as any other server on the
network. Multiple servers that are represented by the WAE appear as multiple servers to the
user without having multiple WAEs. When using name publishing, the published name can be
the original name with a prefix or a suffix, or an alias name.
In transparent mode, the WAE does not appear as a file server on the remote office LAN, and
clients map network drives to shares that are being made accessible by file servers or NAS
devices in the data center. In transparent mode, the WAE relies on the network, via Web Cache
Communication Protocol version 2 (WCCPv2) or Policy-Based Routing (PBR), to provide the
WAE with flows to optimize, and the WAE provides the optimizations transparently.
CIFS traffic that is accessed through WAAS file services optimizations receives the additional
benefit of DRE, TFO, and LZ. Additionally, CIFS traffic that does not use the file services
optimizations can still leverage the benefit of DRE, TFO, and LZ, but this does not provide
performance similar to using the file services optimization capabilities of WAAS.

1-96

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

R/O Disconnected Mode of Operation


The file services application adaptor handles network
outages in the following ways:
Intermittent disconnection: For periods of less than 90
seconds, user operations are buffered, with no impact to the
user.
Prolonged disconnection: In this situation, sessions are
disconnected by the edge WAE and the core WAE.

In prolonged disconnection mode, user sessions can


be re-established to access cached files in read-only
mode, assuming a domain controller is reachable. The
WAE must be configured to join the Windows domain.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-17

WAAS is designed to be resilient during periods of WAN disconnection. Two types of WAN
outages are identified by Cisco WAAS, and each is handled in a different manner:

Intermittent disconnection: This term refers to periods of loss of WAN connectivity


lasting less than 90 seconds, in which case the WAFS Edge WAE buffers user operations.
If the WAN returns to service and the WAFS Edge WAE is able to successfully reconnect
to the WAFS Core WAE, the user sees no impact. The WAFS Edge WAE always attempts
to reconnect to the Core WAE to which it was originally connected. If the WAFS Edge
WAE is unable to reconnect to the original WAFS Core WAE, the session is broken and
regenerated. The user might see a disconnection to the file server in this case. If this
happens, the user can save data locally and merge the changes back into the document on
the file server after reconnection.

Prolonged disconnection: This term refers to periods of loss of WAN connectivity lasting
longer than 90 seconds, in which case the WAE enters a prolonged disconnection mode,
and all state is cleaned up on the Edge WAE and the Core WAE. At this point, the Edge
WAE can enter into read-only disconnected mode, assuming the file server is configured
for this mode in Central Manager. If this mode is not configured, the file server is no longer
accessible through WAAS, although offline files and folders within Windows can be
configured.

WAASv4 provides a R/O disconnected mode of operation that allows users to have read-only
access to fully-cached files during periods of prolonged WAN disconnection. A series of
functions is implemented specifically to support servers and shares defined for R/O
disconnected access:

Aggressive file caching of files accessed on-demand (read-ahead and file read-ahead):
This function ensures that files are fully cached in the Edge WAE so they can be available
if the WAE enters a prolonged disconnection mode lasting more than 90seconds.

Metadata and access list prefetch: This function ensures that access control information
is cached by the Edge WAE for the purposes of authorization during disconnection.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-97

Preposition: This optional function is used to continually update the Edge WAE cache and
ensure that files are available in the Edge WAE cache if the WAE enters a prolonged
disconnection mode.

When WAAS file services enters prolonged disconnected mode, all CIFS sessions are
disconnected in both the remote office and the data center. If read-only disconnected mode is
not configured, the user does not have access to the file server. Windows Offline Files and
Folders can be configured as an alternative to read-only disconnected mode, providing users
with the ability to continue working during the period of disconnection, and resynchronizing
changes back to the origin file server when the connection is re-established.
If read-only disconnected mode is configured, the WAEs still enter prolonged disconnected
mode, which destroys user sessions in the remote office and in the data center. User sessions
must be restarted, which requires authentication with a domain controller, which must be
reachable on the network. The WAFS Edge WAE can self-authorize the user based on cached
ACLs from the origin file server. After the user re-authenticates successfully, the Edge WAE
exports the server and acts on its behalf, providing read-only access to cached files and folders
based on the cached access control information. With read-only disconnected mode, the last set
of cached files and last set of cached ACLs is used. If the file server is unreachable through
WAAS for a long period of time and files or access control information has changed, the
contents in the Edge WAE will not be the same as those on the origin file server.
For read-only disconnected mode to work properly, the WAE must be configured for Windows
authentication and be successfully joined to the domain.

1-98

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Integrated Print Services


This topic provides a high-level description of Cisco WAAS print services, which help to
minimize the administrative traffic that must traverse the WAN.

Cisco WAAS Integrated Print Services


Many organizations have difficulty consolidating file services
because of the WAN burden created by print services traffic.
Cisco WAAS provides Windows-compatible print services to
eliminate the need for print jobs to traverse the WAN.

Data Center

Branch Office
Router

NAS

WAN

Driver
Distribution

2007 Cisco Systems, Inc. All rights reserved.

JOB
FILE

Print
FILE.DOC

WAAS v4.0.71-19

Cisco WAAS provides Windows-compatible print services to keep print jobs local to the
branch office. This keeps bandwidth-intensive print traffic from consuming precious WAN
capacity and ensures that print performance is predictable. Almost any printer is supported, as
the WAE does not require special software to support a particular printer due to the use of Raw
mode queues (client handles the rendering). Cisco WAAS printing provides printing to any user
regardless of whether the WAN is connected or disconnected, as it does not need to integrate
into a Windows domain. Cisco WAAS print services uses guest mode printing and does not
allow for the definition of access control for print queues. Each user is able to manage their
own jobs, and the print administrator is able to manage any job. This is identical to the behavior
of interacting with a Windows print server.
Cisco WAAS allocates 1GB of data to the PRINTSPOOL file system. This storage capacity can
not be manually allocated and is shared by all of the print queues. Although this storage
capacity can support a recommended maximum of 100 concurrent queues, 20-25 is the
recommended number for adequate storage allocation per queue, and there is no hard limit or
enforced maximum number of queues that can be defined. Cisco WAEs acting as print servers
support up to a maximum of 100 concurrent printing users and up to a maximum of 500
concurrent print jobs. The print job timeout is 60 seconds. Cisco WAAS print services requires
the WAFS Edge service be configured.
Cisco WAAS print services eliminates the need to leave a server in the branch office to provide
local printing capabilities. Cisco WAAS print services leverages SAMBA and Common Unix
Printing System (CUPS) to provide print services to the branch office. By using Cisco WAAS,
Windows-compatible print services can remain in the branch, keeping print jobs from needing
to traverse the WAN.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-99

WAAS print services rely on users configured through the command-line interface (CLI) for
print queue administration and print driver repository administration. No authentication or
authorization is required for print services (users will still attempt to authenticate and authorize
and any request will be responded to as a successful request), so any user in the remote office
can print to a print queue that is configured on the WAE regardless of whether the WAN is
connected or disconnected.
Cisco WAAS self-authenticates users that are attempting to print, and usernames are
maintained with the active job set. As such, a user can only modify or manipulate their own
jobs using standard Windows printer management tools. Users that authenticate to the print
server using administrative credentials can manipulate any job running on the WAE.

1-100

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Acceleration for Centralized Print


Services
Cisco WAAS can also help accelerate access to centralized print
servers and minimize bandwidth consumption for print jobs
traversing the WAN.

Data Center

Branch Office

WAN

JOB

JOB
FILE

2007 Cisco Systems, Inc. All rights reserved.

Optimized
Connections

FILE

Print
FILE.DOC

WAAS v4.0.71-20

Cisco WAAS also provides optimizations to improve performance for print jobs that traverse
the WAN from the branch office client workstation to a data center print server, as shown in the
slide. By employing CIFS acceleration, DRE, TFO, and persistent LZ compression, access to
centralized print servers and spooling of jobs consumes far less WAN bandwidth capacity and
completes in a much shorter period of time. While Cisco WAAS can be used as a print server
replacement by employing print services on WAEs deployed in the branch, IT organizations
that want to centralize print services find that Cisco WAAS provides a solution for that
environment as well.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-101

Printer Driver Distribution


Printer drivers are uploaded to the Central Manager WAE and
then distributed to edge print servers or groups of devices.
Printer drivers are then accessible at the edge of the network for
local download from PRINT$ share to support click-and-print
functionality.

Data Center
WAN

DC

Upload
Drivers

Print

Download
driver
and print

Branch Office

Distribute HP
LaserJet
Driver

2007 Cisco Systems, Inc. All rights reserved.

JOB

Print

FILE

WAAS v4.0.71-21

Central Manager can be configured as a repository for print drivers. After it is configured as a
repository, Central Manager can be accessed directly and print drivers can be uploaded to it.
After the drivers have been uploaded, they can then be distributed to Edge print server WAEs.
Printer driver distribution need be employed only when branch office WAEs are acting as print
servers in remote offices.

1-102

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Summary
This topic summarizes the key points that were discussed in this lesson.

Summary
Cisco WAAS provides CIFS acceleration services such as latency
reduction, data caching, and metadata caching to provide
LAN-like access to centralized file server or NAS storage.
Prepositioning distributes content to an edge cache based on a
schedule. This tool is useful for software packaging and
distribution, and for imaging and multimedia environments.
Cisco WAAS provides edge print services with central driver
distribution to allow print traffic to remain local to a remote office.
Alternatively, Cisco WAAS can optimize access to centralized
print servers.

2007 Cisco Systems, Inc. All rights reserved.

2007 Cisco Systems, Inc.

WAAS v4.0.71-22

Cisco Wide Area Application Services

1-103

1-104

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Summary
This topic summarizes the key points that were discussed in this module.

Module Summary
Cisco WAAS combines powerful technologies to improve
application delivery, consolidate costly infrastructure, and to
minimize WAN bandwidth consumption.
WAN optimization, including TCP optimization and advanced
compression, helps Cisco WAAS overcome network barriers to
application performance, including bandwidth, latency, and packet
loss.
Application acceleration components help Cisco WAAS improve
the performance of specific protocols such as CIFS while enabling
disconnected mode of operation, thereby permitting file server
consolidation.
Local print services deployed on the Cisco WAE help to keep
administrative traffic from needing to traverse the WAN, which
helps maintain performance expectations.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.71-2

In this module, you learned how Cisco WAAS can bridge the gap between centralized IT
infrastructures and the service needs of remote users. You learned how Cisco WAAS provides
powerful WAN optimization and application acceleration technologies to optimize application
performance over the WAN, and to allow IT to consolidate costly remote office infrastructure.
You also learned about the WAE appliance platform and router-integrated network module, and
the licensing model for Cisco WAAS.

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-105

1-106

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Self-Check
Use the questions here to review what you learned in this module. The correct answers and
solutions are found in the Module Self-Check Answer Key.
Q1)

What are three factors that hinder application performance over the WAN? (Choose 3.)
(Source: Applications and the WAN)
A)
B)
C)
D)
E)

Q2)

Which network features can be crippled by technologies that use tunnels between
acceleration devices? (Source: Cisco WAAS Introduction)
A)
B)
C)
D)
E)

Q3)

network module content engine (NM-CE)


WAE-512
WAE-612
WAE-7326

Why are application-specific optimizations needed to complement WAN optimization?


(Source: Application Specific Optimizations)
A)
B)
C)
D)

Q5)

QoS
Firewall policies
NetFlow
Access Lists
All of the above

Which Cisco WAAS hardware platform would be recommended for a data center
deployment supporting 7,500 concurrent optimized TCP connections? (Source: Cisco
WAE Performance and Scalability)
A)
B)
C)
D)

Q4)

QoS
Latency
Bandwidth
Packet Loss
TCP fragmentation

Application-specific latency
Bandwidth constraints
Sensitivity to WAN conditions
All of the above

What optimizations for file services protocols does WAAS provide? (Source: WAAS
Optimizations for File Protocols)
A)
B)
C)
D)
E)
F)
G)
H)
I)

2007 Cisco Systems, Inc.

Protocol proxy
Data cache
Meta data cache
Intelligent read-ahead
Operation batching
Message prediction
Preposition
Disconnected mode
All of the above

Cisco Wide Area Application Services

1-107

Q6)

How does WAAS handle a brief WAN outage of less than 90 seconds? (Source:
Disconnected Mode of Operations)
A)
B)
C)
D)

Q7)

What level of access does WAAS file server disconnected mode provide? (Source:
Disconnected Mode of Operations)
A)
B)
C)
D)

Q8)

Read-only
Read-write
Local file server
Asynchronous write-back

What is a common usage scenario for WAAS file preposition? (Source: Using
Prepositioning)
A)
B)
C)
D)
E)

Software distribution environments


CAD/CAM environments
Medical imaging environments
Software development environments
All of the above

Q9)

How can Cisco WAAS optimize print services?


Local print services
Centralized driver distribution
Optimize remote print server access
All of the above

Q10)

Which three of the following messages are exchanged during establishment of a TCP
connection? (Choose 3.) (Source: Introduction to TCP)
A)
B)
C)
D)

Q11)

Q12)

SYN
ACK
Finish (FIN)
Synchronize acknowledge (SYN-ACK)

What is the BDP of a DS3 with 100 ms of latency? (Source: Introduction to TCP)
A)
B)
C)
D)
E)

562 Kb
450 Kb
562 KB
4.5 MB
4.5 Mbps

What are the primary optimizations provided by Cisco WAAS TFO? (Source:
Introduction to TFO)
A)
B)
C)
D)
E)

1-108

The user session is disconnected and reconnected immediately


The user session is disconnected and reconnected after the document change
The user session is disconnected and reconnected manually
The disruption is masked and the user is not impacted

Window Scaling
Selective Acknowledgement
Large Initial Windows
Binary Increase Congestion Avoidance
All of the above

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Q13)

In what scenarios is window scaling effective? (Source: Window Scaling)


A)
B)
C)
D)

Q14)

After packet loss is detected, which TCP extension helps minimize the amount of data
that must be retransmitted? (Source: Selective Acknowledgements)
A)
B)
C)
D)

Q15)

To minimize bandwidth
To consume bandwidth
To make links virtually larger
To improve latency

What is the function of DRE? (Source: WAAS Compression Architecture)


A)
B)
C)
D)

Q19)

Improves packet loss


Improves congestion
Masks problematic WAN conditions
Compresses data

What is the purpose of WAN compression? (Source: Need for WAN Compression)
A)
B)
C)
D)

Q18)

Selective Acknowledgement
Slow-start
Congestion avoidance
Large initial windows

Which statement accurately describes the purpose of WAAS TFO? (Source: Binary
Increase Congestion)
A)
B)
C)
D)

Q17)

Selective Acknowledgement
Slow-start
Congestion avoidance
Large initial windows

What feature circumvents bandwidth starvation for mouse connections? (Source: Large
Initial Windows)
A)
B)
C)
D)

Q16)

When MWS < BDP


When MWS > BDP
When MWS = BDP
When MWS > 2X BDP

Suppress unnecessary TCP control messages


Suppress unnecessary TCP data messages
Suppress unnecessary User Datagram Protocol (UDP) control messages
Suppress unnecessary UDP data messages

Which three of the following are components of DRE? Choose three. (Source: Data
Redundancy Elimination)
A)
B)
C)
D)
E)

2007 Cisco Systems, Inc.

Fingerprint and chunk identification


Pattern matching
LZ compression
Synchronization
TCP proxy

Cisco Wide Area Application Services

1-109

Q20)

Which two of the following represent primary functions of a DRE encoder? Choose
two. (Source: DRE Encoding)
A)
B)
C)
D)

Q21)

Which two of the following represent primary functions of a DRE decoder? (Source:
DRE Decoding)
A)
B)
C)
D)

Q22)

Rebuild messages
Synchronize contexts
Eliminate redundancy
Verify message validity

Which two of the following are forms of DRE data integrity verification? (Choose
two.) (Source: DRE Decoding)
A)
B)
C)
D)

1-110

Identify chunks
Synchronize contexts
Eliminate redundancy
Verify message validity

Signature ACK/NACK
Message validity verification
Context synchronization
Shared context

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Self-Check Answer Key


Q1)

B,C,D

Q2)

Q3)

Q4)

Q5)

Q6)

Q7)

Q8)

Q9)

Q10)

A,B,D

Q11)

Q12)

Q13)

Q14)

Q15)

Q16)

Q17)

Q18)

Q19)

A,B,D

Q20)

A,C

Q21)

A,D

Q22)

A,B

2007 Cisco Systems, Inc.

Cisco Wide Area Application Services

1-111

1-112

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module 2

Designing Cisco WAAS


Solutions
Overview
This module describes how to design Cisco Wide Area Application Services (WAAS)
solutions.

Module Objectives
Upon completing this module, you will be able to design Cisco WAAS solutions, including
network design, interception method, and solution sizing. This includes being able to meet
these objectives:

Describe how Cisco WAAS is integrated into an existing network infrastructure

Describe how to size Cisco WAAS solutions based on performance, scalability, and
capacity sizing metrics

2-2

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Lesson 1

Network Design, Interception,


and Interoperability
Overview
This lesson explains how Cisco Wide Area Application Services (WAAS) and the Wide Area
Application Engine (WAE) are integrated in an existing network infrastructure, how traffic is
intercepted and redirected to the WAE, and how WAAS interoperates with existing network
functionality.

Objectives
Upon completing this lesson, you will be able to describe how Cisco WAAS is integrated into
an existing network infrastructure. This includes being able to meet these objectives:

Describe how Cisco WAEs can be deployed physically in-path within the network to
provide a simple method of network integration

Describe how Cisco WAAS can be deployed in an off-path configuration

Describe WCCPv2 and explain how it functions as a network interception option

Describe how Cisco WAAS can leverage PBR as a network interception option

Describe how Cisco WAAS can be integrated in the enterprise data center with the
Application Control Engine (ACE) line card for the Catalyst 6500 series switch

Define how devices automatically discover each other in the optimization path

Discuss how Cisco WAAS handles situations where asymmetric routing is encountered

Discuss how Cisco WAAS maintains network transparency and minimizes impact to
features that rely on packet header visibility

Physical Inline Deployment


This topic describes how Cisco WAEs can be deployed physically in-path within the network
to provide a simple method of network integration.

Network Integration Overview: In-Path


Cisco WAEs can be deployed physically in-path:
WAE devices sit physically in-path between two network elements
(such as the router and the switch).
Inspect all traffic passing through and determines which traffic to
optimize:
Interception in both directions of packet flow.
Pass through uninteresting (non-TCP) traffic at low layer.
Transparent optimizations maintain compatibility with most IOS
features and other platforms.
IP
Network

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-4

Cisco WAE devices can be integrated into the network using either an off-path deployment
mechanism (relying on the network to redirect flows) or an in-path deployment mechanism,
whereby the WAE itself is physically in the network path of all traffic. In most cases, a physical
in-path deployment is preferable only in environments where an off-path deployment is not
possible.
With physical in-path deployments, the Cisco WAE sits between the LAN switch and the next
adjacent device (generally a firewall or a router) and selectively optimizes flows that are based
on the configured policy of the traffic being seen. For instance, all non-TCP traffic is
immediately passed through the device without modification. For TCP traffic that is traversing
the device, the WAE examines the configured policy to see if there is a match and if
optimization is configured.
Cisco WAE in-path deployment requires use of the four-port in-path network card. This in-path
network card provides fail-to-wire functionality to ensure that if the WAE fails (loss of power,
hardware failure, software failure, other conditions), then the path between the two adjacent
network devices is effectively bridged. In this way, a WAE failure does not block traffic
between the LAN and the WAN.

2-4

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAE Physical Inline Deployment


Physical inline interception:
Physical in-path deployment between switch,
and router or firewall
Mechanical fail-to-wire upon hardware,
software, or power failure
Requires no router configuration

Scalability and high availability:


Two two-port groups
Serial clustering with load-sharing and fail-over
Redundant network paths and asymmetric
routing

Seamless integration:
Transparency and automatic discovery

Cisco WAE
4-Port Inline Card

802.1q support, configurable VLANs


Supported on all WAE appliances
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-5

Physical inline interception allows the WAE to be physically deployed in the path between the
LAN switch and the next upstream device, which is generally a WAN router or firewall. In this
position, all traffic passes through the WAE, and optimizations can be applied based on the
configured policy. The WAE physical inline card includes a mechanical fail-to-wire that is
triggered upon any hardware error, nonrecoverable software error, or power failure, and thereby
ensuring network connectivity is not permanently interrupted in the event of a WAE failure.
The WAE inline card provides four 1000BaseTX copper Gigabit Ethernet interfaces. These
interfaces are grouped in two two-port groups, and each two-port group represents an inline
pair. If a WAE is deployed between a single router and a single switch, for example, only one
inline pair is used. With two two-port groups, the WAE can be deployed between two switches
and two routers (or two firewalls). Additionally, WAEs can be serially clustered back to back in
the physical path to provide load sharing and failover.
The Cisco WAE inline card supports 802.1q and allows for the explicit configuration of
VLANs to be examined for optimization. The card is supported in all WAE appliance models.
Cisco WAAS is transparent to the underlying network, and WAEs have the ability to
automatically discover each other.
For high-availability environments, WAEs with inline cards can also be clustered in a serial
(daisy-chain) fashion. Serial clustering is compatible with WAE automatic discovery, and
during the process of automatic discovery, the outermost WAE endpoints always take on
ownership of optimization. In the case of a serially-clustered pair of WAEs, the outermost
WAE always is the first to take on ownership of optimization, and only in the case of the
outermost WAE reaching TCP connection capacity does the inner WAE begin applying
optimization. At this point, the outermost WAE is unable to take on additional connections to
optimize and handles them as pass-through connections. This form of clustering and load
sharing is commonly referred to as spill-over load balancing.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-5

Inline Interception Deployment Modes


In-Path, Single WAE, Single WAN Connection
MGMT
WAN
WAE1

In-Path Cluster, Single WAN Connection


MGMT
WAN
WAE1

WAE2

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-6

This figure shows two examples of inline deployments; one with a single WAE device, and the
other with a serial cluster of two WAE devices. In both configurations, the WAEs
management port (generally GigabitEthernet 1/0) must also be attached to the LAN switch and
assigned an IP address in a routable VLAN for management and other purposes. The WAEs
must be deployed between two LAN-capable devices, that is, between a switch and a router, or
between a switch and a firewall.
If a firewall is present in the network and the firewall is providing VPN tunnel termination,
then the WAE is deployed between the switch and the firewall. If a firewall is present in the
network, and the firewall is providing security services only (no tunnel termination), then the
WAE is deployed between the firewall and the router.

2-6

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Inline Interception Deployment Modes (Cont.)


In-Path, Single WAE, Redundant WAN Links
WAN
MGMT

WAN
WAE1

In-Path Cluster, Redundant WAN Links


WAN
MGMT

WAN
WAE1

2007 Cisco Systems, Inc. All rights reserved.

WAE2

WAAS v4.0.72-7

This figure shows the same deployment situations as the previous slide, but in this example,
two WAN connections are present. In this example, both inline port groups on each WAEs
inline card are used, one inline port group per LAN-WAN connection.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-7

In-path WAE Configurations


WAE In-line Adapter

IP
Network

IP
Network

Router LAN
Interface
LAN

LAN

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-8

The figure shows a basic in-path configuration. Notice that in this basic configuration, only one
inline group is used.
When connecting a WAE in between two devices using GigabitEthernet, either a straightthrough cable or crossover cable can be used on either side or any combination thereof.
When connecting a WAE in between two devices using FastEthernet, a straight-through cable
is used on one side, and the other cable is the opposite of the cable type used to connect the two
devices natively without a WAE in between them. For example, if connecting a WAE with
FastEthernet between a router and a switch, a straight-through cable is used on one side, and a
cross-over cable is used on the other, because the crossover cable is the opposite of what is
normally used to connect the switch to the router.
If a firewall appliance is physically in-path between the router and the switch, the WAE inline
card WAN interface is connected to the firewall LAN interface.

2-8

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

In-path WAE Configurations


IP
Network

IP
Network

WAE In-Line Adapter

IP
Network
IP
Network

2007 Cisco Systems, Inc. All rights reserved.

Router LAN
Interface

Router LAN
Interface

WAAS v4.0.72-9

This example shows a basic in-path configuration with redundant WAN connections and
redundant LAN. Notice that both inline groups are used, and the original network paths are
retained by injecting the WAE into the path with the same inline group for each of the original
connections.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-9

In-path WAE Configurations (Cont.)


IP
Network

Router LAN
Interface

WAE
In-Line Adapter

WAE
In-Line Adapter

LAN
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-10

WAE appliances can also be clustered inline. This kind of clustering is called serial clustering
or spill-over clustering. The WAEs must be configured connected back-to-back. With serial
clustering, if one WAE fails, or otherwise becomes overloaded, then the other is able to provide
optimization. The figure shows a basic serial cluster configuration.

2-10

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

In-path WAE Configurations (Cont.)


IP
Network

IP
Network

Router LAN
Interface

Router LAN
Interface

WAE
In-Line Adapter

WAE
In-Line Adapter

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-11

The example in the figure shows a serial cluster with redundant WAN connections, routers, and
switches. For single-switch situations, the two LAN interfaces of the inline card of the WAE
closest to the LAN both can be connected to the same LAN switch.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-11

Hierarchical Network Placement


The Core layer is typically
reserved for highperformance forwarding.
The Distribution layer
provides an optimal
deployment location for
WAAS.
The Access layer can be
used, but it is too contained
to be used for large scale
optimizations.

2007 Cisco Systems, Inc. All rights reserved.

Core

Distribution

Access

WAAS v4.0.72-12

WAEs should be deployed in a location conducive to allowing Cisco WAAS to provide a high
degree of optimization. LANs are typically designed with three autonomous layers, each
serving a different purpose:

Core: The Core Layer provides high-speed forwarding between Distribution Layers and
the aggregation of multiple-Distribution Layers. The Core Layer often provides a gateway
to the Internet edge, which is generally another Distribution Layer. Network services are
typically not implemented in the Core, because high-speed forwarding is desired.

Distribution: The Distribution Layer provides aggregation for connected Access Layer
networks. The Distribution Layer is commonly responsible for routing and other network
services, such as security, server load balancing, and wireless switching. End nodes such as
servers, workstations, and wireless access points are rarely attached to the Distribution
Layer.

Access: The Access Layer is typically departmentalized or deployed within a small


physical boundary and provides network connectivity to end nodes, including servers,
workstations, and wireless access points.

In single subnet or small network scenarios, WAEs are typically deployed in proximity to the
network boundary router. In multisubnet or large network situations, WAEs are typically
deployed in the network Distribution Layer.
WAE deployment in the Access Layer is not recommended unless WAAS is being used for a
specific set of nodes or for a specific application. Access Layer deployment makes termination
of optimization difficult for nodes connected to other Access or Distribution Layers. If a WAE
is deployed in the Access Layer, modifications must be made to the routing topology to
intentionally send traffic to the WAE if it is desired that traffic to other Access Layer nodes be
optimized.
This level of optimization relative to deployment locale is called locality. High locality to a
reference point means that the device in question is deployed very close (generally, directly
attached) to that reference point. Low locality to a reference point means that the device in
2-12

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

question is deployed in a way that is not in proximity (generally, not directly attached to) of
that reference point. Generally speaking, a high degree of locality to the WAN entry-exit point
is most preferred, because it provides the broadest level of optimization.
High locality to the core yields a more global level of optimization:

Intercept traffic going to and coming from the WAN exclusively (based on placement of
interception).

Closer to the WAN entry-exit prevents intrasite access from traversing the WAEs.

Provides optimization for all attached Distribution and Access Layers.

High locality to the Access Layer yields a more focused level of optimization:

Optimization is restricted to a specific Access Layer unless significant changes to network


routing are introduced.

Can cause intrasite access to traverse the WAEs, which causes unnecessary WAE resource
utilization.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-13

Hierarchical Network Placement: In-Path

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-13

The same principals regarding locality and physical location in the network hold true for both
methods of Cisco WAAS integration (in-path and off-path). For in-path deployments, it is
recommended that Cisco WAEs be deployed as close to the WAN boundary point as possible,
as shown in the example in this figure.

2-14

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WAE Placement With Respect to


Firewalls
If the firewall is providing VPN tunnel termination:

WAN
WAE

Firewall

If the firewall is providing security services but no VPN tunnels:

WAN
Firewall

2007 Cisco Systems, Inc. All rights reserved.

WAE

WAAS v4.0.72-14

If a firewall is present in the network, and the firewall is providing VPN tunnel termination,
then the WAE should be deployed between the switch and the firewall. If a firewall is present
in the network, and the firewall is providing security services only (no VPN tunnel
termination), then the WAE should be deployed between the firewall and the router.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-15

Off-Path Network Deployment


This topic examines how the Cisco WAE devices can be physically deployed within the
network as off-path nodes.

Network Integration Overview: Off-Path


Cisco WAE devices attach to the LAN as an appliance.
WAE devices rely on packet interception and redirection to enable
application acceleration and WAN optimization:
Interception in each site where deployed.
Interception in both directions of packet flow.

Transparent optimizations maintain compatibility with most IOS


features and other platforms.

IP
Network

Cisco WAE

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-16

Cisco WAE devices can be deployed as in-path (using the inline card) or off-path devices. As
off-path devices, the Cisco WAE device attaches to the LAN just like any other node on the
network. Packet interception and redirection methods within the network itself are used to
forward packets to the WAE for examination and optimization. Network interception must be
established for both directions of traffic flow and be available in two or more locations between
communicating nodes. This positioning is necessary to facilitate WAE optimizations, which
require decoding at the distant end of the network. The optimizations applied by the WAE are
transparent to ensure compatibility and compliance with features that are already configured in
the network routers, switches, and firewalls.

2-16

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Network Integration Overview: Integrated


Or Cisco WAE devices attach as service modules in a
Cisco ISRs:
WAE network module deployed physically in the ISR
Deployed as an off-path service module (using network
interception) using an internal network or external network interface
Identical in configuration, management, monitoring, and behavior to
an off-path appliance using WCCPv2 or PBR

IP
Network

Cisco WAE
Network
Module

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-17

Another form of off-path deployment is physical integration, that is, the NME-WAE (network
module enhanced WAE) router-integrated network module. The NME-WAE is actually an offpath device deployed physically in the router and leverages WCCPv2 or PBR for traffic
interception and redirection. Other than the physical integration aspects of the service module,
the NME-WAE acts and behaves like any off-path appliance does.
Note

2007 Cisco Systems, Inc.

The NME-WAE must be configured with WCCPv2 or PBR for traffic interception. Inline
interception is not possible with the NME-WAE, and the ACE module can not be used to
integrate a large number of NME-WAEs.

Designing Cisco WAAS Solutions

2-17

Router Integration: Internal Architecture


Cisco Integrated Services Router
LAN
LAN
interface
interface

WAN
WAN
interface
interface

WAN

Service
Service
Module
Module
Internal
Internal
interface
interface

Service
Service
Module
Module
interface
interface

Cisco NME-WAE Network Module

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-18

The internal architecture of the NME-WAE router-integrated network module is identical to the
architecture of an off-path appliance. The module itself has two Gigabit Ethernet interfaces
(GigabitEthernet1/0, which is internal, and GigabitEthernet2/0, which is external). The internal
Ethernet interface connects directly to an internal interface on the router itself.
When configuring the router to support the NME-WAE, a separate subnet is required. An IP
address within this subnet must be assigned to the routers internal IP address (that the NMEWAE directly connects to) and the NME-WAEs internal interface must also be assigned an IP
address within this subnet.
Note

2-18

The NME-WAE subnet must be reachable throughout the network.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Off-Path Deployment
When deployed off-path, the WAE must not be
attached to the same segment as the interface
performing redirection.
This allows traffic handled by the WAE to return to the
router where interception is not configured, preventing
an infinite loop.
PBR or
WCCPv2
IP
Network

Infinite Loop
WAE Device
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-19

Cisco WAAS is a transparent solution for WAN optimization and application acceleration,
meaning that key information is preserved to allow network functions to continue operation,
including packet header information. Given that the router has no way to distinguish whether
traffic has already been redirected to a WAE for optimization, the WAE must be deployed on a
subnet that is separate from the nodes that are optimized by the WAE (separate physical
interface or separate logical interface, that is, a separate VLAN). Traffic interception is
configured on the router interface adjacent to the client or server, which redirects traffic to the
WAE. After optimization, the WAE returns traffic on an interface where interception is not
configured and explicitly configured to be excluded from future redirection operations. This
process prevents redirection of the traffic that has already been handled by a WAE and
eliminates the possibility of infinite interception loops. Generally, the WAE is deployed on a
routable subnet that is shared only by routers with no clients or servers.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-19

Off-Path Deployment Options


PBR or
WCCPv2

Tertiary Interface

Fa0/0

IP
Network

Fa1/0
Redirect
Exclude

PBR or
WCCPv2

Subinterface

Fa0/0.10

IP
Network

Fa0/0.20

Redirect
Exclude

WAE Device
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-20

Two examples of an off-path deployment are shown in the figure:

Tertiary interface: In a tertiary interface deployment, the WAE is installed on a subnet


that is connected to a physical router interface that is not shared. Redirection is configured
on the user interface using PBR or WCCP, and a redirect exclude command is applied to
the router interface to prevent packets sent by the WAE from being redirected, again, by
the router (if WCCP is being used). The tertiary interface represents the recommended
deployment method.

Note

2-20

The NME-WAE uses a tertiary interface internally , because the routers internal interface is
dedicated to connectivity to the NME-WAE.

Subinterface: In a subinterface deployment, the WAE is installed on a VLAN that is


separate from the user VLAN but sharing the same interface through 802.1q VLAN
trunking. In this method, the WAE shares the same physical router interface with the users
but not the same logical interface as VLANs and subinterfaces that are being used. From
the routers perspective, a logical interface exists for each VLAN, and each is treated as an
autonomous interface. Redirection is configured on the subinterface for the user VLAN,
and a redirect exclude command is applied to the adjacent router VLAN subinterface. The
subinterface provides the same level of isolation as the tertiary interface but uses only a
single, shared physical interface.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

One-Arm Off-Path Deployment


1
4

IP
Network

2
3

Pros:

Cons:

Simplicity

Performance constrained

Single interface

Higher router utilization

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-21

Each WAE (appliances and the NME-WAE) have a minimum of two Gigabit Ethernet
interfaces. The simplest form of off-path deployment, called one-arm, uses only one of those
two interfaces. In one-arm off-path deployments, one of the WAEs Gigabit Ethernet interfaces
is attached directly to the router or to the dedicated WAE VLAN on the switch. The second
interface is not used.
In one-arm off-path deployments, all traffic to and from the WAE must pass through the router,
even if the traffic is destined for a user in the same site. While one-arm off-path deployment is
the simplest deployment mode (one interface to configure), it can create additional workload on
the router, which may translate to additional CPU utilization. The amount of extra workload
and CPU utilization is subjective and based on throughput and number of packets per second.
Note

2007 Cisco Systems, Inc.

One-arm off-path is the deployment mode that is used only when configuring the internal
interface of the NME-WAE network module.

Designing Cisco WAAS Solutions

2-21

One-Arm Off-Path Deployment


IP
Network

The WAE interface


must be routable.
This is the primary interface. The default
gateway is the attached router interface IP.

On-router one-arm deployment:


Single interface for optimizations and management.
Directly attached to the router.
WAE must be reachable throughout the network.
The default gateway is the router interface that WAE is adjacent to.
Primary interface must be set to physical interface.
PortChannel is not permitted in this configuration (router support).
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-22

The example in the figure shows the details and requirements of the on-router one-arm
deployment mode. In this mode, the WAE has an interface directly attached to an interface on
the router. A single interface is used for both management and optimization traffic.
The WAE must have its default gateway configured as the adjacent router interface IP address,
and the primary interface must be set to the interface that is attached to the router. PortChannels
are probably not permitted in this situation, because most routers do not support PortChannel.
In all configurations, the WAE VLAN must be routable and the WAE must be reachable
throughout the network.
This is the default configuration for the NME-WAE. The internal interface is always on router,
that is, always directly connected to the router through the backplane to the internal router
interface dedicated to the NME-WAE.

2-22

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

One-Arm Off-Path Deployment (Cont.)


IP
Network

Primary interface or PortChannel


Default GW is router WAE VLAN IP
VLAN must be routable

dot1q
Black = client VLAN
Red = WAE VLAN

Off-router one-arm deployment:


Single interface or PortChannel for optimizations and
management
WAE VLAN must be routable and WAE must be reachable
Default gateway is router WAE VLAN subinterface IP address
Primary interface must be set to physical interface or
PortChannel
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-23

An equally common one-arm deployment mechanism, as shown in the figure, is with the WAE
connected off-router, i.e. attached to the LAN switch. In this mode, the WAE has one or both
interfaces attached to a switch port in a VLAN that is connected through an 802.1q trunk to the
router. The VLAN that the WAE is attached to is separate from the VLAN that the users or
servers are attached to. In this mode, a single interface (or PortChannel of both interfaces) is
used for both management and optimization traffic.
The WAE must have its default gateway configured as the adjacent router VLAN subinterface
IP address, and the primary interface must be set to the interface (or PortChannel) that is
attached to the optimization VLAN. PortChannels are permitted in this scenario, assuming the
switch that the WAE is attached to supports PortChannels.
This configuration is not applicable to the NME-WAE, as the internal interface is directly
connected to the router via the backplane.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-23

Two-Arm Off-Path Deployment


1
IP
Network

Pros:

Cons:

Better performance

Additional switch port consumed

Lower router utilization

Additional configuration
Interface adjacency to node
Usually feasible in branch only

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-24

An alternative to the one-arm off-path deployment mode is to use the second interface (the
external interface on the NME-WAE). The first interface is connected directly to the router or
to the switch in the WAE VLAN, and the second interface connects to the switch in the VLAN
with the users or the servers. With this deployment mode, all traffic from the WAE has to go
through the router with the exception of any traffic going to a node adjacent to the WAEs
second interface (users or servers that are in the same VLAN). Two-arm off-path mode can
provide a slight improvement in performance and less overall router CPU workload.
Note

2-24

The NME-WAE can support this configuration if the external interface is used. The internal
interface is always directly connected to the router, so there is only one cable going from the
NME-WAE to the switch.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Two-Arm Off-Path Deployment (Cont.)


Client VLAN
IP
Network

The secondary interface resides


in the client or server access
VLAN for improved performance:
The VLAN must be routable
No default gateway is needed

WAE interface
Management interface
Primary interface
Optimization interface
Default gateway router WAE VLAN IP
WAE must be reachable

On-router two-arm deployment:


The primary interface is also used for management and optimizations.
The secondary interface resides on the user VLAN to improve
performance.
All WAE interfaces must be reachable throughout the network.
PortChannels are not supported in this configuration.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-25

The figure shows a configuration that is on-router in two-arm deployment mode. In this mode,
the WAE has an interface directly attached to an interface on the router to support optimization
and management, and an interface attached to the client VLAN. This mechanism is commonly
used to enable improved performance, assuming users or servers are Layer 2 adjacent to the
WAEs secondary interface, because only one default gateway can be configured on the WAE.
Performance is improved because traffic can be returned to the user or server directly (Layer 2
adjacency is required) without crossing the router.
As with all deployment situations, the WAE interfaces must be attached to routable VLANs
and reachable throughout the entire network. The WAE must have its default gateway
configured as the directly connected router interface IP address, and the primary interface must
be set to that interface as well. PortChannels are not permitted in this situation, because the
interfaces are connected to two separate devices.
This configuration is applicable to the NME-WAE when using the external interface (shown as
the secondary interface in this figure). The internal interface is always directly connected to the
router through the backplane, and the default-gateway of the NME-WAE always points back to
the routers internal interface IP address.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-25

Two-Arm Off-Path Deployment (Cont.)


Dot1q trunking
Black = client VLAN
Red = WAE VLAN
IP
Network

The secondary interface resides


in the client or server access
VLAN for improved performance:
The VLAN must be routable
No default-gateway is needed

WAE VLAN
Management interface
Primary interface
Optimization interface
Default gateway router WAE VLAN IP
The VLAN must be routable

Off-router two-arm deployment:


The primary interface is also used for management and optimizations.
The secondary interface resides on the user VLAN to improve
performance.
All WAE interfaces must be reachable throughout the network.
PortChannels are not supported in this configuration.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-26

This configuration is off-router in two-arm deployment mode. In this mode, the WAE has both
interfaces directly attached to the LAN switch, but both interfaces are in separate VLANs. One
interface provides optimization and management, and is configured as the primary interface.
The other interface is used to improve performance, assuming users or servers are Layer 2
adjacent to the WAEs secondary interface, because only one default gateway can be
configured on the WAE. In this mode, traffic can be returned directly to the user or server
without crossing the router.
This deployment situation is common when higher levels of performance are desired and the
WAE can not physically attach one of its interfaces to the router. As with all deployment
situations, the WAE interfaces must be attached to routable VLANs and reachable throughout
the entire network. The WAE must have its default gateway configured as the IP address of the
routers WAE VLAN subinterface, and the primary interface must be set to that interface as
well. PortChannels are not permitted in this situation, because the interfaces are connected to
two separate VLANs.
The WAE does not support 802.1q trunking in off-path mode.
This configuration is not possible with the NME-WAE, because the internal interface is always
directly attached to the router through the backplane.

2-26

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Hierarchical Network Placement: Off-Path

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-27

For off-path deployments, it is generally recommended that WAEs be deployed with high
locality to the WAN routers or the device performing interception. Generally, this device is one
of the following:

Integrated Services Router in the branch office (WCCPv2, PBR)

A Catalyst 6500 in the data center (WCCPv2, PBR, ACE)

A data center WAN router (WCCPv2, PBR)

WAEs are deployed in a location conducive to allowing Cisco WAAS to provide a high degree
of optimization. LANs are typically designed with three autonomous layers, each serving a
different purpose:

Core: The Core Layer provides high-speed forwarding between Distribution Layers and
the aggregation of multiple Distribution Layers. The Core Layer often provides a gateway
to the Internet edge, which is generally another Distribution Layer. Network services are
typically not implemented on the Core Layer, because high-speed forwarding is desired.

Distribution: The Distribution Layer provides aggregation for connected Access Layer
networks. The Distribution Layer is commonly responsible for routing and other network
services, such as security, server load balancing, and wireless switching. End nodes, such
as servers, workstations, and wireless access points, are rarely attached to the Distribution
Layer.

Access: The Access Layer is typically departmentalized or deployed within a small


physical boundary and provides network connectivity to end nodes, including servers,
workstations, and wireless access points.

In single subnet or small network scenarios, WAEs are typically deployed in close proximity to
the network boundary router. In multisubnet or large network scenarios, WAEs are typically
deployed in the network Distribution Layer.
WAE deployment in the Access Layer is not recommended unless WAAS is being used for a
specific set of nodes or for a specific application. Access Layer deployment makes termination
2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-27

of optimization difficult for nodes connected to other Access or Distribution Layers. If a WAE
is deployed in the Access Layer, modifications must be made to the routing topology to
intentionally send traffic to the WAE should it be desired that traffic to other Access Layer
nodes be optimized.
This level of optimization relative to deployment locale is called locality. High locality to a
reference point means that the device in question is deployed very close (generally, directly
attached) to that reference point. Low locality to a reference point means that the device in
question is deployed in a way that is not in proximity (generally, not directly attached) to that
reference point. Generally speaking, a high degree of locality to the WAN entry-exit point is
most preferred, because it provides the broadest level of optimization.
High locality to the Core Layer yields a more global level of optimization:

Intercept traffic going to and coming from the WAN exclusively (based on placement of
interception)

Closer to the WAN entry-exit point prevents intrasite access from traversing the WAEs

Provides optimization for all attached Distribution and Access Layers

High locality to the Access Layer yields a more focused level of optimization:

2-28

Optimization restricted to a specific Access Layer unless significant changes to network


routing are introduced

Can cause intrasite access to traverse the WAEs which causes unnecessary WAE resource
utilization

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Interception Using WCCPv2


This section introduces Web Cache Communication Protocol version 2 (WCCPv2) and
discusses redirection, load balancing, and failover. WCCPv2 represents the preferred method of
network interception for WAAS.

Cisco WAE WCCPv2 Deployment


WCCPv2 interception:
Out-of-path with redirection of flows
to be optimized (all flows or selective
flows via a redirect-list)
Automatic load-balancing, load
redistribution, fail-over, and failthrough operation

Original
Flow

Scalability and high availability:


Up to 32 WAEs within a service
group and up to 32 routers
Linear performance and scalability
increase as devices are added

Seamless integration:

Service
Group
Interception
Redirection

Optimized
Flow

Transparency and automatic


discovery
Supported on all WAE platforms
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-29

WCCPv2 is an out-of-path interception mechanism. WCCPv2 allows all WAE devices to be


physically deployed off-path and relies on the router, switch, or firewall to selectively or
globally redirect traffic for optimization or de-optimization. With WCCPv2, the load is
automatically balanced across available nodes. When a device fails, it is removed from the
group, and its workload is redistributed to another WAE. If no WAEs are available, no traffic is
redirected, and all flows are routed normally.
WCCPv2 supports up to 32 routers and 32 WAEs within a service group and provides linear
performance and scalability as additional devices are added. WCCPv2 is supported on all Cisco
WAE platforms.
It is recommended that the following IOS versions be installed when using WCCPv2 for Cisco
WAAS.

IOS Routing Platforms


The IOS major version; M-train recommended version, and T-train recommended versions are:

IOS 12.1: 12.1(14), 12.1(3)T

IOS 12.2: 12.2(26), 12.2(8)T0c

IOS 12.3: 12.3(13), 12.3(14)T5

IOS 12.4: 12.4(10), 12.4(9)T1

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-29

Note

For IOS routers and the NME-WAE, 12.4(9)T1 must be installed or a later version from the
T-train.

Switching Platforms
The recommended switching platforms are:

Catalyst 6500 with Supervisor 1a, Supervisor 2, or Supervisor 32 with IOS 12.1(27)E

Catalyst 6500 with Supervisor 720 (Native mode) with IOS 12.2(18)SXF5

Catalyst 6500 with Supervisor 720 (Hybrid mode) with IOS 12.2(18)SXF5 and CatOS 8.5

Catalyst 4500 or 4900 with IOS 12.2(31)SG

Security Platforms
The recommended security platform is:

PIX 515/535 or ASA 5500 with version 7.1.2.4

Note

2-30

Earlier software versions can work; however, they are not recommended due to bugs,
stability, compatibility, or other reasons.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Introduction to WCCPv2
WCCPv2 tells the network which packets to redirect to the WAE.
Up to 32 WAEs:

Act as service group clients


Perform traffic optimization

Up to 32 routers:

Act as service group servers


Perform traffic inspection and redirection

WAE1

R1

WAE2

Traffic Flow
R2

IP
Network

R32

WAE32

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-30

Cisco developed WCCP within Cisco IOS software to enable routers and switches to
transparently redirect packets to nearby caching devices. WCCP does not interfere with normal
router or switch operations. Using WCCP, the router redirects requests to configured TCP ports
and sends the redirected traffic to caching devices instead of the intended destination. WCCP
also balances traffic loads across multiple local caches while ensuring fault-tolerant and failsafe operation. As caches are added or deleted from a cluster, the WCCP-aware router, or
switch, dynamically adjusts its redirection map to reflect the currently available caches,
resulting in maximized performance and content availability.
WCCP allows content caching devices to join a service group with a router. A service group
defines the types of protocols that can be optimized by a cache and the types of traffic that can
be processed. The router manages the service group and determines how to forward appropriate
traffic to another service group member (cache). Because WCCP is only an interception
mechanism, nodes must be able to reach one another natively.
WCCPv2 is used with Cisco WAAS to provide the WAE with messages that can be optimized
and to enable high availability, load balancing, and failover to provide an enterprise class
solution. With WCCPv2, the WAE is configured with a router list, and the appropriate services
are enabled. In the case of Cisco WAAS, the TCP promiscuous mode services are enabled
(service group 61 and 62), and the WAE is instructed to join these service groups with the
router. Assuming the service groups are also configured on the router, the WAE joins the
service groups with the router, and the router begins forwarding packets that match the service
group criteria to the WAE for local handling.
With WCCPv2, up to 32 routers can participate in a service group as service group servers, and
up to 32 WAEs can join a service group as service group clients, thereby allowing for
scalability and fault tolerance. WAEs can join service groups without disruption. The function
of the server routers is to manage the service group and examine ingress or egress traffic, based
on redirection configuration, to see if that traffic matches the criteria of the running service
groups. When matching traffic is identified, the server forwards that traffic to a client according
to the load distribution mechanism that is specified. The job of the client WAEs is to specify
the configuration parameters for the service group and receive traffic from the server routers.
2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-31

When this traffic is received, the WAE can apply optimizations or perform specialized handling
of the traffic.
Note

Service group servers are the most commonly used routers, but many switches and firewalls
can also run WCCPv2 as well.

WCCPv2
WCCPv2 is an Internet draft and not a standard. The Internet draft is authored by Cisco and has
been widely accepted as the de facto standard for content routing and network interception. A
copy of the WCCPv2 Internet draft can be found at http://www.wrec.org/Drafts/draft-wilsonwrec-wccp-v2-00.txt.
The following chronology identifies the messages sent and received in a WCCPv2 service
group connection:
1. The cache transmits a WCCP2_HERE_I_AM message to each defined router or multicast
address. This message contains details about the cache including the IP address and the
service groups that the cache wants to participate in. After receiving this message, the
router responds with a WCCP2_I_SEE_YOU message if the cache meets group
membership criteria that was specified by the shared-secret message digest algorithm 5
(MD5) authentication password or access list. Upon receipt of the WCCP2_I_SEE_YOU
message from the router, the cache responds with another WCCP2_HERE_I_AM message
with the Receive ID field matching that of the router message.
2. At this point, the cache becomes usable, and the router begins redirecting traffic to the
cache based on service group assignments. WCCP2_HERE_I_AM and
WCCP2_I_SEE_YOU messages are sent every 10 seconds as a service heart beat. WCCP
forwards traffic to an available cache using either Layer 2 redirection or generic routing
encapsulation (GRE), the default. One of the components of the WCCP2_I_SEE_YOU
message is the advertisement of supported forwarding mechanisms. If a method is not
listed, it is assumed that GRE tunneling is to be used by default. Redirection negotiation is
done per service. A cache and a router can use different redirection mechanisms for
different services. Layer 2 redirection specifies that the redirecting router is to rewrite the
Ethernet frame address and forward the updated frames to the cache. GRE tunneling
specifies that a GRE tunnel is to be built between the router and the cache, and the original
packets are encapsulated in this tunnel and delivered to the cache.

2-32

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WCCPv2 Interception
Service group servers monitor interfaces based on
interception configuration criteria to identify traffic to be
redirected to a service group client:
Ingress redirection (inbound): When applied to an interface, the
router monitors traffic entering an interface to see if it matches
criteria for any of the running service groups.
Egress redirection (outbound): When applied to an interface, the
router monitors traffic leaving an interface to see if it matches
criteria for any of the running service groups.

FastEthernet0/0
Redirect In

Serial0
Redirect Out
Router

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-31

Interception is the process of examining packets that are traversing an interface and comparing
them to criteria defined for active service groups. When a packet matches the relevant criteria
(protocol), it is redirected to a service group client (WAE).
Interception is configured on the router in one of two ways:

Ingress redirection: The router examines traffic as it enters an interface from an external
network. Generally, ingress redirection is less resource intensive than egress redirection.

Egress redirection: The router examines traffic as it is leaving an interface toward an


external network. Generally, the traffic has already entered an interface prior to reaching
the interface where egress redirection is configured, and a load has already been placed on
the router resources.

Note

2007 Cisco Systems, Inc.

Ingress redirection is recommended to minimize the impact of WCCPv2 on router resources.

Designing Cisco WAAS Solutions

2-33

WCCPv2 Redirection
Service group servers (routers) can use one of two
methods to redirect traffic to a WAE:
GRE: This method is most commonly used. The entire packet is
encapsulated into a new IP packet that is destined for the WAE.
Layer 2 redirect: Less frequently used, but common with LAN
switches. The original frame header is rewritten with the WAE
MAC address as the destination and then forwarded to the WAE.

Interception monitors for


traffic that matches any
configured service groups

Redirection forwards the


traffic to a service group client
using GRE or Layer 2 redirect

WAE Device
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-32

Redirection occurs after traffic has been identified and intercepted. Redirection occurs when a
router moves an intercepted packet to a service group client WAE, based on the load-balancing
mechanism defined for the service group. WCCPv2 redirection is performed using one of two
redirection mechanisms:

GRE: GRE redirection is the more commonly used redirection mechanism when working
with routers and firewalls as service group servers. With GRE redirection, the WCCP
router maintains a GRE tunnel connection to the WAE and forwards packets that match the
service group criteria to the WAE. With GRE redirection, the WAE does not need to be
adjacent to the network router. Instead, the WAE can be one or more hops away. The
tunnel provides the connectivity and plumbing necessary to carry packets.

Layer 2 redirect: Layer 2 redirect is more commonly used when working with LAN
switches. With Layer 2 redirect, the WCCP router does not maintain a tunnel with the
WAE. Instead, the packets are intercepted and the WCCP router rewrites the Ethernet
frame header and forwards the packets to the WAE. With Layer 2 redirect, the WAE must
be Layer 2 adjacent to the WCCP router.

As a best practice, use GRE encapsulation when working with routers for WCCPv2, as most
routers do not support L2 redirection. L2 redirection is used when working with switches, as
redirection can be performed in hardware. This approach minimizes the CPU workload on the
WAE and the switch, and can improve performance.

2-34

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WCCPv2 Load Balancing


WCCPv2 allows for load balancing based on a number
of parameters, including source or destination
information (IP address, subnet, or port).
Cisco WAE devices running WAAS use two service
groups that request the redirection of any TCP traffic:
Service group 61: All TCP traffic, load-balance on source IP
Service group 62: All TCP traffic, load balance on destination IP

These service groups ensure that traffic is redirected to


the same WAE for both directions of packet flow.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-33

WCCPv2 provides a configurable assignment method that allows load balancing across service
group clients within the service group. This load-balancing scheme can use a hash algorithm of
the parameters specified (IP address, port, and so on) and distribute the load across the service
group clients, either evenly or based on configurable weighting.
With Cisco WAAS, two service group numbers are used to make up the TCP promiscuous
service group. These two service group numbers are 61 and 62. Both of these service groups
notify the router that any traffic that matches IP Protocol 6 (TCP) is to be intercepted and
redirected to a service group client WAE. Service group 61 performs load balancing based on
source IP address, while service group 62 performs load balancing based on destination IP
address.
With WAAS, both service groups must be in the path of traffic flow. For example, if a client is
communicating with a server, one of the two must be in the path for packets traveling from the
client to the server, and the other must be in the path for packets traveling from the server to the
client.
This approach ensures that the same WAE is used as the redirection destination regardless of
direction. As node-to-node communications involve a source IP address and a destination IP
address, which are generally reversed in the reverse path, the source IP address going one way
is the destination IP address for traffic going in the reverse direction.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-35

WCCPv2 Availability Monitoring


WCCPv2 keepalive (heart beat) information is exchanged every
10 seconds between WAEs and the routers.
If a WAE is unresponsive for three consecutive heart beats, it is
removed from the service group.
WCCPv2 heart beat is stateful and process based.
WAE Devices

Router

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-34

WCCPv2 uses HERE_I_AM and I_SEE_YOU messages as heart beat messages. These
messages are exchanged every 10 seconds to verify that nodes are available. If a router does not
see a HERE_I_AM message for 25 seconds, it begins the process of removing the WAE from
the service group. This process begins with a unicast query sent from the router to the WAE to
check if it is ready to be removed from the group. If no response is received, the WAE is
removed. If the WAE chooses to remain in the service group, then it notifies the router of this
determination.
The heart beat is considered stateful in that the service group server and clients are in constant
communication with processes that are tightly coupled with the WAE core optimization
function. If a software interruption occurs on a WAE that prevents it from functioning, the
device remains on the network and the likelihood of a running WCCP process is low. In this
manner, WCCPv2 heart beat and availability monitoring is considered very stable and reliable.

2-36

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WCCPv2 Failover
If a WAE in a service group fails, the portion of the load that it was
handling is automatically distributed to other WAEs in the service
group.
If no additional WAEs are available, the service group is taken
offline, and packets are not redirected.

Buckets 86128
Buckets 185

Buckets 86170

Buckets 129170
Buckets 171255

X
A

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-35

If a WAE fails, or is otherwise removed from a service group, the portion of the load
represented by the buckets that the WAE is handling is distributed to the remaining WAEs in
the service group. If no other WAEs are available, the service group is taken offline, and any
configured instances of interception are negated. These instances of interception still appear in
the router configuration but are not used, because no service group client devices are available.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-37

Interception Using Policy-Based Routing


This topic examines Policy-Based Routing (PBR) as a network interception option and
discusses how it supports high availability.

Cisco WAE PBR Deployment


PBR:
Out-of-path with redirection of flows to be
optimized (all flows or selective flows with an
access list).
WAE is treated as a next-hop router.

Original
Flow

High availability:
Failover capability allows a secondary WAE to
be used if the primary WAE fails.
IP SLAs ensure availability by tracking WAE
liveliness.

Policy Route
WAE = Next Hop

Seamless integration:
Transparency and automatic discovery.

Optimized
Flow

Supported on all WAE platforms.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-37

PBR is another interception mechanism that can be used with Cisco WAAS. With PBR, one or
more WAEs is configured as a next-hop router for TCP traffic. PBR supports failover, but it
does not support load sharing. WAE availability can be tracked through IP Service Level
Agreements (IP SLAs) in Cisco IOS.
PBR is supported on all Cisco WAE platforms.

2-38

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Introduction to PBR
The router can use PBR to forward traffic to a WAE as a next-hop
router.
WAEs:
Perform traffic optimization

WAE1

WAN routers:

Perform traffic inspection


Forward to WAEs as next-hop routers

R1
Traffic Flow

WAE2

WAE3

WAE4

2007 Cisco Systems, Inc. All rights reserved.

IP
Network

R2

R3

R4

WAAS v4.0.72-38

While WCCPv2 is the preferred network interception and redirection mechanism for WAAS,
PBR provides an alternative for situations where WCCPv2 can not be used. With PBR, the
WAE is configured as a next-hop router for specific types of traffic as defined by access lists.
To use PBR with WAAS, PBR route criteria is based on access lists that specify TCP traffic
coming from or going to specific IP subnets.
PBR provides support for an unlimited number of WAEs, but it is generally limited to two
because of the active-standby nature of PBR. An unlimited number (limited only by address
space) of WAN routers can be deployed that use the WAEs as next-hop routers.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-39

PBR Failover
PBR uses only the first available next-hop as
configured; only one WAE is used at a time.
If the previous WAE becomes unavailable, the next
configured next-hop WAE is used.
If no WAEs are available, the policy route is not used
by the router, and traffic is forwarded normally.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-39

While WCCPv2 provides load balancing and scalability, PBR can use only one next-hop route
at a time per route map. In this way, multiple WAEs can be deployed in a location, and each
can be listed as a next-hop router. However, the first WAE is the only one receiving packets,
and the remainder are unused until the first WAE is taken offline or otherwise failed.
If all WAEs go offline, the policy route is considered invalid, and the routing tables built by
routing protocols or static routes are used.

2-40

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

PBR Availability Monitoring


PBR can be configured to use IP SLAs to verify the
availability of next-hop (WAE) devices before using
them as a next-hop router.
IP SLAs can track the availability of a WAE using
ICMP echo on a configurable interval:
A 20-second interval is recommended.
Alternately, IP SLAs can use CDP or TCP connection attempts
(not recommended).

If a WAE does not respond to the ICMP echo in a


timely fashion, the route is considered unavailable and
is not used.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-40

PBR can use IP SLAs to monitor the availability of a next-hop router WAE to ensure that the
next hop is available before sending packets to that device. IP SLAs can monitor the
availability of a WAE with one of three mechanisms:

Cisco Discovery Protocol (CDP) neighbor relationship: This method requires that the
WAE is directly attached to the router with no intermediary device. The router checks the
CDP database to verify that the next-hop router WAE is online before considering it as a
valid next hop.

Internet Control Message Protocol (ICMP) echo: This method does not require a direct
attachment to the router. With this method, the router periodically sends ICMP echo
messages to the WAE to see if the WAE responds. ICMP echo is the preferred mechanism
for next-hop router tracking when using WAAS. The time between ICMP echoes is
configurable, and is configured to use the lowest interval for checking WAE availability
(20 seconds).

TCP connection attempts: This method does not require a direct attachment to the router.
With this method, the router periodically attempts TCP connections to the next-hop device
on a specific TCP port using a specific source TCP port. The period of time between
connection attempts is not configurable.

If the WAE is unresponsive to the configured tracking mechanism, the router considers the
next-hop to be unavailable and proceeds to the next next-hop router in the list. If all next-hop
routers are offline, the router considers the policy route invalid and uses the routing tables built
by standard routing protocols or static routes.
Note

2007 Cisco Systems, Inc.

IP SLAs that use ICMP echo are recommended because the granularity is 20 seconds,
whereas the granularity for CDP and TCP is much higher.

Designing Cisco WAAS Solutions

2-41

WCCPv2 Versus PBR


WCCPv2 is the preferred method of traffic interception
and redirection:
Provides stateful monitoring of WAE availability.
Up to 32 WAEs and 32 routers are supported in a service group.
Provides load balancing and failover.

PBR is used only in situations where WCCPv2 can not


be used:
Provides no stateful (process-based) monitoring.
Provides no scalability or load balancing.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-41

WCCPv2 is the preferred mechanism for traffic interception and redirection. PBR is used only
in cases where WCCPv2 can not be used. WCCPv2 is process based and provides stateful
monitoring of WAE availability, scalability to 32 nodes and 32 routers within a service group,
load distribution, failover, and high availability. PBR provides basic failover capabilities but
relies on nonstateful mechanisms to verify next-hop availability, whereas WCCPv2 failover is
handled by a running process that is closely tied to the optimization framework of the WAE.

2-42

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Data Center Deployment Using ACE


This section describes how Cisco WAAS can be integrated in the enterprise data center with the
Application Control Engine (ACE) line card for the Catalyst 6500 series switch.

Cisco WAE ACE Deployment


ACE:
Industry-leading scalability and performance for
the most demanding data center networks

WAN

Supports up to 16Gbps throughput, 4M


concurrent TCP connections, and 350K
connections-per-second setup

Optimized
Flow

Seamless integration:
Fully integrated with the Catalyst 6500 Series
of intelligent switches

Catalyst
6509 with
ACE

Transparency and automatic discovery


Supported on all WAE appliances

Original
Flow

Industry-leading functionality:
Solution for scaling servers, appliances, and
network devices
Virtual partitions, flexible resource assignment,
security, and control
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-43

Cisco WAAS can be deployed by installing ACE or CSM line cards for the Catalyst 6500
Series of intelligent switches. This solution meets the needs of the most demanding enterprise
data center environments in terms of performance, scalability, and availability. The ACE
module can scale to 4 million TCP connections with a setup rate of 350 thousand TCP
connections per second and provide up to 16Gbps of throughput. Additionally, ACE represents
the industrys leading solution for server load balancing, network device load balancing,
virtualization, and application control.
Note

2007 Cisco Systems, Inc.

Cisco WAAS can also be deployed by using the Cisco Content Services Module (CSM) line
card for the Catalyst 6500 Series switch. The CSM is functionally similar to the ACE module
in the case of Cisco WAAS integration, but it does not provide the same level of
performance or functionality as the ACE module.

Designing Cisco WAAS Solutions

2-43

ACE Data Center Entry Configuration

WAN

Data Center
LAN

DCE configurations allow Cisco WAAS to be used to


accelerate traffic going to any destination within the data
center (can be selectively filtered); providing less locality.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-44

The Cisco ACE module for the Cisco Catalyst 6500 series switch is recommended as an
interception option in the enterprise data center. While WCCPv2 can scale to support 32
devices, the ACE module can scale to support thousands of devices.
The ACE module can be deployed in one of two configurations:

Data center entry (DCE): The DCE configuration provides the highest degree of locality
to the WAN boundary, which allows Cisco WAAS to be used to optimize a broad degree of
applications within the data center. Generally, the DCE configuration is applied in the
Distribution or Core Layer of the data center LAN.

Server Load Balancing (SLB): The SLB configuration provides the highest degree of
locality to a set of servers, which allows Cisco WAAS to be used to optimize those servers
only, or, traffic to any location that happens to pass through that particular Access Layer.
Generally, the SLB configuration is applied in the distribution or Access Layer of the data
center LAN.

The figure shows the DCE configuration of ACE for Cisco WAAS.
Note

2-44

It is generally recommended that the DCE configuration be used, as it provides a broader


degree of optimization applicability. The SLB configuration is generally only recommended in
environments where asymmetric routing can not be overcome in the network and the
optimization needs to be moved as close to the endpoints as possible. Alternatively, source
NAT can be used on the ACE module to circumvent asymmetric routing, but this does not
provide source IP transparency.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

ACE Server Load Balancing Configuration

WAN

Data Center
LAN

SLB configurations allow Cisco WAAS to be used to


accelerate traffic going to a specific set of servers within
the data center (can be selectively filtered), more locality.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-45

The figure shows the SLB configuration of ACE for Cisco WAAS. This configuration is useful
in environments where asymmetric routing can not be overcome within the network by moving
optimization closer to the nodes being optimized.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-45

Which Interception Method to Use?


Inline

PBR

WCCPv2

ACE

Number of
Active WAEs

2
(serial cluster,
tested limit)

32

16000
(not practical but
possible)

Maximum
Number of
WAEs

2
(serial cluster,
tested limit)

IOS Dependent

32

16000
(not practical but
possible)

Maximum
Number of
TCP
Connections
(with WAE7326)

15K

7.5K

240K

4M

Maximum
Throughput

Up to 2Gbps
(two inline pairs)

Up to 1Gbps

Up to 32Gbps
(platform dependent)

Up to 16Gbps
(platform
dependent)

Recommended
Use

Only if WCCPv2
not possible

Only if WCCPv2
not possible and
inline not possible

Generally
Recommended

Very large scale


data center
deployments

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-46

Which interception mechanism is recommended? WCCPv2 is the recommended interception


mechanism for the branch and for the data center. For branch environments where WCCPv2
can not be used and an off-path deployment is required, PBR should be used. For branch
environments where WCCPv2 can not be used and an in-path deployment is permitted,
physical inline should be used. For data center environments where WCCPv2 can not be used,
or where WCCPv2 will not scale to meet design requirements in terms of number of devices or
TCP connections, ACE should be used.

2-46

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Automatic Discovery
This topic examines how Cisco WAE devices automatically discover each other in the
optimization path and describes situations where more than two, or less than two, WAE devices
are located in the optimization path.

Cisco WAAS Auto-Discovery


Cisco WAAS uses TCP connection establishment
packets to automatically discover WAE devices in the
path between communicating nodes.
By applying a small amount of data to the TCP
connection establishment messages, Cisco WAAS can
automatically:
Discover the WAEs closest to the communicating nodes
Define which optimizations to apply, based on policy

Auto-discovery occurs on a connection-by-connection


basis to allow different optimizations to be applied to
different traffic types or nodes.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-48

Cisco WAAS provides an automatic discovery mechanism that allows WAE devices to
automatically identify each other during the course of a TCP connection setup. With autodiscovery, administrators do not need to define the subnets that can be optimized, the devices
that can be optimized, the devices that can terminate optimizations, or the communication
between WAEs.
Auto-discovery uses TCP packets to identify itself and the other WAE devices that are in the
network path. This is accomplished with the TCP option during the setup of the TCP
connection between the communicating nodes. The TCP option allows the WAEs that are
closest to the communicating nodes to establish a peering relationship and then negotiate the
level of optimization to apply to the connection. The TCP option specifies the first node to see
the option (the WAE that marked the packet) and the last node to see the option (which is
constantly overwritten as each additional WAE in the network path sees the marked packets).
The auto-discovery mechanism occurs on a connection-by-connection basis, which allows the
WAE to use different levels of optimizations for different nodes or different applications.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-47

Typical TCP Connection Setup

Attempt connection:
Src port, Dst port
Sequence Number
Window Size, Checksum
Options (MSS, SACK, etc.)
Acknowledge connection:
Sequence Number
Acknowledgement Number
Window Size, Checksum
Options (MSS, SACK, etc.)
GET HTTP/1.1

TCP SYN
TCP SYN, ACK

TCP ACK

Acknowledge connection:
Attempt connection:
Src port, Dst port
Sequence Number
Acknowledgement Number
Window Size, Checksum
Options (MSS, SACK, etc.)

(Applicatio
n Data)

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-49

Cisco WAEs automatically discover each other during the establishment of TCP connections
between communicating nodes as shown in the figure.
A TCP connection setup is called a three-way-handshake. The three-way-handshake establishes
connection parameters between two communicating nodes to ensure guaranteed, reliable
delivery of application data. The TCP synchronize and start (SYN) message is issued when a
node wants to establish a connection with another node. The SYN packet contains information
about the source and destination TCP ports, window size, and TCP options. The receiving node
responds with a TCP SYN acknowledgement (ACK), which establishes the connection for the
reverse direction. After the SYN ACK has been received by the initiated node, it responds with
a TCP ACK, and applications using the connection are able to exchange application data.

2-48

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Auto-Discovery: TCP SYN


When the client sends a TCP-SYN packet, WAE1 applies TCP
options to identify itself and specifies the optimizations that it
wants to apply.
The modified TCP-SYN packet is then forwarded to the server
and intercepted on the other side.

WCCPv2
or PBR

WCCPv2
or PBR

WAN
A:B TCP SYN

A:B TCP SYN


(Marked)

WAE1

2007 Cisco Systems, Inc. All rights reserved.

I would like
to accelerate
this connection!
Here are my details.

WAE2

WAAS v4.0.72-50

When the WAE receives a TCP-SYN packet, it adds a 12-byte Cisco-unique TCP option. This
TCP option includes a request to optimize the connection and a definition of the requested
optimization. In addition to the requested optimizations, the TCP option also includes the
device ID of the WAE that first received the TCP-SYN packet. This device ID is added to a
field that identifies the last WAE to receive the TCP-SYN packet. The packet is then returned
to the network for delivery to the destination. At this point, no optimizations have yet been
applied.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-49

Auto-Discovery: TCP SYN (Cont.)


After WAE2 receives the TCP-SYN packet with the options
marked, it knows WAE1s details and desire to optimize this
connection.
The TCP-SYN packet is then forwarded to the server.

WCCPv2
or PBR

WCCPv2
or PBR
WAN
A:B TCP SYN
(marked)

WAE1

WAE2

2007 Cisco Systems, Inc. All rights reserved.

Now I know about


WAE1 and which
optimizations are
desired

WAAS v4.0.72-51

When the next WAE in the path receives the marked TCP-SYN packet, it knows about the first
WAE in the path and its request to optimize the connection. That WAE then adds its device ID
to the last WAE field in the TCP Options field and forwards the packet to the intended
destination. Notice that the TCP-SYN packet is still marked in case additional WAEs are closer
to the intended destination.

2-50

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Auto-Discovery: TCP SYN ACK


When the server responds with the TCP SYN ACK, WAE2 marks
TCP options to acknowledge the optimization and to identify itself
to WAE1.
The marked TCP SYN-ACK packet is forwarded toward the client
and intercepted on the other side.
A

WCCPv2
or PBR

WAN

B:A TCP SYN ACK

B:A TCP SYN ACK


(marked)

WAE1

2007 Cisco Systems, Inc. All rights reserved.

WCCPv2
or PBR

WAE2

Acknowledge
acceleration!
Here are my details.

WAAS v4.0.72-52

When WAE2 receives an unmarked TCP SYN-ACK message, it knows that it is the closest
WAE to the node. WAE2 then applies a Cisco-specific 12-byte option to the TCP SYN-ACK
packet with its device ID listed as the first WAE and also the last WAE to see the SYN-ACK
packet. WAE2 then specifies the optimizations that it wants to apply to the connection based on
the configured policy and forwards the packet to the intended recipient.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-51

Auto-Discovery: TCP SYN ACK (Cont.)


After WAE1 receives the TCP SYN ACK with the optimization
confirmation and details about WAE2, the defined policy (or
negotiated optimizations) can be acknowledged.
The TCP SYN-ACK packet is forwarded to the client.

WCCPv2
or PBR

WCCPv2
or PBR

WAN

B:A TCP SYN ACK

WAE1

ACCELERATION
CONFIRMED!

WAE2

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-53

When WAE1 receives the marked TCP SYN-ACK packet, it discovers that WAE2 is the
closest WAE to the destination of the connection. Because WAE1 was the first to see the TCPSYN message (it was unmarked), it strips the marking and forwards the TCP SYN-ACK packet
to the destination.

2-52

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Auto-Discovery: TCP ACK


After the SYN ACK is received, the TCP proxy is initiated for the
connection, and WAE1 sends a TCP ACK to WAE2 to
acknowledge optimizations.
WAE2 can then send a TCP ACK to server B.
Client A sends a TCP ACK to WAE1.

WCCPv2
or PBR

A:B TCP ACK

A:B TCP ACK

WAE1
2007 Cisco Systems, Inc. All rights reserved.

WAN

WCCPv2
or PBR

A:B TCP ACK

WAE2

ACCELERATION
CONFIRMED!

WAAS v4.0.72-54

To confirm that the connection should be accelerated, WAE1 sends a marked TCP-ACK packet
back toward the original destination of the connection. When this marked ACK packet is
forwarded to the destination, intercepted, and then sent to WAE2, WAE2 then learns that the
connection is to be optimized. The least common denominator of the application policy
configuration on both WAEs is used. From there, the TCP proxy is started, and separate
connections are maintained between the client and WAE1, between WAE1 and WAE2, and
between WAE2 and the server.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-53

Auto-Discovery: Summary
A

A:D SYN
A:D SYN(OPT)
A:D SYN(OPT)
D:A SYN ACK
D:A SYN ACK(OPT)
D:A SYN ACK
A:D ACK
A:D ACK

A:D ACK

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-55

The figure shows a summary of Cisco WAAS auto-discovery events:


1. As TCP SYN packets are received by WAE B, a small amount of data is added to the TCP
options of the SYN packet to identify WAE B and also the level of optimizations that WAE
B wants to apply (based on application policy).
2. The TCP SYN packet is then forwarded toward the destination and intercepted at the other
side of the network.
3. As WAE C receives the marked TCP SYN packet, it learns about WAE B and the
optimizations that WAE B wants to apply. Given that WAE C does not know if it is the
closest WAE to the destination, it changes a small amount of data in the TCP options
(modifying the last-seen WAE to include its own device ID) and forwards the packet to the
destination.
4. When the TCP-SYN packet with options is received by the destination, the destination
drops the options and responds back with a TCP SYN-ACK message.
5. The SYN ACK is intercepted and redirected to WAE C, and as the SYN-ACK packet is
unmarked, WAE C knows that it is the closest WAE in proximity to the server. WAE C
then applies TCP options to the TCP SYN-ACK packet, identifying itself and the level of
optimizations it wants to apply, and it forwards the packet to the intended recipient.
6. As the SYN ACK is intercepted and redirected to WAE B, WAE B learns about WAE C
and the level of optimizations that it can apply to the connection. WAE B then forwards the
SYN-ACK packet (without options because it knows that it is the closest WAE in the path)
to the client, and then sends a TCP ACK, which is marked with options confirming the
level of acceleration that are to be applied, to WAE C.
7. When the TCP ACK is sent, the TCP proxy service is started, and the WAEs begin
applying optimizations that are based on traffic policy.

2-54

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Auto-Discovery: Three or More WAEs


A

A:E SYN
A:E SYN(OPT)
A:E SYN(OPT)
A:E SYN(OPT)
E:A SYN ACK
E:A SYN ACK(OPT)
E:A SYN ACK(OPT)
E:A SYN ACK
A:E ACK(OPT)
A:E ACK(OPT)
A:E ACK

2007 Cisco Systems, Inc. All rights reserved.

A:E ACK

WAAS v4.0.72-56

If more than two WAEs are in the path, the intermediary WAEs go into pass-through mode for
the connection and allows the two outer WAEs to manage the connection and optimization.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-55

Auto-Discovery: One WAE


If only one WAE is in the path, then auto-discovery does
not complete, and the WAE goes into pass-through for this
connection, and no optimization is performed.
A

A:C TCP SYN


A:C SYN(OPT)
C:A SYN ACK
C:A SYN ACK
A:C ACK
A:C ACK

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-57

If there is only one WAE in the path, auto-discovery fails, and the WAE goes into bypass (passthrough) mode for the connection. No optimizations are applied to this connection. This is
common in deployments where one of the two locations involved in the TCP connection does
not have a WAE.

2-56

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Auto-Discovery and WAE Failure


If a WAE that was optimizing connections fails:
The receiving node sees segments with SEQ-ACK numbers it
was not expecting.
All optimized TCP connections handled by that WAE are reset.
There is no impact to pass-through TCP connections.

The client or server application can then re-establish a


new TCP connection, which goes through the autodiscovery process, again.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-58

The WAE TCP proxy independently manages TCP sequence (SEQ) ACK numbers. The SEQACK numbers used between the client and the WAE are different than those used between the
two WAEs, and different than those being used between the distant WAE and the server. If a
WAE fails, the receiving node sees segments with SEQ-ACK numbers that it was not expecting
and causes a TCP connection reset (RST) to be sent. The application can then choose to reestablish the connection, which begins the auto-discovery process, again.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-57

Asymmetric Routing
This topic examines how Cisco WAAS supports environments where asymmetric routing is
encountered.

Asymmetric Routing Example


In this example, the client request to the server takes the path of
R1-R2, and the server response takes the path of R3-R4-R5-R6:
Server

R2
Client

IP Network

R1

IP Network

R3

R6
R4

R5
IP Network

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-60

Asymmetric routing occurs when traffic takes one path between communicating nodes and a
different path for traffic flowing in the reverse direction. For application acceleration
technologies that do not provide auto-discovery, asymmetric routing makes connection
management difficult. In the figure, three sites are shown with three separate WAN
connections: One connection runs from site 1 to site 2, one connection runs from 2 to 3, and
one connection runs from 3 to 1. A client connection to the server uses one path, and the return
traffic uses an alternate path. Cisco WAAS supports environments that have asymmetric
routing as long as each of the boundary routers within a site (R6 and R1, R2 and R3, and R4
and R5) share the same network interception and redirection configuration. WCCPv2 is highly
recommended in environments where asymmetric routing is possible, because it is more
stateful and process-based than PBR, and because there is a higher likelihood of multiple
WAEs in locations where multiple WAN links exist.

2-58

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Asymmetric Routing and WCCPv2


To support asymmetric routing environments with WCCPv2
deployments, the WAEs must be registered with both routers.
Both routers must maintain states about WAAS WCCPv2 service
groups 61 (src-ip hash) and 62 (dst-ip hash).

Client

R1

IP
Network

WAE

WCCP Service Group


TCP Promiscuous

R2

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-61

This example shows how WAAS supports asymmetric routing when network boundary routers
share a common interception and redirection configuration. In these cases, the WAEs is
configured to register against both routers as TCP promiscuous devices.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-59

Asymmetric Routing and WCCPv2 (Cont.)


Regardless of the link where traffic entered the network,
the entry and exit routers load-balance the traffic to the
WAEs in an identical fashion based on incoming traffic:
R1

WAE

R2

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-62

As the interception configuration is shared and identical across boundary routers, network loadbalancing and distribution processes are the same regardless of the router the traffic was
forwarded to or through. Thus, either router can hash the load and forward to the same WAE.

2-60

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WAAS Asymmetric Routing Deployment


In this example, sites are not well connected (low bandwidth, high
latency), and routes between client and server are asymmetric
(R1-R2 vs R3-R4-R5-R6):
R2

Client

IP
Network

R1
IP
Network
WAE1

Server

WAE3
WAE4
R3

WAE2
WAE5
WAE6

R6

R4

R5
IP Network

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-63

The figure shows WAAS integration with asymmetric routing. Each of the boundary routers
(R1-R6, R2-R3, R4-R5) share a common network interception and redirection configuration.
Regardless of the router encountered, the same WAE is chosen. When the client sends data to
the server, it takes the R1-R2 network path. The return path to the client is R3-R4-R5-R6.
In auto-discovery and asymmetric environments where one direction has more WAEs than the
other direction, the TCP options used by Cisco WAAS also include the first device ID and the
last device ID. During the auto-discovery process, the first WAE to see an unmarked TCP
connection setup packet (SYN, SYN ACK, ACK) adds its device ID to the first device ID and
last device ID fields of the option. At each WAE that is traversed by the packet, the last device
ID field is overwritten with the device ID of the local WAE. Regardless of the number of
intermediary WAEs, the connection setup message contains only two WAE device IDs: the
first WAE to see the message and the last WAE to see the message.
In this example, two WAEs are in the path between the client and the server for one direction,
and three WAEs are in the path between the server and the client. Auto-discovery and sharedinterception and redirection configurations enable WAAS to support such topologies.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-61

Asymmetric Routing and Well-Connected Sites


Well-connected sites can be treated as a single site. This is helpful
for situations involving nearby data centers that are well connected.
This single site is configured with multiple WAN connections and
common service groups.
Server

R2
IP
Network

WCCP Service Group


TCP Promiscuous
WAE
WAE
DWDM
R4
IP
Network

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-64

Sites that are well connected using links that are high bandwidth (>=100Mbps) and low latency
(<2ms) and low packet loss can be treated as a single site from a network interception and
redirection configuration perspective. This approach helps to minimize the challenges
associated with deploying WAAS in environments where two well-connected data centers are
located in a given geographical region. In these situations, you can have router-to-WAE GRE
traffic traversing the high-speed low-latency interconnect without causing problemsunless
the network is already saturated, and a large volume of WAN traffic is to be redirected to a
WAE on the other side of the link.
Note

2-62

This configuration is not recommended for sites that are not well connected or provide low
bandwidth, high latency, or high packet loss.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Asymmetric Routing and Well-Connected Sites


(Cont.)
In this example, sites are well connected (high bandwidth, low
latency), and routes between client and server are asymmetric
(R1-R2 vs R4-R3):
Server

R2
Client

IP
Network

R1
IP
Network

WAE3
WAE1
WAE2

DWDM

R3
WAE4

R4
IP
Network
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-65

The example in the figure shows a deployment using network interception and redirection
configuration sharing across well-connected sites. In some cases, using WCCP can have an
impact on the route taken by return traffic. In the example, the server is trying to use R4 as its
return path to the client. When WCCPv2 on R4 intercepts and redirects the traffic to WAE3,
WAE3 uses R2 as its default gateway.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-63

Asymmetric Routing and Inline Deployment


With physical inline interception, the WAE can sit in-path for up to
two WAN connections concurrently, and thereby address
asymmetric routing in most situations.

Server

R2

Client

WAE2

R1
WAE1

R3
WAE2
R6

R5

R4

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-66

Similar to asymmetric routing environments where WCCPv2 or PBR is used, physical inline
interception can also be deployed in such a way that asymmetric routing is supported. In the
example in the figure, each WAE is deployed physically inline between the local switch and the
two WAN routers. If traffic takes a divergent path on the return, then it still passes through the
same WAE, assuming that the WAE is physically inline for both WAN connections.

2-64

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Asymmetric Routing and ACE

WAE2

2007 Cisco Systems, Inc. All rights reserved.

WAE2

WAAS v4.0.72-67

The Cisco Application Control Engine (ACE) behaves in a similar way to WCCPv2 in terms of
common asymmetric routing situations. ACE can be configured in a high-availability cluster
such that traffic handled by a cluster member on a divergent path can be forwarded to the
appropriate ACE module and then redirected to the appropriate WAE based on the flow state
maintained by the WAE. The WAE then handles the flow accordingly and places the flows
back on the network toward the intended destination with the configured default gateway. In
the example in the figure, if the WAE default gateway is the router on the right, the router on
the right is used, thereby bypassing the configuration on the server, which uses the router on the
left as its default gateway. If the default gateway is an HSRP or VRRP virtual router address,
then the physical router that owned the HSRP or VRRP virtual router address at that time is
used.
Note

2007 Cisco Systems, Inc.

In complex asymmetric routing environments, the ACE module can be configured to use
source IP address network address translation (source NAT). By using source NAT, packets
sent out by the ACE through the WAE have the original IP address masqueraded, replaced
by the IP address of the ACE in that VLAN. In this way, when the end node responds to the
packet, the response goes to the IP address of the ACE as opposed to the IP address of the
other end node.

Designing Cisco WAAS Solutions

2-65

Network Transparency
This topic describes the superiority of Cisco WAAS network transparency over solutions that
masquerade traffic with encapsulation or tunnels, and how friendliness is maintained with
features that are already deployed in the network.

Network Transparency
Cisco WAAS preserves
Layer 3 and Layer 4
header information to
allow the network to
retain visibility in
network flows:

Src MAC AAA Src IP 1.1.1.10 Src TCP 15131


Dst MAC BBB Dst IP 2.2.2.10 Dst TCP 80

Application Data

Source and destination IP


address
Source and destination
TCP port

Maintaining this
visibility minimizes the
impact to other network
features and mitigates
the need for feature
reconfiguration.

Src MAC BBB Src IP 1.1.1.10 Src TCP 15131 Optimized


Optimized
Dst MAC AAA Dst IP 2.2.2.10
Dst TCP 80

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-69

Cisco WAAS supports all three facets of transparency: client transparency (no client software
or configuration changes), server transparency (no server software or configuration changes),
and network transparency (no feature reconfiguration). Network transparency is a feature
unique to Cisco. Network transparency enables Cisco to maintain compliance and compatibility
with the largest set of features already configured in the packet network. Many features require
visibility into Layer 3 (IP) and Layer 4 (TCP) packet header information to make
determinations on how specific types of traffic are to be handled.
Many non-Cisco application acceleration products use a dynamically configured or staticallydefined IP tunnel between devices. Tunnels can masquerade critical packet header information
from the network and force network administrators into drastic feature reconfigurations. For
instance, if TCP port information is masqueraded by a non-Cisco accelerator, the upstream
router is unable to then determine what application the traffic is related to, which can cause
problems for things such as access lists, firewall policies, or quality of service (QoS). If IP
address information is masqueraded, then similar problems can occur.
Only those network features that require visibility to the application payload can be impeded by
WAAS. If Cisco WAAS is configured to compress and suppress redundant data from a flow,
and the network feature is configured to examine the traffic after the WAE optimization, the
feature is not be able to locate the strings it needs to perform an assigned function.

2-66

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Network Feature Compatibility


Feature

Supported?

Importance

Layer 3 and Layer 4 header visibility

Yes

Provides visibility to network features,


monitoring, reporting.

DSCP and TOS preservation

Yes

Preserves QoS markings on packets received by


the WAE.

QoS

Yes

Traffic classification, prioritization, and marking in


the router is fully supported. WAN QoS is not
supported, but it is also not necessary.

CoS

Not necessary

Not necessary because this is typically used in


WAN L2 protocols not LAN.

NBAR

Partial

Classifies packets based on application payload,


assumes NBAR classification results in DSCP
marking, first few packets are not compressed.

Router queuing configuration

Yes

Use different queuing mechanisms for different


traffic types as necessary.

Policing, shaping, and rate limiting

Yes

Rate limiting and policing is fully supported; can


police or rate limit based on compressed data
depending on where deployed. Shaping is
supported.

MPLS

Partial

MPLS decapsulation occurs before packets are


forwarded to the WAE. WCCP and WAE are not
VRF-aware

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-70

The goal of WAASv4 with transparency is to maintain compatibility with a broad set of IOS
capabilities.
The following defines features from the Network Feature Compatibility table:

Layer 3 and Layer 4 header visibility: Cisco WAAS provides full visibility to IP and
TCP header information. This information is critical to the proper operation of network
value-added features such as those found in this table. Non-transparent accelerators do
NOT provide IP and TCP header preservation, which defeats many of these features.

DSCP and TOS preservation: Cisco WAAS preserves existing DiffServ Code Point
(DSCP) and Type of Service (TOS) markings to comply with QoS configurations. For any
packet received by a WAE containing a pre-existing marketing in the ToS byte, the ToS
byte will be applied to the outgoing (optimized or pass-through) packet to preserve end-toend QoS semantics.

QoS: Cisco WAAS preserves DSCP and TOS markings, and the L3/L4 header. As Cisco
WAAS is a transparent acceleration solution, it can leverage existing QoS capabilities in
the network, including classification (identifying the flow based on its characteristics), prequeuing operations (such as policing, marking, dropping), advanced queuing architectures
(LLQ, CBWFQ, PQ, shaping), and post-queuing optimizations (link fragmentation and
interleaving, packet header compression). Non-transparent accelerators break QoS
classification, which renders the remainder of router-based QoS nearly useless.

CoS: Cisco WAAS does not need to support class of service (CoS), as this is commonly
used in WAN Layer 2 protocols. Cisco WAE appliances use Ethernet for Layer 2.

NBAR: Cisco WAAS provides partial support for Network Based Application Recognition
(NBAR), because there can be situations where NBAR needs to find application strings
within optimized (heavily compressed) data. This occurrence would likely be infrequent, as
application strings are generally found at the beginning of the connection before
optimizations are applied.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-67

Router queuing: Cisco WAAS supports router queuing configurations such as priority
queuing, class-based weighted fair queuing, and low-latency queuing, due to packet header
preservation.

Policing, shaping, and ratelimiting: Cisco WAAS supports policing, shaping, and rate
limiting.

MPLS: Cisco WAAS has no need to support Multiprotocol Label Switching (MPLS),
because the WAE has Ethernet-only interfaces and MPLS decapsulation occurs before
packets are sent to the WAE. However, Cisco WAAS, being a network transparent
solution, provides significant value over what is provided with nontransparent accelerator
solutions in that features such as path selection are preserved due to Cisco WAAS
transparency. WCCPv2 is not yet VPN routing and forwarding (VRF) aware, so Cisco
WAAS must be deployed per VRF and WCCPv2 must be configured on a non-VRF router.

Note

2-68

This table is not an all-inclusive list.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Network Feature Compatibility (Cont.)


Feature

Supported?

Importance

ACLs

Yes

Security, filtering, control

NAT

Yes

Address translation. Payload-based NAT does


not work and is rarely used.

IDS

Partial

First few messages are unoptimized, allowing


IDS to detect intrusive strings.

IOS Firewall, PIX, ASA FWSM

Yes

Policies based on Layer 3 and Layer 4 full


support; policies based on Level 7 partial
support, because first few messages are
unoptimized.
Automatic discovery requires that option
scrubbing be disabled.

NetFlow

Yes

NetFlow is supported. Depending on where


statistics are gathered, you might see
compressed values rather than uncompressed.

Asymmetric routing

Yes

Requires use of WCCPv2 services 61 and 62 or


PBR with appropriate configuration.

VPN and IPsec tunnel

Yes

Must occur after redirection to WAE.

VoIP

Yes

VoIP call traffic not redirected to the WAE.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-71

The Network Feature Compatibility table is continued:

ACLs: Cisco WAAS supports the use of access control list to: limit access, ensure security,
utilize filtering, and provide controls.

NATs: Cisco WAAS supports NAT due to Level 3 and Level 4 header visibility.
Nontransparent accelerators do not support many NAT configurations because the original
source and destination IP and TCP information is masqueraded from the router.

IDS: Cisco WAAS provides partial support for Intrusion Detection Systems (IDS)
technologies, because WAAS might have already optimized traffic before IDS has a chance
to examine that traffic. However, the first few messages are nonoptimized to allow IDS to
detect intrusive strings. Alternatively, IDS capabilities can be deployed before Cisco
WAAS. Nontransparent accelerators have significant challenges with IDS systems because
traffic is encapsulated or tunneled between devices, thereby removing complete visibility
even at the beginning of a packet exchange.

IOS firewall, PIX/ASA, FWSM: Cisco WAAS provides support for firewalls and
configured policies continue to function. Nontransparent accelerators masquerade packet
header information, thereby defeating firewall policies. Cisco WAAS automatic discovery
requires that any scrubbing (that is, removal) of TCP options be disabled to allow the
automatic discovery TCP option to propagate. For deployments where firewalls require
consistent sequence numbers, this behavior must be changed on the firewall, because Cisco
WAAS manipulates sequence numbers to ensure optimized packets are not considered
valid by an end node.

NetFlow: Cisco WAAS supports NetFlow due to Level 3 and Level4 header visibility.
Nontransparent accelerators masquerade packet header data and use persistent connections,
thereby defeating NetFlow. In essence, the NetFlow collector sees long-lived flows
between accelerators and not actual flows between users and servers.

Asymmetric Routing: Cisco WAAS supports auto-discovery and optimizations in


environments with asymmetric routing, assuming the network interception and redirection
configuration is identical across network boundary routers.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-69

2-70

(VPN) and IPSec Tunnel: Cisco WAAS supports (Virtual Private Network) VPN and
IPsec technologies. For traffic to be optimized appropriately, the optimizations must be
applied before traffic enters the VPN-IPSec tunnel.

VoIP: Cisco WAAS can optimize VoIP control protocols running over TCP. VoIP data
(phone calls) use User Datagram Protocol (UDP) as a transport, and the traffic is not
redirected to the WAE nor is it processed when a WAE is inline. VoIP and other
noninteresting traffic are passed through at a very low layer (very quickly) whenever they
enter the WAE (for example, when in inline mode).

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Summary
This topic summarizes the key points that were discussed in this lesson.

Summary
Cisco WAE appliances can be deployed either as an in-path device or as
an off-path node on the network, typically in the Distribution layer.
The physical inline interception card provides fail-through operation,
serial clustering, and 802.1q support and allows Cisco WAAS integration
in environments where in-path is preferred or off-path is not possible.
Off-path interception mechanisms for Cisco WAAS include WCCPv2,
PBR, and the ACE line card for the Catalyst 6500 Series switch.
WAE devices automatically discover each other during TCP connection
establishment requests from communicating nodes and negotiate the
optimizations to apply.
Cisco WAAS supports environments with asymmetric routing when
network entry and exit points are configured with common interception
configuration criteria..
Cisco WAAS maintains transparency with clients and servers, and also
with the network, which minimizes costly feature reconfigurations.

2007 Cisco Systems, Inc. All rights reserved.

2007 Cisco Systems, Inc.

WAAS v4.0.72-72

Designing Cisco WAAS Solutions

2-71

2-72

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Lesson 2

Performance, Scalability, and


Capacity Sizing
Overview
This lesson explains how to use Cisco Wide Area Application Services (WAAS) to design
solutions that are based on metrics related to performance (bandwidth and application
throughput), scalability (WAN bandwidth, fan-out, number of users, and TCP connections),
and capacity sizing (physical disk capacity).

Objectives
Upon completing this lesson, you will be able to describe how to size Cisco WAAS solutions
based on performance, scalability, and capacity sizing metrics. This includes being able to meet
these objectives:

Describe the key areas of understanding required to adequately design a Cisco WAAS
solution

Describe the performance and scalability characteristics of each Cisco WAE platform and
understand the positioning of each

Describe how the WCCPv2 configuration can be manipulated to achieve design


requirements

Cisco WAAS Design Fundamentals


This topic describes the characteristics of the IT operating environment that need to be
understood before designing a Cisco WAAS solution, and which product capabilities must be
examined to ensure proper alignment of the correct platforms in a Cisco WAAS design.

Factors Influencing Cisco WAAS Design


Application performance
problems over the WAN
WAN bandwidth
upgrades and network
upgrades
Server, storage, and
application consolidation
Global collaboration
Desktop management,
software distribution,
patch management
Disaster recovery,
business continuity,
replication, backup over
the WAN
Compliance with industry
or government regulation

Business
Challenges

Number of locations and


connectivity to each:
bandwidth, latency, loss
Network architecture of
each location including
hardware, software
version, and
configuration
Network characteristics
per location including
bandwidth, latency, loss,
utilization, and paths
Ownership and
management of network
elements, feature
configurations
High-availability
requirements

Network
Considerations

2007 Cisco Systems, Inc. All rights reserved.

Workstations, servers,
operating systems, patch
revision levels
Applications in use,
protocols, file sizes, usage
characteristics, and
workload patterns
Network utilization per
application or per protocol
Collaboration requirements
Home directory capacity per
user and per location
Software distribution and
desktop management
capacity requirements
Number of users (total vs.
concurrent)

Application
Considerations
WAAS v4.0.72-4

Designing a Cisco WAAS solution is simple and straightforward; however, designers are best
positioned if they have an intimate understanding of their organizations IT infrastructure.
Before any design exercise, be sure to examine the business challenges being experienced to
ensure that Cisco WAAS can help overcome the challenge:

2-74

Application performance over the WAN: Cisco WAAS employs powerful WAN
optimization and application acceleration technologies that help to improve performance
for applications that are already centralized. Cisco WAAS also helps to ensure that
performance metrics are maintained for applications associated with infrastructure that is
being consolidated in such a way that users can access the application over the WAN.

WAN bandwidth upgrades and network upgrades: Cisco WAAS compression


technologies, such as DRE and persistent LZ compression, help to minimize bandwidth
consumption on the WAN and can help mitigate the need for a bandwidth or network
upgrade. Many users find that deploying Cisco WAAS also helps to alleviate bandwidth
utilization to provide network capacity to support other applications, such as VoIP or video.

Server, storage, and application consolidation: Cisco WAAS provides performance


improvement capabilities that allow IT organizations to consolidate costly server, storage,
and application infrastructure from distributed locations back to the data center. This
provides significant cost relief and minimizes management overhead of distributed
systems.

Global collaboration: Environments that share data across locations (for example,
marketing collateral, software development, computer-aided design/computer-aided

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

manufacturing, or CAD/CAM) Cisco WAAS enables consolidation of infrastructure and


performance improvements to make global collaboration mimic local collaboration.

Desktop management, software distribution, patch management: Cisco WAAS enables


consolidation of these services in the data center. By leveraging WAN optimization and
application acceleration capabilities, Cisco WAAS can help provide near-LAN response
times for delivery of packages from such systems while also helping to offload the servers
in the data center.

Disaster recovery/business continuance (DR/BC) and replication: Cisco WAAS helps


to improve performance of replication applications, and backup and restore over WAN
functions. This helps to ensure that geographically distributed storage repositories are kept
more closely synchronized, which helps improve recoverability if a failure is encountered.

Compliance: Cisco WAAS also helps to ensure compliance and regulations that are
focused on data availability and replication are met by enabling centralization and
improving performance for data movement facilities.

The following data about the underlying network infrastructure should be considered when
undergoing a Cisco WAAS design:

Number of locations and connectivity to each: Understand how many locations (branch
offices, regional hubs, data center locations) require optimization capabilities; how much
bandwidth is available to each site; the amount of latency between locations; and the
amount of expected packet loss.

Network architecture: Understand what devices are responsible for moving and
manipulating data throughout the network including switches, routers, firewalls, and other
appliances. Understand the software version on devices as well as the configuration of
these devices.

Per-location network characteristics: Understand how the network is used at each


location, traffic patterns, and workload and if multiple network paths exist (links that are
load balanced, active or standby). Understand path selection mechanisms in the network
and symmetry or asymmetry of network routing.

Ownership and management of network elements: Understand how the network is


managed and what role the network plays in end-to-end application performance and
visibility. For example, quality of service (QoS) configurations and requirements, network
management requirements, and monitoring and analysis requirements.

High-availability requirements: Understand which locations have high-availability


configurations, for example, redundant WAN links, redundant routers, redundant LAN
switches, and redundant connections.

The following data about the applications using the network and the nodes that are using the
applications is also important to understand:

Workstations and servers: Understand what operating systems are in use, patch revision
levels, and configuration.

Applications: Understand what applications are in use and how they are used, and the size
and frequency of use of application objects or content. Understand application protocols
and workload patterns. Identify which applications are candidate for optimization and
acceleration and how Cisco WAAS can provide benefit to these applications, potentially
even enabling centralization and consolidation.

Collaboration requirements: Understand which applications or data sets are shared


among users locally within a site or among users within distributed sites and how Cisco
WAAS can improve collaboration through performance improvements or minimize

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-75

opportunity for coherency challenges with distributed data through centralization and
consolidation.

2-76

Per-user storage capacity: Understand how users distributed throughout the enterprise use
storage capacity per user, per workgroup, and per location. Understand how much capacity
each user requires from an active working set perspective and how much history is
retained, as well as any user quotas in place.

Desktop management: Understand which IT services are in use to manage distributed


workstations, such as desktops and laptops, including software distribution systems, patch
management systems, antivirus systems and definition file updates, and remote
administration. These systems can benefit from the capabilities provided by Cisco WAAS
and might become candidates for consolidation.

User count: Understand how many users are in each location and the application workload
and productivity characteristics.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WAE Sizing Guidelines


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers
High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-5

Designing a Cisco WAAS solution requires understanding of the six components shown in the
figure:

Optimized TCP connections: Each Cisco Wide Area Application Engine (WAE) has an
optimization capacity that is defined by the number of TCP connections that can be
optimized concurrently.

WAN bandwidth capacity: Each Cisco WAE has a WAN throughput capacity.

Disk capacity: Each Cisco WAE has disk storage capacity and sizing should account for
storage requirements for each location.

License: Three licenses are available, and each Cisco WAE should be configured with the
license that is required to support its function on the WAAS topology.

Number of peers: Each Cisco WAE has a fan-out capacity defined by the number of
currently connected peers.

High availability: Any location requiring high availability (for example, data center
locations) should be designed with N+1 design principles.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-77

Optimized TCP Connections


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers

WAEs are sized based on the number of TCP


connections that are being optimized by that WAE:
Most users tend to have upwards of 15-20 TCP
connections open at a given time.
Not all TCP connections need to be optimized (generally
4-7 per user require optimization).
Configurable policies in Central Manager allow filtering
of applications to prevent unnecessary use of WAE
system resources and achieve better scale.

NetFlow can provide rough estimates of number of


connections within a given location over a period of
time.
Microsoft Performance Monitor and also netstat (per
client) can provide more accurate data.

High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-6

Cisco WAAS solutions must be designed based on a number of factors including the number of
TCP connections that are going to be optimized by each WAE within the topology. For WAEs
within branch offices, this is generally a simple calculation, however, for a WAE within a data
center serving as an aggregation point for multiple WAEs in multiple branch offices, more
consideration must be given.
Most productive users tend to have anywhere between 15 to 20 open TCP connections at any
given time. This includes connections for instant messaging, e-mail, web browsing, enterprise
applications, file shares, stock tickers, weather bugs, and many other processes running on the
users PC. Of these 15 to 20 open TCP connections, generally only 4 to 7 TCP connections
require optimization or are related to an application that can be optimized by Cisco WAAS. In
many cases the open connection count per user is less, however, 4 to 7 is a conservative
estimation that can be used safely in most cases.
For an accurate view of how many connections a user has open, two tools can be used. The first
is NetFlow, which can provide the number of completed flows in a given location. The second,
which provides real-time and historical views of connection data, is Microsoft Performance
Monitor, which is included in each Microsoft operating system available today.

2-78

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Optimized TCP Connections


Optimized TCP
Connections

Microsoft Performance Monitor:


TCP Performance Object > Connections Established

WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers
High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-7

The figure shows the output of Microsoft Performance Monitor on a user workstation. To use
Microsoft Performance Monitor (also known as perfmon) to view the real-time count of open
TCP connections, open perfmon in the Microsoft Management Console (MMC) and examine
the TCP Performance object, Connections Established counter. This shows all of the
connections open at a given time on a client workstation. The netstat utility also shows a list of
all of the open connections, including the four-tuple (source IP, destination IP, source TCP
port, destination TCP port), which allows you to discern which connections can and should be
optimized:
C:\Documents and Settings\Administrator>netstat -n
Active Connections
Proto Local Address Foreign Address State
TCP 2.1.1.207:389 2.1.1.207:9804 TIME_WAIT
TCP 2.1.1.207:389 2.1.1.207:9805 TIME_WAIT
TCP 2.1.1.207:3389 10.21.81.179:49154 ESTABLISHED
TCP 2.1.1.207:9782 171.70.145.48:80 ESTABLISHED
TCP 10.10.10.100:135 10.10.10.100:9801 ESTABLISHED
TCP 10.10.10.100:1025 10.10.10.100:1259 ESTABLISHED
TCP 10.10.10.100:1025 10.10.10.100:1261 ESTABLISHED
TCP 10.10.10.100:1025 10.10.10.100:1461 ESTABLISHED
TCP 10.10.10.100:1025 10.10.10.100:9802 ESTABLISHED
TCP 10.10.10.100:1259 10.10.10.100:1025 ESTABLISHED
TCP 10.10.10.100:1261 10.10.10.100:1025 ESTABLISHED
TCP 10.10.10.100:1461 10.10.10.100:1025 ESTABLISHED
TCP 10.10.10.100:9800 10.10.10.100:445 TIME_WAIT
TCP 10.10.10.100:9801 10.10.10.100:135 ESTABLISHED
TCP 10.10.10.100:9802 10.10.10.100:1025 ESTABLISHED
TCP 10.10.10.108:9806 10.10.10.108:445 TIME_WAIT

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-79

WAN Bandwidth Capacity


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License

WAN bandwidth capacity should be measured per


site and sized accordingly.
Size based on the maximum amount of optimized
WAN throughput:
Network tools such as NBAR or NetFlow can provide data
related to percentage of traffic
For instance, on an OC3 (155Mbps) where only 45Mbps
needs to be optimized, can size with a device capable of
supporting a 45Mbps link.
For redundant links, combine capacity if active or active
load-balanced.
Can simply size based on the full capacity of the WAN
connection.

Number of Peers
High Availability

Cisco WAE WAN throughput capacity is not


limited in software:
A device rated for 20Mbps can actually drive a larger degree
of WAN throughput.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-8

Cisco WAAS solutions need to be designed with WAN bandwidth capacity per location in
mind. Each WAE device has a recommended WAN bandwidth capacity that can be supported.
It is only necessary to size the Cisco WAAS solution based on the maximum amount of
optimized WAN throughput. For instance, if sizing for a location with a T3 (45Mbps), but only
20Mbps of that capacity needs to be optimized; then a device supporting 20 Mbps can be used
for that location. Tools found in the network such as Network Based Application Recognition
(NBAR) and NetFlow provide an accurate view of link utilization based on application to help
guide the design. If redundant links are employed in the network that is used in a load-balanced
active/active configuration, the sum total of these links should be considered when designing
for that location. If redundant links are employed in the network that are used in a failover
configuration, that is, active/passive, then the capacity of the largest link in the location should
be considered when designing for that location.
Cisco WAE devices have recommended WAN bandwidth capacity; however, the Cisco WAAS
software does not restrict the optimized output of these devices. For instance, a device rated for
a 20 Mbps link can actually drive more than 20 Mbps of optimized throughput over the WAN.

2-80

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Storage Capacity Requirements: Edge


Optimized TCP
Connections
WAN Bandwidth
Capacity

Dynamic
Dynamic Capacity
Capacity
Utility
Utility Capacity
Capacity
Static
Static Capacity
Capacity

Total
Total Storage
Storage Capacity
Capacity

Disk Capacity
License
Number of Peers
High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-9

Each Cisco WAE device has some amount of storage capacity. This capacity should be aligned
with the storage requirements for a given location. Storage requirements for a given location
can be broken down into four categories: dynamic capacity, utility capacity, static capacity, and
total storage capacity.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-81

Storage Capacity Requirements: Edge (Cont.)


Optimized TCP
Connections
WAN Bandwidth
Capacity

Edge storage-capacity requirements should be sized


based on three attributes:
Dynamic capacity: The amount of disk capacity
necessary to support a history of interactive user
operation; generally one week of compression history.

Disk Capacity

Utility capacity: The amount of disk capacity


necessary to support IT infrastructure related services,
such as software distribution, patch management,
antivirus updates, and desktop management.

License

Static capacity: The amount of disk capacity


necessary to support interactive user access to
historical data such as home directories; generally two
weeks of file history

Number of Peers
High Availability

It is not recommended, or even realistic, to try and


size a WAN optimization and application
acceleration solution based on the sum total storage
capacity of an application or an enterprise IT
organization.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-10

Dynamic capacity is defined as the amount of disk capacity necessary to support a history of
interactive user operation. This is primarily relevant to the traffic that is traversing a given
network connection over a period of time that impacts the compression history for that location.
In most cases, one week of compression history is adequate to ensure high performance access
to the most recently accessed applications, content, and other data.
Utility capacity is defined as the amount of disk capacity necessary to support IT infrastructure
and services at the remote office. These services include software distribution (application
installation files), patch management (hotfix files, service pack files), antivirus definition file
updates, and desktop management functions.
Static capacity is defined as the amount of disk capacity necessary to support interactive user
access to historical data such as home directories. Most branch offices today have file servers,
and these file servers are commonly configured in such a way that users are allowed to store
only a fixed amount of data on them (disk quotas enforce this capacity utilization limit). It is
best to provide up to two weeks of history of previously accessed files for such data in the
remote office depending on the size, type, and nature of the files being stored.
Total capacity is defined as the amount of storage capacity that was immediately available to
the users prior to deploying WAAS. It is not realistic to try and size a location based on the
amount of capacity that was readily available to users prior to deploying WAAS given that
users need interactive access only to the most frequently used pieces of data. Most data that
ages beyond one week are accessed exponentially less than data that is only one to two days
old.

2-82

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Storage Capacity Requirements: Edge (Cont.)


Optimized TCP
Connections

Dynamic capacity is defined as the amount of disk


capacity necessary to support a history of interactive
user operation.

WAN Bandwidth
Capacity

It is desirable to have a compression history that


spans the past week of previously seen network
traffic, does not need to be a scientific calculation:

Disk Capacity
License

For instance, assume a branch office with a T1


(1.544Mbps) where the WAN is 75% used for 50% of
the day for 5 days per week:
1.544Mbps / 8Bytes/bit = 192KB/s
192KB/s * 60 sec/min * 60 min/hr * 24 hr/day =
16.6GB/day
16.6GB/day * 5 days/week = 83GB/week

Number of Peers
High Availability
2007 Cisco Systems, Inc. All rights reserved.

83GB * 75% utilization * 50% of each day = 31.3GB/week

Assuming 75% redundancy, this location requires a


compression history of approximately 7 GB to support
a one week compression history
WAAS v4.0.72-11

While it is generally not necessary to embark on a study of the exact amount of disk required to
support the dynamic capacity requirements (compression history) of a given location, this slide
is provided as a reference. For instance, for a location with a T1 connection (1.544Mbps) where
the WAN is commonly 75 percent used for five days a week for 50 percent of the day, you can
calculate the amount of storage capacity required to support one week worth of compression
history data.
1.544 Mbps is equal to 192 KB/s. Conversion from bits per second (bps) to bytes per second
(Bps) requires that the value in bits per second be divided by 8. This means that the WAN
connection is able to support up to 192 KB per second of data throughput. To convert this value
to bytes per day, multiply by 60 (seconds to minutes), then by 60 again (minutes to hours), and
then by 24 (hours per day). In this example, 192 KB/s x 60 x 60 x 24 becomes 16.6 GB per day.
Using the above metrics (75 percent utilized for five days per week for 50 percent of each day),
you can calculate the disk capacity requirements to support a week assuming all data is net
new. 16.6 GB/day x 5 days per week equates to 83 GB for the five-day week. 83 GB x 75
percent utilization per day x 50 percent of each day equates to 31.3 GB of disk capacity
required to support a given week of data.
The last piece of this equation is making assumptions on the amount of redundancy found in
that data, because the DRE feature of Cisco WAAS removes redundant pieces of data from the
network. Making the assumption that the data is 4:1 compressible (75 percent redundant, which
in many cases is very conservative) yields 31.3 GB x (1 - .75) yields approximately 7 GB of
disk capacity for compression history (dynamic capacity) to support that location.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-83

Storage Capacity Requirements: Edge (Cont.)


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers
High Availability

Utility capacity is defined as the amount of


disk capacity necessary to support IT
infrastructure-related services such as
software distribution, patch management,
antivirus updates, and desktop
management:
Examine the amount of capacity necessary to
support a minimum of one month of previous
software patches, service packs, antivirus
updates, and other infrastructure related
services.
This content is generally not dynamic, remains
unchanged, not interactively accessed or
otherwise changed or updated by users at the
edge, and commonly consumes capacity in the
CIFS acceleration cache at the edge of the
network.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-12

Utility capacity, which is the amount of disk capacity needed to support IT infrastructure
services, such as software distribution, should also be examined. It is best to have the last one
month of previous software patches, services packs, applications, and other infrastructure files
readily available in the edge device. This kind of data is generally not changed by the remote
user; rather it is read and installed on the user workstation.

2-84

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Storage Capacity Requirements: Edge (Cont.)


Optimized TCP
Connections
WAN Bandwidth
Capacity

Static capacity is defined as the amount of disk


capacity necessary to support interactive user
access to historical data, such as home
directories, generally two weeks of file history:
Each user likely has capacity on a file server with
some form of quotas being enforced.

Disk Capacity

Typically files that are stored are significantly less likely


to be accessed after they have been idle for over one
week.

License

This content is generally dynamic at the beginning of


its life and is interactively accessed, or otherwise
changed or updated by a user at the edge, and
commonly consumes capacity in the CIFS acceleration
cache at the edge of the network

Number of Peers
High Availability
2007 Cisco Systems, Inc. All rights reserved.

After a week of being idle, the content is unlikely to be


accessed in a read/write fashion again.

WAAS v4.0.72-13

The static capacity required at the edge of the network is defined as the amount of disk storage
necessary to support interactive user access to historical files. It is best to provide capacity to
support the last two weeks worth of files accessed from the user home directory. These files,
unlike the utility capacity, are likely accessed in a read/write fashion.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-85

Storage Capacity Requirements: Core


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers
High Availability

Core WAE devices must be designed with enough


capacity to provide adequate compression history
for each of the peers while considering fan-out
requirements per platform.
DRE context capacity is dynamic and therefore
shifts to accommodate the most active connected
peers. Therefore, it is generally safe to assume that
a portion of the peers are not talking at a given time
(assume 2:1 disk oversubscription):
For instance, assume a WAE-612 with 300GB of disk
capacity (~140GB assigned to DRE). With 2:1 disk
oversubscription in the core, this yields effectively
~280GB of disk sizing capacity for DRE.
Per the edge defined earlier (7GB compression history
per week), a WAE-612 with 280GB of disk capacity
can support up to 40 connected edges of this type and
provide adequate compression history.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-14

Cisco WAAS core devices must also be sized appropriately. Unlike edge devices, where disk
sizing takes into account factors such as utility capacity and static capacity, the core Cisco
WAAS devices need to be sized only according to dynamic capacity. Cisco WAAS core
devices do not cache files; they cache only segments of TCP data identified by DRE.
It should also be assumed that every peer is not active at the exact same time. In this way, some
level of oversubscription can be designed in the core device storage capacity sizing. For
instance, if each of the branch office locations requires 7GB per week, and the assumption is
made that only one in two peer (edge) devices are active at a given moment, then 3.5GB of
capacity per edge device is required per peer on the core device.

2-86

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

License Requirements
Optimized TCP
Connections
WAN Bandwidth
Capacity

The appropriate licenses should be purchased per


Cisco WAE in the WAAS topology:
Transport: A WAE configured with transport is only providing
WAN optimization capabilities (DRE, TFO, LZ) and no
application-specific acceleration (that is, CIFS/print).

Disk Capacity

Enterprise: A WAE configured with enterprise is providing


WAN optimization capabilities (DRE, TFO, LZ) along with
application-specific acceleration (that is, CIFS/print).

License

Central Manager: A WAE configured with Central Manager is


providing centralized management services for WAEs within
the topology.

Number of Peers

The appropriate memory configuration should also


be purchased based on license and services:
Enterprise with WAFS edge: minimum 1GB memory
Enterprise with WAFS core: minimum 2GB memory

High Availability

Enterprise with both services: minimum 2GB memory

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-15

Three licenses are available for Cisco WAAS:

Transport: This license should be used for WAEs where only WAN optimization features
are used (Data Redundancy Elimination, Transport Flow Optimization, and persistent LZ
compression). Application acceleration capabilities (such as CIFS protocol acceleration,
CIFS server disconnected mode, preposition, and print services) are not available if this
license is used.

Enterprise: This license should be used for WAEs where WAN optimization and
application acceleration features are required.

Central Manager: This license should be used for up to two WAEs within the Cisco
WAAS topology to allow them to act as Central Manager devices. These WAEs provide
configuration synchronization, alerts, and reporting against all of the WAEs in the topology
that are registered with the Central Manager.

Each of the services has a memory requirement, so the appropriate WAE configuration should
be chosen with enough memory to support the configured services:

Transport: The transport license does not have a memory requirement. Any supported
WAE platform can be configured with the transport license.

WAFS Edge service: Allows the remote WAE to perform CIFS protocol acceleration,
CIFS server disconnected mode, preposition, and print services. This service requires that
the WAE be configured with a minimum of 1 GB of memory. The WAFS Edge service
requires the Enterprise license.

WAFS Core service: Allows the core WAE to perform aggregation for WAEs configured
with WAFS Edge for CIFS protocol acceleration capabilities. This service requires that the
WAE be configured with a minimum of 2 GB of memory. The WAFS Core service
requires the Enterprise license.

Concurrent WAFS Edge and WAFS Core services: Allows the WAE to provide
protocol acceleration for local users accessing remote file servers, as well as aggregation

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-87

for remote WAEs acting as WAFS Edge devices. This configuration requires 2 GB of
memory and the Enterprise license.

2-88

Central Manager: Any WAE appliance can be configured as a Central Manager. The
network module enhanced WAE (NME-WAE) can not be configured as a Central
Manager. The scalability of the Central Manager (number of managed nodes) is directly
determined by the WAE model selected and memory configuration.

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Fan-out and Number of Peers


Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers

Each WAE device, when acting as a core device, has


a set of fan-out limits that must be considered:
Devices acting as an edge device only do not need to be
examined; arbitrary cross-site access does not impact fan-out
scalability metrics.
Devices acting as both an edge and a core need to be examined
as a core device.

Each WAE platform, when deployed as a Central


Manager, has a limit on the number of child WAEs it
can manage within the topology:
It is recommended that Central Manager be clustered (active or
standby) in environments with more than 25 managed WAEs.
Central Manager clustering does not change the number of
WAEs that can be managed: It provides only failover capabilities
and not management load-balancing.

High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-16

Each WAE device that is acting as an aggregation point for remote WAEs has a limit in terms
of the number of peers it can manage. This is called the fan-out, that is, the number of edge
devices that a core device can support. When examining fan-out, arbitrary cross-site interactive
access does not need to be considered, unless an edge device is also serving as a dedicated core
for a large number of remote users that also is sitting behind a WAE in their respective
locations.
The Central Manager also has a fan-out limit based on the hardware platform and memory
configuration of the WAE that is acting as a Central Manager. It is best to dedicate (but not
required) two WAEs to be configured as Central Manager WAEs. It is recommended that
redundant Central Manager WAEs be deployed in any medium-to-large scale deployment (25
locations or more) or where high availability is a requirement.
Using two Central Manager WAEs does not increase the number of WAEs that can be
managed, because the Central Manager is clustered in an active or standby configuration.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-89

High Availability
Optimized TCP
Connections
WAN Bandwidth
Capacity
Disk Capacity
License
Number of Peers

Sites that require high availability should be


designed with N+1 WAE devices where:
N is the number of devices necessary to support the
workload requirements of the given location.
+1 is an additional WAE to provide headroom in case a
failure of any WAE within that location is encountered.
For load-sharing interception mechanisms (WCCPv2,
ACE, inline) load is shared among all WAEs.

Central Manager should be deployed with high


availability in environments where:
High availability of management and reporting is of
importance.
25 or more managed WAEs are present.

High Availability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-17

Any locations that require high availability (generally large branch offices, regional offices, and
data center locations) should be configured with the appropriate number of WAEs and one
extra WAE. This configuration is called N+1, where N is the number of WAEs necessary to
support the workload requirements of the location and +1 is an extra WAE device that provides
enough extra headroom to continue providing adequate levels of service should a single device
fail. When integrated with load-sharing interception mechanisms, such as WCCPv2, ACE, or
inline, the load is shared among all WAEs in some capacity. When integrated with non-loadsharing interception mechanisms, such as PBR, only the first WAE is used until it fails.

2-90

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Cisco WAE Device Positioning


This topic describes the Cisco WAE platform metrics that must be examined when designing a
Cisco WAAS solution and where each Cisco WAE platform is intended to be deployed.

Cisco WAE Family Positioning


Enterprise
Data Center

Performance

ACE

WAE-7326
Regional
Office or Small
Data Center

WAE-612

Branch or Remote
Office
WAE-512

NME-WAE-502
NME-WAE-302

Scalability
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-19

The Cisco WAE family of devices is comprised of devices that range from the small branch
office all the way up to the enterprise data center:

NME-WAE-302: A router-integrated network module for Cisco Integrated Services


Router (ISR) models 2811, 2821, 2851, 3825, and 3845. This module is targeted to the
small branch office where only WAN optimization capabilities are required.

NME-WAE-502: A router-integrated network module for the Cisco ISR models 2811,
2821, 2851, 3825, and 3845. This module is targeted at the small and medium branch
offices where WAN optimization and application acceleration capabilities are required.

WAE-512: 1RU appliance for small or medium branch office locations, regional office
locations, or small data center deployments.

WAE-612: 1RU appliance for medium or large branch office locations, regional office
locations, or data center deployments.

WAE-7326: 2RU appliance for the very large branch, regional office, or enterprise data
center.

The Cisco ACE module provides exceptional scalability in the data center for Cisco WAAS.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-91

Cisco WAE Family Performance and


Scalability
Max
Optimized
Throughp
ut
(Mbps)

CM
Scalability
(Devices
Managed)

Core
Fan-Out
(Number
of Peers)

90

n/a

150

n/a

750

100

500

1500

1500

20

150

1000

10

300/300

2000

2000

45

250

2000

30

300/300

6000

2500

155

350

2500

50

300/900

7500

2500

310

450

n/a

90

100

100

Platform

Mem
(GB)

Max
Drives

Drive
Capacity/
Max
Capacity

Max
Optimized
TCP
Conns

NME-WAE-302

.5

80/80

250

n/a

NME-WAE-502

120/120

500

500

WAE-512-1GB

250/250

750

WAE-512-2GB

250/250

WAE-612-2GB

WAE-612-4GB

WAE-7326

Max Edge
CIFS
Sessions

WAN Link
Capacity
(Mbps)

Current Platforms

Legacy Models
WAE-511-512MB

.5

250/250

250

250

WAE-511-1GB

250/250

750

750

100

500

WAE-611

300/300

2000

2000

12

150

1000

10

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-20

This table shows the performance, scalability, and capacity metrics necessary for designing a
WAAS solution. Each WAE platform has a variety of static system limits that should be
considered, including:

2-92

Memory: This figure represents the memory configuration of each of the WAEs. Memory
directly impacts each of the performance and scalability metrics of the platform and also
defines which services can be run on a given platform.

Drive unit capacity, system capacity, and maximum number of drives: This is the size
of the drives that can be installed in the WAE and the maximum usable system capacity
(RAID-1 protected for systems with two or more drives). The amount of disk capacity
determines the size of the compression history and cached data on each platform.

Maximum optimized TCP connections: This is the maximum number of TCP


connections that can be optimized by a given WAE. When this value is exceeded, the WAE
passes through new connections without optimizations until the connection count falls
below 95 percent of maximum system capacity.

Maximum CIFS sessions: This is the maximum number of Common Internet File System
(CIFS) sessions that can be accelerated by a given WAE acting as a Wide Area File
Services (WAFS) Edge. CIFS sessions count against the TCP connection count. For a
WAE acting as a WAFS core, 2GB of memory, or more, must be installed. CIFS sessions
count as optimized TCP connections for the WAFS core and do not need to be counted
separately.

WAN link capacity: This is the recommended amount of WAN link capacity that the
WAE can fully optimize. This is not an enforced number, and in many cases, the WAE can
drive beyond the capacity limits shown in the figure. With compression and DRE disabled
(using just TFO only), the WAE can drive far larger amounts of capacity, up to 450Mbps.

Maximum optimized throughput: This is the maximum amount of optimized throughput


that can be driven through the WAE platform. This figure represents perceived application
performance, not WAN bandwidth. For very high bandwidth deployments, Cisco WAAS

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

can provide reduction of bandwidth utilization but might not show throughput
improvement if the amount of WAN bandwidth is larger than the maximum optimized
throughput of the system. Maximum throughput is probably not reachable through a single
client connection and likely requires multiple concurrent high-performance flows
traversing the WAE.

Central Manager scalability: This is the maximum number of devices a WAE can
manage when acting as a WAAS Central Manager.

Core fan-out: This figure represents the maximum number of peer WAEs that a WAE can
support when acting as an aggregation device. Fan-out numbers below 50:1 are not
enforced limits, but 50:1 is an enforced limit. The numbers here are recommendations to
accommodate a beneficial compression history per connected peer. Fan-out numbers do not
account for arbitrary cross-site access.

The numbers shown in this table are static system limits. When the WAE reaches these static
system limits, the following occurs:

Overload on number of concurrent TCP connections: After the optimized connection


count reaches 100 percent of the maximum system value, the WAE begins passing through
new TCP connections. After the value falls below 95 percent, the WAE again accepts new
connections to optimize. WAEs have been tested with hundreds of thousands of TCP
connections beyond those that are being optimized with little effect on performance.

Overload on number of concurrent CIFS sessions for WAFS Edge WAEs: The WAE
blocks additional CIFS sessions.

Cisco SEs and Cisco partners have access to a WAE sizing tool that can help simplify the
design of a WAAS solution for your network. Please consult with your Cisco SE or partner SE
on the use of this tool.
Note

Some legacy platforms, such as the File Engine (FE) and Content Engine (CE) support
Cisco WAAS. These include the FE/CE-511 (identical to WAE-511), FE/CE-611 and CE-566
(identical to WAE-611), and FE/CE-7326 (identical to WAE-7326).

Note

The NME-WAE-302 is a transport-only platform. It does not support CIFS acceleration.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-93

WCCPv2 Design Considerations


This topic examines how the Web Cache Communication Protocol version 2 (WCCPv2)
configuration can be manipulated to ensure higher levels of performance and optimization.

WCCPv2 Interception Considerations


WAAS uses service groups 61 and 62 for traffic
interception and redirection:
Service group 61: hash bucket assignment based on source IP
address of the packet
Service group 62: hash bucket assignment based on destination
IP address of the packet

One service group needs to be in the path of traffic for


each direction of traffic flow:
Ingress interception (preferred): analyze, intercept, and redirect as
packets enter an interface (less CPU utilization)
Egress interception: analyze, intercept, and redirect as packets
prepare to exit an interface (higher CPU utilization)
Do not overlook placement of the services.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-22

WCCPv2 is a powerful interception protocol that provides users the ability to deploy Cisco
WAAS in an off-path mode while providing fail-through operation, failover, load-balancing,
and warm removal and insertion of new devices. WCCPv2 is also flexible in terms of how it is
configured, providing users with a means of ensuring optimal system operation.
Cisco WAAS uses two WCCP service groups: 61 and 62. The service groups are identical in
that they instruct the router to promiscuously intercept and redirect all TCP traffic to one of the
WCCP child devices (WAEs). Where they differ, however, is what predictor is used to
determine how traffic is load balanced across WAEs within a given location. Service group 61
uses the source IP address as the predictor for load balancing, and service group 62 uses the
destination IP address as the predictor for load balancing.
Given that one service group needs to be in the path of traffic for each direction of traffic flow,
placement of these services determines how traffic is pinned to a WAE. For instance, in a
branch office, if service group 61 is applied on the LAN side of the router for ingress traffic
(traffic going toward the WAN from the branch office), the router load balances the flows
across the child WAE devices based on the source IP address of the flow. With that in mind,
traffic from a particular user is redirected to the same WAE each time that user has traffic
leaving the network. In the reverse direction, when using service group 62 on the WAN side of
the router for ingress traffic (traffic coming into the branch office from the WAN), the router
load balances based on the destination IP address. This means that traffic destined to a
particular user is always redirected to the same WAE.

2-94

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Ingress interception is always preferred over egress interception. Ingress interception on a


network device performs interception and redirection as traffic is entering an interface from the
outside. Using ingress interception equates to lower utilization of network device resources
such as CPUs. Egress interception, on the other hand, performs interception after the traffic has
passed through the route processor and is getting ready to exit an interface. This means that the
traffic has already consumed additional cycles on the device, and in some cases can lead to
higher CPU utilization.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-95

WCCPv2 Configuration: Routers


62/in LAN and 61/in WAN keep flows to a
particular server pinned to the same WAE
in both directions of traffic flow yielding
better likelihood of compression per server.

LAN

61/in
62/in

Load balancing is based on nodes outside


of the location.
61/in LAN and 62/in WAN keep flows from
a particular client pinned to the same WAE
in both directions of traffic flow yielding
better likelihood of compression per
client.
Load balancing is based on nodes
within the location.

WAN

LAN

WAN

62/in
61/in
62/in

Most routers support only GRE-redirect, GRE-return, and hash assignment, which are
default WCCP service configuration parameters.
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-23

This figure shows two examples based on the text in the previous slide about placement of
service groups 61 and 62. In the top example, traffic going to a particular server is pinned to the
same WAE for each direction of traffic flow. This effectively yields higher overall compression
when users access a common server. In the lower example, traffic from a particular client is
pinned to the same WAE for each direction of traffic flow. This effectively yields higher
overall compression for the user if data being accessed across multiple servers has
commonality.
In general, it is recommended that the top example be employed in the data center and the
bottom example be employed in the branch office.

2-96

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

WCCPv2 Configuration: Service Isolation


Branch: 62/in LAN and 61/in WAN keep
flows to a particular server pinned to the
same WAE in both directions of traffic flow
yielding better likelihood of compression
per server.

DC: 62/in WAN1 and 61/out WAN1 keep


flows to a particular server pinned to the
same WAE in both directions of traffic flow
yielding better likelihood of compression
per server.

Load balancing is based on nodes


outside of the location.

No ACLs are required to not redirect flows


to and from the unoptimized branch.

LAN

Load balancing is based on nodes outside


of the location.

WAN

WAN1

LAN

IP
Network
WAN2

61/in
62/in
2007 Cisco Systems, Inc. All rights reserved.

61/out

WAAS v4.0.72-24

It is generally recommended that only ingress interception and redirection be employed when
using WCCPv2; however, there are times when it is desirable to use egress interception and
redirection, as well. The figure shows a configuration in the data center where a combination of
ingress and egress interception and redirection is used. In this case, the primary reason for
configuring WCCPv2 this way is to prevent traffic from the second branch office (bottom left)
from having its traffic intercepted and redirected to a Cisco WAE device. This type of
configuration is applicable in environments where the routers have multiple WAN interfaces
and is commonly referred to as service isolation mode. With service isolation mode, all
interception and redirection is confined to a single interface, and only traffic traversing that
interface is intercepted and redirected to Cisco WAAS.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-97

WCCPv2 Configuration: Switches


Recommended
61/in LAN and 62/in WAN keep
flows from a particular server
pinned to the same WAE in both
directions of traffic flow yielding
better likelihood of compression.
Load balancing is based on nodes
within the location.
62/in LAN and 61/in WAN keep flows
to a particular client pinned to the
same WAE in both directions of
traffic flow.

IP
Network

62/in
61/in

Load balancing is based on nodes


outside of the location.

Configuration on switches is configured on Layer 3 interfaces or SVIs only.


Configure with appropriate parameters (Layer 2-redirect, Layer 2-return,
mask assignment).
2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-25

When configuring WCCPv2 in the data center on the data center switches, it is important to
remember that routed interfaces (that is, Layer 3 interfaces such as Virtual Switch Interfaces, or
SVIs, otherwise known as VLAN interfaces) must be the location where WCCPv2 interception
and redirection is configured. WCCPv2 can not be configured on a layer 2 interface. Much like
with a router configuration, WCCPv2 must be configured such that traffic going out toward the
WAN and traffic coming in from the WAN is intercepted and redirected to the local WAE
devices. If multiple Layer 3 paths exist (for instance, if the connections between the switches
are Layer 3) then each path should be evaluated to ensure that flows do not bypass redirection.
An alternative to using WCCPv2 configuration on the switches is to employ the Cisco Content
Services Module (CSM) or the Application Control Engine (ACE) module for the Catalyst
6500 series. These modules allow for data center scalability and integration without requiring
Layer 3 interfaces.

2-98

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Summary
This topic summarizes the key points that were discussed in this lesson.

Summary
Cisco WAAS solution design requires an understanding of the
holistic IT environment, including the network, nodes,
applications, and utilization characteristics.
The Cisco WAE family of appliances and ISR network modules
support environments of all sizes ranging from small branch
offices up to enterprise data centers. Each Cisco WAE model has
a set of characteristics that should be considered during design,
including number of optimized TCP connections, bandwidth, disk
capacity, memory, and fan-out.
WCCPv2 provides flexible configuration capabilities to ensure
higher levels of optimization through service group placement.

2007 Cisco Systems, Inc. All rights reserved.

2007 Cisco Systems, Inc.

WAAS v4.0.72-26

Designing Cisco WAAS Solutions

2-99

2-100

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Summary
This topic summarizes the key points that were discussed in this module.

Module Summary
Cisco WAE appliances and network modules integrate into the network
using in-path interception or off-path interception techniques such as
WCCPv2, PBR, or ACE.
Cisco WAE devices automatically discover one another and negotiate
optimization policy.
Cisco WAAS provides service transparency, which supports any network
feature that requires visibility to packet header information, and also
supports environments with asymmetric routing.
Designing a Cisco WAAS solution involves understanding of the business
challenges, network environment and configuration, and applications and
utilization characteristics.
WAN bandwidth capacity, disk capacity, and fan-out ratios that should be
considered when designing a Cisco WAAS solution.
Cisco WAEs have static system characteristics based on the supported
number of optimized TCP connections.

2007 Cisco Systems, Inc. All rights reserved.

WAAS v4.0.72-2

This module described how to design Cisco WAAS solutions.

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-101

2-102

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Self-Check
Use the questions here to review what you learned in this module. The correct answers and
solutions are found in the Module Self-Check Answer Key.
Q1)

In a data center deployment, where is the best place in the network to deploy WAE
devices? (Source: Deploying the WAEs in the Network)
A)
B)
C)
D)

Q2)

Which two service groups are used by Cisco WAAS and WCCPv2? (Choose 2.)
(Source: Web Cache Communication Protocol v2)
A)
B)
C)
D)
E)

Q3)

60
61
62
89
90

Which TCP connection establishment message is used to identify the desire to apply
optimization? (Source: Auto-Discovery)
A)
B)
C)
D)

Q4)

Access
Distribution
Core
WAN

TCP SYN
TCP SYN ACK
TCP ACK
TCP FIN

Which network features are compatible with Cisco WAAS? (Source: Network
Transparency)
A)
B)
C)
D)
E)

QoS
Policing and rate-limiting
NetFlow
Access lists
All of the above

Q5)

What form of load-balancing is provided when serially clustering WAEs using inline?
Spillover
Round-robin
Least recently used
None

Q6)

How many WAN routers can an inline WAE be directly connected to?
A)
B)
C)
D)

2007 Cisco Systems, Inc.

1
2
3
4

Designing Cisco WAAS Solutions

2-103

Q7)

What interception mechanism is recommended for the enterprise data center?


A)
B)
C)
D)

Q8)

Within a Cisco WAAS solution, which four factors should be considered when
designing a solution for a particular location? ?
A)
B)
C)
D)
E)
F)

Q9)

61
62
80
89

Which TCP promiscuous service group provides load-balancing based on destination


IP address?
A)
B)
C)
D)

2-104

N+N
N+1
active/passive failover
N+2

Which TCP promiscuous service group provides load-balancing based on source IP


address?
A)
B)
C)
D)

Q13)

Transport
Enterprise
CIFS Accelerator
Central Manager

Which deployment methodology should be followed to provide a highly availability


configuration?
A)
B)
C)
D)

Q12)

1 day
5 days
1 week
1 month
2 months

Which license is required for a WAE that is providing CIFS acceleration?


A)
B)
C)
D)

Q11)

TCP connection count


Processor speed
Memory
WAN bandwidth
Packet loss
Disk capacity

For a given location, what is the recommended length of compression history that
should be provided?
A)
B)
C)
D)
E)

Q10)

WCCPv2
PBR
Inline
ACE

61
62
80
89

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

Module Self-Check Answer Key


Q1)

Q2)

B,C

Q3)

Q4)

Q5)

Q6)

Q7)

Q8)

ACDF

Q9)

Q10)

Q11)

Q12)

Q13)

2007 Cisco Systems, Inc.

Designing Cisco WAAS Solutions

2-105

2-106

Cisco Wide Area Application Services Technical Training (WAAS) v4.0.7

2007 Cisco Systems, Inc.

You might also like