After reading this chapter, you should be able to perform the following tasks:
Explain the Internet architecture process.
Define a Content Delivery Network (CDN).
Define the components of a CDN.
Explain some of the CDN Architectures that apply CDN components.
As the Internet continues to grow and evolve, more and more data must be delivered from place to place. Many feature-rich applications are also being deployed to enhance productivity and accomplish even more with technology. The Internet Business applications that were discussed earlier, such as e-commerce and e-learning, are some of the applications that companies can now deploy to set new standards for their business operations. The challenge is that as more data (applications and files) is added to the network, the greater the potential for a slower user experience.
Content Delivery Networks (CDNs) enhance the end user's experience by overcoming some inherent flaws in the Internet's anatomy and provide a solution to what was once a major challenge. CDNs are architectures in and of themselves that sit on top of, and complement, a Layer 2/Layer 3 architecture. This chapter covers how CDNs increase overall network performance by solving existing challenges, presents the components of a CDN solution, and provides recipe-based solutions that help in deploying a CDN architecture.
Traditional Web Growth
Internet web sites of the world have evolved over time to meet the needs of the end customer. As traffic grew, web sites also had to grow to serve their customer base. In the beginning, to minimize complexity and cost, most web sites began with a Centralized Architecture with one point of presence. A tremendous growth in Internet traffic followed. As the amount of traffic grew, so did the size of content. To accommodate the growth as the web sites' traffic increased, servers and bandwidth were added. Figure 5-1 reflects the central storage growth of a company's World Wide Web presence.
Figure 5-1 Traditional World Wide Web Growth
The next section addresses the challenges raised by web site growth.
Anatomy of a Network
Figure 5-2 depicts the Internet's architecture from the ISP network to the home or enterprise network.
Figure 5-2 Anatomy of a Network
Some have referred to this architecture as the dumbbell or the Q-Tip because both ends are large while the center is very skinny. This is much like the Internet's architecture. The Internet service provider's (ISP's) network is OC-12/48, Gigabit Ethernet, or 10 Gigabit Ethernet on its backbone while the interconnections with the other ISPs are many times oversubscribed smaller DS-3,OC-3, or OC-12 connections. Because of the oversubscription and smaller bandwidth sizes, these connections, known as peering points, represent the middle skinny areas. As you look to the right in figure 5-2 you can see that the next area of interconnection is between the ISP and the enterprise or home. These connections, known as the last mile connections from the incumbent local exchange carrier (ILEC) or competitive local exchange carrier (CLEC), range from 56-kbps dialup to multiple T1 (1.544-Mbps) speed. After the last mile connection is terminated, there is once again more bandwidth on the Ethernet side of the premises enterprise or home network. These Ethernet speeds range from 10 to 100 to 1000 to soon 10,000-Mbps.
Bottleneck Points
What does the connection speed data mean? It means that low bandwidth, oversubscription, and bandwidth congestion points exist where there are skinny points in the picture, at the peering points and in the last mile. Those are the areas where the congestion on the Internet exists. Most of the time, slower infrastructure is cost prohibitive. So, seeing the world as it is and taking a look at what's in place shows an Internet infrastructure with known areas of congestion and a web site/content store architecture that has ballooned in size. Many times this store is in just one location and many users are trying to access it from slower, last mile connections. What if you could alleviate the known congestion points without spending the resources to upgrade the bandwidth in the last mile or the peering points? What if you could also improve the end users' performance significantly at the same time? You can do both with a CDN.
Cisco Content Delivery Networks are the Solution
The number one barrier for e-business applications is bandwidth bottlenecks. The solution is a CDN because it gets the content to the user faster. A CDN can solve both of the bandwidth congestion problems previously discussed by sitting on top of a Layer 2/Layer 3 infrastructure while locating content near the end user and routing user requests to the best source for content delivery to the end user. Large streaming media files, streaming audio, and images are some of the file types that are being pushed close to the end user so they don't have to go over the slower connections to retrieve the data. Instead, that data can be retrieved from a faster local connection. Figure 5-3 shows what happens to the network bottlenecks when a CDN is put in place. As you can see, those bottlenecks have disappeared so that everyone benefits from newly optimized Internet applications with higher performance at a lower cost.
Figure 5-3 Demonstrating the Pushing of Content to the Edge
Cisco's Content Networking Architecture: Components
Through internal developments and acquisitions of breakthrough content networking technologies, Cisco has built a comprehensive architecture for optimizing Web site performance and content delivery. This architecture comprises the five essential technology building blocks along with two complementary technologies that provide the foundation for all of Cisco's existing and future content networking solutions. Table 5-1 describes each one of the components and complementary technologies that make up CDNs. This table provides an overview of each of the components. A more detailed explanation appears later in the chapter.
Table 5-1 CDN Components
|
CDN Component |
|
Purpose |
|
Building Block |
Content Distribution and Management |
Control the distribution and management of content from the enterprise; the goal is to get the content as close to each user as possible. |
|
|
Content Routing |
Route the users request for content to the closest available content store or content engine. |
|
|
Content Edge-Delivery |
Deliver the content quickly from the content engine. |
|
|
Content Switching |
Receive the user's request and make an intelligent decision as to how to distribute the request across one or multiple servers. |
|
|
Intelligent Network Services |
Provide more intelligence to the inherent infrastructure components to add more overall value to the infrastructure. |
|
CDN Component |
|
Purpose |
|
Complementary Technology |
Origin Web Servers |
Provide the content that the user is requesting. |
|
|
Core Networking |
Provide the foundation for all CDNsthe reliable, fast, scalable and manageable Layer 2/Layer 3 infrastructure that CDNs utilize. |
Cisco is the only vendor that offers all five CDN solution components. By adding these components to their existing IP networks, organizations can quickly benefit from a variety of new, high-value e-services.
The foundation of a CDN, as listed in Table 5-1, is a highly available Layer 2 and 3 infrastructure. This infrastructure is made reliable by many of the protocols that organizations already run such as HSRP, Spanning Tree, OSPF, EIGRP, and BGP.
After this infrastructure is in place, a content delivery network can be overlaid onto the foundation to begin to allow intelligent networking to take place at the upper layers of the OSI reference model.
Intelligent Network Services
The first component, Intelligent Network Services, is the most readily available and, in many cases, is the enabler for the other components. Intelligent Network Services are tightly integrated with the Layer 2 and Layer 3 infrastructure because the software features for Intelligent Network Services such as QoS, multicast, VPN, and security are available on Cisco routers and switches. For added functionality and performance, external appliances can be purchased to give a company even more intelligence and increased performance.
For example, multicast is a tremendous enabler for an enterprise network because it allows the distribution of one large video file to be made with one stream rather than multiple unicast streams. Companies can easily deploy an e-learning or corporate communications solution without requiring an upgrade of all connections to support the multiple streams because a popular e-learning presentation can be sent once (multicasted) to many people simultaneously.
Content Switching
Content Switching is an essential category to many CDN architectures because of its capability to increase the entire architecture's performance, scalability, and redundancy. You'll hear the term "load balancing" used almost as a synonym with content switching. There is only a small technical difference between load balancing and content switching. However, both accomplish the same outcomenamely distributing requests across multiple servers.
The distinction to be made for load balancing is that the distribution of requests is done with less intelligence than content switching. This intelligence is based on the 7-layer OSI model. Load balancing is handled at Layer 3 and Layer 4, which means that the forwarding decisions are made on the IP address/port number relationship. In most cases, for web switches load balancing occurs on HTTP TCP port 80 and SSL TCP port 443. Figure 5-4 depicts how the load balancer distributes requests across multiple servers, creating more scalability by reducing the load on one server.
Figure 5-4 Load Balancer Reduces Load Through More Even Request
Distribution
The growing volume and sophistication of e-business-related traffic required a solution that provided increased performance and features. Content switching answered the challenge because of its capability to gather knowledge of the user, device, network, and content. While doing all of this, it is still able to provide deep packet inspection and parsing by using an appliance- or module-based architecture with the increased hardware performance coming from application-specific integrated circuit (ASIC) or network processor-based packet handling.
Content switches are capable of answering these questions:
Who is requesting the content?
What content the user is requesting?
Which server has the content?
Which server that has the content is best available to serve the content back to the user?
With policies that are implemented at Layer 5 through 7 of the OSI model, the content switch is capable of specifying the policy-based web traffic direction based on full visibility of URLs, host tags, and cookies. This gives a company the flexibility to make forwarding decisions with even more intelligence. Figure 5-5 shows how the Layer 7 content switch is able to distribute the request based on more information, adding more intelligence into the network.
Figure 5-5 Content Switching Basics
Cisco Content Services Switches
The Cisco CSS 11000 series content services switches are Layer 5/7 aware and provide a robust front-end for web server farms and cache clusters. With unique features for e-commerce, web hosting, and content delivery, the Cisco CSS 11000 series is an important piece in Cisco's end-to-end content networking solution. Cisco's CSS 11000 series switches, powered by Cisco Web Network Services (Web NS) software, offer unique services, including the following:
Directing web traffic based on full visibility of URLs, host tags, and cookies
Enabling premium services for e-commerce and web hosting providers
Strengthening DoS protection, cache/firewall load-balancing, and "flash-crowd" management
Cisco Content Switch Module
The Cisco Content Switching Module (CSM) is a Catalyst 6500 line card that balances client traffic to farms of servers, firewalls, SSL devices, or VPN termination devices. The CSM provides a high-performance, cost-effective load-balancing solution for enterprise and ISP networks. The CSM meets the demands of high-speed Content Delivery Networks, tracking network sessions and server load conditions in real time and directing each session to the most appropriate server. Fault-tolerant CSM configurations maintain full state information and provide true hitless failover required for mission-critical functions.
The CSM provides the following key benefits:
Market-leading performanceEstablishes up to 200,000 Layer 4 connections per second (performance might vary based on software versions being run), and provides high-speed content switching while maintaining one million concurrent connections.
Outstanding price/performance value for enterprises and ISPsFeatures a low connection cost and occupies a small footprint. The CSM slides into a slot in a new or existing Catalyst 6500 and enables all ports in the Catalyst 6500 for Layer 4 through Layer 7 content switching. Multiple CSMs can be installed in the same Catalyst 6500.
Ease of configurationUses the same Cisco IOS Software command-line interface (CLI) that is used to configure the Catalyst 6500 Switch.
Features and Benefits of the CSM
Some of the features and benefits of the Cisco Content Switching Module are as follows:
Firewall Load BalancingThe CSM allows you to scale firewall protection by distributing traffic across multiple firewalls on a per-connection basis while ensuring that all packets belonging to a particular connection go through the same firewall. Both stealth and regular firewalls are supported.
URL and Cookie-based Load BalancingThe CSM allows full regular expression pattern matching for policies based on URLs, cookies, and HTTP header fields. The CSM supports any URL or cookie format allowing it to load balance existing World Wide Web content without requiring URL/cookie format changes.
High PerformanceThe CSM performs up to 200,000 new Layer 4 TCP connection setups per second (Performance might vary based on software versions being run.). These connections can be spread across 4096 virtual services (16,384 real servers) and all of the ports in a Catalyst 6500 or they can be focused on a single port. This provides a benefit over competitors who use distributed architectures that require all of the ports to be used in order to gain maximum performance.
Network configurationsThe CSM supports many different network topologies. A CSM can operate in a mixed bridged and routed configuration allowing traffic to flow from the client side to the server side on the same or on different IP subnets.
IP Protocol SupportThe CSM accommodates a wide range of common IP protocols including TCP and User Datagram Protocol (UDP). Additionally, the CSM supports higher-level protocols including Hypertext Transfer Protocol (HTTP), File Transfer Protocol (FTP), Telnet, Dynamic Name Server (DNS), and Simple Mail Transfer Protocol (SMTP).
User Session StickinessWhenever encryption or e-commerce is involved, the end user must be consistently directed to the same server (that is, the server where the user's shopping cart is located or the encryption tunnel terminates). The CSM's user session stickiness provides the ability to consistently bring users back to the same server based on Secure Socket Layer (SSL) session ID, IP address, cookie, or HTTP redirection.
Load-Balancing AlgorithmsThe CSM supports the following load-balancing algorithms:
Round robin
Weighted round robin
Least connections
Weighted least connections
Source and/or destination IP hash (subnet mask also configurable)
URL hashing
Quality of Service (QoS)Providing differentiated levels of service to end users is important when generating revenue from content. The CSM leverages the Catalyst 6500's robust QoS, enabling traffic differentiation as follows:
Correctly prioritizes packets based on Layer 7 rules
Directs users who are paying more for services to faster or less loaded servers
High availabilityThe CSM continually monitors server and application availability using health monitoring probes, inband health monitoring, return code checking, and Dynamic Feedback Protocol (DFP). SNMP server health traps can also be configured. When a real server or gateway failure occurs, the CSM redirects traffic to a different location. Servers can be added and removed without disrupting service and systems can easily be scaled up or down.
Connection RedundancyOptionally, two CSMs can be configured in a fault-tolerant configuration to share state information about user sessions and provide connection redundancy. In the event the active CSM fails, open connections are handled by the standby CSM without interruption. Users will experience hitless failover which is an important requirement for e-commerce sites and sites where encryption is used.
Content Edge Delivery
Content Edge Delivery consists of edge content engines that seamlessly deliver web, streaming media, or static content faster than traditional servers to clients. This cached content is delivered by Cisco Content Engines with specialized software optimized for content delivery. Content Edge Delivery accomplished from the edge localizes traffic patterns. Content requests are then filled in an optimal manner, improving the end user experience and allowing rich media content over the Internet or corporate network. This is done by having that traffic retrieved from the local corporate network and not over the corporate WAN. This is known as transparent caching. Delivering a client to the most optimal content engine requires the CDN's intelligence to optimize traffic flows. Content Edge Delivery is tightly integrated with the other CDN traffic components discussed throughout this chapter. Today, Cisco Content Engines handle static and streaming media content. In the future, dynamic database content and applications will also be cached.
When a content engine receives a request, it answers the following questions before delivering the content to the end user:
What's the best resource for this request?
Is my current copy of this content fresh?
If not, go to the origin server and refresh it.
Is there QoS information that I should use or alter for these packets?
Do I have sufficient processing cycles to handle more requests? If not, tell everyone.
Cisco's Content Engine products include the following:
- 507
- 507-AV
- 560
- 560-AV
- 590
- 7320
NOTE
The main difference between each of these Content Engines are disk storage space, throughput, and interface support.
Content Routing or Global Server Load Balancing
Remember Figure 5-1 earlier in the chapter that shows how the size of the server store has grown? The solution, Content Delivery Networking, allows that content to be distributed close to the user so his experience of the content is outstanding rather than good. The category of Content Routing or Global Server Load Balancing, as it's also known, takes care of the intelligent routing of a user request to the best available content engine (this device stores the content). The request is taken in by the content router and the end user is redirected to the best site based on a set of metrics such as delay, topology, server load, and a set of policies such as location of content, enabling the accelerated delivery of web content and streaming media.
As the number of users accessing content on the Internet grows, providing a high level of availability and rapid response from a single location becomes increasingly difficult. The solution to this problem is content routing between multiple data centers or Points of Presence (POPs). Content routing ensures the fastest delivery of content regardless of location, providing the highest availability and site response because the content router is able to sense which content engine, content switch, or point of presence (POP) is the best available to deliver the user's request. This is where a great part of Cisco's CDN intelligence comes from. When a content router takes a look at a request from a user it is essentially asking and answering the following questions:
What content is requested?
Where is the user?
What's the best site now based on site and network health-checks?
Are there other rules to factor into this decision?
Cisco's content routing products include the following:
Content Services Switch 11000
Content Router 4450
Content Router 4400
Cisco 7200 Distributed Director
Content Distribution and Management
How does the content get to all the content engines or content POPs? The answer to that question is through the Content Distribution and Management category. The Content Distribution Manager (CDM) proactively distributes cacheable content from origin servers to content servers at the edges of the network, and keeps the content fresh. This device is a company's central store of content. The CDM enables you to automatically import, maintain copies, and configure content at the edge of the network. From the CDM, an enterprise can globally and centrally provision content to remote locations by using either unicast or multicast, the Intelligent Network Service previously discussed.
Some critical functions include the following:
Implementing content-specific policies across a global network infrastructure
Creating virtual CDNs to support multiple discrete clients on a shared services infrastructure
Transaction logging for automated billing based on actual content service usage
The CDM is continually able to answer the following questions:
What content can be pre-populated at network edge?
How should it be distributed (Self-Organizing Distributed Architecture [SODA] or multicast)?
Which locations should be pre-populated with which content?
When should content be refreshed?
How should the service be provisioned and billed?
Cisco's CDM products include the following:
- CDM 4670
- CDM 4650
- CDM 4630
NOTE
The main difference between these models is content storage space. The content must first be loaded into the CDM and then distributed to the content engines, so a CDM with sufficient data storage space is critical.
Figure 5-6 illustrates what happens in Content Delivery architecture from Step 1 to Step 7, and which CDN categories (listed in Table 5-1) take care of each function.
Figure 5-6 Cisco Content Networking in Action
The following detailed list is a description of how the main components can be used.
The origin content is pulled to the CDM.
The CDM pushes the content to content engines at the edge of the network.
The end user makes a request for Content.
The request is intercepted by the Content Router, which finds the best site to service that content.
The content engine retrieves the content for the end user, if necessary.
The Content Switch receives the request and finds the best server to serve that content to the user.
The end user receives the content.
Applications of Content Delivery Networking
Transparent Caching
To fully realize the value of caches, enterprises and service providers can implement transparent caching. In transparent caching, the cache intercepts the connection to an origin World Wide Web server, handling all HTTP traffic without requiring explicit browser configuration, easing IT pressures at enterprises and service providers by eliminating the need to preconfigure user browsers. It also removes the problem of users covertly changing their browser settings to bypass the cache because traffic is sent through the cache transparently.
Cisco improved transparent caching by developing Web Cach Communication Protocol (WCCP) and high-performance caches to take full advantage of this Cisco IOS Software-based protocol. WCCP is the open, defacto standard protocol from Cisco that enables a router to redirect content requests to a cache, fully leveraging cache performance and performing basic load balancing across multiple caches in a cluster. Cisco Content Engines are network devices optimized to accelerate content delivery by locally storing frequently accessed content (such as GIFs, JPGs, and MPEGs) and quickly responding to user requests instead of going to the origin server. The combination of Cisco WCCP-Enabled Routers and Content Engines provides a powerful Network Caching Services Solution for accelerating content delivery and reducing WAN bandwidth costs for enterprises and service providers. Figure 5-7 and the associated list that follows illustrate the solution.
Figure 5-7 Mechanics of Transparent Caching
A user requests a World Wide Web page from a browser.
The Cisco WCCP-Enabled Router analyzes the request and, based on TCP port number, determines if it should transparently redirect the request to a Cisco Content Engine. Access lists can be applied to control which requests are redirected.
If a Cisco Content Engine does not have the requested content, it does one of the following things:
It sets up a separate TCP connection to the end server to retrieve the content.
The content returns to, and is stored on, the Cisco Content Engine.
The Cisco Content Engine sends the content to the end user. From its local storage, the Cisco Content Engine transparently fulfills subsequent requests for the same content.
E-CDN
Cisco System's Enterprise Content Delivery Network (E-CDN) solution enables high-quality, high-bandwidth streaming video, rich audio, large graphics, presentations, and documents to be delivered over corporate wide-area and local-area networks in an efficient manner. It utilizes the Self Organizing Distributed Architecture (SODA) protocol, which takes advantage of existing bandwidth to efficiently route high-bandwidth content over enterprise networks or even over the Internet. This process eliminates network bottlenecks that often accompany the distribution and retrieval of large multimedia and streaming files.
Cisco E-CDN components include a Cisco CDM for centralized media distribution, a Content Router for redirecting client requests to the closest media source, and Cisco content engines (CEs) which poll the Cisco CDM for new content and ultimately fulfill the client browser requests.
Implementing rich media distribution and retrieval in an enterprise environment can enable opportunities across the organization. These opportunities include delivery of rich media (audio/video, large files of any type) across last mile constraints for remote offices. Businesses can use this technology to quickly and cost-effectively distribute e-learning programs, corporate and field communications, software downloads, and so on from one centralized location (corporate HQ) to any number of remote offices or storefronts. This all leads to improved productivity, competitiveness, and a decreased time to market. The E-CDN architecture enhances an enterprise network's scalability, reliability, and built-in redundancy by supporting a rich knowledge exchange environment where employees, partners, and customers can contribute and learn from each other.
Content Filtering Solution
Employees using the Internet for non-business reasons results in lost productivity and wasted network bandwidth and can lead to "hostile work environment" lawsuits. Consequently, many enterprises must monitor, manage, and restrict employee access to non-business and objectionable content on the Internet.
The Children's Internet Protection Act (CIPA) requires that any school or library receiving federal E-Rate funds must have a filtering solution in place. By filtering irrelevant and unwanted Internet content, the Cisco Content Engine can be an effective tool for implementing and enforcing Internet usage policies as well as caching web content to improve delivery and WAN utilization. By deploying a Cisco Content Engine to cache content and using the SmartFilter option to filter non-productive URLs, enterprises can realize a quick return on investment as a result of increased productivity while recapturing network bandwidth and reducing legal liability. The Cisco Content Engine can filter URLs either by a limited internal deny and permit list, which the customer builds and configures, or with the SmartFilter subscription-based URL filtering software.
SmartFilter's Control List is currently composed of two million international URLs in 30 categories. Using automated methods to find and classify these URLs, the list is built and maintained by multi-lingual, multi-cultural Control List Technicians (CLTs), who physically review and categorize World Wide Web pages.
As Figure 5-8 demonstrates, an organization can filter at a local branch office or at headquarters.
Figure 5-8 Content Filtering
Summary
This chapter shares that Content Delivery Networks are networks in and of themselves. The ingredients of a CDN (the five components that were listed earlier) can be put together in a multitude of ways to form some great recipes, like the three discussed: Transparent Caching, E-CDN, and Content Filtering. As time goes on, new recipes will be created that use different varieties of these ingredients; there might even be another ingredient that comes about. For the near future though, these are the ingredients used to build great content delivery solutions.
Review Questions
What is the goal of Content Delivery Networking?
Accelerate content delivery
Reduce the last mile bottlenecks
Improve the user's experience of the content he is retrieving
All of the Above
What are the five components of a Content Delivery Network?
What is content switching/load balancing able to provide a company?
How does content routing help a company scale?
