Web acceleration: PoP Infrastructures

Slides:



Advertisements
Similar presentations
Scheduling in Web Server Clusters CS 260 LECTURE 3 From: IBM Technical Report.
Advertisements

Multi-Layer Switching Layers 1, 2, and 3. Cisco Hierarchical Model Access Layer –Workgroup –Access layer aggregation and L3/L4 services Distribution Layer.
Working with Proxy Servers and Application-Level Firewalls Chapter 5.
1 Routing and Scheduling in Web Server Clusters. 2 Reference The State of the Art in Locally Distributed Web-server Systems Valeria Cardellini, Emiliano.
1 Internet Networking Spring 2004 Tutorial 13 LSNAT - Load Sharing NAT (RFC 2391)
Internet Networking Spring 2006 Tutorial 12 Web Caching Protocols ICP, CARP.
11/2/2000Weihong Wang/Content Switch Page 1 Content Switch. Introduction of content web switch.. Some content switch products in the market.. Design of.
Content Switch. Introduction of content web switch.. Some content switch products in the market.. Design of a content switch.
Page: 1 Director 1.0 TECHNION Department of Computer Science The Computer Communication Lab (236340) Summer 2002 Submitted by: David Schwartz Idan Zak.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #13 Web Caching Protocols ICP, CARP.
SERVER LOAD BALANCING Presented By : Priya Palanivelu.
1 Web Proxies Dr. Rocky K. C. Chang 6 November 2005.
Design and Implementation of a Server Director Project for the LCCN Lab at the Technion.
Lesson 1: Configuring Network Load Balancing
Application Layer  We will learn about protocols by examining popular application-level protocols  HTTP  FTP  SMTP / POP3 / IMAP  Focus on client-server.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #12 LSNAT - Load Sharing NAT (RFC 2391)
(part 3).  Switches, also known as switching hubs, have become an increasingly important part of our networking today, because when working with hubs,
FIREWALL TECHNOLOGIES Tahani al jehani. Firewall benefits  A firewall functions as a choke point – all traffic in and out must pass through this single.
Network Address Translation (NAT) CS-480b Dick Steflik.
Server Load Balancing. Introduction Why is load balancing of servers needed? If there is only one web server responding to all the incoming HTTP requests.
Redirection and Load Balancing
Chapter 6: Packet Filtering
1 Semester 2 Module 10 Intermediate TCP/IP Yuda college of business James Chen
Implementing ISA Server Publishing. Introduction What Are Web Publishing Rules? ISA Server uses Web publishing rules to make Web sites on protected networks.
Web Cache Redirection using a Layer-4 switch: Architecture, issues, tradeoffs, and trends Shirish Sathaye Vice-President of Engineering.
Network Security. 2 SECURITY REQUIREMENTS Privacy (Confidentiality) Data only be accessible by authorized parties Authenticity A host or service be able.
Proxy Servers.
® IBM Software Group © 2007 IBM Corporation Best Practices for Session Management
TCP/IP (Transmission Control Protocol / Internet Protocol)
Switch Features Most enterprise-capable switches have a number of features that make the switch attractive for large organizations. The following is a.
TCP/IP1 Address Resolution Protocol Internet uses IP address to recognize a computer. But IP address needs to be translated to physical address (NIC).
Polytechnic University Firewall and Trusted Systems Presented by, Lekshmi. V. S cos
1 © 1999, Cisco Systems, Inc. 1293_07F9_c1 LocalDirector Version3.1.
1 Chapter 1 INTRODUCTION TO WEB. 2 Objectives In this chapter, you will: Become familiar with the architecture of the World Wide Web Learn about communication.
Security fundamentals
SDN and Security Security as a service in the cloud
CompTIA Security+ Study Guide (SY0-401)
Layer 3 Redundancy 1. Hot Standby Router Protocol (HSRP)
Lab A: Planning an Installation
Scaling Network Load Balancing Clusters
How HTTP Works Made by Manish Kushwaha.
F5 Internet Quality Control Products and Services
Mobile IP.
WWW and HTTP King Fahd University of Petroleum & Minerals
Content Distribution Networks
F5 BIGIP V 9 Training.
Reddy Mainampati Udit Parikh Alex Kardomateas
Securing the Network Perimeter with ISA 2004
Anonymous Communication
Network Load Balancing
VIRTUAL SERVERS Presented By: Ravi Joshi IV Year (IT)
Web Caching? Web Caching:.
Introduction to Networking
Introduction to Networks
Switching Techniques In large networks there might be multiple paths linking sender and receiver. Information may be switched as it travels through various.
Inside of a computer… What happens when you turn your computer on? What loads? Where are applications stored? How are do they run? In what form is information.
Internet Networking recitation #12
CompTIA Security+ Study Guide (SY0-401)
Distributed Content in the Network: A Backbone View
TCP/IP Networking An Example
Cisco Content Delivery Solutions
Web Privacy Chapter 6 – pp 125 – /12/9 Y K Choi.
Anonymous Communication
70-293: MCSE Guide to Planning a Microsoft Windows Server 2003 Network, Enhanced Chapter 4: Planning and Configuring Routing and Switching.
AWS Cloud Computing Masaki.
Inside of a computer… What happens when you turn your computer on? What loads? Where are applications stored? How are do they run? In what form is information.
دیواره ی آتش.
EE 122: HyperText Transfer Protocol (HTTP)
AbbottLink™ - IP Address Overview
Anonymous Communication
Presentation transcript:

Web acceleration: PoP Infrastructures Erv Johnson Director of Technical Marketing ArrowPoint Communications May 9, 2000 www.arrowpoint.com ejohnson@arrowpoint.com (978) 692-3020

Summary Looking at the big picture Analysis of network topologies and web interactions Potential sources of delay Technologies that can mitigate this delay.

How Delayed Binding Works Step 1: User clicks: www.urlxyz.com/info.asp - Browser talks to DNS for IP Address - Browser sends TCP SYN (connect?) Internet Step 2: Switch Sends TCP SYN ACK to browser Step 3: Browser sends URL: www.urlxyz.com/info.asp Step 4: Switch determines Best Server for the content being requested Step 5: Switch connects to Best Server and “splices” TCP connection Current Layer 4 switches and load balancers, route incoming requests based on the combination of destination IP address, and TCP port. They immediately “bind” a Web browser to a Web server based on the initial TCP SYN request. Therefore the content request is routed before the switch receives the actual HTTP request, making it incapable of optimizing flows based on URL. This can be problematic in a Web environment. To a Layer 4 load balancer, all Web applications appear to be using TCP port 80 (well-known port for HTTP), making them indistinguishable from one another. Therefore, a CGI request looks no different from a Web-enabled SAP application or streaming audio request, even though all of these requests have very different requirements or may be found on different servers. In contrast, Web switches use URLs to route incoming requests to target servers. By looking deep into the HTTP payload, down to the URL and cookie, a Web switch “knows” what content is being requested. This provides unprecedented flexibility in defining policies for security and traffic prioritization – enabling tiered services and ensuring Service Level Agreements are met. Further, the ability to use cookies enables sticky connections – a critical requirement for e-commerce. There are 5 basic steps involved in web switching: 1. User makes a content request by typing a URL into a browser. 2. Web switch with virtual IP of the requested URL intercepts the request. 3a. Web switch spoofs TCP ACK back to client. 3b. The Web switch examines HTTP header and URL and compares to current content rules to select best server or cache to satisfy the request. 4. A flow is created between the switch and the optimal server and “snaps” together with the flow from the client to the switch. 5. All subsequent packets for that Web flow are forwarded without intervention by the switch controllers. Step 6: With HTTP 1.0 there is one HTTP request/response per TCP session With version 1.1, there can be many GETs per TCP session

What is Content Intelligence? Content Routing based on: Host Tag Entire URL Dynamic Cookie location File extension # of rules # of services # of services per content rule Layer 5-7 (content) Session load balancing based on IP address and TCP port Network Address Translation (NAT) Policies based on TCP port Layer 4 (TCP) Switching on MAC address, VLANs IP Routing 802.1 P/Q policy Layer 3 (IP) ArrowPoint Web switches are unique in their ability to perform content discovery, the key to achieving content or name-based switching. Content discovery requires: 1) spoofing the TCP using the entire URL and cookie; 2) providing content based keep-alives to detect changes in content; 3) probing the server automatically to determine content attributes then dynamically selecting the best connection for the keep alive. Conventional load balancers and Layer 4 switches were designed for address-based switching, and differentiate applications based on identity of well-known TCP ports: Port 80 for HTTP and Port 21 for FTP. However, even as more and more information is based on HTTP requests, load balancers cannot differentiate between multiple HTTP requests for different content. ArrowPoint’s Web switches were designed for name-based switching and are the only switches to use the entire URL and cookie to select the best site and server.

Applications Local and Distributed Content Routing Internet Select Best site/server Route HTTP request to best location Deliver content from best location Push/pull content from origin server Customer Web Site Internet Boston PoP London PoP Atlanta PoP San Jose CS-150 CS-50

Local Server Selection Methods Select server group based on URL/cookie, other Select “best” server based on response time Build normalized load factor for every server Send requests to fastest server having the requested content Internet With a combination of high speed HTTP flow setups and URL routing, ArrowPoint goes beyond traditional load balancing. ArrowPoint provides the broadest selection of server selection rules, but only ACA automatically balances the load based on actual server response time at any given point in time, with no user intervention. Because content requests are directed based on the URL, redirection of HTTP requests can be off loaded from the Web server. For example, a CGI request can be sent directly to the CGI server rather than being sent to the main HTTP server and then redirected. In fact, Web servers are further optimized by only handing TCP connections and HTTP requests once the switch determines it to be the best source for the requested content. ACA monitors the flow until it sees a TCP FIN message from the server, at which point the switch creates a tear down report containing statistics for the flow. This allows the switch to aggregate statistics which can be the basis for service level agreements and billing. There is a dedicated slide on SLAs and management statistics later in the presentation. Traditional load balancers and L4 switches use simple static methods for distributing server load, including Round Robin and Weighted Round Robin. They can also use (Least) session connection count to measure server load, but this is limited in several respects. Least connections misses the significant difference in actual server performance between processing 100 HTTP requests for the home page and processing 100 HTTP requests for CGI or ASP pages. Also, in the Internet, session connection count is misleading and inaccurate for TCP traffic where the TCP inactivity timer is very long, leaving most TCP circuits in WAIT-state when the connection is idle. ArrowPoint supports all of these algorithms, and in certain situations they can be useful. For instance, servers processing ASP requests may perform with consistent response time up to a point, but then hit the wall. In this case distributing load using a Least Connections rule would be appropriate with a MAX connections limitation. *.html Other methods Weighted Round Robin Least Connections Max connections limits *.ram *.cgi

Content and Application Verification Web Server (IP Address) IP ICMP Ping/Echo Application (TCP port) TCP connect/acknowledge Scripting: TCP connect - request - reply - verify Content (URL) HTTP request, response verification Verifies back-end resources Secure content HTTPs request, response verification Internet ArrowPoint can test server availability using Layer 3 and Layer 4 techniques, plus ArrowPoint can test HTTP GETs, POSTs, and HEADs, comparing the complete response and detecting the most minute change in content. And in the event of a failed HTTP keep alive, the ArrowPoint switch will direct only requests for that particular content or application to another server, continuing to utilize the server for surviving content, application, or services. For example, if a back end cgi process is not responding there is no need to remove the server from rotation for serving HTTP requests for the Home page.

E-transaction Assurance Using cookies optimizes E-transactions Cookie-based sticky connections for non-encrypted transactions Cookie-based priority for users and transactions SSL session ID can be used for sticky connections Must be able to ensure seamless transition between secure and non-secure Internet In order for an e-commerce transaction to be successful, clients must be bound, or "stuck" to a specific server until the transaction is complete. Maintaining persistent connections to a server, called "sticky" connections, is the key to any money-generating e-commerce Web site. ArrowPoint's competitors provide load-balancing solutions that base server connections on IP address parameters. This approach does not scale. ArrowPoint provides the ability to configure persistent connections based on cookies or SSL session IDs, thereby maintaining sticky connections. Having the ability to maintain persistence on cookies, called "cookie-switching," optimizes non-encrypted e-commerce transactions. ArrowPoint Web switches can access information deep in the TCP and HTTP headers, including "mobile" cookies that change location within the header between requests. The most common method of securing Web-based transactions is the use of the popular SSL protocol. Maintaining persistence based on SSL session ID optimizes the integrity of e-commerce transactions and provides a balance between security and performance. In a secure transaction, the Content Smart™ Web switches maintain state during the transition from the user cookie (shopping) to the SSL session ID (checkout), ensuring a successful and secure user transaction. Server 1 Server 3 Server 2

Transparent Caching with Content-based rules Bypass Internet Origin Server Network Cache Transparent - no browser configuration required Improves performance by bypassing cache for non-cacheable content or cache failures Content policy allows include, exclude (bypass), or block actions based on Access Control Lists Based on IP address, TCP port or URL First, instead of just diverting all HTTP (port 80) to the cache, content aware switches inspect the HTTP header to learn about content, so they can bypass the caches entirely for non-cacheable URLs and transmit them directly to the origin server. This maximizes cache hit rate and request/response latency because the cache is caching only cache-able content, and is bypassed completely for non-cacheable requests, or if all of the caches are overloaded. The ArrowPoint switch can also be configured to bypass the cache for any particular domain name and URL, providing more granular control to exclude all or part of a particular customer's content from being diverted to the cache. Proxy caches can be clustered for scalability and redundancy, but the cache cannot be bypassed, since the browser is explicitly defined to talk to the cache. Only ArrowPoint can also bypass Reverse Proxy can be bypassed. Read on...

Web Caching Performance with Content-based Cache Bypass More than 4 times Faster URL rules Layer 4 rules Connections per second In the real world, a significant amount of HTTP requests are for non-cacheable content (as much as 35-45%), including URLs for objects which are dynamically, generated such as CGI scripts and Active Server Pages (ASP), or URLs carrying cookies. Every time these requests are sent to the cache, the cache introduces significant latency and unnecessary processing in evaluating the request, fetching the content from the origin server, and then sending the response to the client. Thus, the hit rate and performance of the cache is inversely proportional to the amount of non-cacheable content sent to the cache. ArrowPoint's content-aware cache bypass mechanism ensures minimum latency and maximum cache hit rate and efficiency, by bypassing the cache for non-cacheable URLs, and sending them directly to the origin server with the client's source IP address. The cache then can apply all resources to serving requests for cacheable content. This fact is dramatically illustrated by a test conducted by ArrowPoint which compared the performance of a layer 4 switch and a ArrowPoint CS-100 Content Smart Web Switch in a transparent caching configuration. The test simulated 180 clients generating HTTP requests and measured the performance impact on the “virtual POP” as the percentage of cacheable content was decreased. % cacheable content

Digital Product Distribution and Delivery San Jose Web hosting Data Center “mymusic.com” HQ Web site Boston Distribution Center Internet Client Requests MP3 file London Distribution Center Paris Distribution Center HQ switch determines best site “Best” distribution site delivers MP3 file Switch delivers New MP3 inventory