Latency-sensitive hashing for collaborative Web caching Presented by: Xin Qi Yong Yang 09/04/2002.

Slides:



Advertisements
Similar presentations
A Survey of Web Cache Replacement Strategies Stefan Podlipnig, Laszlo Boszormenyl University Klagenfurt ACM Computing Surveys, December 2003 Presenter:
Advertisements

Scheduling in Web Server Clusters CS 260 LECTURE 3 From: IBM Technical Report.
Dynamic Task Assignment Load Index for Geographically Distributed Web Services PhD Research Proposal By: Dhiah Al-Shammary Supervised.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
LASTor: A Low-Latency AS-Aware Tor Client
Cloud Download : Using Cloud Utilities to Achieve High-quality Content Distribution for Unpopular Videos Yan Huang, Tencent Research, Shanghai, China Zhenhua.
The War Between Mice and Elephants LIANG GUO, IBRAHIM MATTA Computer Science Department Boston University ICNP (International Conference on Network Protocols)
1 Routing and Scheduling in Web Server Clusters. 2 Reference The State of the Art in Locally Distributed Web-server Systems Valeria Cardellini, Emiliano.
The War Between Mice and Elephants Presented By Eric Wang Liang Guo and Ibrahim Matta Boston University ICNP
Spring 2003CS 4611 Content Distribution Networks Outline Implementation Techniques Hashing Schemes Redirection Strategies.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Internet Networking Spring 2006 Tutorial 12 Web Caching Protocols ICP, CARP.
EEC-484/584 Computer Networks Discussion Session for HTTP and DNS Wenbing Zhao
1 A Comparison of Load Balancing Techniques for Scalable Web Servers Haakon Bryhni, University of Oslo Espen Klovning and Øivind Kure, Telenor Reserch.
Improving Proxy Cache Performance: Analysis of Three Replacement Policies Dilley, J.; Arlitt, M. A journal paper of IEEE Internet Computing, Volume: 3.
Locality-Aware Request Distribution in Cluster-based Network Servers 1. Introduction and Motivation --- Why have this idea? 2. Strategies --- How to implement?
Analysis of Web Caching Architectures: Hierarchical and Distributed Caching Pablo Rodriguez, Christian Spanner, and Ernst W. Biersack IEEE/ACM TRANSACTIONS.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #13 Web Caching Protocols ICP, CARP.
Submitting: Barak Pinhas Gil Fiss Laurent Levy
A Distributed Proxy Server for Wireless Mobile Web Service Kisup Kim, Hyukjoon Lee, and Kwangsue Chung Information Network 2001, 15 th Conference.
Flash Crowds And Denial of Service Attacks: Characterization and Implications for CDNs and Web Sites Aaron Beach Cs395 network security.
Internet Networking Spring 2002 Tutorial 13 Web Caching Protocols ICP, CARP.
Detecting SYN-Flooding Attacks Aaron Beach CS 395 Network Secu rity Spring 2004.
Squirrel: A decentralized peer- to-peer web cache Paul Burstein 10/27/2003.
Web-Conscious Storage Management for Web Proxies Evangelos P. Markatos, Dionisios N. Pnevmatikatos, Member, IEEE, Michail D. Flouris, and Manolis G. H.
Caching And Prefetching For Web Content Distribution Presented By:- Harpreet Singh Sidong Zeng ECE Fall 2007.
Web Caching Schemes For The Internet – cont. By Jia Wang.
Web Caching and CDNs March 3, Content Distribution Motivation –Network path from server to client is slow/congested –Web server is overloaded Web.
The Medusa Proxy A Tool For Exploring User- Perceived Web Performance Mimika Koletsou and Geoffrey M. Voelker University of California, San Diego Proceeding.
1 ENHANCHING THE WEB’S INFRASTUCTURE: FROM CACHING TO REPLICATION ECE 7995 Presented By: Pooja Swami and Usha Parashetti.
World Wide Web Caching: Trends and Technology Greg Barish and Katia Obraczka USC Information Science Institute IEEE Communications Magazine, May 2000 Presented.
Locality-Aware Request Distribution in Cluster-based Network Servers Presented by: Kevin Boos Authors: Vivek S. Pai, Mohit Aron, et al. Rice University.
Web Caching and Content Delivery. Caching for a Better Web Performance is a major concern in the Web Proxy caching is the most widely used method to improve.
1 Content Distribution Networks. 2 Replication Issues Request distribution: how to transparently distribute requests for content among replication servers.
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
Active Network Applications Tom Anderson University of Washington.
Achieving Load Balance and Effective Caching in Clustered Web Servers Richard B. Bunt Derek L. Eager Gregory M. Oster Carey L. Williamson Department of.
Server Load Balancing. Introduction Why is load balancing of servers needed? If there is only one web server responding to all the incoming HTTP requests.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
On the Scale and Performance of Cooperative Web Proxy Caching University of Washington Alec Wolman, Geoff Voelker, Nitin Sharma, Neal Cardwell, Anna Karlin,
Web Cache Replacement Policies: Properties, Limitations and Implications Fabrício Benevenuto, Fernando Duarte, Virgílio Almeida, Jussara Almeida Computer.
EXPOSE GOOGLE APP ENGINE AS TASKTRACKER NODES AND DATA NODES.
Web Prefetching Between Low-Bandwidth Clients and Proxies : Potential and Performance Li Fan, Pei Cao and Wei Lin Quinn Jacobson (University of Wisconsin-Madsion)
World Wide Web Caching: Trends and Technologys Gerg Barish & Katia Obraczka USC Information Sciences Institute, USA,2000.
Segment-Based Proxy Caching of Multimedia Streams Authors: Kun-Lung Wu, Philip S. Yu, and Joel L. Wolf IBM T.J. Watson Research Center Proceedings of The.
1 On the Placement of Web Server Replicas Lili Qiu, Microsoft Research Venkata N. Padmanabhan, Microsoft Research Geoffrey M. Voelker, UCSD IEEE INFOCOM’2001,
« Performance of Compressed Inverted List Caching in Search Engines » Proceedings of the International World Wide Web Conference Commitee, Beijing 2008)
Scalable Web Server on Heterogeneous Cluster CHEN Ge.
Understanding the Performance of Web Caching System with an Analysis Model and Simulation Xiaosong Hu Nur Zincir-Heywood Sep
1 On the Placement of Web Server Replicas Lili Qiu, Microsoft Research Venkata N. Padmanabhan, Microsoft Research Geoffrey M. Voelker, UCSD IEEE INFOCOM’2001,
Kiew-Hong Chua a.k.a Francis Computer Network Presentation 12/5/00.
Load-Balancing Routing in Multichannel Hybrid Wireless Networks With Single Network Interface So, J.; Vaidya, N. H.; Vehicular Technology, IEEE Transactions.
Efficient P2P Search by Exploiting Localities in Peer Community and Individual Peers A DISC’04 paper Lei Guo 1 Song Jiang 2 Li Xiao 3 and Xiaodong Zhang.
PROP: A Scalable and Reliable P2P Assisted Proxy Streaming System Computer Science Department College of William and Mary Lei Guo, Songqing Chen, and Xiaodong.
Efficient Cache Structures of IP Routers to Provide Policy-Based Services Graduate School of Engineering Osaka City University
ICP and the Squid Web Cache Duane Wessels and K. Claffy 산업공학과 조희권.
 Cachet Technologies 1998 Cachet Technologies Technology Overview February 1998.
Distributed Computing Systems CSCI 4780/6780. Scalability ConceptExample Centralized servicesA single server for all users Centralized dataA single on-line.
Content Delivery Networks: Status and Trends Speaker: Shao-Fen Chou Advisor: Dr. Ho-Ting Wu 5/8/
Hiearchial Caching in Traffic Server. Hiearchial Caching  A set of techniques and mechanisms to increase the size and performance of network caches.
09/13/04 CDA 6506 Network Architecture and Client/Server Computing Peer-to-Peer Computing and Content Distribution Networks by Zornitza Genova Prodanoff.
On the Placement of Web Server Replicas Yu Cai. Paper On the Placement of Web Server Replicas Lili Qiu, Venkata N. Padmanabhan, Geoffrey M. Voelker Infocom.
Overview on Web Caching COSC 513 Class Presentation Instructor: Prof. M. Anvari Student name: Wei Wei ID:
On the scale and performance of cooperative Web proxy caching 2/3/06.
Coral: A Peer-to-peer Content Distribution Network
Advanced Topics in Concurrency and Reactive Programming: Case Study – Google Cluster Majeed Kassis.
An Analysis of Facebook photo Caching
Internet Networking recitation #12
Kalyan Boggavarapu Lehigh University
Presentation transcript:

Latency-sensitive hashing for collaborative Web caching Presented by: Xin Qi Yong Yang 09/04/2002

About Authors: Kun-Lung Wu –Ph.D in University of Illinois at Urbana-Champaign –IBM Thomas J. Watson Research Center and currently manager of the Software Tools and Techniques group. –published more than 50 papers in refereed journals and conferences. Holds or has applied for 15 US patents. Philip S Yu –Ph.D in Stanford University –IBM Thomas J. Watson Research Center and currently manager of the Software Tools and Techniques group. –Published more than 270 papers in refereed journals and conferences. Holds or has applied for 92 US patents.

Content table Introduction New latency-sensitive hashing scheme New approach to evaluating the LSH and simulation model Trace-driven simulations Results analysis Conclusion Questions

Why Caching? Internet grows very quick. The problems are network congestion and server overloading. User response times for accessing the web have become increasingly unsatisfactory. Web caching is needed to reduce network traffic. Three ways to cache: Caching at client, Caching at proxy and Caching at Servers.

Why Proxy? Proxy was firstly used to allow accesses to the internet from users within a firewall. Proxy served a previous request and cached document for next one. Web caching at proxy server can not only save network bandwidth but also lower access latency for the clients.

Collaborative Web Caching A single server is single point of failure A single server cache is always a bottleneck Multiple proxies are used.

A generic WWW caching system

Geographically distributed proxies Response times tend to be negatively impacted for those requests hashed into geographically distant proxies or overloaded proxies. Distant proxies tend to incur longer network latency delays Overloaded proxies can cause significant delays too. Strong need to consider the latency issue in hashing based web caching among geographically distributed proxies.

Geographically distributed proxies Geographically clustered hashing (GCH) Requests are served only by proxies in a geographically close region. Work well if the proxies within a region can adequately service all the requests originated within the same region. However, proxies in one region may be overloaded while those in another region are under loaded Geographically distributed hashing (GDH) Requests are hashed into all cooperating proxy caches regardless of geographical location. Load tends to be more balanced among all the geographically distributed cooperating caches compared with GCH However, GDH did not take into account network latency delays due to geographical distances.

Geographically distributed proxies (cont’d) Latency-sensitive hashing (LSH) –It hashes requests into all proxies –It counts latency delays and potential overloaded proxies. –Firstly, a request is hashed into an anchor hash bucket. Each hash bucket is mapped to one of the geographically distributed proxies. –Secondly, a selection algorithm is used to pick a proxy among a small number of hash buckets adjacent to the anchor hash bucket. –The selection is based on objective to reduce network latency and to avoid creating over-loaded proxies.

An example Three Geographically distributed clusters of proxies

An example of a geographically distributed Hashing

An example of a geographically distributed Hashing(cont’d) This example shows the potential problem of hashing requests into more and more geographically distributed proxies. The network latency can be a problem for those that are hashed into geographically distant proxies.

An example of latency-sensitive hashing

An example of latency-sensitive hashing (cont’d) Compared with GDH, the proxy with lowest latency will be chosen. Mapping of hash buckets to proxies and the selection of window size are important to its performance. It is not obvious to do so if requests are evenly distributed to all proxies when there are different numbers of proxies within a cluster.

Indirect Mapping Scheme Map each hash bucket to an index of a proxy ID array instead of directly mapping each hash bucket into a proxy ID. From this proxy ID array, we then obtain the proxy ID for the hash bucket. Two parts for indirect mapping scheme: 1. Construction of proxy ID array. 2. The assignment of the indices of the proxy ID array to hash buckets.

An example of an indirect mapping scheme for LSH

Indirect Mapping Construction of Proxy ID array Two-level round-robin fashion Size of PA is N*LCMc, N is the number of clusters and LCM c is the l.c.m. of C i. Construction of hash bucket segment LCM p is the l.c.m. of n j The total size of the hash bucket segment is

Load Balance Without considering Load Balance, the LSH degenerates into GCH. If the load of a proxy is too high, this proxy should not be selected. DNS is easy to detect the load condition of all proxies DNS is a better place to implement the LSH.

Performance Evaluation Trace driven simulator that models the three hashing schemes, GCH, GDH, and LSH. Nine proxies organized into three geographical clusters, each cluster has three proxies. Each Proxy has the same amount of computing resources.

Performance Evaluation(cont’d) For each proxy, implemented: –A CPU server FIFO service queue –A cache manager LRU stack Response time for a request whose object can be found locally = L + T http + T cache + T http + L + Q L: latency delay T http : service time for processing an HTTP request or reply T cache: service time for looking up an object from its cache or storing an object into its cache Q: the queue delay the request incurs waiting for the CPU

Performance Evaluation(cont’d) Response time for a request whose object is a cache miss = L + T http + T cache + C miss +T cache +T http + L + Q –C miss : A cache miss delay if the requested object can not found locally. –Assume T cache = 0.5*Thttp Zipf-like distribution –Zipf(x, M) is a parametric distribution where the probability of selecting the ith item is proportional to 1/i (1-x), where x is a parameter and i belongs to {1, …, M}

Distributions of clients around the proximity of each proxy cache

The impact of request origination skew on average response time

The impact of request origination skew on coefficient of variation

The level of load imbalance with no skew More balancedLess balanced

The level of load imbalance with high skew

Simulation Results GCH is very sensitive to skew in request origination –GCH can not effectively utilize proxies in other clusters to help balance the load GDH is immune to the skew in request origination –Hashing is based on URL and thus the load distribution among the proxies remains the same regardless of skew in request origination. LSH can distribute requests among all the proxies, but it is slightly less balanced compared with GDH –In order to lower latency delays, LSH tends to choose a proxy within the same cluster as the browser originating the request.

The impact of hot-spot references on average response time

The level of load imbalance with hot- spot references

Simulation Results GDH can become quite unbalanced in the presence of hot-spot references –Each UTL is hashed into the same proxy cache no matter which browser issues the request. GCH is less susceptible to 9-4 trace hot-spot references, but highly sensitive to 9-6 trace. LSH handles is almost insensitive to hot-spot references. –LSH can select different proxies to offload the hot-spot references originating from different browsers.

The impact of selection window size when the system is lightly loaded and balanced

The impact of selection window size when the system is moderately loaded and unbalanced

The level of load imbalance when the system is lightly loaded and well balanced

The level of load imbalance when the system is moderately loaded and unbalanced

Simulation Results For light load and relatively well balanced system, a larger w enables more requests to be hashed into geographically closer proxies. The average response time is better. For a moderately loaded and unbalanced system, w=3 may cause too many requests to be hashed into geographically closer proxies, resulting in slightly less balanced system compared with w=2. When W=1, system is highly unbalanced.

Conclusion GCH hashes requests originated from one region into proxies within the same region. It’s performance is poor. GDH hashes requests to all proxies regardless of geographical locations. It fails in the presence of hot-spot references. LSH effectively handles both skew in request origination and hot spot references by hashing requests among geographically distributed proxies. Overall system is lightly loaded, LSH effectively reduces latency delays by hashing requests in to geographically closer proxies.

Questions?