A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Presented by: Raymond Leung Wai Tak Supervisor:

Slides:



Advertisements
Similar presentations
Disk Arrays COEN 180. Large Storage Systems Collection of disks to store large amount of data. Performance advantage: Each drive can satisfy only so many.
Advertisements

Scheduling in Web Server Clusters CS 260 LECTURE 3 From: IBM Technical Report.
CSCE430/830 Computer Architecture
1 NCFS: On the Practicality and Extensibility of a Network-Coding-Based Distributed File System Yuchong Hu 1, Chiu-Man Yu 2, Yan-Kit Li 2 Patrick P. C.
Continuous Media 1 Differs significantly from textual and numeric data because of two fundamental characteristics: –Real-time storage and retrieval –High.
Lava: A Reality Check of Network Coding in Peer-to-Peer Live Streaming Mea Wang, Baochun Li Department of Electrical and Computer Engineering University.
Slice–and–Patch An Algorithm to Support VBR Video Streaming in a Multicast– based Video–on–Demand System.
CSE521: Introduction to Computer Architecture Mazin Yousif I/O Subsystem RAID (Redundant Array of Independent Disks)
CS Spring 2011 CS 414 – Multimedia Systems Design Lecture 27 – Media Server (Part 3) Klara Nahrstedt Spring 2011.
A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Jack Lee Yiu-bun, Raymond Leung Wai Tak Department.
Efficient and Flexible Parallel Retrieval using Priority Encoded Transmission(2004) CMPT 886 Represented By: Lilong Shi.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Peer-to-peer Multimedia Streaming and Caching Service Jie WEI, Zhen MA May. 29.
Network Coding for Large Scale Content Distribution Christos Gkantsidis Georgia Institute of Technology Pablo Rodriguez Microsoft Research IEEE INFOCOM.
End-to-End Analysis of Distributed Video-on-Demand Systems Padmavathi Mundur, Robert Simon, and Arun K. Sood IEEE Transactions on Multimedia, February.
Scalable and Continuous Media Streaming on Peer-to-Peer Networks M. Sasabe, N. Wakamiya, M. Murata, H. Miyahara Osaka University, Japan Presented By Tsz.
Lecture 17 I/O Optimization. Disk Organization Tracks: concentric rings around disk surface Sectors: arc of track, minimum unit of transfer Cylinder:
Peer-to-Peer Based Multimedia Distribution Service Zhe Xiang, Qian Zhang, Wenwu Zhu, Zhensheng Zhang IEEE Transactions on Multimedia, Vol. 6, No. 2, April.
VCR-oriented Video Broadcasting for Near Video-On- Demand Services Jin B. Kwon and Heon Y. Yeon Appears in IEEE Transactions on Consumer Electronics, vol.
Energy Efficient Prefetching – from models to Implementation 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software Engineering.
Energy Efficient Prefetching with Buffer Disks for Cluster File Systems 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software.
Locality-Aware Request Distribution in Cluster-based Network Servers 1. Introduction and Motivation --- Why have this idea? 2. Strategies --- How to implement?
Prefix Caching assisted Periodic Broadcast for Streaming Popular Videos Yang Guo, Subhabrata Sen, and Don Towsley.
Multiple Sender Distributed Video Streaming Thinh Nguyen, Avideh Zakhor appears on “IEEE Transactions On Multimedia, vol. 6, no. 2, April, 2004”
End-to-End Analysis of Distributed Video-on-Demand Systems P. Mundur, R. Simon, and A. K. Sood IEEE Transactions on Multimedia, Vol. 6, No. 1, Feb 2004.
PROMISE: Peer-to-Peer Media Streaming Using CollectCast M. Hefeeda, A. Habib, B. Botev, D. Xu, and B. Bhargava ACM Multimedia 2003, November 2003.
A Novel Video Layout Strategy for Near-Video-on- Demand Servers Shenze Chen & Manu Thapar Hewlett-Packard Labs 1501 Page Mill Rd. Palo Alto, CA
Distributed Servers Architecture for Networked Video Services S. H. Gary Chan, Member IEEE, and Fouad Tobagi, Fellow IEEE.
Redundant Data Update in Server-less Video-on-Demand Systems Presented by Ho Tsz Kin.
Performance Evaluation of Peer-to-Peer Video Streaming Systems Wilson, W.F. Poon The Chinese University of Hong Kong.
Efficient Support for Interactive Browsing Operations in Clustered CBR Video Servers IEEE Transactions on Multimedia, Vol. 4, No.1, March 2002 Min-You.
Server-Based Smoothing of Variable Bit-Rate Streams Stergios V. Anastasiadis, Kenneth C. Sevcik, and Michael Stumm ACM Multimedia 2001.
Streaming Video Gabriel Nell UC Berkeley. Outline Scalable MPEG-4 video – Layered coding method – Integrated transport-decoder buffer model RAP streaming.
Peer-to-peer Multimedia Streaming and Caching Service by Won J. Jeon and Klara Nahrstedt University of Illinois at Urbana-Champaign, Urbana, USA.
Design of an Interactive Video- on-Demand System Yiu-Wing Leung, Senior Member, IEEE, and Tony K. C. Chan IEEE Transactions on multimedia March 2003.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 34 – Media Server (Part 3) Klara Nahrstedt Spring 2012.
Lecture 39: Review Session #1 Reminders –Final exam, Thursday 3:10pm Sloan 150 –Course evaluation (Blue Course Evaluation) Access through.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Storage System: RAID Questions answered in this lecture: What is RAID? How does one trade-off between: performance, capacity, and reliability? What is.
A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Raymond Leung and Jack Y.B. Lee Department of Information.
Storage Systems CSE 598d, Spring 2007 Lecture 5: Redundant Arrays of Inexpensive Disks Feb 8, 2007.
1 Recitation 8 Disk & File System. 2 Disk Scheduling Disks are at least four orders of magnitude slower than main memory –The performance of disk I/O.
©2001 Pål HalvorsenINFOCOM 2001, Anchorage, April 2001 Integrated Error Management in MoD Services Pål Halvorsen, Thomas Plagemann, and Vera Goebel University.
Exploring VoD in P2P Swarming Systems By Siddhartha Annapureddy, Saikat Guha, Christos Gkantsidis, Dinan Gunawardena, Pablo Rodriguez Presented by Svetlana.
B. Prabhakaran1 Multimedia Storage & Retrieval Large sizes as well as real-time requirements of multimedia objects influence their storage and retrieval.
E0262 MIS - Multimedia Playback Systems Anandi Giridharan Electrical Communication Engineering, Indian Institute of Science, Bangalore – , India.
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS 2007 (TPDS 2007)
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
RAID COP 5611 Advanced Operating Systems Adapted from Andy Wang’s slides at FSU.
1 Cache Me If You Can. NUS.SOC.CS5248 OOI WEI TSANG 2 You Are Here Network Encoder Sender Middlebox Receiver Decoder.
QoS Support in High-Speed, Wormhole Routing Networks Mario Gerla, B. Kannan, Bruce Kwan, Prasasth Palanti,Simon Walton.
A novel approach of gateway selection and placement in cellular Wi-Fi system Presented By Rajesh Prasad.
Paper # – 2009 A Comparison of Heterogeneous Video Multicast schemes: Layered encoding or Stream Replication Authors: Taehyun Kim and Mostafa H.
App. TypeApp. Name Distributed or Parallel A parallel version of the Gaussian elimination application SAGE (SAIC's Adaptive Grid Eulerian hydrocode) Adaptive.
1 Push-to-Peer Video-on-Demand System. 2 Abstract Content is proactively push to peers, and persistently stored before the actual peer-to-peer transfers.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 31 – Media Server (Part 1) Klara Nahrstedt Spring 2012.
Multimedia Retrieval Architecture Electrical Communication Engineering, Indian Institute of Science, Bangalore – , India Multimedia Retrieval Architecture.
CS Spring 2009 CS 414 – Multimedia Systems Design Lecture 27 – Media Server (Part 2) Klara Nahrstedt Spring 2009.
Experimental Study on Wireless Multicast Scalability using Merged Hybrid ARQ with Staggered Adaptive FEC S. Makharia, D. Raychaudhuri, M. Wu*, H. Liu*,
Geethanjali College Of Engineering and Technology Cheeryal( V), Keesara ( M), Ranga Reddy District. I I Internal Guide Mrs.CH.V.Anupama Assistant Professor.
RAID Redundant Arrays of Independent Disks
Repair Pipelining for Erasure-Coded Storage
A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Raymond Leung and Jack Y.B. Lee Department of Information.
Video On Demand.
RAID RAID Mukesh N Tekwani
Xiaoyang Zhang1, Yuchong Hu1, Patrick P. C. Lee2, Pan Zhou1
Data Orgnization Frequently accessed data on the same storage device?
CMPE 252A : Computer Networks
RAID RAID Mukesh N Tekwani April 23, 2019
Presentation transcript:

A Server-less Architecture for Building Scalable, Reliable, and Cost-Effective Video-on-demand Systems Presented by: Raymond Leung Wai Tak Supervisor: Prof. Jack Lee Yiu-bun Department of Information Engineering The Chinese University of Hong Kong

Contents  1. Introduction  2. Challenges  3. Server-less Architecture  4. Reliability Analysis  5. Performance Modeling  6. System Dimensioning  7. Multiple Parity Groups  8. Conclusion

1. Introduction  Traditional Client-server Architecture Clients connect to server and request for video Server capacity limits the system capacity Cost increases with system scale

1. Introduction  Server-less Architecture Motivated by the availability of powerful user devices Each user node (STB) serves both as a client and as a mini-server Each user node contributes to the system Memory Processing power Network bandwidth Storage Costs shared by users

1. Introduction  Architecture Overview Composed of clusters

2. Challenges  Video Data Storage Policy  Retrieval and Transmission Scheduling  Fault Tolerance  Distributed Directory Service  Heterogeneous User Nodes  System Adaptation – node joining/leaving

3. Server-less Architecture  Storage Policy Video data is divided into fixed-size blocks (Q bytes) Data blocks are distributed among nodes in the cluster (data striping) Low storage requirement and load balancing Capable of fault tolerance using redundant blocks (discussed later)

3. Server-less Architecture  Retrieval and Transmission Scheduling Round-based scheduler Grouped Sweeping Scheduling 1 (GSS) Composed of macro rounds and micro rounds Tradeoff between disk efficiency and buffer requirement 1 P.S. Yu, M.S. Chen & D.D. Kandlur, “Grouped Sweeping Scheduling for DASD-based Multimedia Storage Management”, ACM Multimedia Systems, vol. 1, pp. 99 –109, 1993

3. Server-less Architecture  Retrieval and Transmission Scheduling Data retrieved in current micro round will be transmitted immediately in next micro round Each retrieval block is divided into b transmission blocks for transmission Transmission block size: Transmission lasts for one macro round

3. Server-less Architecture  Retrieval and Transmission Scheduling Macro round length Defined as the time required by all nodes transmitting one retrieval block Number of requests served: N Macro round length: Micro round length Each macro round is divided into g micro rounds Number of requests served: N/g Micro round length:

3. Server-less Architecture  Modification in Storage Policy As the retrieval blocks are divided into transmission blocks for transmission Video data is striped across transmission blocks, instead of retrieval blocks

3. Server-less Architecture  Fault Tolerance Recover from not a single node failure, but multiple simultaneously node failures as well Redundancy by Forward Error Correction (FEC) Code e.g. Reed-Solomon Erasure Code (REC)

3. Server-less Architecture  Impact of Fault Tolerance on Block Size Tolerate up to h simultaneous failures To maintain the same amount of video data transmitted in each macro round, the block size is increased to Q r. Similarly, the transmission block size is increased to U r.

4. Reliability Analysis  Reliability Analysis Find out the system mean time to failure (MTTF) Assuming independent node failure/repair rate Tolerate up to h failures by redundancy Analysis by Markov chain model

4. Reliability Analysis  Reliability Analysis With the assumption of independent failure and repair rate Let T i be the expected time the system takes to reach state h+1 from state i

4. Reliability Analysis  Reliability Analysis By solving the above set of equations, the system MTTF (T 0 ) is With a target system MTTF, we can find the redundancy (h) required

4. Reliability Analysis  Redundancy Level Defined as the proportion of nodes serving redundant data (h/N) Redundancy level versus number of nodes on achieving the target system MTTF

5. Performance Modeling  Storage Requirement  Network Bandwidth Requirement  Buffer Requirement  System Response Time  Assumptions: Zero network delay Zero processing delay Bounded clock jitters among nodes

5. Performance Modeling  Storage Requirement Let S A be the combined size of all video titles to be stored in the cluster With redundancy h, additional storage is required The storage requirement per node (S N )

5. Performance Modeling  Bandwidth Requirement Assume video bitrate of R v bps Without redundancy, each node transmits (N  1) streams of video data to other nodes in the cluster, Each stream consuming a bitrate of R v /N bps With redundancy h, additional bandwidth is required The bandwidth requirement per node (C R )

5. Performance Modeling  Buffer Requirement Composed of sender buffer requirement and receiver buffer requirement  Sender Buffer Requirement Under GSS scheduling

5. Performance Modeling  Receiver Buffer Requirement Store the data temporarily before playback Absorb the deviations in data arrival time caused by clock jitter  Total Buffer Requirement One data stream is for local playback rather than transmission Buffer sharing for this local playback stream Subtract b buffer blocks of size U r from the receiver buffer

5. Performance Modeling  System Response Time Time required from sending out request to playback begins Scheduling delay + pre-fetch delay  Scheduling delay under GSS Time required from sending out request to data retrieval starts Can be analyzed using urns model Detailed derivation available in Lee’s work 2 2 Lee, J.Y.B., “Concurrent push-A scheduling algorithm for push-based parallel video servers”, IEEE Transactions on Circuits and Systems for Video Technology, Volume: 9 Issue: 3, April 1999, Page(s):

5. Performance Modeling  Prefetch delay Time required from retrieving data to playback begins One micro round to retrieve a data block and buffering time to fill up the prefetch buffer of the receiver Additional delay will be incurred due to clock jitter among nodes

6. System Dimensioning  Storage Requirement What is the minimum number of nodes required to store a given amount of video data? For example: video bitrate: 4 Mb/s video length: 2 hours storage required for 100 videos: GB If each node can allocate 2 GB for video storage, then 176 nodes are needed (without redundancy); or 209 nodes are needed (with 33 nodes added for redundancy) This sets the lower limit on the cluster size

6. System Dimensioning  Network Capacity How many nodes can be connected given a certain network switching capacity? For example: video bitrate: 4 Mb/s If the network switching capacity is 32Gbps, and assume 60% utilization up to 2412 nodes (without redundancy) Network switching capacity is not a bottleneck

6. System Dimensioning  Disk Access Bandwidth Determine the value of Q and g to evaluate the buffer requirement and the system response time Finite disk access bandwidth limits the value of Q and g  Disk Model on Disk Service Time Time required to retrieve data blocks for transmission Depends on seeking overhead, rotational latency and data block size Suppose k requests per GSS group The maximum service time in worst case scenario – maximum round service time -- fixed overhead – maximum seek time for k requests W -1 – rotational latency r min – minimum transfer rate Q r – data block size

6. System Dimensioning  Constraint for Smooth Data Flow Disk service round to be finished before transmission Disk service time shorter than micro round length

6. System Dimensioning  Buffer Requirement Decreasing block size (Q r ) and increasing number of groups (g) to achieve minimum system response time, provided that the smooth data flow constraint is satisfied

6. System Dimensioning  System Response Time System response time versus number of nodes in the cluster

6. System Dimensioning  Scheduling Delay Relatively constant while system scales up  Prefetch Delay Time required to receive the first group of blocks from all nodes Increases linearly with system scale – not scalable Ultimately limits the cluster size  What is the Solution? Multiple parity groups

7. Multiple Parity Groups  Primary Limit in Cluster Scalability Prefetch delay in system response time  Multiple Parity Groups Instead of single parity group, the redundancy is encoded with multiple parity groups Decrease the number of blocks required to receive before playback Playback begins after receiving the data of first parity group Reduce the prefetch delay

7. Multiple Parity Groups  Multiple Parity Groups Transmission of different parity groups are staggered

7. Multiple Parity Groups  Impact on Performance Buffer requirement System response time Redundancy requirement  Buffer Requirement The number of blocks within same parity group is reduced Receiver buffer requirement is reduced

7. Multiple Parity Groups  System Response Time Playback begins after receiving the data of first parity group System response time is reduced

7. Multiple Parity Groups  Redundancy Requirement Cluster is divided into parity groups with less number of nodes Higher redundancy level to maintain the same system MTTF Tradeoff between response time and redundancy level

7. Multiple Parity Groups  Performance Evaluation Buffer requirement and system response versus redundancy level at a cluster size of 1500 nodes Both system response time and buffer requirement decrease with more redundancy (i.e. more parity groups)

7. Multiple Parity Groups  Cluster Scalability What are the system configurations if the system a. achieves a MTTF of 10,000 hours, and b. keeps under a response time constraint of 5 seconds, and c. keeps under a buffer requirement of 8/16 MB?

7. Multiple Parity Groups  Cluster Scalability The cluster is divided into more parity groups if it exceeds either the response time constraint, or the buffer constraint The redundancy level keeps relatively constant as the increased cluster size results in improved redundancy efficiency that compensates for the increased redundancy overhead incurred by the multiple parity group scheme (eg. 16 MB buffer constraint)

7. Multiple Parity Groups  Shifted bottleneck in Cluster Scalability Transmission buffer increases linearly with cluster scale and cannot be reduced by multiple parity group scheme The system is forced to divided into more parity groups to reduce the receiver buffer requirement to stay within the buffer constraint The redundancy overhead is sharply increased and the system response system is sharply reduced (eg. 8 MB buffer constraint) Eventually the total buffer requirement exceeds the buffer constraint even the cluster is further divided into more parity groups  Scalability Bottleneck Shifted to the Buffer Requirement System can be further scaled up by forming autonomous clusters

8. Conclusion  Server-less Architecture Scalable Acceptable redundancy level to achieve reasonable response time in a cluster Further scale up by forming new autonomous clusters Reliable Fault tolerance by redundancy Comparable reliability as high-end server by the analysis using Markov chain Cost-Effective Dedicated server is eliminated Costs shared by all users

8. Conclusion  Future Work Distributed Directory Service Heterogeneous User Nodes Dynamic System Adaptation Node joining/leaving Data re-distribution

End of Presentation Thank you Question & Answer Session.