Presentation is loading. Please wait.

Presentation is loading. Please wait.

Institute of Computer Science Chair of Communication Networks Prof. Dr.-Ing. P. Tran-Gia Modeling YouTube QoE based on Crowdsourcing and Laboratory User.

Similar presentations


Presentation on theme: "Institute of Computer Science Chair of Communication Networks Prof. Dr.-Ing. P. Tran-Gia Modeling YouTube QoE based on Crowdsourcing and Laboratory User."— Presentation transcript:

1 Institute of Computer Science Chair of Communication Networks Prof. Dr.-Ing. P. Tran-Gia Modeling YouTube QoE based on Crowdsourcing and Laboratory User Studies Tobias Hoßfeld, Raimund Schatz STSM 15.8.-30.9.2011 http://www3.informatik.uni-wuerzburg.de/research/fia http://www3.informatik.uni-wuerzburg.de/staff/hossfeld

2 2 2 Tobias Hoßfeld QoE Issue: Waiting, Waiting, Waiting… Stalling Waiting Time Perception

3 3 3 Tobias Hoßfeld Research Activities Related to STSM Application-Level Measurements bottleneck scenario with constant bandwidth video characteristics realistic stalling patterns Monitoring and Stalling Detector heuristics fit QoS information extraction approach leads to exact QoE results Optimization and Dimensioning initial delay (GI/GI/1): T 0 /D<5% bandwidth provisioning: 120%V TCP better UDP in bottleneck QoE Modeling only stalling relevant, not content, demographics, etc. users “accept” almost no or only short stalling crowdsourcing supports i:lab video player parameter, initial buffer 2sec variable video bit rate V; high stalling frequency for V=B QoE management used stalling length in tests: 1-6sec mapping between QoS (e.g. bandwidth B) and QoE stalling as key influence factor

4 4 4 Tobias Hoßfeld Executive Summary of STSM Developed Test Design Conducted Crowd- sourcing Tests Derived QoE Model Application Measurements Laboratory Study Remote users ‘Reliability’ questions App./user monitoring Preloading of data Data analysis Identification of reliable users Key influences factors via machine learning Fitting with fundamental relationships Mapping function: stalling and QoE Acceptance vs. perception Comparison crowdsourcing with laboratory results Realistic parameters for temporal stimuli Reliable users Different demographics Different test setting, e.g. longer user tests

5 5 5 Tobias Hoßfeld Crowdsourcing Workflow  Challenge: identify unreliable QoE results Countermeasures:  proper test design (gold standard data, consistency questions, content questions, application monitoring)  filtering data and analyzing QoE results 12 3 4 5  Methods also applicable to e.g. field trials!

6 6 6 Tobias Hoßfeld Crowdsourcing: Unreliable workers LEVEL 1: ‘reliability’ questions - wrong answers to content questions - different answers to the same questions - always selected same option - consistency questions: specified the wrong country/continent LEVEL 1: ‘reliability’ questions - wrong answers to content questions - different answers to the same questions - always selected same option - consistency questions: specified the wrong country/continent LEVEL 2: ‘QoE’ question - did not notice stalling - perceived non-existent stalling LEVEL 2: ‘QoE’ question - did not notice stalling - perceived non-existent stalling LEVEL 3: ‘application/user’ monitoring - did not watch all videos completely LEVEL 3: ‘application/user’ monitoring - did not watch all videos completely C1 C2 C3 C4 C5 C6 C7 Facebook SOS hypothesis indicates unreliable test Many user ratings rejected  further improvements required User warnings („Test not done carefully“)  rejection rate decreased about 50% Filtering may be too strict  application layer monitoring not reliable

7 7 7 Tobias Hoßfeld Crowdsourcing vs. Laboratory Studies 4 seconds of stalling  Key influence factors on YouTube QoE  stalling frequency and stalling duration determine the user perceived quality  Lab studies within ACE 2.0 at FTW’s i:Lab  Similar shapes of curves in laboratory and crowdsourcing study

8 8 8 Tobias Hoßfeld Conclusions  Most of relevant stimuli of Internet applications are of temporal nature  QoE models have to be extended in temporal dimension: stalling, waiting times, service interruptions  Gap between user perception and user acceptance, differences in lab and crowdsourcing (WG3)  ‘Failed’ subjective studies for analysis of reliability (WG4)  Standards to detect unreliable subjects (WG5)  Crowdsourcing appears promising  Tests are conducted fast at low costs  Possibility to access different user groups (in terms of expectations/social background)  But new challenges are imposed WG1: “Web and cloud apps” WG2: “Crowd- sourcing”

9 9 9 Tobias Hoßfeld Outcome of STSM  “Quantification of YouTube QoE via Crowdsourcing” by Tobias Hoßfeld, Raimund Schatz, Michael Seufert, Matthias Hirth, Thomas Zinner, Phuoc Tran-Gia, IEEE International Workshop on Multimedia Quality of Experience - Modeling, Evaluation, and Directions (MQoE 2011), Dana Point, CA, USA, December 2011.  “FoG and Clouds: On Optimizing QoE for YouTube” by Tobias Hoßfeld, Florian Liers, Thomas Volkert, Raimund Schatz, accepted at 5th KuVS GI/ITG Workshop "NG Service Delivery Platforms", at DOCOMO Euro-Labs, Munich, Germany  “Quality of Experience of YouTube Video Streaming for Current Internet Transport Protocols” by Tobias Hoßfeld and Raimund Schatz, currently under submission at ACM Computer Communications Review; a technical report of University of Würzburg is available containing the numerical results, Technical Report No. 482: “Transport Protocol Influences on YouTube QoE”, July 2011.  " ‘Time is Bandwidth’? Narrowing the Gap between Subjective Time Perception and Quality of Experience” by Sebastian Egger, Peter Reichl, Tobias Hoßfeld, Raimund Schatz, submitted to IEEE ICC 2012 - Communication QoS, Reliability and Modeling Symposium  “Challenges of QoE Management for Cloud Applications” by Tobias Hoßfeld, Raimund Schatz, Martin Varela, Christian Timmerer, submitted to IEEE Communications Magazine, Special Issues on QoE management in emerging multimedia services  “Recommendations and Comparison of Subjective User Tests via Crowdsourcing and Laboratories for online video streaming”, intended for submission  “Impact of Fake User Ratings on QoE”, intended for Journal submission.


Download ppt "Institute of Computer Science Chair of Communication Networks Prof. Dr.-Ing. P. Tran-Gia Modeling YouTube QoE based on Crowdsourcing and Laboratory User."

Similar presentations


Ads by Google