CIS 581 Course Project Heshan Lin

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Learning Techniques for Video Shot Detection Under the guidance of Prof. Sharat Chandran by M. Nithya.
Automatic Video Shot Detection from MPEG Bit Stream Jianping Fan Department of Computer Science University of North Carolina at Charlotte Charlotte, NC.
Automated Shot Boundary Detection in VIRS DJ Park Computer Science Department The University of Iowa.
Taxonomic classification for web- based videos Author: Yang Song et al. (Google) Presenters: Phuc Bui & Rahul Dhamecha.
1 Video Processing Lecture on the image part (8+9) Automatic Perception Volker Krüger Aalborg Media Lab Aalborg University Copenhagen
Chapter 4: Image Enhancement
Segmentation Course web page: vision.cis.udel.edu/~cv May 2, 2003  Lecture 29.
Video Shot Boundary Detection at RMIT University Timo Volkmer, Saied Tahaghoghi, and Hugh E. Williams School of Computer Science & IT, RMIT University.
EE 7730 Image Segmentation.
Computer Vision Group Edge Detection Giacomo Boracchi 5/12/2007
Chapter 10 Image Segmentation.
Introduction to Computer Vision
1 Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or.
Segmentation Divide the image into segments. Each segment:
A Robust Scene-Change Detection Method for Video Segmentation Chung-Lin Huang and Bing-Yao Liao IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY.
Segmentation by Clustering Reading: Chapter 14 (skip 14.5) Data reduction - obtain a compact representation for interesting image data in terms of a set.
1 Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or.
CSE (c) S. Tanimoto, 2007 Segmentation and Labeling 1 Segmentation and Labeling Outline: Edge detection Chain coding of curves Segmentation into.
Computer Vision - A Modern Approach Set: Segmentation Slides by D.A. Forsyth Segmentation and Grouping Motivation: not information is evidence Obtain a.
Image Segmentation by Clustering using Moments by, Dhiraj Sakumalla.
Chapter 10: Image Segmentation
Information Extraction from Cricket Videos Syed Ahsan Ishtiaque Kumar Srijan.
Scripting Once the proposal is accepted, it’s time to script it. The script is the master plan that details the elements of your program, from the razzle-dazzle.
Chapter 9.  Mathematical morphology: ◦ A useful tool for extracting image components in the representation of region shape.  Boundaries, skeletons,
Interactive Graph Cuts for Optimal Boundary & Region Segmentation of Objects in N-D Images (Fri) Young Ki Baik, Computer Vision Lab.
Computer Vision, Robert Pless Lecture 11 our goal is to understand the process of multi-camera vision. Last time, we studies the “Essential” and “Fundamental”
Chap. 9: Image Segmentation Jen-Chang Liu, Motivation Segmentation subdivides an image into its constituent regions or objects Example: 生物細胞在影像序列中的追蹤.
Digital Image Processing CCS331 Relationships of Pixel 1.
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
Image Segmentation and Edge Detection Digital Image Processing Instructor: Dr. Cheng-Chien LiuCheng-Chien Liu Department of Earth Sciences National Cheng.
KRISTIN LAM REU WEEKS 1 & 2. MATERIAL COVERED MATLAB basics Edge Detection Harris Corner Detector Adaboost Face Detection Optical Flow Lucas-Kanade Method.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities Prof. Charlene Tsai.
Magic Camera Master’s Project Defense By Adam Meadows Project Committee: Dr. Eamonn Keogh Dr. Doug Tolbert.
Content-Based Image Retrieval QBIC Homepage The State Hermitage Museum db2www/qbicSearch.mac/qbic?selLang=English.
Image and Video Retrieval INST 734 Doug Oard Module 13.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Text From Corners: A Novel Approach to Detect Text and Caption in Videos Xu Zhao, Kai-Hsiang Lin, Yun Fu, Member, IEEE, Yuxiao Hu, Member, IEEE, Yuncai.
– Area presented with Camera Shots – The Viewpoint (Point of View) – Camera Angle 3 Important Considerations in Shooting TV Commercials.
Machine Vision ENT 273 Regions and Segmentation in Images Hema C.R. Lecture 4.
Wonjun Kim and Changick Kim, Member, IEEE
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities May 2, 2005 Prof. Charlene Tsai.
Course 5 Edge Detection. Image Features: local, meaningful, detectable parts of an image. edge corner texture … Edges: Edges points, or simply edges,
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Working with Cinematic Techniques English 9 Perry High School.
Introduction To Computational and Biological Vision Max Binshtok Ohad Greenshpan March 2006 Shot Detection in video.
Shadow Detection in Remotely Sensed Images Based on Self-Adaptive Feature Selection Jiahang Liu, Tao Fang, and Deren Li IEEE TRANSACTIONS ON GEOSCIENCE.
The Grammar of Editing. What is an Edit? ed·it –assembly of 2 or more shots in a linear fashion using cuts, fades, dissolves, and other common transitions.
Medical Image Analysis
Automatic Video Shot Detection from MPEG Bit Stream
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities Prof. Charlene Tsai.
Presenter: Ibrahim A. Zedan
Computer Vision Lecture 13: Image Segmentation III
Image Segmentation – Edge Detection
Video Titles The Titler is a versatile tool enabling you to create not just titles and credits, but animated graphics and text as well. Titles may.
Introduction to Morphological Operators
: Chapter 9: Image Segmentation
Computer Vision Lecture 12: Image Segmentation II
Outline Announcement Texture modeling - continued Some remarks
DC Image Extraction and Shot Segmentation
Video Titles The Titler is a versatile tool enabling you to create not just titles and credits, but animated graphics and text as well. Titles may be placed.
Histogram Probability distribution of the different grays in an image.
پردازش تصاویر دیجیتال- احمدی فرد
Spatial operations and transformations
Digital Image Processing Week III
CSSE463: Image Recognition Day 29
Segmentation and Edge Detection
Spatial operations and transformations
Presentation transcript:

CIS 581 Course Project Heshan Lin Video Shot Detection CIS 581 Course Project Heshan Lin

Agenda What’s shot detection? Classification of shot detection Close look to hard cuts detection Experiments and Results

What’s Shot Detection Problem definition – shot detection: given a video V consisting of n shots, find the beginning and end of each shot. Also known as shot boundary detection or transition detection. It is fundamental to any kind of video analysis and video application since it enables segmentation of a video into its basic components: the shots.

Classification Hard cuts: A cut is an instantaneous transition from one scene to the next. There are no transitional frames between 2 shots. Fades: A fade is a gradual transition between a scene and a constant image (fade-out) or between a constant image and a scene (fade-in). Classified based on transition types

Fades During a fade, images have their intensities multiplied by some value α. During a fade-in, α increases from 0 to 1, while during a fade-out α decreases from 1 to 0. Qiang’s movie

Classification Hard cuts: A cut is an instantaneous transition from one scene to the next. Fades: A fade is a gradual transition between a scene and a constant image (fade-out) or between a constant image and a scene (fade-in). Dissolves: A dissolve is a gradual transition from one scene to another, in which the first scene fades out and the second scene fades in. Classified based on transition types

Dissolves Combination of fade-in and fade-out.

Classification Hard cuts: A cut is an instantaneous transition from one scene to the next. Fades: A fade is a gradual transition between a scene and a constant image (fade-out) or between a constant image and a scene (fade-in). Dissolves: A dissolve is a gradual transition from one scene to another, in which the first scene fades out and the second scene fades in. Wipe: another common scene break is a wipe, in which a line moves across the screen, with the new scene appearing behind the line. Germy’s movie

Schema of Cut Detection Calculate a time series of discontinuity feature values f(n) for each frame. Suppose we use function d(x,y) to measure the dissimilarity between frame x and y. The discontinuity feature value for frame n is f(n)=d(n-1,n). Pick the cuts position from f(n) based on some threshold techniques. My project only cover cuts detection

Example

Features to Measure Dissimilarity Intensity/color histogram Edges/contours: Based on edge change ratio (ECR). Let σn be the number of edge pixels in frame n, and Xnin and Xn-1out the number of entering and exiting edge pixels in frames in frames n and n-1, respectively. The edge change ratio ECRn between frames n-1 and n is defined as: Our approach is based on a simple observation: during a cut or a dissolve, new intensity edges appear far from the locations of old edges. Similarly, old edges disappear far from the location of new edges. We define an edge pixel that appears far from an existing edge pixel as an entering edge pixel, and an edge pixel that disappears far from an existing edge pixel as an exiting edge pixel. By counting the entering and exiting edge pixels, we can detect and classify cuts, fades and dissolves. By analyzing the spatial distribution of entering and exiting edge pixels, we can detect and classify wipes.

Edges/contours (cont.) How to define the entering and exiting edge pixels Xnin and Xn-1out? Suppose we have 2 binary images en-1 and en. The entering edge pixels Xnin are the fraction of edge pixels in en which are more than a fixed distance r from the closest edge pixel in en-1. Similarly the exiting edge pixels are the fraction of edge pixels in en-1 which are farther than r away from the closest edge pixel in en. Not entering edge En-1 En Impose En to En-1 Entering edge

We can set the distance r by specify the Dilate parameter imd1 = rgb2gray(im1); Imd2 = rgb2gray(im2); % black background image bw1 = edge(imd1, 'sobel'); bw2 = edge(imd2, 'sobel'); % invert image to white background ibw2 = 1-bw2; ibw1 = 1-bw1; s1 = size(find(bw1),1); s2 = size(find(bw1),1); % dilate se = strel('square',3); dbw1 = imdilate(bw1, se); dbw2 = imdilate(bw2, se); imIn = dbw1 & ibw2; imOut = dbw2 & ibw1; ECRIn = size(find(imIn),1)/s2; ECROut = size(find(imOut),1)/s1; ECR = max(ECRIn, ECROut); We can set the distance r by specify the Dilate parameter

Thresholding Global threshold A hard cut is declared each time the discontinuity value f(n) surpasses a global thresholds. Adaptive threshold A hard cut is detected based on the difference of the current feature values f(n) from its local neighborhood. Generally this kind of method has 2 criteria for a hard cut declaration: - F(n) takes the maximum value inside the neighborhood. - The difference between f(n) and its neighbors’ feature values is bigger than a given threshold. A common problem of global thresholding is that in practice it is impossible to find a single global threshold that works with all kinds of video material. Should be avoided. Use slide window, an example of adaptive threshold.

Experiments Input: Mr. Beans movie. (80*112, 2363 frames) Dissimilarity function - Intensity histogram - Edge change ratio (ECR) Thresholding - Adaptive threshold based on statistics model.

Thresholding Use a slide window with size 2w+1. The middle frame in the window is detected as a cut if: - Its feature value is the maximum in the window. - Its feature value is greater than where Td is a parameter given a value of 5 in this experiment. Draw some graph to explain windows size here.

The statistics model is based on following assumption: The dissimilarity feature values f(n) for a frame comes from two distributions: one for shot boundaries(S) and one for “not-a-shot-boundary”(N). In general, S has a considerably larger mean and standard deviation than N. Threshold

Results Intensity histogram dissimilarity + adaptive thresholding

Results(cont.) ECR dissimilarity + adaptive thresholding

Compare We compare the cut positions detected by these 2 methods in the following table. From the results we can see the cut detected by these 2 methods are pretty stable. Frame# Cut1 Cut2 Cut3 Cut4 Cut5 Cut6 Cut7 Intensity Histogram 998 1167 1292 1359 2081 2184 ECR 86 2129 2312

Cut detected in frame 998

Comments