Evaluating the Quality of Image Synthesis and Analysis Techniques Matthew O. Ward Computer Science Department Worcester Polytechnic Institute.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Active Shape Models Suppose we have a statistical shape model –Trained from sets of examples How do we use it to interpret new images? Use an “Active Shape.
Design of Experiments Lecture I
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Robust statistical method for background extraction in image segmentation Doug Keen March 29, 2001.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Evaluation of segmentation. Example Reference standard & segmentation.
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
GrabCut Interactive Image (and Stereo) Segmentation Joon Jae Lee Keimyung University Welcome. I will present Grabcut – an Interactive tool for foreground.
Mapping Nominal Values to Numbers for Effective Visualization Presented by Matthew O. Ward Geraldine Rosario, Elke Rundensteiner, David Brown, Matthew.
ASTER image – one of the fastest changing places in the U.S. Where??
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
EE 7730 Image Segmentation.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
RETRIEVAL OF MULTIMEDIA OBJECTS USING COLOR SEGMENTATION AND DIMENSION REDUCTION OF FEATURES Mingming Lu, Qiyu Zhang, Wei-Hung Cheng, Cheng-Chang Lu Department.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Segmentation Divide the image into segments. Each segment:
© Tefko Saracevic, Rutgers University 1 EVALUATION in searching IR systems Digital libraries Reference sources Web sources.
Minimum Classification Error Networks Based on book chapter 9, by Shigeru Katagiri Jaakko Peltonen, 28 th February, 2002.
Image Classification.
Fitting a Model to Data Reading: 15.1,
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Visual Analytics and the Geometry of Thought— Spatial Intelligence through Sapient Interfaces Alexander Klippel & Frank Hardisty Department of Geography,
Thresholding Thresholding is usually the first step in any segmentation approach We have talked about simple single value thresholding already Single value.
1 New York State Mathematics Core Curriculum 2005.
HELSINKI UNIVERSITY OF TECHNOLOGY LABORATORY OF COMPUTER AND INFORMATION SCIENCE NEURAL NETWORKS RESEACH CENTRE Variability of Independent Components.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
1 CSI5388 Data Sets: Running Proper Comparative Studies with Large Data Repositories [Based on Salzberg, S.L., 1997 “On Comparing Classifiers: Pitfalls.
כמה מהתעשייה? מבנה הקורס השתנה Computer vision.
Tal Mor  Create an automatic system that given an image of a room and a color, will color the room walls  Maintaining the original texture.
Accuracy Assessment. 2 Because it is not practical to test every pixel in the classification image, a representative sample of reference points in the.
Xiaojiang Ling CSE 668, Animate Vision Principles for 3D Image Sequences CSE Department, SUNY Buffalo
NERCOMP Workshop, Dec. 2, 2008 Information Visualization: the Other Half of Data Analysis Dr. Matthew Ward Computer Science Department Worcester Polytechnic.
CSE 185 Introduction to Computer Vision Pattern Recognition.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Environmental Remote Sensing Lecture 5: Image Classification " Purpose: – categorising data – data abstraction / simplification – data interpretation –
Classroom Assessment A Practical Guide for Educators by Craig A
Perception Introduction Pattern Recognition Image Formation
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Cost drivers, cost behaviour and cost estimation
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
© Fraunhofer MEVIS Meeting of the Working Group VCBM, 3. September 2013, Vienna Frank Heckel Fraunhofer MEVIS, Bremen, Germany | Innovation Center Computer.
1 Ch. 1: Software Development (Read) 5 Phases of Software Life Cycle: Problem Analysis and Specification Design Implementation (Coding) Testing, Execution.
Computer Graphics and Image Processing (CIS-601).
D. M. J. Tax and R. P. W. Duin. Presented by Mihajlo Grbovic Support Vector Data Description.
Software Development Problem Analysis and Specification Design Implementation (Coding) Testing, Execution and Debugging Maintenance.
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
Image Classification for Automatic Annotation
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Brian Lukoff Stanford University October 13, 2006.
URBDP 591 A Lecture 16: Research Validity and Replication Objectives Guidelines for Writing Final Paper Statistical Conclusion Validity Montecarlo Simulation/Randomization.
Comparative legal studies (Zinkovskiy Sergey, associate professor, PhD Department of the Theory and History of State and Law) Topic 2 Methodology of comparative.
Visual Information Processing. Human Perception V.S. Machine Perception  Human perception: pictorial information improvement for human interpretation.
Introduction to Machine Learning, its potential usage in network area,
Machine Learning with Spark MLlib
What Is Cluster Analysis?
Mean Shift Segmentation
Dynamical Statistical Shape Priors for Level Set Based Tracking
Project Implementation for ITCS4122
Unsupervised Classification
Ying Dai Faculty of software and information science,
Image Segmentation.
EE 492 ENGINEERING PROJECT
Visual Recognition of American Sign Language Using Hidden Markov Models 문현구 문현구.
Detecting Digital Forgeries using Blind Noise Estimation
Presentation transcript:

Evaluating the Quality of Image Synthesis and Analysis Techniques Matthew O. Ward Computer Science Department Worcester Polytechnic Institute

Evaluation is Important Has my modification improved the results? Which method works best for my data? What are limitations of my technique? Is my method better than XXX?

Evaluation is Difficult What aspects to test? How to measure? What are limitations of evaluation procedure? How to recruit evaluators?

Evaluation Often Avoided Majority of papers show no substantive evaluation Most common approach is subjective, by authors, on 1-3 test cases Quantitative measures exist for computational performance, but not quality of results Not much “glory” in evaluation

Case Study 1: Data Visualization

The Problem..... Lack of rigorous assessment of visualization techniques Lack of good test cases Limited comparison with other techniques Lack of guidelines for selection of appropriate techniques

A Possible Solution..... Create a list of goals of visualization –what is the overall task? –what is desired/acceptable level of accuracy? –what are we looking for? Locate/create data sets which contain desired features Test users on a wide range of tasks using different visualization methods

Goals of Visualization Identification - is there some interesting feature in the image? Classification - what is it? Quantification - how many? how big? how close? Understanding - are there correlations conveyed by the image? Comparison - does the image have characteristics similar to one generated with a different set of data?

Advantages of Synthetic Data Easy to adjust characteristics Less ambiguous than real data Easy to create data which contains a single structure or phenomena Real data can be noisy Hard to find real data with desired characteristics

Advantages of Real Data Results using real data are more “believable” Reality is hard to simulate accurately Real data has context which can help justify usefulness of tasks

Our Experiments Select two data characteristics of interest (outliers and clusters) Locate real data sets containing these features (validate with statistical analysis) Create synthetic data sets containing these features (also validate) Select three visualization techniques to test (scatterplots, parallel coordinates, principal components analysis with glyphs)

Our Experiments (continued) Train subjects on interpreting different display techniques Train subjects on the desired data characteristics Test subjects on each characteristic, varying –number of outliers/clusters –degree or size –amount of noise in synthetic sets –location of outlier/clusters

Visualization Techniques Tested Scatterplot MatrixParallel Coordinates Glyphs

Outlier Example

Cluster Example OriginalAdded Noise

Assessing the Results Detection - did subjects identify some structure in the image? Classification - did subjects correctly classify structure? Measurement - –number of clusters or outliers –outlier and cluster degree of separation –size of cluster Errors - false positives, missed structure, measurement accuracy

Summary of Experiments Scatterplot matrix –best overall –weak on overlapping clusters, size estimation for large clusters, interior outliers Glyphs –best for interior outliers –good for conveying outlier separation, overlapping clusters, measuring cluster size –poor for differentiating non-outliers Parallel coordinates –generally worse than others –good for differentiating non-outliers

Future Work Test alternate data characteristics (e.g. repeated patterns) Test alternate perceptual tasks (e.g. correlation) Test other visualization techniques (e.g. alternate glyphs, VisDB, dimensional stacking…..) Create publicly available benchmark suite for data sets and analysis tools (submissions from other researchers always welcome) Compare other multivariate visualization assessment methods as they arise.

Case Study 2: Image Segmentation

Problem Statement Image segmentation algorithms traditionally classified as model-based or context-free Model-based methods highly effective, but expensive to design and execute Context-free methods are fast, but quality of results often poor Is there some way to improve the results of context-free systems without incurring costs of model-based methods?

Conjecture In most image analysis domains, expectations can be placed on the likely occurrence of certain shapes, colors, and region/segment sizes. –Objects in an office scene mostly planar and non-specular –In medical images, boundaries are mostly smooth, and regions are usually small or moderate in size –Outdoor scenes contain a lot of fine texture We should be able to use high-level domain constraint knowledge to improve the segmentation process by: –Selecting a segmentation method likely to produce good results –Set the segmentation parameters to their most effective values

Defining a Good Segmentation All physical object boundaries should be isolated False boundaries should be minimized Boundary shape should be comparable to internal model of object in scene Precision in shape and position needed varies based on application and importance of individual objects to task at hand

Defining a Good Evaluation Procedure Should be based on real images Influence of human subjectivity minimized Errors categorized by type, severity, and significance Magnitude of error should accurately reflect difference from ideal Tolerance must be permitted

Problems with Pixel Counting 2 images with similar error counts, uniform dilation (left) and bad merge (right).

Procedure Acquire representative set of images for multiple domains Approximate constraints on edge/region features in domain Interactively segment and label edges/region tolerance and priority to create ideal segmentation Compute errors between ideal and algorithmically generated segmentation If error > acceptable, adjust parameters (simplex algorithm) and recompute errors Associate segmentation parameters with domain constraints

Creating the Ideal Segmentation Start with initial region- based segmentation Click on a region of interest Merge, split, set tolerance level, set priority level Iterate until all significant regions labeled Results are domain and task specific

Comparing Ideal to Computed Results Edge Detection: 78% detected essentials 209% oversegmentation Region Growing: 67% detected essentials 120% oversegmentation

More Results Split and Merge: 79% detected essentials 73% oversegmentation Rule-based System: 88% detected essentials 93% oversegmentation

Summary and Future Work Domain constraints produced better segmentations than context-free methods (after training) Future work includes investigating other types of constraints (e.g., texture) and improve the tolerance specification and error calculation

General Procedure for Assessing Image-Based Algorithms Determine the task to be performed by user of image Determine image features most relevant to this task, and ascertain level of accuracy needed in detection, classification, and measuring Create benchmark suite of data containing these features in varying degrees (real and synthetic data) Create and administer user tests to evaluate effectiveness of algorithm to accurately and reliably convey desired data features, or Develop image processing algorithms to identify desired data features and calculate error types and severities in images generated by algorithm being assessed

Summary of Presentation Formal assessment has proven useful in both visualization and image processing applications Results can be used to guide algorithm development and selection Quantitative and qualitative approaches can provide many insights into effectiveness of image analysis and synthesis tasks