Hack, Math and Stat --- call for a reconciliation between extreme views Song-Chun Zhu The Frontiers of Vision Workshop, August 20-23, 2011.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Local Discriminative Distance Metrics and Their Real World Applications Local Discriminative Distance Metrics and Their Real World Applications Yang Mu,
Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute.
Structured Sparse Principal Component Analysis Reading Group Presenter: Peng Zhang Cognitive Radio Institute Friday, October 01, 2010 Authors: Rodolphe.
The loss function, the normal equation,
Image Parsing: Unifying Segmentation and Detection Z. Tu, X. Chen, A.L. Yuille and S-C. Hz ICCV 2003 (Marr Prize) & IJCV 2005 Sanketh Shetty.
1 Hierarchical Image-Motion Segmentation using Swendsen-Wang Cuts Adrian Barbu Siemens Corporate Research Princeton, NJ Acknowledgements: S.C. Zhu, Y.N.
Graph Based Semi- Supervised Learning Fei Wang Department of Statistical Science Cornell University.
Primal Sketch Integrating Structure and Texture Ying Nian Wu UCLA Department of Statistics Keck Meeting April 28, 2006 Guo, Zhu, Wu (ICCV, 2003; GMBV,
Margins, support vectors, and linear programming, oh my! Reading: Bishop, 4.0, 4.1, 7.0, 7.1 Burges tutorial (on class resources page)
Conceptual model Examples: Heat flow on a wire Air flow around a plane’s wing Response to medication.
Optical Flow Brightness Constancy The Aperture problem Regularization Lucas-Kanade Coarse-to-fine Parametric motion models Direct depth SSD tracking.
Stereopsis Mark Twain at Pool Table", no date, UCR Museum of Photography.
Object Detection and Tracking Mike Knowles 11 th January 2005
Challenges in Computer Vision Understanding the ”seeing machine” The input (images) The output (shapes, actions?, diagnosis?) The mapping (statistics,
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from:  David Lowe, UBC  Jiri Matas, CMP Prague
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Linear Methods, cont’d; SVMs intro. Straw poll Which would you rather do first? Unsupervised learning Clustering Structure of data Scientific discovery.
Support Vector Machines
Data-Driven Markov Chain Monte Carlo Presented by Tomasz MalisiewiczTomasz Malisiewicz for Advanced PerceptionAdvanced Perception 3/1/2006.
Introduction to machine learning
More Machine Learning Linear Regression Squared Error L1 and L2 Regularization Gradient Descent.
CSE 788 X.14 Topics in Computational Topology: --- An Algorithmic View Lecture 1: Introduction Instructor: Yusu Wang.
Seeing Math in the World. Application Use functions to model real-life problems Project trends based on models Finances Population Data analysis.
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
Sampletalk Technology Presentation Andrew Gleibman
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Markov Random Fields Probabilistic Models for Images
28 February, 2003University of Glasgow1 Cluster Variation Method and Probabilistic Image Processing -- Loopy Belief Propagation -- Kazuyuki Tanaka Graduate.
Learning to Perceive Transparency from the Statistics of Natural Scenes Anat Levin School of Computer Science and Engineering The Hebrew University of.
Human pose recognition from depth image MS Research Cambridge.
Neural Heuristics For Problem Solving: Using ANNs to Develop Heuristics for the 8-Puzzle by Bambridge E. Peterson.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Ensemble Methods in Machine Learning
Machine Learning Saarland University, SS 2007 Holger Bast Marjan Celikik Kevin Chang Stefan Funke Joachim Giesen Max-Planck-Institut für Informatik Saarbrücken,
COMP24111: Machine Learning Ensemble Models Gavin Brown
Markov Random Fields & Conditional Random Fields
CSE 5559 Computational Topology: Theory, algorithms, and applications to data analysis Lecture 0: Introduction Instructor: Yusu Wang.
Introduction to Machine Learning © Roni Rosenfeld,
1 Review and Summary We have covered a LOT of material, spending more time and more detail on 2D image segmentation and analysis, but hopefully giving.
WLD: A Robust Local Image Descriptor Jie Chen, Shiguang Shan, Chu He, Guoying Zhao, Matti Pietikäinen, Xilin Chen, Wen Gao 报告人:蒲薇榄.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Introduction to Sampling based inference and MCMC
Introduction to Machine Learning
Action-Grounded Push Affordance Bootstrapping of Unknown Objects
Hongfang Wang and Edwin R. Hancock Department of Computer Science
Performance of Computer Vision
STEREO MATCHING USING POPULATION-BASED MCMC
Nonparametric Semantic Segmentation
COMP61011 : Machine Learning Ensemble Models
Linear regression project
Machine Learning & Data Science
UC San Diego Department of Mathematics
Outline Statistical Modeling and Conceptualization of Visual Patterns
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Analysis of Contour Motions
Outline Texture modeling - continued Julesz ensemble.
Expectation-Maximization & Belief Propagation
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
A Block Based MAP Segmentation for Image Compression
Ensemble Methods: Bagging.
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Week 5 Cecilia La Place.
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Hack, Math and Stat --- call for a reconciliation between extreme views Song-Chun Zhu The Frontiers of Vision Workshop, August 20-23, 2011.

Math : under certain conditions, things can be said analytically. Hack, Math and Stat: co-existing ingredients Stat : is, essentially, regression, fitting by weighted features. Hack : something, somehow, works better somewhere. Hack Stat Math breadth depth

* Berry 1898, A Short History of Astronomy, p.240. ** Stigler 1986,The History of Statistics, p.26. Hack : Celestial navigation For example, when people looked at the sky Math : Newtonian Gravitational Theory. 1680s Newton reportedly told Halley that lunar theory (to represent mathematically the motion of moon ) “made his head ache and kept him awake so often that he would think of it no more”.* e.g. the Chinese expedition in Ming Dynasty, Stat : Rescue the solar system by lsq (Mayer, Legendre, Laplace, Euler) n x 1, x 2, …, x k y ** Euler 1749

Math : --- 3D geometry, lighting models, PDEs, … Pattern Theory. When people look at images Stat : --- Boosting, Support vector machines (Regression with various loss functions and algorithms.) Hack : --- good features (corner, SIFT, HoG, LBP, …), algorithms that are not supposed to apply … (such as Loopy Belief Propagation)… BTW, true for most vision algorithms, it is unclear what state spaces they are operating on, and whether or how it may converge. What makes research exciting are the “discoveries” that connection them !

Julesz s Daugman1985 Bergen, Adelson Heeger … 1991, 95 Julesz ensemble, 1999 Gibbs 1902 equivalence, 1999 Lewis etc. 95 Case study I: Development of MRF/texture models Besag 1973 Ising 1920 Potts 1957 Cross and Jain 1983 FRAME, 1996

Case study II: Dev. MCMC/optimizing algorithms Metropolis 1946 Hastings 1970 Waltz 1972 (labeling) Rosenfeld, Hummel, Zucker 1976 (relaxation-labeling ) Geman brothers 1984, (Gibbs sampler) Swendsen-Wang 1987 (cluster sampling) Jump-diffusion, Miller & Grenander,1994 Kirkpatrick, 1983 Reversible jump, Green 1995 Swendsen-Wang Cut 2003 DDMCMC C4: Clustering w. +/- Constraints, 2009

Historic trends in vision, based on my perception HackStatMath HackStat Math Before HackStat Math

The field is out of balance and polarized HackStatMath 2006-now Not only that hacks have become dominating, but original work on stat and math have been diminishing. A longer term problem is its impacts on student training. Vision = Classification +  = Datasets + Features + Classifiers + Curves + 

Quotes from review comments From the far-right wing: Why not you compare against X1, X2, X3 on datasets Y1, Y2, Y3 ??? Mathematically rigorous methods usually do not work ! I have no interest reading your theory until you show me performance better than the state-of-the-art. If you are smart, why aren’t you rich?

Quotes from review comments From the far-left wing. Your method is trained and tested only on a finite, artificial dataset, therefore it does not contribute to the solution of any real vision problem I am talking with mathematical certainty that it is impossible to reconstruct a 3D scene from a single image. So this paper must be rejected. With so many sub-problems unsolved, it makes no sense to solve a problem as complex as image parsing. “Professor, your question was wrong …”

Finally: How to reach the moon, in 20 years?