(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)

Slides:



Advertisements
Similar presentations
3D Geometry for Computer Graphics
Advertisements

The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Lecture 2 Math & Probability Background ch. 1-2 of Machine Vision by Wesley E. Snyder & Hairong Qi Spring 2014 BioE 2630 (Pitt) : (CMU RI)
Determinants Bases, linear Indep., etc Gram-Schmidt Eigenvalue and Eigenvectors Misc
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Principal Component Analysis
Motion Analysis (contd.) Slides are from RPI Registration Class.
CSci 6971: Image Registration Lecture 4: First Examples January 23, 2004 Prof. Chuck Stewart, RPI Dr. Luis Ibanez, Kitware Prof. Chuck Stewart, RPI Dr.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
Linear Algebra and Image Processing
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Unsupervised Learning II Feature Extraction
Lecture 16: Image alignment
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
Eigenfaces (for Face Recognition)
Review of Linear Algebra
CS479/679 Pattern Recognition Dr. George Bebis
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
Matrices and vector spaces
University of Ioannina
Postulates of Quantum Mechanics
Search Engines and Link Analysis on the Web
The Chinese University of Hong Kong
Intelligent Information System Lab
Lecture 13 Shape ch. 9, sec. 1-8, of Machine Vision by Wesley E
Methods In Medical Image Analysis
Lecture 12 Theory of Registration ch
Lecture 12 Level Sets & Parametric Transforms sec & ch
Lecture 6 Linear Processing ch. 5 of Machine Vision by Wesley E
An Introduction to Support Vector Machines
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
REMOTE SENSING Multispectral Image Classification
REMOTE SENSING Multispectral Image Classification
Lecture 5 Image Characterization ch. 4 of Machine Vision by Wesley E
Lecture 19 ITK’s Path Framework
Lecture 7—Image Relaxation: Restoration and Feature Extraction ch
2D transformations (a.k.a. warping)
3D Transformation CS380: Computer Graphics Sung-Eui Yoon (윤성의)
Lecture 15 Active Shape Models
Lecture 16: Morphology (ch 7) & Image Matching (ch 13) ch. 7 and ch
CONTEXT DEPENDENT CLASSIFICATION
Lecture 10 Segmentation, Part II (ch 8) Active Contours (Snakes) ch
The loss function, the normal equation,
Maths for Signals and Systems Linear Algebra in Engineering Lecture 18, Friday 18th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
Mathematical Foundations of BME Reza Shadmehr
Maths for Signals and Systems Linear Algebra in Engineering Lecture 6, Friday 21st October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
CIS 4350 Image ENHANCEMENT SPATIAL DOMAIN
Lecture 3 Math & Probability Background ch
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
Lecture 14 Shape ch. 9, sec. 1-8, of Machine Vision by Wesley E
Lecture 3 Math & Probability Background ch
Lecture 10 Segmentation, Part II (ch 8) Active Contours (Snakes) ch
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
Methods In Medical Image Analysis
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

18-791 (CMU ECE) : 42-735 (CMU BME) : BioE 2630 (Pitt) Lecture 3 Math & Probability Background ch. 1-2 of Machine Vision by Wesley E. Snyder & Hairong Qi Spring 2016 18-791 (CMU ECE) : 42-735 (CMU BME) : BioE 2630 (Pitt) Dr. John Galeotti Note: To print these slides in grayscale (e.g., on a laser printer), first change the Theme Background to “Style 1” (i.e., dark text on white background), and then tell PowerPoint’s print dialog that the “Output” is “Grayscale.” Everything should be clearly legible then. This is how I generate the .pdf handouts.

General notes about the book The book is an overview of many concepts Top quality design requires: Reading the cited literature Reading more literature Experimentation & validation ch. 1-2 of Machine Vision by Wesley E. Snyder & Hairong Qi

Two themes Consistency Optimization A conceptual tool implemented in many/most algorithms Often must fuse information from many local measurements and prior knowledge to make global conclusions about the image Optimization Mathematical mechanism The “workhorse” of machine vision

Image Processing Topics Enhancement Coding Compression Restoration “Fix” an image Requires model of image degradation Reconstruction

Classification & Further Analysis Machine Vision Topics Feature Extraction Classification & Further Analysis Original Image AKA: Computer vision Image analysis Image understanding Pattern recognition: Measurement of features Features characterize the image, or some part of it Pattern classification Requires knowledge about the possible classes Our Focus

Feature measurement Varies Greatly Original Image Restoration Ch. 6-7 Noise removal Segmentation Shape Analysis Consistency Analysis What about Registration? Matching Features

Probability Probability of an event a occurring: Independence Pr(a) Independence Pr(a) does not depend on the outcome of event b, and vice-versa Joint probability Pr(a,b) = Prob. of both a and b occurring Conditional probability Pr(a|b) = Prob. of a if we already know the outcome of event b Read “probability of a given b” Independence -> joint probability: Pr(a,b) = Pr(a|b)Pr(b) = Pr(a)Pr(b) = Pr(b|a)Pr(a)

Probability for continuously-valued functions Probability distribution function: P(x) = Pr(z<x) Probability density function:

Linear algebra Unit vector: |x| = 1 Orthogonal vectors: xTy = 0 Orthonormal: orthogonal unit vectors Inner product of continuous functions Orthogonality & orthonormality apply here too

Linear independence No one vector is a linear combination of the others xj  ai xi for any ai across all i  j Any linearly independent set of d vectors {xi=1…d} is a basis set that spans the space d Any other vector in d may be written as a linear combination of {xi} Often convenient to use orthonormal basis sets Projection: if y=ai xi then ai=yTxi Liner independence geometrically?

Linear transforms = a matrix, denoted e.g. A Quadratic form: Positive definite: Applies to A if Matrix ex: rotation matrices Positive definite: Matrix always gives a vector pointing “somewhat” in the same direction as the original Positive definite: Always have a global minima: popular for optimization

More derivatives Of a scalar function of x: Of a vector function of x Called the gradient Really important! Of a vector function of x Called the Jacobian Hessian = matrix of 2nd derivatives of a scalar function Discuss derivative on prev. slide

Misc. linear algebra Derivative operators Eigenvalues & eigenvectors Translates “most important vectors” Of a linear transform (e.g., the matrix A) Characteristic equation: A maps x onto itself with only a change in length  is an eigenvalue x is its corresponding eigenvector

Function minimization Find the vector x which produces a minimum of some function f (x) x is a parameter vector f(x) is a scalar function of x The “objective function” The minimum value of f is denoted: The minimizing value of x is denoted: Why a scalar function?

Numerical minimization Gradient descent The derivative points away from the minimum Take small steps, each one in the “down-hill” direction Local vs. global minima Combinatorial optimization: Use simulated annealing Image optimization: Use mean field annealing

Markov models For temporal processes: For spatial processes The probability of something happening is dependent on a thing that just recently happened. For spatial processes The probability of something being in a certain state is dependent on the state of something nearby. Example: The value of a pixel is dependent on the values of its neighboring pixels.

Markov chain Simplest Markov model Example: symbols transmitted one at a time What is the probability that the next symbol will be w? For a “simple” (i.e. first order) Markov chain: “The probability conditioned on all of history is identical to the probability conditioned on the last symbol received.”

Hidden Markov models (HMMs) 1st Markov Process 2nd Markov f (t) Viewed as process that switches randomly b/t diff. signals Each graph is just 1 possibility from the underlying Markov pdfs. f (t)

HMM switching Governed by a finite state machine (FSM) Output 1st Process Output 2nd Process

The HMM Task Given only the output f (t), determine: The most likely state sequence of the switching FSM Use the Viterbi algorithm (much better than brute force) Computational Complexity of: Viterbi: (# state values)2 * (# state changes) Brute force: (# state values)(# state changes) The parameters of each hidden Markov model Use the iterative process in the book Better, use someone else’s debugged code that they’ve shared