Detection theory 1. Definition of the problematic

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Maximum likelihood (ML) and likelihood ratio (LR) test
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Machine Learning CMPT 726 Simon Fraser University
1 Bayesian Restoration Using a New Nonstationary Edge-Preserving Image Prior Giannis K. Chantas, Nikolaos P. Galatsanos, and Aristidis C. Likas IEEE Transactions.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Inferences About Process Quality
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter 3 (part 1): Maximum-Likelihood & Bayesian Parameter Estimation  Introduction  Maximum-Likelihood Estimation  Example of a Specific Case  The.
Today Wrap up of probability Vectors, Matrices. Calculus
Data Selection In Ad-Hoc Wireless Sensor Networks Olawoye Oyeyele 11/24/2003.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Chapter 8 Introduction to Hypothesis Testing
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 13 Oct 14, 2005 Nanjing University of Science & Technology.
Dept. of EE, NDHU 1 Chapter Three Baseband Demodulation/Detection.
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
REVISED CONTEXTUAL LRT FOR VOICE ACTIVITY DETECTION Javier Ram’ırez, Jos’e C. Segura and J.M. G’orriz Dept. of Signal Theory Networking and Communications.
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Baseband Demodulation/Detection
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
1 E. Fatemizadeh Statistical Pattern Recognition.
1 University of Texas at Austin Machine Learning Group 图像与视频处理 计算机学院 Motion Detection and Estimation.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
BCS547 Neural Decoding.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
In the name of Allah. Scientific & Technical Presentation Leila Sharif Sharif University of Technology
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Labwork 3.
Applied statistics Usman Roshan.
Lecture 1.31 Criteria for optimal reception of radio signals.
Chapter 3: Maximum-Likelihood Parameter Estimation
ELEC4600 Radar and Navigation Engineering
A 2 veto for Continuous Wave Searches
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Deep Feedforward Networks
12. Principles of Parameter Estimation
Fundamentals of estimation and detection in signals and images
LECTURE 03: DECISION SURFACES
LECTURE 01: COURSE OVERVIEW
LECTURE 05: THRESHOLD DECODING
REMOTE SENSING Multispectral Image Classification
Ying shen Sse, tongji university Sep. 2016
Slides for Introduction to Stochastic Search and Optimization (ISSO) by J. C. Spall CHAPTER 15 SIMULATION-BASED OPTIMIZATION II: STOCHASTIC GRADIENT AND.
LECTURE 05: THRESHOLD DECODING
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
ELEC6111: Detection and Estimation Theory Course Objective
Solving an estimation problem
EE513 Audio Signals and Systems
LECTURE 01: COURSE OVERVIEW
Image and Video Processing
Parametric Methods Berlin Chen, 2005 References:
LECTURE 05: THRESHOLD DECODING
12. Principles of Parameter Estimation
Bayesian Decision Theory
Presentation transcript:

Detection theory 1. Definition of the problematic 2. Criterion for evaluating detection quality 3. Optimal detector : the likelihood ratio 4. Nuisance parameters : the GLRT 5. Application to edge detection in images

Lagrange multipliers One searches points x* in which a scalar function of a vectorial parameter reaches an extremum under the constraint that x verifies : The solution of this problem is found in the following way : One defines the « Lagrange function» : One determines the extrema of this function, for example by annulling it gradient One obtains solutions that depend on l : The values of l that correspond to the solutions of the problem are those which verify the constraint: The solutions are : 4 4 4 4 2 4

Lagrange multipliers Example : in the plane, maximize under the constraint 4 4 4 4 2 4

4 4 4 4 4 2

IV. Detection theory 1. Definition of the problematic 2. Criterion for evaluating detection quality 3. Optimal detector : the likelihood ratio 4. Nuisance parameters : the GLRT 5. Application to edge detection in images

Problematic Binary decision Multiple decision Hypothesis 1 : Object present Hypothesis 0 : Object absent Binary decision Multiple decision Detection consists in choosing (making a decision) between two hypotheses -> hypothesis testing

In this course, one will only address the problem of detection Problematic In practice, one often encounters detection / localization problems 1) One sweeps the image with the « window » F 2) For each position (i,j) of the window, one performs a hypothesis test H1: object present in F (i,j) H0: no object in F In this course, one will only address the problem of detection

Decision theory The measured signal is a vector s that belongs to a P-dimensional space W. One assumes that it can belong to two classes go ou g1. The goal of every detection algorithm is to partition the space W into 2 regions w0 et w1 so that the decision rule is: W w1 w0 x s 4 4 4 4 2 4

Decision theory Equivalently, determining a detection algorithm consists in defining a « discriminant surface » between regions w1 et w0. In practice, one obtains a discriminant function : function g(s) such that : Discriminant surface W w1 w0 Question : what is the optimal discriminant function ? x s 4 4 4 4 4 2

Statistical detection theory One models the signal s as a random vector -> statistical decision theory. One knows the following information about the problem : : a priori probability of class K NB: « a priori » means « before performing the measure » Ex : Radar detection : Property : : « Likelihood » NB: probability of obtaining a given value of the vector if one knows its class 4 4 4 4 4 2

Example Detection of a signal r perturbed by an additive white Gaussian noise with zero mean and variance s2. g0 : signal absent / g1 : signal present If the signal belongs to class g1 : If the signal belongs to class g0 : 4 4 4 2 4 4

a posteriori probability Bayes relations: Joint probability density a posteriori probability: It is the probability that signal s belongs to class gk. Useful relations: 4 4 4 4 2 4

Two types of detection errors Probability of non-detection : W w0 w1 Probability of false alarm : Can we minimize simultaneously Pnd et Pfa ? 4 4 4 4 4 2

Two types of detection errors One can normalize these probabilities with respect to the a priori probability of each class : One can also define the detection probability : Objective : maximize Pd for a given value of Pfa 4 4 4 2 4 4

Lagrange multipliers One searches points x* in which a scalar function of a vectorial parameter reaches an extremum under the constraint that x verifies : The solution of this problem is found in the following way : One defines the « Lagrange function» : One determines the extrema of this function, for example by annulling it gradient One obtains solutions that depend on l : To determine the values of l that correspond to the true solutions, one applies the constraint: The solutions are : 4 4 4 4 2 4

Optimal detector Lagrange function: Optimal decision rule ? w1 w0 4 4 2 4 4

Optimal detector Optimal decision rule ? Likelihood ratio (LR) : avec 4 4 4 2 4 4

Optimal detector Detection of a signal r perturbed by an additive white Gaussian noise with zero mean and variance s2. One compares the scalar product between the reference signal and the measured signal to a threshold. 4 4 4 4 2 4

Statistical distribution of the likelihood ratio The likelihood ratio is a random variable. Its probability density depends on whether the signal s belongs to class g0 or to class g1: 4 4 4 4 2 4

Statistical distribution of the likelihood ratio 4 4 4 4 2 4

It is easier to work with Example Detection of a signal r perturbed by an additive white Gaussian noise with zero mean and variance s2. It is easier to work with

Example Detection of a signal r perturbed by an additive white Gaussian noise with zero mean and variance s2. One obtains :

Performance of a detector Pd 1 Receiver Operating Characteristic (ROC). Pfa 1

Performance of a detector Optimality of the LR : its ROC is above the ROC of any other detector The ROC depends on two parameters : Pd LR 1 - The nature of the detector (the algorithm) Other detector - The intrinsic difficulty of the detection problem (RSB, …) “Tossing a coin” Pfa 1

Nuisance parameters In each class, the signal may depend on unknown parameters pk. The expression of the likelihood is thus : One estimates their values in the ML sense : One defines the Generalized Likelihood Ratio (GLR) 4 4 4 2 4 4

Nuisance parameters Example : detection in the presence of additive Gaussian noise Nuisance parameter ML estimate of the nuisance parameter : Expression of the GLR : 4 4 4 2 4 4

Performance of the GLR Pd LR 1 GLR “Tossing a coin” Pfa 1

The article is to be handed in, PDF format, on October 28th. Project The goal of the project is To study the resolution that can be reached by PALM imagery, and the optimal parameters of the imaging system parameter. To simplify, this study is to be done on 1 dimensional signals. To implement the algorithm for reconstruction a higher resolution image from a sequence of images of isolated fluorophores, and apply it to a test image. This work is to be reported under the form of a scientific journal paper. To this end, a template of the journal “Optics Express” is on Libres Savoirs (in MS Word and Latex). The article is to be handed in, PDF format, on October 28th.

Project The features that must be reported in this work are: Precision analysis (1D): Analysis of the CRLB when parameter a is unknown, especially its variation with q. Expression of the ML estimator, and estimation of its bias and efficiency Optimal value of the PSF width w   Implementation of the algorithm (2D) Algorithm for detection and rough estimation of the position of single fluorophores in an image. Expression and implementation of ML position estimator in 2D. Application of the method to the proposed test image. The previous points are only guidelines. You are free order them as you want. The only requirement is that the “flow” of the paper is logical.

Project Low resolution image

Project PALM sequence

Project Data (available on Libres Savoirs) ImageTest.mat: reference image for testing your algorithm. The PSF is an isotropic Gaussian pulse with the same FWHM w=2 in the two directions and maximal value a=1. CoordinatesTest.mat: exact positions of the fluorophores in image ImageTest.mat ImagesPALM.mat: sequence of 784 PALM images. BlurredImage.png: low resolution image of the pattern to reconstruct from ImagesPALM.mat sequence.

Structure of a scientific paper Title, authors, institutions, abstract 1. Introduction Context, bibliography (literature review) Description the addressed problem (why is this article worth publishing ?) Outline of the paper 2. Derivation of the estimator -> ML estimator 3. Implementation of the algorithm -> optimization algorithm 4. Validation of the algorithm on simulated images Influence of data and algorithm parameters on convergence and reconstructed image quality 5. Application of the algorithm to real data 6. Conclusion Summary of main results Perspectives (improvements, other application domains, …) Appendices Detail of calculations, technical details of an experiment, … References : publications cited in the paper (numbered)

Reference (publication date !) Title Authors and institutions Abstract Context and bibliography Introduction

Context and bibliography Original contribution and content of the paper Derivation of the estimator

Derivation of the estimator Implementation of the algorithm

Validation of the algorithm on simulated images Parameter study

Convergence of the algorithm Quality of reconstruction

Quality of reconstruction Results on experimental data CONCLUSION

CONCLUSION Results on experimental data APPENDIX

APPENDIX REFERENCES

Structure of a scientific paper Figures Always numbered and cited in the paper Caption understandable without reading the main text

Structure of a scientific paper Figures Always numbered and cited in the paper Caption understandable without reading the main text Equations Numbered, and all notations are explained

Structure of a scientific paper Figures Always numbered and cited in the paper Caption understandable without reading the main text Equations Numbered, and all notations are explained Bibliographical references Always cited in the text Always numbered Respect the format of the review

These advices are also useful for your 3rd year internship report ! Tips for writing Decide the outline of the paper before beginning to write : What the important results I want to report ? What are the figures I want to show ? Start writing the core of the paper, then the introduction, the conclusion, and finally the abstract. Length of the text is not proportional to the time spent doing the work ! The outline of the paper is chosen for clarity : it does not always follow the chronology of the work ! These advices are also useful for your 3rd year internship report !