. Algorithms in Computational Biology – 236522 אלגוריתמים בביולוגיה חישובית – 236522 (http://webcourse.cs.technion.ac.il/236522)http://webcourse.cs.technion.ac.il/236522.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Basics of Statistical Estimation
Probabilistic models Haixu Tang School of Informatics.
1 Essential Probability & Statistics (Lecture for CS598CXZ Advanced Topics in Information Retrieval ) ChengXiang Zhai Department of Computer Science University.
Learning: Parameter Estimation
Scores and substitution matrices in sequence alignment Sushmita Roy BMI/CS 576 Sushmita Roy Sep 11 th,
Hidden Markov Model.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio LECTURE 11 (Lab): Probability reminder.
The Occasionally Dishonest Casino Narrated by: Shoko Asei Alexander Eng.
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
.. . Parameter Estimation using likelihood functions Tutorial #1 This class has been cut and slightly edited from Nir Friedman’s full course of 12 lectures.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Intro to Bayesian Learning Exercise Solutions Ata Kaban The University of Birmingham 2005.
Probabilities and Probabilistic Models
Parameter Estimation using likelihood functions Tutorial #1
. Learning – EM in ABO locus Tutorial #08 © Ydo Wexler & Dan Geiger.
. Learning – EM in The ABO locus Tutorial #8 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
HMM for CpG Islands Parameter Estimation For HMM Maximum Likelihood and the Information Inequality Lecture #7 Background Readings: Chapter 3.3 in the.
. Algorithms in Computational Biology – אלגוריתמים בביולוגיה חישובית – (
. Learning Hidden Markov Models Tutorial #7 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
This presentation has been cut and slightly edited from Nir Friedman’s full course of 12 lectures which is available at Changes.
. Learning – EM in The ABO locus Tutorial #9 © Ilan Gronau.
Probabilistic Graphical Models Tool for representing complex systems and performing sophisticated reasoning tasks Fundamental notion: Modularity Complex.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Introduction to Bayesian Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán)
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
. Maximum Likelihood (ML) Parameter Estimation with applications to reconstructing phylogenetic trees Comput. Genomics, lecture 6b Presentation taken from.
Modeling biological data and structure with probabilistic networks I Yuan Gao, Ph.D. 11/05/2002 Slides prepared from text material by Simon Kasif and Arthur.
. Inference in HMM Tutorial #6 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
CS5263 Bioinformatics Lecture 9: Motif finding Biological & Statistical background.
Bayesian Learning and Learning Bayesian Networks.
Introduction to Bayesian Learning Ata Kaban School of Computer Science University of Birmingham.
Class 3: Estimating Scoring Rules for Sequence Alignment.
Thanks to Nir Friedman, HU
. Learning – EM in The ABO locus Tutorial #9 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
. Parameter Estimation For HMM Lecture #7 Background Readings: Chapter 3.3 in the text book, Biological Sequence Analysis, Durbin et al.,  Shlomo.
. Markov Chains Tutorial #5 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
Recitation 1 Probability Review
.. . Maximum Likelihood (ML) Parameter Estimation with applications to inferring phylogenetic trees Comput. Genomics, lecture 6a Presentation taken from.
Bayesian Inference Ekaterina Lomakina TNU seminar: Bayesian inference 1 March 2013.
BINF6201/8201 Hidden Markov Models for Sequence Analysis
Naive Bayes Classifier
Introduction to Bayesian statistics Yves Moreau. Overview The Cox-Jaynes axioms Bayes’ rule Probabilistic models Maximum likelihood Maximum a posteriori.
Course Introduction What these courses are about What I expect What you can expect.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Learning In Bayesian Networks. General Learning Problem Set of random variables X = {X 1, X 2, X 3, X 4, …} Training set D = { X (1), X (2), …, X (N)
CS498-EA Reasoning in AI Lecture #10 Instructor: Eyal Amir Fall Semester 2009 Some slides in this set were adopted from Eran Segal.
Maximum Likelihood Estimation
1 Learning P-maps Param. Learning Graphical Models – Carlos Guestrin Carnegie Mellon University September 24 th, 2008 Readings: K&F: 3.3, 3.4, 16.1,
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
The Uniform Prior and the Laplace Correction Supplemental Material not on exam.
Probability. Probability Probability is fundamental to scientific inference Probability is fundamental to scientific inference Deterministic vs. Probabilistic.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Essential Probability & Statistics (Lecture for CS397-CXZ Algorithms in Bioinformatics) Jan. 23, 2004 ChengXiang Zhai Department of Computer Science University.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Probability and Probability Distributions. Probability Concepts Probability: –We now assume the population parameters are known and calculate the chances.
Bayesian Estimation and Confidence Intervals Lecture XXII.
Markov Chains Tutorial #5
Tutorial #3 by Ma’ayan Fishelson
Important Distinctions in Learning BNs
CS498-EA Reasoning in AI Lecture #20
Statistical NLP: Lecture 4
CSCI 5822 Probabilistic Models of Human and Machine Learning
Important Distinctions in Learning BNs
Markov Chains Tutorial #5
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

. Algorithms in Computational Biology – אלגוריתמים בביולוגיה חישובית – ( תרגול : יום ג ' 9:30 – 10:20 טאוב 4 אילן גרונאו משרד : טאוב 700 טל :829(4894) דוא " ל : מצגות התרגול / הרצאה : ניתן להוריד יום לפני התרגול / הרצאה מאתר הקורס המצגות הן חומר עזר בלבד !! תרגילי בית : 5 תרגילים הגשה חובה בתא המתרגל (110# בקומה 5) איחורים ובעיות אחרות : להודיע למתרגל זמן מספיק לפני מועד ההגשה

. Algorithms in Computational Biology – The full story Best biological explanaiton Biological data Hypotheses space Problems: size of the space complex scoring functions

. Parameter Estimation Using Likelihood Functions Tutorial #1 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger

. The Problem: Data set Probabilistic Model Find the best explanation for the observed data Helps predict behavior of similar data sets Parameters: Θ = θ 1, θ 2, θ 3, …

. An example: Binomial experiments Heads - P(H) Tails - 1-P(H) The unknown parameter: θ=P(H) The data: series of experiment results, e.g. D = H H T H T T T H H … Main assumption: each experiment is independent of others Data set Model Parameters: Θ

. Maximum Likelihood Estimation (MLE): The likelihood of a given value for θ : L D (θ) = P(D| θ)  We wish to find a value for θ which maximizes the likelihood Example: The likelihood of ‘HTTHH’ is: L HTTHH (θ) = P(HTTHH | θ)= θ (1-θ) (1- θ) θ θ = θ 3 (1-θ) 2 We only need to know N(H) and N(T) (number of Heads and Tails). These are sufficient statistics : L D (θ) = θ N(H) (1-θ) N(T) An example: Binomial experiments

7 Sufficient Statistics A sufficient statistic is a function of the data that summarizes the relevant information for the likelihood. Formally, s(D) is a sufficient statistics if for any two datasets D and D’: s(D) = s(D’ )  L D (  ) = L D’ (  ) Likelihood may be calculated on the statistics.

. Maximum Likelihood Estimation (MLE): Reminder: L D (θ) = θ N(H) (1-θ) N(T) We wish to maximize the likelihood (or log-likelihood). l D (θ) = log(L D (θ)) = N(H)·log(θ) + N(T)·log(1-θ) Maximization: l D ’(θ) = 0   L()L() D = H T T H H An example: Binomial experiments

. Still – a series of independent experiments Each experiment has K possible results Examples: - die toss ( K=6 ) - DNA sequence ( K=4 ) - protein sequence ( K=20 ) We want to learn the parameters  1,  2. …,  K Sufficient statistics: N 1, N 2, …, N K - the number of times each outcome is observed Likelihood: MLE: Multinomial experiments

. Data set Model Parameters: Θ Likelihood – P(D|Θ) What we would like is to maximize is P(Θ|D). Bayes’ rule: Due to: In our case: The prior probability captures our subjective prior knowledge (prejudice) of the model parameters. Is ML all we need? prior probability likelihood posterior probability

. The prior probability captures our subjective prior knowledge (prejudice) of the model parameters. Example: After a set of coin tosses: ‘HTTHH’ what would you bet on for the next toss?  we a priori assume the coin is balanced with high probability. Possible assumptions in molecular biology: some amino acids have the similar frequencies some amino acids have low frequencies … Bayesian inference

. Example: dishonest casino A casino uses 2 kind of dice: 99% are fair 1% is loaded: 6 comes up 50% of the times We pick a die at random and roll it 3 times We get 3 consecutive sixes  What is the probability the die is loaded? Bayesian inference

. Example: dishonest casino  Pr[ loaded|3 sixes ] = ? Use Bayes’ rule !  Pr[ 3 sixes|loaded ] = (0.5) 3 (likelihood of ‘loaded’)  Pr[ loaded ] = 0.01 (prior)  Pr[ 3 sixes ] = Pr[ 3 sixes|loaded ] * Pr[ loaded ] + Pr[ 3 sixes|unloaded ] * Pr[ unloaded ] = (0.5) (0.1666) = Unlikely loaded!! Bayesian inference

. Biological Example: protein amino-acid composition Extracellular proteins have a slightly different amino acid composition than intracellular proteins. From a large enough protein database (e.g. SwissProt ) we can get the following statistics: - P(int) - the probability that an amino-acid sequence is intracellular - P(ext) - the probability that an amino-acid sequence is extracellular - P(a i |int) - the frequency of amino acid a i for intracellular proteins - P(a i |ext) - the frequency of amino acid a i for extracellular proteins  What is the probability that a given new protein sequence: X=x 1 x 2 … x n is extracellular? Bayesian inference

. What is the probability that a given new protein sequence: X=x 1 x 2 ….x n is extracellular? Assuming that every sequence is either extracellular or intracellular (but not both), we have: P(ext) = 1-P(int) P(X) = P(X|ext) ·P(ext) + P(X|int)·P(int) Furthermore: By Bayes’ theorem: Bayesian inference - Biological Example: Proteins (cont.) prior posterior

. Summary: Data set Model Parameters: Θ = θ 1, θ 2, θ 3, … The basic paradigm: Estimate model parameters maximizing:  Likelihood - P(D|Θ)  Posterior probability - P(Θ |D) α P(D|Θ) P(Θ) In the absence of a significant prior they are equivalent