Presentation is loading. Please wait.

Presentation is loading. Please wait.

Bayesian Machine learning and its application Alan Qi Feb. 23, 2009.

Similar presentations


Presentation on theme: "Bayesian Machine learning and its application Alan Qi Feb. 23, 2009."— Presentation transcript:

1 Bayesian Machine learning and its application Alan Qi Feb. 23, 2009

2 Motivation massive data from various sources: web pages, facebook, high-throughput biological data, high- throughput chemical data, etc. Challenging goal: how to model complex systems and extract knowledge from data.

3 Bayesian machine learning  Bayesian learning method Principled way to fuse prior knowledge and new evidence in data  Key issues  Model Design  Computation  Wide-range applications

4 Bayesian learning in practice  Applications: Recommendation systems (Amazon, NetFlix) Text Parsing (Finding latent topics in documents) Systems biology (where computations meets biology) Computer vision (parsing handwritten diagram automatically) Wireless communications Computational finance....

5 Learning for biology: understanding gene regulation during organism development Protein, product of Gene B DNA Gene A  Learning functionalities of genes for development  Inferring high-resolution protein- DNA binding locations from low- resolution measurement  Learning regulatory cascades during embryonic stem cell development

6 6 Wild-type lineageNo C lineageExtra ‘C’ lineages Data: gene expression profiles from wide-types & mutants (Baugh et al, 2005)

7 Bayesian semisupervised classification for finding tissue-specific genes BGEN: (Bayesian GENeralization from examples, Qi et al., Bioinformatics 2006) Labeled expression Labeled expression Classifier  Graph-based kernels (F. Chung, 1997, Zhu et al., 2003, Zhou et al. 2004)  Gaussian process classifier that is trained by EP and classifies the whole genome efficiently  Estimating noise and probe quality by approximate leave-one-out error Gene expression

8 Biological experiments support our predictions CNon C Muscle Epidermis CNon C Muscle Epidermis K01A2.5 R11A5.4 Ge’s lab

9 Data: genomic sequences

10

11 RNA: messager

12 Consensus Sequences Useful for publication IUPAC symbols for degenerate sites Not very amenable to computation Nature Biotechnology 24, 423 - 425 (2006)

13 Probabilistic Model.2.5.1.7.2.1.3.1.2.4.5.4.1.2.1.4.1.2.1 ACGTACGT M1M1 MKMK M1M1 P k (S|M) Position Frequency Matrix (PFM) 1 K Count frequencies Add pseudocounts

14 Bayesian learning: Estimating motif models by Gibbs sampling P(Sequences|params1,params2) Parameter1 Parameter2 In theory, Gibbs Sampling less likely to get stuck a local maxima

15 Bayesian learning: Estimating motif models by expectation maximization P(Sequences|params1,params2) Parameter1 Parameter2 To minimize the effects of local maxima, you should search multiple times from different starting points

16 Scoring A Sequence To score a sequence, we compare to a null model Background DNA (B).2.5.1.7.2.1.3.1.2.4.5.4.1.2.1.4.1.2.1 ACGTACGT Log likelihood ratio -0.3 1 -1.3 1.4-0.3 -1.30.3 -1.3-0.30.61 -1.3-0.30.3-0.3 -1.30.6-1.3-0.3-1.3 ACGTACGT Position Weight Matrix (PWM) PFM

17 Scoring a Sequence MacIsaac & Fraenkel (2006) PLoS Comp Bio Common threshold = 60% of maximum score

18 Visualizing Motifs – Motif Logos Represent both base frequency and conservation at each position Height of letter proportional to frequency of base at that position Height of stack proportional to conservation at that position

19 Software implemenation: AlignACE http://atlas.med.harvard.edu/cgi-bin/alignace.pl Implements Gibbs sampling for motif discovery – Several enhancements ScanAce – look for motifs in a sequence given a model CompareAce – calculate “similarity” between two motifs (i.e. for clustering motifs)

20 Data: biological networks

21 Network Decomposition Infinite Non-negative Matrix Factorization 1.Formulate the discovery of network legos as a non- negative factorization problem 2.Develop a novel Bayesian model which automatically learns the number of the bases.

22 Network Decomposition Synthetic Network Decomposition

23 Network Decomposition

24 Data: Movie rating User-item Matrix of Ratings Recommend: 5 Not Recommend: 1 X =

25 Task: how to predict user preference “Based on the premise that people looking for information should be able to make use of what others have already found and evaluated.” (Maltz & Ehrlich, 1995) E.g., if you like movies A, B, C, D, and E. And I like A, B, C, D but have not seen E yet. What would be my possible rating on E?

26 Collaborative filtering for recommendation systems Matrix factorization as an collaborative filtering approach: X ≈ Z A where X is N by D, Z is N by K and A is K by D. x i,j : user i’s rating on movie j z i,k : user i’s interests in movie category k (e.g., action, thriller, comedy, romance, etc.) A k,j : how likely movie j belong to movie category k Such that x i,j ≈ z i,1 A 1,j + z i,2 A2 2,j + … + z i,K A K,j

27 Bayesian learning of matrix factorization Training: Use probability theory, in particular, Bayeisan inference, to learn the model parameters Z, A given data X, which contains missing elements, i.e., unknown ratings Prediction: use estimated Z and A to predict unkown ratings in X

28 Test resutls ‘Jester’ dataset: Map from [-10,10] to [0,20] 10 random chosen datasets, each with 1000 users. For each user we randomly hold out 10 ratings for testing IMF, INMF and NMF(K=2…9)

29 Collaborative Filtering

30 Task How to find latent topics and group documents, such as emails, papers, or news into different clusters?

31 Data: text documents X = Computer science papersBiology papers

32 Assumptions 1.The keywords are shared in different documents of one topic. 2.The more important the keyword is, the more frequent it appears.

33 Matrix factorization models (again) X = Z A x i,j : the frequency word j appears in document z i,k : how much content in document i is related to topic k (e.g., biology, computer science, etc.) A k,j : how important word j to topic k

34 Bayesian Matrix Factorization We will use Bayesian methods again to estimate Z and A. Once we can identify hidden topics by examining A and cluster documents.

35 Text Clustering ‘20 newsgroup’ dataset A subset of 815 articles and 477 words.

36 Discovered hidden topics

37 Summary Bayesian machine learning: A powerful tool enables computers to learn hidden relations from massive data and make sensible predictions. Applications in computational biology, e.g., gene expression analysis and motif discovery, and information extraction, e.g., text modeling.


Download ppt "Bayesian Machine learning and its application Alan Qi Feb. 23, 2009."

Similar presentations


Ads by Google