CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis.

Slides:



Advertisements
Similar presentations
Ch 7.7: Fundamental Matrices
Advertisements

Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
Linear Regression.
Separating Hyperplanes
CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis.
Principal Component Analysis
Chapter 5: Linear Discriminant Functions
L15:Microarray analysis (Classification) The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Discrimination Methods As Used In Gene Array Analysis.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Discriminant Functions Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Greg GrudicIntro AI1 Introduction to Artificial Intelligence CSCI 3202: The Perceptron Algorithm Greg Grudic.
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Fa 05CSE182 CSE182-L8 Mass Spectrometry. Fa 05CSE182 Bio. quiz What is a gene? What is a transcript? What is translation? What are microarrays? What is.
Whole Genome Assembly Microarray analysis. Mate Pairs Mate-pairs allow you to merge islands (contigs) into super-contigs.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Linear Discriminant Functions Chapter 5 (Duda et al.)
Introduction to Bioinformatics Algorithms Clustering and Microarray Analysis.
Radial Basis Function Networks
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Sp’10Bafna/Ideker Classification (SVMs / Kernel method)
Whole Genome Expression Analysis
Classification (Supervised Clustering) Naomi Altman Nov '06.
CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis.
Scenario 6 Distinguishing different types of leukemia to target treatment.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Class Prediction and Discovery Using Gene Expression Data Donna K. Slonim, Pablo Tamayo, Jill P. Mesirov, Todd R. Golub, Eric S. Lander 발표자 : 이인희.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Extracting quantitative information from proteomic 2-D gels Lecture in the bioinformatics course ”Gene expression and cell models” April 20, 2005 John.
Non-Bayes classifiers. Linear discriminants, neural networks.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
Lecture 4 Linear machine
Linear Models for Classification
Class 23, 2001 CBCl/AI MIT Bioinformatics Applications and Feature Selection for SVMs S. Mukherjee.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 7: Linear and Generalized Discriminant Functions.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
A Short and Simple Introduction to Linear Discriminants (with almost no math) Jennifer Listgarten, November 2002.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Linear Classifiers Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
SUPPORT VECTOR MACHINES Presented by: Naman Fatehpuria Sumana Venkatesh.
Computational Intelligence: Methods and Applications Lecture 22 Linear discrimination - variants Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
EQTLs.
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
LECTURE 11: Advanced Discriminant Analysis
Molecular Classification of Cancer
Statistical Learning Dong Liu Dept. EEIS, USTC.
Whole Genome Assembly.
Clustering.
Generally Discriminant Analysis
Parametric Methods Berlin Chen, 2005 References:
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Clustering.
Presentation transcript:

CSE182 L14 Mass Spec Quantitation MS applications Microarray analysis

CSE182 Counting transcripts cDNA from the cell hybridizes to complementary DNA fixed on a ‘chip’. The intensity of the signal is a ‘count’ of the number of copies of the transcript

CSE182 Quantitation: transcript versus Protein Expression mRNA Protein 1 Protein 2 Protein 3 Sample 1Sample 2 Sample 1Sample2 Our Goal is to construct a matrix as shown for proteins, and RNA, and use it to identify differentially expressed transcripts/proteins

CSE182 Gene Expression Measuring expression at transcript level is done by micro-arrays and other tools Expression at the protein level is being done using mass spectrometry. Two problems arise: –Data: How to populate the matrices on the previous slide? (‘easy’ for mRNA, difficult for proteins) –Analysis: Is a change in expression significant? (Identical for both mRNA, and proteins). We will consider the data problem here. The analysis problem will be considered when we discuss micro-arrays.

CSE182 The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute Lymphocytic Leukemia) & AML (Acute Myelogenous Leukima) Possibly, the set of expressed genes is different in the two conditions

CSE182 Supplementary fig. 2. Expression levels of predictive genes in independent dataset. The expression levels of the 50 genes most highly correlated with the ALL-AML distinction in the initial dataset were determined in the independent dataset. Each row corresponds to a gene, with the columns corresponding to expression levels in different samples. The expression level of each gene in the independent dataset is shown relative to the mean of expression levels for that gene in the initial dataset. Expression levels greater than the mean are shaded in red, and those below the mean are shaded in blue. The scale indicates standard deviations above or below the mean. The top panel shows genes highly expressed in ALL, the bottom panel shows genes more highly expressed in AML.

CSE182 Gene Expression Data Gene Expression data: –Each row corresponds to a gene –Each column corresponds to an expression value Can we separate the experiments into two or more classes? Given a training set of two classes, can we build a classifier that places a new experiment in one of the two classes. g s1s1 s2s2 s

CSE182 Three types of analysis problems Cluster analysis/unsupervised learning Classification into known classes (Supervised) Identification of “marker” genes that characterize different tumor classes

CSE182 Supervised Classification: Basics Consider genes g 1 and g 2 –g 1 is up-regulated in class A, and down-regulated in class B. –g 2 is up-regulated in class A, and down-regulated in class B. Intuitively, g1 and g2 are effective in classifying the two samples. The samples are linearly separable. g1g1 g2g

CSE182 Basics With 3 genes, a plane is used to separate (linearly separable samples). In higher dimensions, a hyperplane is used.

CSE182 Non-linear separability Sometimes, the data is not linearly separable, but can be separated by some other function In general, the linearly separable problem is computationally easier.

CSE182 Formalizing of the classification problem for micro-arrays Each experiment (sample) is a vector of expression values. –By default, all vectors v are column vectors. –v T is the transpose of a vector The genes are the dimension of a vector. Classification problem: Find a surface that will separate the classes v vTvT

CSE182 Formalizing Classification Classification problem: Find a surface (hyperplane) that will separate the classes Given a new sample point, its class is then determined by which side of the surface it lies on. How do we find the hyperplane? How do we find the side that a point lies on? g1g1 g2g

CSE182 Basic geometry What is ||x|| 2 ? What is x/||x|| Dot product? x=(x 1,x 2 ) y

CSE182 Dot Product Let  be a unit vector. –||  || = 1 Recall that –  T x = ||x|| cos  What is  T x if x is orthogonal (perpendicular) to  ?  x   T x = ||x|| cos 

CSE182 Hyperplane How can we define a hyperplane L? Find the unit vector that is perpendicular (normal to the hyperplane)

CSE182 Points on the hyperplane Consider a hyperplane L defined by unit vector , and distance  0 Notes; –For all x  L, x T  must be the same, x T  =  0 –For any two points x 1, x 2, (x 1 - x 2 ) T  =0 x1x1 x2x2

CSE182 Hyperplane properties Given an arbitrary point x, what is the distance from x to the plane L? –D(x,L) = (  T x -  0 ) When are points x1 and x2 on different sides of the hyperplane? x 00

CSE182 Separating by a hyperplane Input: A training set of +ve & - ve examples Goal: Find a hyperplane that separates the two classes. Classification: A new point x is +ve if it lies on the +ve side of the hyperplane, -ve otherwise. The hyperplane is represented by the line {x:-  0 +  1 x 1 +  2 x 2 =0} x2x2 x1x1 + -

CSE182 Error in classification An arbitrarily chosen hyperplane might not separate the test. We need to minimize a mis- classification error Error: sum of distances of the misclassified points. Let y i =-1 for +ve example i, –y i =1 otherwise. Other definitions are also possible. x2x2 x1x1 + - 

EXTRA SLIDES CSE182

Gradient Descent The function D(  ) defines the error. We follow an iterative refinement. In each step, refine  so the error is reduced. Gradient descent is an approach to such iterative refinement. D(  )  D’(  )

CSE182 Rosenblatt’s perceptron learning algorithm

CSE182 Classification based on perceptron learning Use Rosenblatt’s algorithm to compute the hyperplane L=( ,  0 ). Assign x to class 1 if f(x) >= 0, and to class 2 otherwise.

CSE182 Perceptron learning If many solutions are possible, it does no choose between solutions If data is not linearly separable, it does not terminate, and it is hard to detect. Time of convergence is not well understood

CSE182 Linear Discriminant analysis Provides an alternative approach to classification with a linear function. Project all points, including the means, onto vector . We want to choose  such that –Difference of projected means is large. –Variance within group is small x2x2 x1x1 + - 

CSE182 LDA Cont’d Fisher Criterion

CSE182 Maximum Likelihood discrimination Suppose we knew the distribution of points in each class. –We can compute Pr(x|  i ) for all classes i, and take the maximum

CSE182 ML discrimination Suppose all the points were in 1 dimension, and all classes were normally distributed.

CSE182 ML discrimination recipe We know the distribution for each class, but not the parameters Estimate the mean and variance for each class. For a new point x, compute the discrimination function g i (x) for each class i. Choose argmax i g i (x) as the class for x