Presentation is loading. Please wait.

Presentation is loading. Please wait.

Machine Learning Ali Ghodsi Department of Statistics

Similar presentations


Presentation on theme: "Machine Learning Ali Ghodsi Department of Statistics"— Presentation transcript:

1 Machine Learning Ali Ghodsi Department of Statistics
School of Computer Science University of Waterloo Winter 2009

2 Two Problems Classical Statistics
Infer information from small data sets (Not enough data) Machine Learning Infer information from large data sets (Too many data)

3 We are drowning in information and starving for knowledge.
Rutherford D. Roger

4 are all closely related to the second problem.
Other Names for ML Data mining, Applied statistics Adaptive (stochastic) signal processing Probabilistic planning or reasoning are all closely related to the second problem.

5 Applications Machine Learning is most useful when the structure of the task is not well understood but can be characterized by a dataset with strong statistical regularity. Search and recommendation (e.g. Google, Amazon) Automatic speech recognition and speaker verification Text parsing Face identification Tracking objects in video Financial prediction, fraud detection (e.g. credit cards) Medical diagnosis

6 More Applications More science and technology applications:
handwritten identification drug discovery (to identify the biological activity of chemical compounds using features describing the chemical structures) Gene expression analysis ( thousands of features with only dozens of observations)

7 Tasks Supervised Learning: given examples of inputs and corresponding desired outputs, predict outputs on future inputs. e.g.: classification, regression Unsupervised Learning: given only inputs, automatically discover representations, features, structure, etc. e.g.: clustering, dimensionality reduction, Feature extraction

8 The Big Picture Weinberger et al 2004

9 Manifold and Hidden Variables

10 Dimensionality Reduction
Dimensionality: The number of measurements available for each item in a data set. The dimensionality of real world items is very high. For example: The dimensionality of a 600 by 600 image is 360,000. The Key to analyzing data is comparing these measurements to find relationships among this plethora of data points. Usually these measurements are highly redundant, and relationships among data points are predictable.

11 Dimensionality Reduction
Knowing the value of a pixel in an image, it is easy to predict the value of nearby pixels since they tend to be similar. Knowing that the word “corporation” occurs often in articles about economics, but not very often in articles about art and poetry then it is easy to predict that it will not occur very often in articles about love. Although there are lots of measurements per item, there are far fewer that are likely to vary. Using a data set that only includes the items likely to vary allows humans to quickly and easily recognize changes in high dimensionality data.

12 Data Representation

13 Data Representation

14 Data Representation 1 0.5

15 Hastie et, al 2001

16 2 by 103 644 by 103 644 by 2 -2.19 -0.02 -3.19 1.02 23 by 28 23 by 28 2 by 1 2 by 1

17

18

19

20

21

22 Hastie et, al 2001

23 Tenenbaum et, al 2000

24 Roweis & Saul 2000

25 Arranging words: Each word was initially represented by a high-dimensional vector that counted the number of times it appeared in different encyclopedia articles. Words with similar contexts are collocated Roweis & Saul 2000

26 Roweis & Hinton 2001


Download ppt "Machine Learning Ali Ghodsi Department of Statistics"

Similar presentations


Ads by Google