Overview of the final test for CSC 2515. Overview PART A: 7 easy questions –You should answer 5 of them. If you answer more we will select 5 at random.

Slides:



Advertisements
Similar presentations
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
Advertisements

An Overview of Machine Learning
Supervised Learning Recap
Linear Models for Classification: Probabilistic Methods
Computer vision: models, learning and inference
Artificial Intelligence Lecture 2 Dr. Bo Yuan, Professor Department of Computer Science and Engineering Shanghai Jiaotong University
Visual Recognition Tutorial
COMP 328: Final Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Classification and risk prediction
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Clustering.
Visual Recognition Tutorial
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Sparse vs. Ensemble Approaches to Supervised Learning
Bayes Classifier, Linear Regression 10701/15781 Recitation January 29, 2008 Parts of the slides are from previous years’ recitation and lecture notes,
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Crash Course on Machine Learning
Incomplete Graphical Models Nan Hu. Outline Motivation K-means clustering Coordinate Descending algorithm Density estimation EM on unconditional mixture.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
EM and expected complete log-likelihood Mixture of Experts
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
University of Southern California Department Computer Science Bayesian Logistic Regression Model (Final Report) Graduate Student Teawon Han Professor Schweighofer,
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Learning Theory Reza Shadmehr logistic regression, iterative re-weighted least squares.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
CS Statistical Machine learning Lecture 10 Yuan (Alan) Qi Purdue CS Sept
1 Generative and Discriminative Models Jie Tang Department of Computer Science & Technology Tsinghua University 2012.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
Lecture 17 Gaussian Mixture Models and Expectation Maximization
Classification Derek Hoiem CS 598, Spring 2009 Jan 27, 2009.
Sparse Kernel Methods 1 Sparse Kernel Methods for Classification and Regression October 17, 2007 Kyungchul Park SKKU.
Lecture 2: Statistical learning primer for biologists
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall DM Finals Study Guide Rodney Nielsen.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Perceptual and Sensory Augmented Computing Machine Learning, WS 13/14 Machine Learning – Lecture 5 Linear Discriminant Functions Bastian Leibe.
Hierarchical Mixture of Experts Presented by Qi An Machine learning reading group Duke University 07/15/2005.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Page 1 CS 546 Machine Learning in NLP Review 2: Loss minimization, SVM and Logistic Regression Dan Roth Department of Computer Science University of Illinois.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Usman Roshan Dept. of Computer Science NJIT
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
A Brief Introduction to Bayesian networks
Learning Deep Generative Models by Ruslan Salakhutdinov
Deep Feedforward Networks
Machine Learning & Deep Learning
Learning Coordination Classifiers
The Elements of Statistical Learning
Data Mining Lecture 11.
Probabilistic Models for Linear Regression
Neural Networks and Backpropagation
دانشگاه صنعتی امیرکبیر Instructor : Saeed Shiry
Probabilistic Models with Latent Variables
Mathematical Foundations of BME Reza Shadmehr
Overview of the final test for CSC 2515
Linear Discrimination
Usman Roshan Dept. of Computer Science NJIT
Presentation transcript:

Overview of the final test for CSC 2515

Overview PART A: 7 easy questions –You should answer 5 of them. If you answer more we will select 5 at random. So cross out any extra answers. –Worth 8 points each. –Each question should take 3-5 minutes. Typical easy question: –a) Write down the softmax function –b) What is the purpose of the denominator?

Overview (cont.) PART B: 4 hard questions. –You must only answer 3 of them. If you answer more we will select 3 of your answers at random. So cross out any extra answers. –Worth 20 points each. –They should take ~10 minutes each. So you’ll have one hour to do the exam. Typical hard question –If the input vectors from two classes are drawn from a Gaussian distribution, and the classes share a covariance matrix, then what is the form of the posterior distribution, and why?

Summary What to study? –Material covered in lectures and tutorial –Use the textbook as back-up, to help understand the methods and derivations The exam is closed book and notes –Do not focus on memorizing formulas, but instead main ideas and methods

Topics to Study Week 2 –Lecture : Linear Regression When is minimizing the squared error equivalent to Maximum Likelihood Learning? Online vs. Batch learning Regularized least squares – alternative regularizers Bayesian approach – integrating over posterior over parameters Week 3 –Lecture: Classification Compare methods of performing multi-class discrimination Probabilistic interpretation of LR; loss function Regularizer: form, rationale Naïve Bayes: key assumption, cost function Week 4 –Lecture : Neural Networks Backprop: when applicable (constraints on cost, activation functions)? Weight constraints: how and why What is a trigram, and why is it inadequate as a model of sentences? Methods to prevent over-fitting

Topics to Study (cont.) Week 5 –Lecture: Deep Belief Networks only responsible for basic RBM slides main idea – RBMs: energies and probabilities; learning rule Week 6 –Lecture : Clustering & EM Hard vs. soft k-means EM: definition of steps in algorithm Free energy – relationship to EM, convergence proof Mixture models Week 7 –Lecture: Continuous Latent Variable Models PCA: motivation, basic algorithm Probabilistic vs. standard PCA, and compare to full Gaussian PPCA, FA, ICA: Generative models, compare/contrast

Topics to Study (cont.) Week 8 –Lecture : Combining Models When is ensemble a win? Bagging vs. Boosting Boosting: main steps in algorithm; 2 different types of weights, rationale Decision trees: algorithm; choosing nodes via MI; decision boundaries, stumps Mixture of experts Week 9 – Lecture: Support Vector Machines Understand max margin, kernel trick Compare to logistic regression Week 10 –Lecture: Bayesian Reasoning Not responsible for details of Bayesian regression main idea - style of approach vs. earlier methods Should understand conjugacy Week 11 Directed vs. undirected graphical models