Classification of melody by composer using hidden Markov models Greg Eustace MUMT 614: Music Information Acquisition, Preservation, and Retrieval.

Slides:



Advertisements
Similar presentations
1 Gesture recognition Using HMMs and size functions.
Advertisements

Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Hidden Markov Models (HMM) Rabiner’s Paper
Angelo Dalli Department of Intelligent Computing Systems
Automatic Speech Recognition II  Hidden Markov Models  Neural Network.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Tutorial on Hidden Markov Models.
2004/11/161 A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition LAWRENCE R. RABINER, FELLOW, IEEE Presented by: Chi-Chun.
Page 1 Hidden Markov Models for Automatic Speech Recognition Dr. Mike Johnson Marquette University, EECE Dept.
An Introduction to Hidden Markov Models and Gesture Recognition Troy L. McDaniel Research Assistant Center for Cognitive Ubiquitous Computing Arizona State.
Hidden Markov Models Adapted from Dr Catherine Sweeney-Reed’s slides.
Ch 9. Markov Models 고려대학교 자연어처리연구실 한 경 수
Ch-9: Markov Models Prepared by Qaiser Abbas ( )
Hidden Markov Models Theory By Johan Walters (SR 2003)
Hidden Markov Models Fundamentals and applications to bioinformatics.
Hidden Markov Models in NLP
SPEECH RECOGNITION Kunal Shalia and Dima Smirnov.
ECE 8527 Homework Final: Common Evaluations By Andrew Powell.
Albert Gatt Corpora and Statistical Methods Lecture 8.
INTRODUCTION TO Machine Learning 3rd Edition
Part II. Statistical NLP Advanced Artificial Intelligence (Hidden) Markov Models Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme Most.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PatReco: Hidden Markov Models Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 14: Introduction to Hidden Markov Models Martin Russell.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Hidden Markov Models K 1 … 2. Outline Hidden Markov Models – Formalism The Three Basic Problems of HMMs Solutions Applications of HMMs for Automatic Speech.
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
Learning HMM parameters Sushmita Roy BMI/CS 576 Oct 21 st, 2014.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
JSymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada.
Ch10 HMM Model 10.1 Discrete-Time Markov Process 10.2 Hidden Markov Models 10.3 The three Basic Problems for HMMS and the solutions 10.4 Types of HMMS.
Isolated-Word Speech Recognition Using Hidden Markov Models
Graphical models for part of speech tagging
BINF6201/8201 Hidden Markov Models for Sequence Analysis
HMM - Basics.
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
Hidden Markov Models Yves Moreau Katholieke Universiteit Leuven.
Hidden Markov Models Usman Roshan CS 675 Machine Learning.
MUMT611: Music Information Acquisition, Preservation, and Retrieval Presentation on Timbre Similarity Alexandre Savard March 2006.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CS Statistical Machine learning Lecture 24
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Hidden Markov Models (HMMs) –probabilistic models for learning patterns in sequences (e.g. DNA, speech, weather, cards...) (2 nd order model)
1 Hidden Markov Models Hsin-min Wang References: 1.L. R. Rabiner and B. H. Juang, (1993) Fundamentals of Speech Recognition, Chapter.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
1 Hidden Markov Model: Overview and Applications in MIR MUMT 611, March 2005 Paul Kolesnik MUMT 611, March 2005 Paul Kolesnik.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida March 31,
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Other Models for Time Series. The Hidden Markov Model (HMM)
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture
Statistical Models for Automatic Speech Recognition
Computational NeuroEngineering Lab
Statistical Models for Automatic Speech Recognition
Hidden Markov Model LR Rabiner
CONTEXT DEPENDENT CLASSIFICATION
Handwritten Characters Recognition Based on an HMM Model
Music Signal Processing
Presentation transcript:

Classification of melody by composer using hidden Markov models Greg Eustace MUMT 614: Music Information Acquisition, Preservation, and Retrieval

Overview Project description Software and dataset Representations of melody HMM parameters The learning process Training and testing HMMs Summary

Project description The goal of this project is to use Hidden Markov Models (HMM) for the automatic classification of symbolic melodic data by composer. Research questions: Are there significant statistical differences between melodies written by different composers? How do different representations of melody affect the performance of the classifier? How do different types of HMMs affect the performance?

Representations of melody C major scale: 1. Absolute pitch (normalised to the octave). (e.g. 1, 3, 5, 6, 8, 10, 11, 12) 2. Absolute pitch with rhythm (e.g. 1, 1, 3, 5, 6, 8, 10, 10, 11, 11, 12, 12) The note number is given once if equal to a quarter note, twice for a half note, etc. (Chai, and Vercoe 2001). 3. Interval (e.g. 2, 2, 1, 2, 2, 2, 1)

Representations of melody 4. Melodic Contour (e.g. 2, 2, 2, 2, 2, 2, 2) The convention is that intervals of a unison = 1, intervals of second = 2 and all other intervals are Alternative contour representations?

Software and dataset Extraction of melodic data uses ad hoc algorithms developed in MAX/MSP and MATLAB. Classification uses the HMM Toolbox by Kevin Murphy. A large collection of links to MIDI files has been compiled by Cory McKay: The melodic data is the soprano line extracted from a type 1 MIDI file. It is hoped that composers of contrasting style will show the greatest statistical differences.

Hidden Markov models “A Hidden Markov model (HMM) is a doubly imbedded stochastic process with an underlying stochastic process that is not observable (i.e. is hidden) but can only be observed through another set of stochastic processes that produce the sequence of observations” (Rabiner 1989). An HMM is defined by the parameters: M = number of distinct observation symbols (e.g. for absolute pitch these are the numbers 1-13 corresponding to the 12 notes of the octave). N = number of states in the model (these may not have physical significance). A = {a ij } = state transition probability distribution B = {b j (k)} = observation symbol probability distribution π = {π i } = the initial state distribution The set of hidden model parameters are given by λ = (A, B, π).

The learning process The Baum-Welch learning algorithm is used to find the hidden parameters (λ) of the HMM. This process uses maximum likelihood parameter estimation. In general, the likelihood is maximized when a given test sequence corresponds to a specific model. It is also common to attempt to maximize the logarithm of the likelihood.

Training and testing HMMs The training procedure corresponds to one of the three fundamental problems associated with HMMs as defined by Rabiner (1989). That is, for a given observation sequence O = {O1, O2,… On} and a model with parameters λ = (A, B, π) what is the value of λ that maximizes the probability of the observation sequence P(O| λ)? The first fundamental problem is essential to the classification of an unknown sequence. Given an observation sequence O = {O1, O2,… On} and a model λ = (A, B, π) how do we efficiently compute P(O|λ) (Rabiner 1989)?

Training and testing HMMs The process can be summarised as: Train a different model on data from each composer. Once the models have been trained compute the log-likelihood for a test sequence for each model. Classify the training data according to the model which gives the highest value for the log-likelihood. Repeat the process for all representations of melody

Training and testing HMMs Using different types of HMMs may affect classification Fully connected (ergodic) models: every state is connected to every other state. Left-right (Bakis) models: states are connected to themselves and to the adjacent state (proceeding from left to right). These are typically used for modelling time varying signals.

Summary A summary of variables which may affect classification: Composer Size of the dataset Nature of the pieces used MIDI transcription Melodic extraction Melodic representation Type of HMM Number of model states

References Chai, W., and B. Vercoe. Folk music classification using hidden Markov models Proceedings of the international conference on artificial intelligence. Rabiner, L A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE 77: 257–86.