Automatic music classification and the importance of instrument identification Cory McKay and Ichiro Fujinaga Music Technology Area Faculty of Music McGill.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Automatic Music Classification Cory McKay. 2/47 Introduction Many areas of research in music information retrieval (MIR) involve using computers to classify.
KARAOKE FORMATION Pratik Bhanawat (10bec113) Gunjan Gupta Gunjan Gupta (10bec112)
Content-based retrieval of audio Francois Thibault MUMT 614B McGill University.
Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification.
Overview What : Stroke type Transformation: Timbre Rhythm When: Stroke timing Resynthesis.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Introduction to Automatic Classification Shih-Wen (George) Ke 7 th Dec 2005.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Improving Musical Genre Classification with RBF Networks Douglas Turnbull Department of Computer Science and Engineering University of California, San.
Postgraduate Department of Electrical Engineering PPGEE UFPR - Federal University of Paraná Luis Gustavo Weigert Machado
Semantic Web Technologies Lecture # 2 Faculty of Computer Science, IBA.
JSymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada.
CHAMELEON : A Hierarchical Clustering Algorithm Using Dynamic Modeling
Water Contamination Detection – Methodology and Empirical Results IPN-ISRAEL WATER WEEK (I 2 W 2 ) Eyal Brill Holon institute of Technology, Faculty of.
Polyphonic Queries A Review of Recent Research by Cory Mckay.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Theoretic and artistic research studying opportunities of symbiosis of Western and non-Western musical idiom Olmo Cornelis.
Educational Software using Audio to Score Alignment Antoine Gomas supervised by Dr. Tim Collins & Pr. Corinne Mailhes 7 th of September, 2007.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
by B. Zadrozny and C. Elkan
Usability Fujinaga Uitdenbogerd, A., and R. Schyndel A review of factors affecting music recommender success. International Symposium on.
5 March 2005Cory McKay 1 Approaches to Overcoming Problems in Interactive Musical Performance Systems Cory McKay Music Technology Area Faculty of Music.
JSymbolic Cedar Wingate MUMT 621 Professor Ichiro Fujinaga 22 October 2009.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
ACE: A Framework for optimizing music classification Cory McKay Rebecca Fiebrink Daniel McEnnis Beinan Li Ichiro Fujinaga Music Technology Area Faculty.
Music Is The Art Of Expressing Yourself Through Sound.
11 Applications of Machine Learning to Music Research: Empirical Investigations into the Phenomenon of Musical Expression 이 인 복.
MUMT611: Music Information Acquisition, Preservation, and Retrieval Presentation on Timbre Similarity Alexandre Savard March 2006.
TEMPLATE DESIGN © Zhiyao Duan 1,2, Lie Lu 1, and Changshui Zhang 2 1. Microsoft Research Asia (MSRA), Beijing, China.2.
Modeling Expressive Performances of the Singing Voice Maria-Cristina Marinescu (Universidad Carlos III de Madrid) Rafael Ramirez (Universitat Pompeu Fabra)
Overview of Part I, CMSC5707 Advanced Topics in Artificial Intelligence KH Wong (6 weeks) Audio signal processing – Signals in time & frequency domains.
Polyphonic Transcription Bruno Angeles McGill University - Schulich School of Music MUMT-621 Fall /14.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
Prostate Cancer CAD Michael Feldman, MD, PhD Assistant Professor Pathology University Pennsylvania.
Audio Tempo Extraction Presenter: Simon de Leon Date: February 9, 2006 Course: MUMT611.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Ivica Dimitrovski 1, Dragi Kocev 2, Suzana Loskovska 1, Sašo Džeroski 2 1 Faculty of Electrical Engineering and Information Technologies, Department of.
1 CHUKWUEMEKA DURUAMAKU.  Machine learning, a branch of artificial intelligence, concerns the construction and study of systems that can learn from data.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Realtime Recognition of Orchestral Instruments Ichiro Fujinaga McGill University.
Issues in Automatic Musical Genre Classification Cory McKay.
 6 th Musical Literacy 1.1 All students will be able to use a steady tone when performing.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Improving Music Genre Classification Using Collaborative Tagging Data Ling Chen, Phillip Wright *, Wolfgang Nejdl Leibniz University Hannover * Georgia.
Classification of melody by composer using hidden Markov models Greg Eustace MUMT 614: Music Information Acquisition, Preservation, and Retrieval.
1 / 22 jSymbolic Jordan Smith – MUMT 611 – 6 March 2008.
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
BASS TRACK SELECTION IN MIDI FILES AND MULTIMODAL IMPLICATIONS TO MELODY gPRAI Pattern Recognition and Artificial Intelligence Group Computer Music Laboratory.
Metamidi: a tool for automatic metadata extraction from MIDI files Tomás Pérez-García, Jose M. Iñesta, and David Rizo Computer Music Laboratory University.
Research Methodology Proposal Prepared by: Norhasmizawati Ibrahim (813750)
Ontology Engineering and Feature Construction for Predicting Friendship Links in the Live Journal Social Network Author:Vikas Bahirwani 、 Doina Caragea.
A shallow description framework for musical style recognition Pedro J. Ponce de León, Carlos Pérez-Sancho and José Manuel Iñesta Departamento de Lenguajes.
Automatic Classification of Audio Data by Carlos H. L. Costa, Jaime D. Valle, Ro L. Koerich IEEE International Conference on Systems, Man, and Cybernetics.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
Usability Fujinaga 2003.
Preface to the special issue on context-aware recommender systems
Erasmus University Rotterdam
Can Computer Algorithms Guess Your Age and Gender?
Automatic Classification of Music Genre
Presented by Steven Lewis
Realtime Recognition of Orchestral Instruments
Realtime Recognition of Orchestral Instruments
Understanding Standards An overview of course assessment
Presentation transcript:

Automatic music classification and the importance of instrument identification Cory McKay and Ichiro Fujinaga Music Technology Area Faculty of Music McGill University Montreal, Canada

10 March 2005 CIM Montreal / McKay & Fujinaga 2/25 Overview Examination of the relative importance of different high-level features in automatic music classification Performed an experiment involving automatic genre classification of MIDI files Found that features based on instrumentation (an abstraction of timbre) were of particular importance

10 March 2005 CIM Montreal / McKay & Fujinaga 3/25 Topics Introduction to automatic music classification Related research Details of experiment performed Features used Feature weighting Taxonomies used Classifiers and training data used ResultsConclusions

10 March 2005 CIM Montreal / McKay & Fujinaga 4/25 Introduction to automatic music classification There are many ways in which computers can classify music GenreComposerPerformer Geographical/temporal/cultural origin etc. Music classification can be difficult for both humans and computers Rarely have precise, clear and consistent guidelines delineating the musical characteristics of categories

10 March 2005 CIM Montreal / McKay & Fujinaga 5/25 Applications of automatic music classification Discovery of probable authorship of anonymous compositions Sociological and psychological research into how humans construct the notion of musical similarity and form musical groupings Automatic sorting of large databases Music recommendation systems Sorting of personal music collections e.g. based on mood or listening scenarios Automated transcription Detection of pirated recordings

10 March 2005 CIM Montreal / McKay & Fujinaga 6/25 Advantages of automatic music classification Computers can perform classifications faster and more consistently than humans Computers can analyze music in novel and non-intuitive ways that might not occur to humans Computers can avoid human preconceptions that might contaminate experimental results

10 March 2005 CIM Montreal / McKay & Fujinaga 7/25 How automatic classification works “Feature” extraction Properties or characteristics of recordings Percepts that classifiers base decisions on Can be extracted from audio (e.g. MP3) or symbolic (e.g. MIDI) recordings Good features are essential to successful classification Classification can be done using Expert systems: utilize pre-set heuristics Machine learning (AI): supervised or unsupervised learning

10 March 2005 CIM Montreal / McKay & Fujinaga 8/25 Features Low-level features Signal processing quantities e.g. spectral centroid and spectral flux Can be effective practically Can have psychoacoustic significance Have little direct theoretical meaning musicologically or sociologically High-level features Based on musical abstractions e.g. tempo and meter Currently difficult or impossible to extract from audio recordings Have more theoretical relevance than low-level features

10 March 2005 CIM Montreal / McKay & Fujinaga 9/25 Overview of this experiment Empirical examination of which features are most useful to classifiers Used high-level features because of their theoretical significance Used test task of genre classification A particularly difficult type of classification Related to many other types of classification Features useful for this task likely to be particularly robust

10 March 2005 CIM Montreal / McKay & Fujinaga 10/25 Related research Relatively little work has been done on features that could be useful for arbitrary types of music Cantometrics project (Lomax 1968) Tagg (1982) Cope (1991) Arden and Huron (2001) Studied the correlation between musical features and geographical regions Automatic genre classification has received considerable attention recently Audio classification work of Tzanetakis and Cook (2002) is often cited Best results to date with symbolic data have been achieved by McKay and Fujinaga (2004)

10 March 2005 CIM Montreal / McKay & Fujinaga 11/25 Bodhidharma Experiments carried out with the Bodhidharma system A general-purpose symbolic feature extraction and classification system Easy-to-usePortable Applicable to a wide range of research tasks

10 March 2005 CIM Montreal / McKay & Fujinaga 12/25 Features studied 111 high-level features implemented: Instrumentation e.g. whether modern instruments are present Musical Texture e.g. standard deviation of the average melodic leap of different lines Rhythm e.g. standard deviation of note durations Dynamics e.g. average note to note change in loudness Pitch Statistics e.g. fraction of notes in the bass register Melody e.g. fraction of melodic intervals comprising a tritone Largest available set of implemented high-level features 42 more features have been proposed, but have not been implemented yet More information available in Cory McKay’s master’s thesis (2004)

10 March 2005 CIM Montreal / McKay & Fujinaga 13/25 Features to use An insufficient number of features can fail to provide classifiers with enough information to make good decisions Too many features can overwhelm and confuse classifiers Can be difficult to predict in advance which features will work well together Individual performance of a feature is not necessarily indicative of its performance in combination with other features

10 March 2005 CIM Montreal / McKay & Fujinaga 14/25 Feature weighting Feature weighting is a technique for experimentally determining the importance of various features by assigning weights to them Used genetic algorithms here “Evolves” a good set of weights The weights produced by the genetic algorithm provides an indication of the importance of particular features in particular contexts

10 March 2005 CIM Montreal / McKay & Fujinaga 15/25 Types of classification performed The choice of “best” features is context-dependant e.g. best features for distinguishing between Baroque and Romantic different than when comparing Punk and Heavy Metal Performed three types of classification: FlatHierarchicalRound-robin Hierarchical and round-robin feature weighting allowed classifiers to use specialized weightings in order to improve performance

10 March 2005 CIM Montreal / McKay & Fujinaga 16/25 Taxonomies used Used hierarchical taxonomies A recording could belong to more than one category A category could be a child of multiple parents in the taxonomical hierarchy Performed experiments with two taxonomies: Large (38 leaf categories): Used to test system under realistic conditions Small (9 leaf categories): Used to loosely compare system to existing sytems

10 March 2005 CIM Montreal / McKay & Fujinaga 17/25 Large taxonomy

10 March 2005 CIM Montreal / McKay & Fujinaga 18/25 Small taxonomy JazzBebop Jazz Soul SwingPopularRapPunkCountry Western Classical Baroque Modern Classical Romantic

10 March 2005 CIM Montreal / McKay & Fujinaga 19/25 Training and testing Used ensembles of k-nearest neighbour and neural network classifiers 950 MIDI files Hand-classified for training based on a variety of on-line databases 5 fold cross-validation 80% training, 20% testing

10 March 2005 CIM Montreal / McKay & Fujinaga 20/25 Average success rates 9 Category Taxonomy Leaf: 90% Root: 98% 38 Category Taxonomy Leaf: 57% Root: 81%

10 March 2005 CIM Montreal / McKay & Fujinaga 21/25 Success rates achieved in previous research Audio results: Many systems have been implemented Generally only used 10 categories or less Success rates generally below 80% for more than 5 categories Symbolic results: 84% for 2-way classifications (Shan & Kuo 2003) 89% for 2-way classifications (Ponce de Leon & Inesta 2004) 63% for 3-way classifications (Chai & Vercoe 2001) 60-70% for 6-way classifications (Basili, Serafini & Stellato 2004)

10 March 2005 CIM Montreal / McKay & Fujinaga 22/25 Feature performance Feature Group Number of Features Weighting Scaled by Number of Features (%) Instrumentation Pitch Rhythm Melody Texture141.7 Dynamics41.6 Features based on instrumentation were assigned 46.1% of all weightings (after scaling) At least one instrumentation feature played a major role in almost every classifier Two of the top three features were based on instrumentation

10 March 2005 CIM Montreal / McKay & Fujinaga 23/25 Importance of instrumentation Features based on instrumentation clearly dominant A high-level abstraction of timbre Implies that audio classification systems could benefit from instrument identification modules Caveat: These results present the overall averages of weightings Other features played a dominant role in certain stages of classification The best results were achieved by including a wide variety of features and applying feature weighting

10 March 2005 CIM Montreal / McKay & Fujinaga 24/25 Conclusions Features based on instrumentation can play an essential role in automatic music classification, and should be used if possible High-level features can produce good results, and should not be neglected in favour of low- level features Bodhidharma’s large feature library combined with feature weighting is an effective approach Very good genre classification success rates can be achieved with small taxonomies, and we are at least approaching a point where large taxonomies can be dealt with effectively

10 March 2005 CIM Montreal / McKay & Fujinaga 25/25