MIR workshop 2008
- 1 CCRMA Workshop: Music Information Retrieval
- 1.1 logistics
- 1.2 Abstract
- 1.3 Workshop syllabus
- 1.4 software, libraries, examples
- 1.5 Lectures
- 1.6 Labs
- 1.7 Final Projects from MIR 2008 Workshop
- 1.8 References for additional info
- 1.9 Audio Source Material
- 1.10 THE WIKI VALUE ADD - Notes can be added here by anyone
- 1.11 MATLAB Utility Scripts
- 1.12 ChucK
CCRMA Workshop: Music Information Retrieval
NOTE: THIS PAGE IS A WORK IN PROGRESS -- IT IS BY NO MEANS, COMPLETE YET. Thanks!
Workshop Title: "Intelligent Audio Systems: Foundations and Applications of Music Information Retrieval"
This workshop will teach the underlying ideas, approaches, technologies, and practical design of intelligent audio systems using Music Information Retrieval (MIR) algorithms.
MIR is a highly-interdisciplinary field bridging the domains of digital audio signal processing, pattern recognition, software system design, and machine learning. Simply put, MIR algorithms allow a computer to “listen” and “understand or make sense of” audio data, such as MP3s in a personal music collection, live streaming audio, or gigabytes of sound effects, in an effort to reduce the semantic gap between high-level musical information and low-level audio data. In the same way that listeners can recognize the characteristics of sound and music – tempo, key, chord progressions, genre, or song structure – MIR algorithms are capable of recognizing and extracting this information, enabling systems to perform extensive sorting, searching, music recommendation, metadata generation, transcription, and even aiding/generating real-time performance.
This workshop will target students, researchers, and industry audio engineers who are unfamiliar with the field of Music Information Retrieval (MIR). We will demonstrate the myriad of exciting technologies enabled by the fusion of basic signal processing techniques with machine learning and pattern recognition. Lectures will cover topics such as low-level feature extraction, generation of higher-level features such as onset timings and chord estimations, audio similarity clustering, search, and retrieval techniques, and design and evaluation of machine classification systems. The presentations will be applied, multimedia-rich, overview of the building blocks of modern MIR systems. Our goal is to make the understanding and application of highly-interdisciplinary technologies and complex algorithms approachable.
The workshop will consist of half-day lectures, half-day supervised lab sessions, classroom exercises, demonstrations, and discussions.
Labs will allow students to design basic ground-up "intelligent audio systems", leveraging existing MIR toolboxes, programming environments, and applications. Labs will include creation and evaluation of basic instrument recognition, transcription, and real-time audio analysis systems.
Knowledge of basic digital audio principles and familiarity with basic programming (Matlab, C/C++, and/or ChucK) will be useful. Students are highly encouraged to bring their own audio source material for course labs and demonstrations.
- CCRMA Overview
- Final Projects
- Introduction to Capabilities and Applications of MIR
- Why MIR?
- Overview of potential research and commercial applications
- Basic System Overview and Architecture
- Survey of the field, real-world applications, MIR research, and challenges
- Current commercial applications
- Searching, Similarity, and Seed Query Systems
- Recommendation and Playlisting systems
- How stuff works: Select Commercial MIR projects
- Academic MIR research projects
- Music Transcription
- Real-time machine listening and audio analysis
Timing and Segmentation
- Frames and Windows
- Onset Detection
- Beat & Tempo Extraction
- Low Level Features
- Zero Crossing
- Temporal centroid, Log Attack time, Attack slope)
- Spectral features (Centroid, Flux, RMS, Rolloff, Flatness, Kurtosis, Brightness)
- Spectral bands
- Log spectrogram
- Chroma bins
- Higher-level features
- Key Estimation
- Chord Estimation
- Genre (genre, artist ID, similarity)
Analysis / Decision Making
- Heuristic Analysis
- Distance measures (Euclidean, Manhattan, etc.)
- SVM / One-class SVM
- Clustering and probability density models
- Density distance measures (centroid distance, EMD, KL-divergence, etc)
- Nested classifier / Anchor-space / template-based systems
Model / Data Preparation Techniques
- Data Preparation
- PCA / LDA
- Scaling data
- Model organization
- concept and design
- Data set construction and organization
- Walk through example: audio similarity retrieval
- Walk through example: instrument/speaker/source identification
- Feature selection
- Cross Validation
- Information Retrieval metrics (precision, recall, F-Measure)
Plus guest lectures/visits from academic experts and real-world folks.
software, libraries, examples
Applications & Environments
- ChucK / UAna
- Weka Machine Learning and Data Mining Toolbox (Standalone app / Java)
- Sonic Visualizer
Machine Learning Libraries & Toolboxes
- Netlab Pattern Recognition and Clustering Toolbox (Matlab)
- libsvm SVM toolbox (Matlab)
- MIR Toolboxes (Matlab)
- MA Toolbox
- MIDI Toolbox
- [see also below references]
- Genetic Algorithm: http://www.ise.ncsu.edu/mirage/GAToolBox/gaot/
- Spider http://www.kyb.tuebingen.mpg.de/bs/people/spider/
- HTK http://htk.eng.cam.ac.uk/
Abstract: This lab will introduce you to the practice of analyzing, segmenting, feature extracting, and applying basic classifications to audio files. Our future labs will build upon this essential work -but will use more sophisticated training sets, features, and classifiers.
Abstract: My first audio classifier: introducing K-NN! We can now appreciate why we need additional intelligence in our systems - heuristics can't very far in the world of complex audio signals. We'll be using Netlab's implementation of the k-NN for our work here. It proves be a straight-forward and easy to use implementation. The steps and skills of working with one classifier will scale nicely to working with other, more complex classifiers. We're also going to be using the new features in our arsenal: cherishing those "spectral moments" (centroid, bandwidth, skewness, kurtosis) and also examining other spectral statistics.
Abstract: Sometimes, an unsupervised learning technique is preferred. Perhaps you do not have access to adequate training data, the classifications for the training data's labels events are not completely clear, or you just want to quickly sort real-world, unseen, data into groups based on it's feature similarity. Regardless of your situation, clustering is a great option! Lab also introduces MFCCs as a main measure of timbral similarity.
Abstract: By the end of this lab, you will understand the how to use GMM models - a probabilistic clustering and "soft classification" technique.
Abstract: By the end of this lab, you should have a skeletal outline of where your project is going to go, what data you will use, what features, classifiers / techniques you will use, and what your metric of success / goal is. Additionally, the lab includes a walk through of cross-validation techniques for measuring your accuracy and a list of helpful HMM functions.
Final Projects from MIR 2008 Workshop
Beat and segment detection comparisons using Harmonic Change Detection Function, MFCC, key strength, spectrum, and novelty. (Cynthia Maxwell) Link slides.
Vocal onset detection, segmentation, analysis, and classification. (Chuck Cooper)
Using audio similarity metrics to remix songs from similar sounding audio snippets. (Mike Williams) Mike to post his temporal centroid code.
Measuring audio similarity between recordings of Beethoven and Haydn quartets, using shifted chromagram and delta chromagram similarity matrices. (Jason Sundram)
Real-time decomposition and reassembly of audio via similarity features and k-nearest neighbor. (Joseph Rosenzweig)
Classification of orchestral instrument classes using timbral features and an array of SVMs. (Alberto Pinto)
References for additional info
- Data Mining: Practical Machine Learning Tools and Techniques, Second Edition by Ian H. Witten , Eibe Frank (includes software)
- Netlab by Ian T. Nabney (includes software)
- Signal Processing Methods for Music Transcription, Klapuri, A. and Davy, M. (Editors)
- Computational Auditory Scene Analysis: Principles, Algorithms, and Applications, DeLiang Wang (Editor), Guy J. Brown (Editor)
Prerequisite / background material:
- Check out the references listed at the end of the Klapuri & Davy book
- Check out Papers listed on Pg 136-7 of MIR Toolbox: http://www.jyu.fi/hum/laitokset/musiikki/en/research/coe/materials/mirtoolbox/userguide1.1
Other books (not necessary reviewed by the instructors yet):
- Neural Networks for Pattern Recognition, Christopher M. Bishop, Oxford University Press, 1995.
- Pattern Classification, 2nd edition, R Duda, P Hart and D Stork, Wiley Interscience, 2001.
- "Artificial Intelligence: A Modern Approach" Second Edition, Russell R & Norvig P, Prentice Hall, 2003.
- Machine Learning, Tom Mitchell, McGraw Hill, 1997.
Audio Source Material
OLPC Sound Sample Archive (8.5 GB) 
RWC Music Database (n DVDs) [available in Stanford Music library]
THE WIKI VALUE ADD - Notes can be added here by anyone
More information to be added re: this section You will be able to add your thoughts, explanation, lecture notes, code demos, recommended papers here - for each topic....
MATLAB Utility Scripts
ChucK is a strongly-timed audio programming language that we will be using for real-time audio analysis.