Jump to Navigation

Main menu

  • Login
Home

Secondary menu

  • [Room Booking]
  • [Wiki]
  • [Webmail]

Deep Learning for MIR II: State-of-the-art Algorithms

Workshop Date: 
Mon, 08/15/2022 - Fri, 08/19/2022

 

 

A survey of cutting-edge research in MIR using Deep Learning presented by instructors and a lineup of guest speakers leading research in industry and academia. This workshop is meant for individuals who want to gain experience applying Deep Learning to solve a problem of their interest in MIR. Instructors will explain and demonstrate concepts in models that are used in cutting-edge industry and academic research. Students will build and train state-of-the-art models using tensorflow and GPU computing, adapting them to a MIR-related problem of their interest. Instructors will serve as advisors to students in the course on-demand.

In-person (CCRMA, Stanford) and online enrollment options available during registration (see red button above). Students will receive the same teaching materials and have access to the same tutorials in either format. In-person students will gain access to more in-depth, hands-on 1:1 instructor discussion and feedback when taking the course in-person. 

Theory includes: Generative models. Self-supervised feature learning. Attention mechanisms. 

Models covered includes: DeepSpeech, Transformer, Crepe, GrFNNs

Practice: music and speech recognition/synthesis, beat-tracking, music-recommendation, and semantic analysis. 

Prerequisites: 

- Deep Learning for MIR I Workshop (August 8 - 12, 2022) 


About the instructors:

Camille Noufi is a PhD student and researcher at the Center for Computer Research in Music and Acoustics (CCRMA) at Stanford University. Camille studies machine generation of expressive communication, and acoustic impact of the environment on the voice. Her interdisciplinary research utilizes signal processing (DSP), machine learning (ML) and human-computer-interaction (HCI) in combination with psychology and vocal science. She was a research intern in the Audio Team at Meta Reality Labs in 2020. Before coming to CCRMA, she worked on audio scene analysis and vocal biomarker research at MIT Lincon Laboratory. Her research has been presented at the Interspeech, ISMIR and ICML conferences. camillenoufi.com


Iran R. Roman is a theoretical neuroscientist and machine listening scientist at New York University’s Music and Audio Research Laboratory. Iran is a passionate instructor, with extensive experience teaching artificial intelligence and deep learning. His industry experience includes deep learning engineering internships at Plantronics in 2017, Apple in 2018 and 2019, Oscilloscape in 2020, and Tesla in 2021. Iran’s research has focused on using deep learning for speech recognition and auditory scene analysis. iranroman.github.io


IMPORTANT: Contact the instructor before registering to confirm your eligibility. Attach a copy of your registration or diploma for the CCRMA Deep Learning for MIR I workshop.

scholarship opportunity: https://docs.google.com/forms/d/e/1FAIpQLSdL4LWoX5EpYUEp0UMFUhhmgMWOHkd8...
  • Home
  • News and Events
    • All Events
      • CCRMA Concerts
      • Colloquium Series
      • DSP Seminars
      • Hearing Seminars
      • Guest Lectures
    • Event Calendar
    • Events Mailing List
    • Recent News
  • Academics
    • Courses
    • Current Year Course Schedule
    • Undergraduate
    • Masters
    • PhD Program
    • Visiting Scholar
    • Visiting Student Researcher
    • Workshops 2022
  • Research
    • Publications
      • Authors
      • Keywords
      • STAN-M
      • Max Mathews Portrait
    • Research Groups
    • Software
  • People
    • Faculty and Staff
    • Students
    • Alumni
    • All Users
  • User Guides
    • New Documentation
    • Booking Events
    • Common Areas
    • Rooms
    • System
  • Resources
    • Planet CCRMA
    • MARL
  • Blogs
  • Opportunities
    • CFPs
  • About
    • The Knoll
      • Renovation
    • Directions
    • Contact

Search this site:

Winter Quarter 2023

101 Introduction to Creating Electronic Sound
158/258D Musical Acoustics
220B Compositional Algorithms, Psychoacoustics, and Computational Music
222 Sound in Space
250C Interaction - Intermedia - Immersion
251 Psychophysics and Music Cognition
253 Symbolic Musical Information
264 Musical Engagement
285 Intermedia Lab
319 Research Seminar on Computational Models of Sound
320B Introduction to Audio Signal Processing Part II: Digital Filters
356 Music and AI
422 Perceptual Audio Coding
451B Neuroscience of Auditory Perception and Music Cognition II: Neural Oscillations

 

 

 

   

CCRMA
Department of Music
Stanford University
Stanford, CA 94305-8180 USA
tel: (650) 723-4971
fax: (650) 723-8468
info@ccrma.stanford.edu

 
Web Issues: webteam@ccrma

site copyright © 2009 
Stanford University

site design: 
Linnea A. Williams