Jump to Navigation

Main menu

  • Login
Home

Secondary menu

  • [Room Booking]
  • [Wiki]
  • [Webmail]

MUS423 Research Seminars

The CCRMA Music 423 Research Seminar brings graduate students and supervising faculty together for planning and discussion of original research. Students and faculty meet either in small groups or individually, as appropriate for the research topics and interests of the participants. Research carried out is typically presented at the weekly CCRMA Colloquium (if it is of general interest to the CCRMA community) or at a Special DSP Seminar scheduled for that purpose.  In either case, announcements appear on the CCRMA Home Page as Upcoming Events.

‹ Colloquium Series up Hearing Seminars ›
  • Printer-friendly version

Recent DSP Seminars

  • Concepts and Control: Understanding Creativity in Deep Music Generation

    Date: 
    Fri, 11/15/2024 - 2:30pm - 3:30pm
    Location: 
    CCRMA Classroom [Knoll 217] (ZOOM Link Below)
    Event Type: 
    DSP Seminar

    Abstract: Recently, generative AI has achieved impressive results in music generation. Yet, the challenge remains: how can these models be meaningfully applied in real-world music creation, for both professional and amateur musicians? We argue that what’s missing is an interpretable generative architecture—one that captures music concepts and their relations, which can be so finely nuanced that they defy straightforward description. In this talk, I will explore various approaches to creating such an architecture, demonstrating how it enhances control and interaction in music generation.

    FREE
    Open to the Public
  • Jin Woo Lee on "Differentiable Physical Modeling for Sound Synthesis: From Design to Inverse Problems"

    Date: 
    Fri, 10/04/2024 - 3:30pm - 5:00pm
    Location: 
    CCRMA Classroom [Knoll 217] (ZOOM Link Below)
    Event Type: 
    DSP Seminar
    Jin Woo Lee is a PhD Candidate at Seoul National University advised by Prof. Kyogu Lee (CCRMA PhD 2008). His research interests are focused on (1) physical modeling for musical instrument sound synthesis and (2) differentiable rendering for immersive and efficient sound simulation. His recent works broadly cover topics in musical sound synthesis, spatial audio rendering, loudspeaker control, and speech quality analysis. He has interned at Meta Reality Labs Research and Supertone. Prior to his PhD, Jin conducted research in computational fluid dynamics during his undergraduate years in Mechanical Engineering from POSTECH. For more information, please visit personal website (http://jnwoo.com/).


    Abstract:
    Open to the Public
  • AI-based Digital Synthesizer Preset Programming: Parameter Estimation for Sound Matching

    Date: 
    Fri, 05/31/2024 - 3:30pm - 5:00pm
    Location: 
    CCRMA Classroom [Knoll 217] (ZOOM Link Below)
    Event Type: 
    DSP Seminar

    Presenter: Soohyun Kim

    FREE
    Open to the Public
  • Generative AI for Music and Audio

    Date: 
    Fri, 11/10/2023 - 3:30pm - 5:00pm
    Location: 
    CCRMA Classroon [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: Generative AI has been transforming the way we interact with technology and consume content. In this talk, I will briefly introduce the three main directions of my research centered around generative AI for music and audio: 1) multitrack music generation, 2) assistive music creation tools, and 3) multimodal learning for audio and music. I will then zoom into my recent work on learning text-queried sound separation and text-to-audio synthesis from videos using pretrained language-vision models. Finally, I will close this talk by discussing the challenges and future directions of generative AI for music and audio.
    FREE
    Open to the Public
  • Adaptive and interactive machine listening with minimal supervision

    Date: 
    Fri, 02/10/2023 - 4:30pm - 5:20pm
    Location: 
    CCRMA Classroom [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: Nowadays deep learning-based approaches have become popular tools and achieved promising results in machine listening. However, a deep model that generalizes well needs to be trained on a large amount of labeled data. Rare, fine-grained, or newly emerged classes (e.g. a rare musical instrument or a new sound effect) where large-scale data collection is hard or simply impossible are often considered out-of-vocabulary and unsupported by machine listening systems. In this thesis work, we aim to provide new perspectives and approaches to machine listening tasks with limited labeled data. Specifically, we focus on algorithms that are designed to work with few labeled data (e.g. few-shot learning) and incorporate human input to guide the machine.
    FREE
    Open to the Public
  • Meta-AF: Meta-Learning for Adaptive Filters

    Date: 
    Fri, 11/18/2022 - 3:30pm - 4:20pm
    Location: 
    CCRMA Classroom [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: Adaptive filtering algorithms are pervasive throughout modern society and have had a significant impact on a wide variety of domains including audio processing, telecommunications, biomedical sensing, astrophysics and cosmology, seismology, and many more. Adaptive filters typically operate via specialized online, iterative optimization methods such as least-mean squares or recursive least squares and aim to process signals in unknown or nonstationary environments. Such algorithms, however, can be slow and laborious to develop, require domain expertise to create, and necessitate mathematical insight for improvement.
    FREE
    Open to the Public
  • Feedback Delay Networks for Artificial Reverberation

    Date: 
    Fri, 11/11/2022 - 12:00pm - 12:50pm
    Location: 
    Zoom
    Event Type: 
    DSP Seminar
    Abstract: Feedback delay networks (FDNs) are recursive filters widely used for artificial reverberation and decorrelation. While vast literature exists on a wide variety of reverb topologies, FDNs provide a unifying framework to design and analyze delay-based reverberators. This talk reviews recent advancements in the FDN theory, such as losslessness, modal and echo representations, and MIMO allpass properties and decorrelation. Many extensions to the FDN were proposed, including time-varying matrices, scattering matrices, high-order attenuation filters, directional reverberation, and coupled room reverberators.

    Presentation Recording
    FREE
    Open to the Public
  • DeepAFx-ST: Style Transfer of Audio Effects with Differentiable Signal Processing

    Date: 
    Fri, 11/04/2022 - 3:30pm - 4:20pm
    Location: 
    CCRMA Classroom [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: We present a framework that can impose the audio effects and production style from one recording to another by example with the goal of simplifying the audio production process. We train a deep neural network to analyze an input recording and a style reference recording and predict the control parameters of audio effects used to render the output. In contrast to past work, we integrate audio effects as differentiable operators in our framework, perform backpropagation through audio effects, and optimize end-to-end using an audio-domain loss. We use a self-supervised training strategy enabling automatic control of audio effects without the use of any labeled or paired training data.
    FREE
    Open to the Public
  • Tanguy Risset -- Compiling Audio DSP for FPGAs Using the Faust Programming Language and High Level Synthesis

    Date: 
    Fri, 10/28/2022 - 3:30pm - 4:20pm
    Location: 
    CCRMA Classroom [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: In this talk, we give a detailed presentation of Syfala (https://github.com/inria-emeraude/syfala), a new "audio DSP to FPGA" compiler based on the Faust programming language (https://faust.grame.fr/ ) and Xilinx/AMD's High level Synthesis (HLS) technology. Our open-source system compiles automatically audio DSP programs to FPGA hardware up to actual sound production (Zynq-based platforms). With this compiler, much smaller audio latency (i.e., one sample at a high sampling rate) can be achieved than with regular "software-based" digital audio systems. This presentation also introduces FPGA architecture in general as well as recent HLS technologies.
    FREE
    Open to the Public
  • Audio Understanding and Room Acoustics in the Era of AI

    Date: 
    Fri, 10/14/2022 - 3:30pm - 4:20pm
    Location: 
    CCRMA Classroom [Knoll 217]
    Event Type: 
    DSP Seminar
    Abstract: This talk will aim to bridge the gap between signal processing and the latest machine learning research by discussing several applications in music and audio. In the first part of the talk, we will discuss how classic signal processing properties can be used to spoon-feed powerful neural architectures such as Transformers to tackle a difficult signal processing task: To do re-reverberation(system identification) at scale. This work now enables hearing music in any concert hall/virtual environment for any music. We use arbitrary audio recorded as an approximate proxy for a balloon pop, thus removing the need for them to measure room acoustics. This work has enormous applications in Virtual/Augmented Reality and the Metaverse if it happens!
    FREE
    Open to the Public
  •  
  • 1 of 10
  • ››
  • Home
  • News and Events
    • All Events
      • CCRMA Concerts
      • Colloquium Series
      • DSP Seminars
      • Hearing Seminars
      • Guest Lectures
    • Event Calendar
    • Events Mailing List
    • Recent News
  • Academics
    • Courses
    • Current Year Course Schedule
    • Undergraduate
    • Masters
    • PhD Program
    • Visiting Scholar
    • Visiting Student Researcher
    • Workshops 2024
  • Research
    • Publications
      • Authors
      • Keywords
      • STAN-M
      • Max Mathews Portrait
    • Research Groups
    • Software
  • People
    • Faculty and Staff
    • Students
    • Alumni
    • All Users
  • User Guides
    • New Documentation
    • Booking Events
    • Common Areas
    • Rooms
    • System
  • Resources
    • Planet CCRMA
    • MARL
  • Blogs
  • Opportunities
    • CFPs
  • About
    • The Knoll
      • Renovation
    • Directions
    • Contact

Search this site:

Fall Courses at CCRMA

Music 1A Music, Mind, and Human Behavior
Music 101
Introduction to Creating Electronic Sounds
Music 192A Foundations in Sound Recording Technology
Music 201 CCRMA Colloquium
Music 220A Foundations of Computer-Generated Sound
Music 223A Composing Electronic Sound Poetry
Music 256A Music, Computing, and Design I: Software Paradigms for Computer Music
Music 319 Research Seminar on Computational Models of Sound Perception
Music 320 Introduction to Audio Signal Processing
Music 351A Research Seminar in Music Perception and Cognition I
Music 451A Auditory EEG Research I

 

 

 

   

CCRMA
Department of Music
Stanford University
Stanford, CA 94305-8180 USA
tel: (650) 723-4971
fax: (650) 723-8468
info@ccrma.stanford.edu

 
Stanford Digital Accessibility
Web Issues: webteam@ccrma
site copyright © 2009-2023
Stanford University

site design: 
Linnea A. Williams