Jump to Navigation

Main menu

  • Login
Home

Secondary menu

  • [Room Booking]
  • [Wiki]
  • [Webmail]

Equalization matching of speech recordings in real-world environments

Submitted by francois on Sat, 07/07/2018 - 11:31am
TitleEqualization matching of speech recordings in real-world environments
Publication TypeConference Paper
Year of Publication2016
AuthorsGermain, F. G., G. J. Mysore, and T. Fujioka
Conference NameIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)
Date Published03/2016
PublisherIEEE
Conference LocationShanghai, China
ISBN Number978-1-4799-9988-0
Accession Number16021757
AbstractWhen different parts of speech content such as voice-overs and narration are recorded in real-world environments with different acoustic properties and background noise, the difference in sound quality between the recordings is typically quite audible and therefore undesirable. We propose an algorithm to equalize multiple such speech recordings so that they sound like they were recorded in the same environment. As the timbral content of the speech and background noise typically differ considerably, a simple equalization matching results in a noticeable mismatch in the output signals. A single equalization filter affects both timbres equally and thus cannot disambiguate the competing matching equations of each source. We propose leveraging speech enhancement methods in order to separate speech and background noise, independently apply equalization filtering to each source, and recombine the outputs. By independently equalizing the separated sources, our method is able to better disambiguate the matching equations associated with each source. Therefore the resulting matched signals are perceptually very similar. Additionally, by retaining the background noise in the final output signals, most artifacts from speech enhancement methods are considerably reduced and in general perceptually masked. Subjective listening tests show that our approach significantly outperforms simple equalization matching.
Audio examples
URLhttp://ieeexplore.ieee.org/document/7471747/
DOI10.1109/ICASSP.2016.7471747
Refereed DesignationRefereed
Full Texthttps://ccrma.stanford.edu/~gautham/Site/Publications_files/EQMatching-i...
  • Tagged
  • XML
  • BibTex
  • Google Scholar
  • Home
  • News and Events
    • All Events
      • CCRMA Concerts
      • Colloquium Series
      • DSP Seminars
      • Hearing Seminars
      • Guest Lectures
    • Event Calendar
    • Events Mailing List
    • Recent News
  • Academics
    • Courses
    • Current Year Course Schedule
    • Undergraduate
    • Masters
    • PhD Program
    • Visiting Scholar
    • Visiting Student Researcher
    • Workshops 2020
  • Research
    • Publications
      • Authors
      • Keywords
      • STAN-M
      • Max Mathews Portrait
    • Research Groups
    • Software
  • People
    • Faculty and Staff
    • Students
    • Alumni
    • All Users
  • User Guides
    • New Documentation
    • Booking Events
    • Common Areas
    • Rooms
    • System
  • Resources
    • Planet CCRMA
    • MARL
  • Blogs
  • Opportunities
    • CFPs
  • About
    • The Knoll
      • Renovation
    • Directions
    • Contact

Search this site:

2021 Spring Quarter Courses

Music 70 Stories and Music of Refugees
Music 220A
Fundamentals of Computer-Generated Sound
Music 220C Research Seminar in Computer-Generated Music
Music 223D Sound Practice: Embodiment and the Social
Music 250A Physical Interaction Design
Music 254 Computational Music Analysis (CS275B)
Music 257 Neuroplasticity and Musical Gaming
Music 285 Intermedia Lab
Music 320C Software Projects in Music/Audio Signal Processing
Music 424 Signal Processing Techniques for Digital Audio Effects

 

 

 

   

CCRMA
Department of Music
Stanford University
Stanford, CA 94305-8180 USA
tel: (650) 723-4971
fax: (650) 723-8468
info@ccrma.stanford.edu

 
Web Issues: webteam@ccrma

site copyright © 2009 
Stanford University

site design: 
Linnea A. Williams