Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Scene-Aware Audio Rendering via Deep Acoustic Analysis (1911.06245v2)

Published 14 Nov 2019 in cs.SD, cs.GR, cs.MM, and eess.AS

Abstract: We present a new method to capture the acoustic characteristics of real-world rooms using commodity devices, and use the captured characteristics to generate similar sounding sources with virtual models. Given the captured audio and an approximate geometric model of a real-world room, we present a novel learning-based method to estimate its acoustic material properties. Our approach is based on deep neural networks that estimate the reverberation time and equalization of the room from recorded audio. These estimates are used to compute material properties related to room reverberation using a novel material optimization objective. We use the estimated acoustic material characteristics for audio rendering using interactive geometric sound propagation and highlight the performance on many real-world scenarios. We also perform a user study to evaluate the perceptual similarity between the recorded sounds and our rendered audio.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zhenyu Tang (40 papers)
  2. Nicholas J. Bryan (23 papers)
  3. Dingzeyu Li (18 papers)
  4. Timothy R. Langlois (3 papers)
  5. Dinesh Manocha (366 papers)
Citations (38)

Summary

We haven't generated a summary for this paper yet.