Sound field reconstruction using neural processes with dynamic kernels

Zining Liang, Wen Zhang, Thushara D. Abhayapala

科研成果: 期刊稿件文章同行评审

6 引用 (Scopus)

摘要

Accurately representing the sound field with high spatial resolution is crucial for immersive and interactive sound field reproduction technology. In recent studies, there has been a notable emphasis on efficiently estimating sound fields from a limited number of discrete observations. In particular, kernel-based methods using Gaussian processes (GPs) with a covariance function to model spatial correlations have been proposed. However, the current methods rely on pre-defined kernels for modeling, requiring the manual identification of optimal kernels and their parameters for different sound fields. In this work, we propose a novel approach that parameterizes GPs using a deep neural network based on neural processes (NPs) to reconstruct the magnitude of the sound field. This method has the advantage of dynamically learning kernels from data using an attention mechanism, allowing for greater flexibility and adaptability to the acoustic properties of the sound field. Numerical experiments demonstrate that our proposed approach outperforms current methods in reconstructing accuracy, providing a promising alternative for sound field reconstruction.

源语言英语
文章编号13
期刊Eurasip Journal on Audio, Speech, and Music Processing
2024
1
DOI
出版状态已出版 - 12月 2024

指纹

探究 'Sound field reconstruction using neural processes with dynamic kernels' 的科研主题。它们共同构成独一无二的指纹。

引用此