The spherical harmonic decomposition can be applied to present realistically localized sound sources over headphones. The acoustic field, measured by a spherical microphone array, is first decomposed into a weighted sum of spherical harmonics evaluated at the microphone positions. The resulting decomposition is used to generate a set of virtual sources at various angles. The virtual sources are thus binaurally presented by applying the corresponding head-related transfer functions (HRTF). Reproduction accuracy is heavily dependent on the spatial distribution of microphones and virtual sources. Nearly-uniform sphere samplings are used in positioning the microphones so as to improve spatial accuracy. However, no previous studies have looked into the optimal arrangement for the virtual sources. We evaluate the effects of the virtual source distribution on the accuracy of the synthesized HRTF. Furthermore, our study considers the impact of spatial aliasing for a 252-channel spherical microphone array. The microphone's body is modeled as a human-head-sized rigid sphere. We evaluate the synthesis error by comparison with the target HRTF using the logarithmic spectral distance. Our study finds that 362 virtual sources, distributed on an icosahedral grid, can synthesize the HRTF in the horizontal plane up to 9 kHz with a log-spectral distance below 5 dB.
|Journal||Proceedings of Meetings on Acoustics|
|Publication status||Published - 2013|
|Event||21st International Congress on Acoustics, ICA 2013 - 165th Meeting of the Acoustical Society of America - Montreal, QC, Canada|
Duration: 2013 Jun 2 → 2013 Jun 7
ASJC Scopus subject areas
- Acoustics and Ultrasonics