- Research Article
- Open Access
Acoustic Event Detection Based on Feature-Level Fusion of Audio and Video Modalities
EURASIP Journal on Advances in Signal Processing volume 2011, Article number: 485738 (2011)
Acoustic event detection (AED) aims at determining the identity of sounds and their temporal position in audio signals. When applied to spontaneously generated acoustic events, AED based only on audio information shows a large amount of errors, which are mostly due to temporal overlaps. Actually, temporal overlaps accounted for more than 70% of errors in the real-world interactive seminar recordings used in CLEAR 2007 evaluations. In this paper, we improve the recognition rate of acoustic events using information from both audio and video modalities. First, the acoustic data are processed to obtain both a set of spectrotemporal features and the 3D localization coordinates of the sound source. Second, a number of features are extracted from video recordings by means of object detection, motion analysis, and multicamera person tracking to represent the visual counterpart of several acoustic events. A feature-level fusion strategy is used, and a parallel structure of binary HMM-based detectors is employed in our work. The experimental results show that information from both the microphone array and video cameras is useful to improve the detection rate of isolated as well as spontaneously generated acoustic events.
To access the full article, please see PDF.
About this article
Cite this article
Butko, T., Canton-Ferrer, C., Segura, C. et al. Acoustic Event Detection Based on Feature-Level Fusion of Audio and Video Modalities. EURASIP J. Adv. Signal Process. 2011, 485738 (2011). https://doi.org/10.1155/2011/485738
- Recognition Rate
- Sound Source
- Object Detection
- Audio Signal
- Fusion Strategy