Exploiting Temporal Feature Integration for Generalized Sound Recognition

This paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 au...

Full description

Bibliographic Details
Main Authors: Nikos Fakotakis, Stavros Ntalampiras, Ilyas Potamitis
Format: Article
Language:English
Published: SpringerOpen 2009-01-01
Series:EURASIP Journal on Advances in Signal Processing
Online Access:http://dx.doi.org/10.1155/2009/807162
id doaj-8265464c7542441d9387dd0fd9a2486e
record_format Article
spelling doaj-8265464c7542441d9387dd0fd9a2486e2020-11-25T00:09:33ZengSpringerOpenEURASIP Journal on Advances in Signal Processing1687-61721687-61802009-01-01200910.1155/2009/807162Exploiting Temporal Feature Integration for Generalized Sound RecognitionNikos FakotakisStavros NtalampirasIlyas PotamitisThis paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 audio protocol, and wavelet decomposition is assessed. Furthermore we explore the application of temporal integration using the following three different strategies: (a) short-term statistics, (b) spectral moments, and (c) autoregressive models. The experimental setup is thoroughly explained and based on the concurrent usage of professional sound effects collections. In this way we try to form a representative picture of the characteristics of ten sound classes. During the first phase of our implementation, the process of audio classification is achieved through statistical models (HMMs) while a fusion scheme that exploits the models constructed by various feature sets provided the highest average recognition rate. The proposed system not only uses diverse groups of sound parameters but also employs the advantages of temporal feature integration. http://dx.doi.org/10.1155/2009/807162
collection DOAJ
language English
format Article
sources DOAJ
author Nikos Fakotakis
Stavros Ntalampiras
Ilyas Potamitis
spellingShingle Nikos Fakotakis
Stavros Ntalampiras
Ilyas Potamitis
Exploiting Temporal Feature Integration for Generalized Sound Recognition
EURASIP Journal on Advances in Signal Processing
author_facet Nikos Fakotakis
Stavros Ntalampiras
Ilyas Potamitis
author_sort Nikos Fakotakis
title Exploiting Temporal Feature Integration for Generalized Sound Recognition
title_short Exploiting Temporal Feature Integration for Generalized Sound Recognition
title_full Exploiting Temporal Feature Integration for Generalized Sound Recognition
title_fullStr Exploiting Temporal Feature Integration for Generalized Sound Recognition
title_full_unstemmed Exploiting Temporal Feature Integration for Generalized Sound Recognition
title_sort exploiting temporal feature integration for generalized sound recognition
publisher SpringerOpen
series EURASIP Journal on Advances in Signal Processing
issn 1687-6172
1687-6180
publishDate 2009-01-01
description This paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 audio protocol, and wavelet decomposition is assessed. Furthermore we explore the application of temporal integration using the following three different strategies: (a) short-term statistics, (b) spectral moments, and (c) autoregressive models. The experimental setup is thoroughly explained and based on the concurrent usage of professional sound effects collections. In this way we try to form a representative picture of the characteristics of ten sound classes. During the first phase of our implementation, the process of audio classification is achieved through statistical models (HMMs) while a fusion scheme that exploits the models constructed by various feature sets provided the highest average recognition rate. The proposed system not only uses diverse groups of sound parameters but also employs the advantages of temporal feature integration.
url http://dx.doi.org/10.1155/2009/807162
work_keys_str_mv AT nikosfakotakis exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition
AT stavrosntalampiras exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition
AT ilyaspotamitis exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition
_version_ 1725411325830496256