Exploiting Temporal Feature Integration for Generalized Sound Recognition
This paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 au...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
SpringerOpen
2009-01-01
|
Series: | EURASIP Journal on Advances in Signal Processing |
Online Access: | http://dx.doi.org/10.1155/2009/807162 |
id |
doaj-8265464c7542441d9387dd0fd9a2486e |
---|---|
record_format |
Article |
spelling |
doaj-8265464c7542441d9387dd0fd9a2486e2020-11-25T00:09:33ZengSpringerOpenEURASIP Journal on Advances in Signal Processing1687-61721687-61802009-01-01200910.1155/2009/807162Exploiting Temporal Feature Integration for Generalized Sound RecognitionNikos FakotakisStavros NtalampirasIlyas PotamitisThis paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 audio protocol, and wavelet decomposition is assessed. Furthermore we explore the application of temporal integration using the following three different strategies: (a) short-term statistics, (b) spectral moments, and (c) autoregressive models. The experimental setup is thoroughly explained and based on the concurrent usage of professional sound effects collections. In this way we try to form a representative picture of the characteristics of ten sound classes. During the first phase of our implementation, the process of audio classification is achieved through statistical models (HMMs) while a fusion scheme that exploits the models constructed by various feature sets provided the highest average recognition rate. The proposed system not only uses diverse groups of sound parameters but also employs the advantages of temporal feature integration. http://dx.doi.org/10.1155/2009/807162 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Nikos Fakotakis Stavros Ntalampiras Ilyas Potamitis |
spellingShingle |
Nikos Fakotakis Stavros Ntalampiras Ilyas Potamitis Exploiting Temporal Feature Integration for Generalized Sound Recognition EURASIP Journal on Advances in Signal Processing |
author_facet |
Nikos Fakotakis Stavros Ntalampiras Ilyas Potamitis |
author_sort |
Nikos Fakotakis |
title |
Exploiting Temporal Feature Integration for Generalized Sound Recognition |
title_short |
Exploiting Temporal Feature Integration for Generalized Sound Recognition |
title_full |
Exploiting Temporal Feature Integration for Generalized Sound Recognition |
title_fullStr |
Exploiting Temporal Feature Integration for Generalized Sound Recognition |
title_full_unstemmed |
Exploiting Temporal Feature Integration for Generalized Sound Recognition |
title_sort |
exploiting temporal feature integration for generalized sound recognition |
publisher |
SpringerOpen |
series |
EURASIP Journal on Advances in Signal Processing |
issn |
1687-6172 1687-6180 |
publishDate |
2009-01-01 |
description |
This paper presents a methodology that incorporates temporal feature integration for automated generalized sound recognition. Such a system can be of great use to scene analysis and understanding based on the acoustic modality. The performance of three feature sets based on Mel filterbank, MPEG-7 audio protocol, and wavelet decomposition is assessed. Furthermore we explore the application of temporal integration using the following three different strategies: (a) short-term statistics, (b) spectral moments, and (c) autoregressive models. The experimental setup is thoroughly explained and based on the concurrent usage of professional sound effects collections. In this way we try to form a representative picture of the characteristics of ten sound classes. During the first phase of our implementation, the process of audio classification is achieved through statistical models (HMMs) while a fusion scheme that exploits the models constructed by various feature sets provided the highest average recognition rate. The proposed system not only uses diverse groups of sound parameters but also employs the advantages of temporal feature integration. |
url |
http://dx.doi.org/10.1155/2009/807162 |
work_keys_str_mv |
AT nikosfakotakis exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition AT stavrosntalampiras exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition AT ilyaspotamitis exploitingtemporalfeatureintegrationforgeneralizedsoundrecognition |
_version_ |
1725411325830496256 |