A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks
碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === Music tags are a set of descriptive keywords that convey high-level information about a music clip, such as emotions(sadness, happiness), genres(jazz, classical), and instruments(guitar, vocal). Since tags provide high-level information from the listener’s per...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/vagbse |
id |
ndltd-TW-106NCHU5394074 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-106NCHU53940742019-05-16T01:24:30Z http://ndltd.ncl.edu.tw/handle/vagbse A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks 使用音訊指紋與深度卷積神經網路的強健音樂自動標記技術 Jia-Hong Yang 楊佳虹 碩士 國立中興大學 資訊科學與工程學系 106 Music tags are a set of descriptive keywords that convey high-level information about a music clip, such as emotions(sadness, happiness), genres(jazz, classical), and instruments(guitar, vocal). Since tags provide high-level information from the listener’s perspectives, they can be used for music discovery and recommendation. However, in music information retrieval (MIR), researchers need to have expertise based on acoustics or engineering design in order to analyze and organize music informations, classify them according to music forms, and then provide music information retrieval. In recent years, people have been paying more attention to the feature learning and deep architecture, thus reducing the required of the engineering works and the need for prior knowledge. The use of deep convolutional neural networks has been successfully explored in the image, text and speech field. However, previous methods for music auto-tagging can’t accurately discriminate the type of music for the distortion and noise audio, it will have the bad results in the auto-tagging. Therefore, we will propose a robust method to implement auto-music tagging. First, convert the music into a spectrogram, and find out the important information from the spectrogram, that is, the audio fingerprint. Then use it as the input of convolutional neural networks to learn the features, in this way to get a good music search result. Experimental results demonstrate the robustness of the proposed method. 吳俊霖 2018 學位論文 ; thesis 35 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === Music tags are a set of descriptive keywords that convey high-level information about a music clip, such as emotions(sadness, happiness), genres(jazz, classical), and instruments(guitar, vocal). Since tags provide high-level information from the listener’s perspectives, they can be used for music discovery and recommendation.
However, in music information retrieval (MIR), researchers need to have expertise based on acoustics or engineering design in order to analyze and organize music informations, classify them according to music forms, and then provide music information retrieval.
In recent years, people have been paying more attention to the feature learning and deep architecture, thus reducing the required of the engineering works and the need for prior knowledge. The use of deep convolutional neural networks has been successfully explored in the image, text and speech field. However, previous methods for music auto-tagging can’t accurately discriminate the type of music for the distortion and noise audio, it will have the bad results in the auto-tagging. Therefore, we will propose a robust method to implement auto-music tagging. First, convert the music into a spectrogram, and find out the important information from the spectrogram, that is, the audio fingerprint. Then use it as the input of convolutional neural networks to learn the features, in this way to get a good music search result. Experimental results demonstrate the robustness of the proposed method.
|
author2 |
吳俊霖 |
author_facet |
吳俊霖 Jia-Hong Yang 楊佳虹 |
author |
Jia-Hong Yang 楊佳虹 |
spellingShingle |
Jia-Hong Yang 楊佳虹 A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
author_sort |
Jia-Hong Yang |
title |
A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
title_short |
A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
title_full |
A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
title_fullStr |
A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
title_full_unstemmed |
A Robust Music Auto-Tagging Technique Using Audio Fingerprinting and Deep Convolutional Neural Networks |
title_sort |
robust music auto-tagging technique using audio fingerprinting and deep convolutional neural networks |
publishDate |
2018 |
url |
http://ndltd.ncl.edu.tw/handle/vagbse |
work_keys_str_mv |
AT jiahongyang arobustmusicautotaggingtechniqueusingaudiofingerprintinganddeepconvolutionalneuralnetworks AT yángjiāhóng arobustmusicautotaggingtechniqueusingaudiofingerprintinganddeepconvolutionalneuralnetworks AT jiahongyang shǐyòngyīnxùnzhǐwényǔshēndùjuǎnjīshénjīngwǎnglùdeqiángjiànyīnlèzìdòngbiāojìjìshù AT yángjiāhóng shǐyòngyīnxùnzhǐwényǔshēndùjuǎnjīshénjīngwǎnglùdeqiángjiànyīnlèzìdòngbiāojìjìshù AT jiahongyang robustmusicautotaggingtechniqueusingaudiofingerprintinganddeepconvolutionalneuralnetworks AT yángjiāhóng robustmusicautotaggingtechniqueusingaudiofingerprintinganddeepconvolutionalneuralnetworks |
_version_ |
1719175220292485120 |