Deep learning of mean field annealing and gradient descent methods
碩士 === 國立東華大學 === 應用數學系 === 105 === Deep neural networks are widely used in machine learning. The deep neural network contains many hidden layers, with the powerful ability of corresponding input to output, through the adjustable neural interconnections. Deep neural network in different artificial i...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Published: |
2017
|
Online Access: | http://ndltd.ncl.edu.tw/handle/85527454990153589583 |
id |
ndltd-TW-105NDHU5507010 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-105NDHU55070102017-11-10T04:25:29Z http://ndltd.ncl.edu.tw/handle/85527454990153589583 Deep learning of mean field annealing and gradient descent methods 深度學習的均場退火與梯度遞減的混合模式 JIA-WEI JIEN 簡嘉緯 碩士 國立東華大學 應用數學系 105 Deep neural networks are widely used in machine learning. The deep neural network contains many hidden layers, with the powerful ability of corresponding input to output, through the adjustable neural interconnections. Deep neural network in different artificial intelligence framework, plays a different role, including feature extraction, dimensionality reduction, function approximation, and so on. The existing method of deep learning have been combined with a variety of methods, such as gradient descent, mini-batch, dropout, momentum, etc. Hinton’s two-stage learning method has been successfully applied in image recognition and speech recognition. The first stage is a restricted Boltzmann machine, RBMs[1], and the second stage is backpropagation learning[2] . The first stage of RBM learning gives a better initial parameter for the deep neural network. The second stage of learning is based on the initial parameters obtained in the first stage to optimize the built-in parameters of the neural network. This paper mainly proposes a new method of deep learning. The purpose is to avoid using the restricted Boltzmann method, rewrite Hinton’s two-stage method in one-stage method, and effectively achieve the goal of reducing mean square error and training error. The new method is based on the hybrid mean field annealing and gradient descent learning. In the S-type activation function, a parameter β represents the reciprocal of the annealing temperature. When the degree of entropy is greater, β value is smaller. When the degree of entropy is smaller, β value is greater. Let β grow from small to large gradually. The process is known as the annealing process. The optimal parameters in the process of annealing are recorded. The new method can be used to strength the learning ability of deep neural network, but also can combine the existing method, such as mini-batch, dropout, momentum to improve efficiency of learning. In this paper, handwriting recognition as an example is resolved to illustrate the effectiveness of new method. Jiann-Ming Wu 吳建銘 2017 學位論文 ; thesis 23 |
collection |
NDLTD |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立東華大學 === 應用數學系 === 105 === Deep neural networks are widely used in machine learning. The deep neural network contains many hidden layers, with the powerful ability of corresponding input to output, through the adjustable neural interconnections. Deep neural network in different artificial intelligence framework, plays a different role, including feature extraction, dimensionality reduction, function approximation, and so on. The existing method of deep learning have been combined with a variety of methods, such as gradient descent, mini-batch, dropout, momentum, etc. Hinton’s two-stage learning method has been successfully applied in image recognition and speech recognition. The first stage is a restricted Boltzmann machine, RBMs[1], and the second stage is backpropagation learning[2] . The first stage of RBM learning gives a better initial parameter for the deep neural network. The second stage of learning is based on the initial parameters obtained in the first stage to optimize the built-in parameters of the neural network. This paper mainly proposes a new method of deep learning. The purpose is to avoid using the restricted Boltzmann method, rewrite Hinton’s two-stage method in one-stage method, and effectively achieve the goal of reducing mean square error and training error. The new method is based on the hybrid mean field annealing and gradient descent learning. In the S-type activation function, a parameter β represents the reciprocal of the annealing temperature. When the degree of entropy is greater, β value is smaller. When the degree of entropy is smaller, β value is greater. Let β grow from small to large gradually. The process is known as the annealing process. The optimal parameters in the process of annealing are recorded. The new method can be used to strength the learning ability of deep neural network, but also can combine the existing method, such as mini-batch, dropout, momentum to improve efficiency of learning. In this paper, handwriting recognition as an example is resolved to illustrate the effectiveness of new method.
|
author2 |
Jiann-Ming Wu |
author_facet |
Jiann-Ming Wu JIA-WEI JIEN 簡嘉緯 |
author |
JIA-WEI JIEN 簡嘉緯 |
spellingShingle |
JIA-WEI JIEN 簡嘉緯 Deep learning of mean field annealing and gradient descent methods |
author_sort |
JIA-WEI JIEN |
title |
Deep learning of mean field annealing and gradient descent methods |
title_short |
Deep learning of mean field annealing and gradient descent methods |
title_full |
Deep learning of mean field annealing and gradient descent methods |
title_fullStr |
Deep learning of mean field annealing and gradient descent methods |
title_full_unstemmed |
Deep learning of mean field annealing and gradient descent methods |
title_sort |
deep learning of mean field annealing and gradient descent methods |
publishDate |
2017 |
url |
http://ndltd.ncl.edu.tw/handle/85527454990153589583 |
work_keys_str_mv |
AT jiaweijien deeplearningofmeanfieldannealingandgradientdescentmethods AT jiǎnjiāwěi deeplearningofmeanfieldannealingandgradientdescentmethods AT jiaweijien shēndùxuéxídejūnchǎngtuìhuǒyǔtīdùdìjiǎndehùnhémóshì AT jiǎnjiāwěi shēndùxuéxídejūnchǎngtuìhuǒyǔtīdùdìjiǎndehùnhémóshì |
_version_ |
1718560741582176256 |