The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems

碩士 === 國立政治大學 === 資訊管理學系 === 107 === Rare Artificial Neural Networks studies address simultaneously the challenges of (1) systematically adjusting the amount of used hidden layer nodes within the learning process, (2) adopting ReLU activation function instead of tanh function for fast learning, and...

Full description

Bibliographic Details
Main Authors: Tsai, Yu-Han, 蔡羽涵
Other Authors: Tsaih, Rua-Huan
Format: Others
Language:zh-TW
Published: 2019
Online Access:http://ndltd.ncl.edu.tw/handle/ymzgt7
id ndltd-TW-107NCCU5396028
record_format oai_dc
spelling ndltd-TW-107NCCU53960282019-08-27T03:42:58Z http://ndltd.ncl.edu.tw/handle/ymzgt7 The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems 強記暨軟化整合演算法:以ReLU激發函數與二元輸入/輸出為例 Tsai, Yu-Han 蔡羽涵 碩士 國立政治大學 資訊管理學系 107 Rare Artificial Neural Networks studies address simultaneously the challenges of (1) systematically adjusting the amount of used hidden layer nodes within the learning process, (2) adopting ReLU activation function instead of tanh function for fast learning, and (3) guaranteeing learning all training data. This study will address these challenges through deriving the CSI (Cramming, Softening and Integrating) learning algorithm for the single-hidden layer feed-forward neural networks with ReLU activation function and the binary input/output, and further making the technical justification. For the purpose of verifying the proposed learning algorithm, this study conducts an empirical experiment using SPECT heart diagnosis data set from UCI Machine Learning repository. The learning algorithm is implemented via the advanced TensorFlow and GPU. Tsaih, Rua-Huan Hsiao, Shun-Wen 蔡瑞煌 蕭舜文 2019 學位論文 ; thesis 58 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立政治大學 === 資訊管理學系 === 107 === Rare Artificial Neural Networks studies address simultaneously the challenges of (1) systematically adjusting the amount of used hidden layer nodes within the learning process, (2) adopting ReLU activation function instead of tanh function for fast learning, and (3) guaranteeing learning all training data. This study will address these challenges through deriving the CSI (Cramming, Softening and Integrating) learning algorithm for the single-hidden layer feed-forward neural networks with ReLU activation function and the binary input/output, and further making the technical justification. For the purpose of verifying the proposed learning algorithm, this study conducts an empirical experiment using SPECT heart diagnosis data set from UCI Machine Learning repository. The learning algorithm is implemented via the advanced TensorFlow and GPU.
author2 Tsaih, Rua-Huan
author_facet Tsaih, Rua-Huan
Tsai, Yu-Han
蔡羽涵
author Tsai, Yu-Han
蔡羽涵
spellingShingle Tsai, Yu-Han
蔡羽涵
The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
author_sort Tsai, Yu-Han
title The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
title_short The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
title_full The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
title_fullStr The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
title_full_unstemmed The Cramming, Softening and Integrating Learning Algorithm with ReLU activation function for Binary Input/Output Problems
title_sort cramming, softening and integrating learning algorithm with relu activation function for binary input/output problems
publishDate 2019
url http://ndltd.ncl.edu.tw/handle/ymzgt7
work_keys_str_mv AT tsaiyuhan thecrammingsofteningandintegratinglearningalgorithmwithreluactivationfunctionforbinaryinputoutputproblems
AT càiyǔhán thecrammingsofteningandintegratinglearningalgorithmwithreluactivationfunctionforbinaryinputoutputproblems
AT tsaiyuhan qiángjìjìruǎnhuàzhěnghéyǎnsuànfǎyǐrelujīfāhánshùyǔèryuánshūrùshūchūwèilì
AT càiyǔhán qiángjìjìruǎnhuàzhěnghéyǎnsuànfǎyǐrelujīfāhánshùyǔèryuánshūrùshūchūwèilì
AT tsaiyuhan crammingsofteningandintegratinglearningalgorithmwithreluactivationfunctionforbinaryinputoutputproblems
AT càiyǔhán crammingsofteningandintegratinglearningalgorithmwithreluactivationfunctionforbinaryinputoutputproblems
_version_ 1719237660085583872