Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
Gradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Hindawi-Wiley
2021-01-01
|
Series: | Security and Communication Networks |
Online Access: | http://dx.doi.org/10.1155/2021/9956773 |
id |
doaj-bb59aa745b154ee19e61e2b121564755 |
---|---|
record_format |
Article |
spelling |
doaj-bb59aa745b154ee19e61e2b1215647552021-08-02T00:01:24ZengHindawi-WileySecurity and Communication Networks1939-01222021-01-01202110.1155/2021/9956773Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination StrategyChuanlei Zhang0Minda Yao1Wei Chen2Shanwen Zhang3Dufeng Chen4Yuliang Wu5College of Artificial IntelligenceCollege of Artificial IntelligenceSchool of Mechanical Electronic and Information EngineeringCollege of Information EngineeringBeijing Geotechnical and Investigation Engineering InsitituteDepartment of Emergency ManagementGradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a hot research topic. Some research efforts have tried to combine multiple methods to assist network training, but these methods seem to be more empirical, without theoretical guides. In this paper, a framework is proposed to illustrate the principle of combining different gradient descent optimization methods by analyzing several adaptive methods and other learning rate methods. Furthermore, inspired by the principle of warmup, CLR, and SGDR, the concept of multistage is introduced into the field of gradient descent optimization, and a gradient descent optimization strategy in deep learning model training based on multistage and method combination strategy is presented. The effectiveness of the proposed strategy is verified on the massive deep learning network training experiments.http://dx.doi.org/10.1155/2021/9956773 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Chuanlei Zhang Minda Yao Wei Chen Shanwen Zhang Dufeng Chen Yuliang Wu |
spellingShingle |
Chuanlei Zhang Minda Yao Wei Chen Shanwen Zhang Dufeng Chen Yuliang Wu Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy Security and Communication Networks |
author_facet |
Chuanlei Zhang Minda Yao Wei Chen Shanwen Zhang Dufeng Chen Yuliang Wu |
author_sort |
Chuanlei Zhang |
title |
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy |
title_short |
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy |
title_full |
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy |
title_fullStr |
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy |
title_full_unstemmed |
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy |
title_sort |
gradient descent optimization in deep learning model training based on multistage and method combination strategy |
publisher |
Hindawi-Wiley |
series |
Security and Communication Networks |
issn |
1939-0122 |
publishDate |
2021-01-01 |
description |
Gradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a hot research topic. Some research efforts have tried to combine multiple methods to assist network training, but these methods seem to be more empirical, without theoretical guides. In this paper, a framework is proposed to illustrate the principle of combining different gradient descent optimization methods by analyzing several adaptive methods and other learning rate methods. Furthermore, inspired by the principle of warmup, CLR, and SGDR, the concept of multistage is introduced into the field of gradient descent optimization, and a gradient descent optimization strategy in deep learning model training based on multistage and method combination strategy is presented. The effectiveness of the proposed strategy is verified on the massive deep learning network training experiments. |
url |
http://dx.doi.org/10.1155/2021/9956773 |
work_keys_str_mv |
AT chuanleizhang gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy AT mindayao gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy AT weichen gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy AT shanwenzhang gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy AT dufengchen gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy AT yuliangwu gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy |
_version_ |
1721245284322770944 |