Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy

Gradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a...

Full description

Bibliographic Details
Main Authors: Chuanlei Zhang, Minda Yao, Wei Chen, Shanwen Zhang, Dufeng Chen, Yuliang Wu
Format: Article
Language:English
Published: Hindawi-Wiley 2021-01-01
Series:Security and Communication Networks
Online Access:http://dx.doi.org/10.1155/2021/9956773
id doaj-bb59aa745b154ee19e61e2b121564755
record_format Article
spelling doaj-bb59aa745b154ee19e61e2b1215647552021-08-02T00:01:24ZengHindawi-WileySecurity and Communication Networks1939-01222021-01-01202110.1155/2021/9956773Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination StrategyChuanlei Zhang0Minda Yao1Wei Chen2Shanwen Zhang3Dufeng Chen4Yuliang Wu5College of Artificial IntelligenceCollege of Artificial IntelligenceSchool of Mechanical Electronic and Information EngineeringCollege of Information EngineeringBeijing Geotechnical and Investigation Engineering InsitituteDepartment of Emergency ManagementGradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a hot research topic. Some research efforts have tried to combine multiple methods to assist network training, but these methods seem to be more empirical, without theoretical guides. In this paper, a framework is proposed to illustrate the principle of combining different gradient descent optimization methods by analyzing several adaptive methods and other learning rate methods. Furthermore, inspired by the principle of warmup, CLR, and SGDR, the concept of multistage is introduced into the field of gradient descent optimization, and a gradient descent optimization strategy in deep learning model training based on multistage and method combination strategy is presented. The effectiveness of the proposed strategy is verified on the massive deep learning network training experiments.http://dx.doi.org/10.1155/2021/9956773
collection DOAJ
language English
format Article
sources DOAJ
author Chuanlei Zhang
Minda Yao
Wei Chen
Shanwen Zhang
Dufeng Chen
Yuliang Wu
spellingShingle Chuanlei Zhang
Minda Yao
Wei Chen
Shanwen Zhang
Dufeng Chen
Yuliang Wu
Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
Security and Communication Networks
author_facet Chuanlei Zhang
Minda Yao
Wei Chen
Shanwen Zhang
Dufeng Chen
Yuliang Wu
author_sort Chuanlei Zhang
title Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
title_short Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
title_full Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
title_fullStr Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
title_full_unstemmed Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
title_sort gradient descent optimization in deep learning model training based on multistage and method combination strategy
publisher Hindawi-Wiley
series Security and Communication Networks
issn 1939-0122
publishDate 2021-01-01
description Gradient descent is the core and foundation of neural networks, and gradient descent optimization heuristics have greatly accelerated progress in deep learning. Although these methods are simple and effective, how they work remains unknown. Gradient descent optimization in deep learning has become a hot research topic. Some research efforts have tried to combine multiple methods to assist network training, but these methods seem to be more empirical, without theoretical guides. In this paper, a framework is proposed to illustrate the principle of combining different gradient descent optimization methods by analyzing several adaptive methods and other learning rate methods. Furthermore, inspired by the principle of warmup, CLR, and SGDR, the concept of multistage is introduced into the field of gradient descent optimization, and a gradient descent optimization strategy in deep learning model training based on multistage and method combination strategy is presented. The effectiveness of the proposed strategy is verified on the massive deep learning network training experiments.
url http://dx.doi.org/10.1155/2021/9956773
work_keys_str_mv AT chuanleizhang gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
AT mindayao gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
AT weichen gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
AT shanwenzhang gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
AT dufengchen gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
AT yuliangwu gradientdescentoptimizationindeeplearningmodeltrainingbasedonmultistageandmethodcombinationstrategy
_version_ 1721245284322770944