2023
DOI: 10.1155/2023/4765891
|View full text |Cite
|
Sign up to set email alerts
|

An Improved Adam Optimization Algorithm Combining Adaptive Coefficients and Composite Gradients Based on Randomized Block Coordinate Descent

Abstract: An improved Adam optimization algorithm combining adaptive coefficients and composite gradients based on randomized block coordinate descent is proposed to address issues of the Adam algorithm such as slow convergence, the tendency to miss the global optimal solution, and the ineffectiveness of processing high-dimensional vectors. The adaptive coefficient is used to adjust the gradient deviation value and correct the search direction firstly. Then, the predicted gradient is introduced, and the current gradient… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
10

Relationship

0
10

Authors

Journals

citations
Cited by 25 publications
(10 citation statements)
references
References 19 publications
0
10
0
Order By: Relevance
“…The input insect pest images and their corresponding labeled images are used to train DMSAU-Net. Five - fold cross-validation (fivefold CV) scheme and stochastic gradient descent (SGD) with an adaptive moment estimator (Adam) are often used to train all models [ 22 ].…”
Section: Dilated Multi-scale Attention U-net (Dmsau-net)mentioning
confidence: 99%
“…The input insect pest images and their corresponding labeled images are used to train DMSAU-Net. Five - fold cross-validation (fivefold CV) scheme and stochastic gradient descent (SGD) with an adaptive moment estimator (Adam) are often used to train all models [ 22 ].…”
Section: Dilated Multi-scale Attention U-net (Dmsau-net)mentioning
confidence: 99%
“…It estimates the first-order and second-order moments of the gradient for each parameter based on the objective function with exponential moving average. By keeping the feature scaling of each parameter's gradient unchanged, high noise and gradient dilution issues can be solved during the parameter space iteration process [22].…”
Section: Input Layermentioning
confidence: 99%
“…To optimize the balanced cross-entropy defined in Equation (1), the Adam optimizer [ 35 ] was used. The model was trained with a learning rate of 0.001 and for 100 epochs.…”
Section: Lightweight Semantic Segmentation Fcn-mobilenetv2mentioning
confidence: 99%