2016
DOI: 10.1109/tnnls.2015.2494361
|View full text |Cite
|
Sign up to set email alerts
|

Is a Complex-Valued Stepsize Advantageous in Complex-Valued Gradient Learning Algorithms?

Abstract: Abstract-Complex gradient methods have been widely used in learning theory, and typically aim to optimize real-valued functions of complex variables. The stepsize of complex gradient learning methods (CGLM) is a positive number, and little is known about how a complex stepsize would affect the learning process. To this end, we undertake a comprehensive analysis of CGLMs with complex stepsize, including the search space, convergence properties and the dynamics near critical points. Furthermore, several adaptive… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 43 publications
(18 citation statements)
references
References 32 publications
0
18
0
Order By: Relevance
“…For example [34], inspired by the theory of widely linear adaptive filters, augments the input to the CVNN with its complex conjugate x * . Additional improvements can be obtained by replacing the real-valued µ with a complex-valued learning rate [35], which can speed up convergence in some scenarios.…”
Section: B Complex-valued Neural Networkmentioning
confidence: 99%
“…For example [34], inspired by the theory of widely linear adaptive filters, augments the input to the CVNN with its complex conjugate x * . Additional improvements can be obtained by replacing the real-valued µ with a complex-valued learning rate [35], which can speed up convergence in some scenarios.…”
Section: B Complex-valued Neural Networkmentioning
confidence: 99%
“…where det(•) denotes the matrix determinant operator, the augmented covariance matrices C x a x a and C xx are defined in Section III.A, and the index ρ is normalized within [0, 1] with the value 0 indicating perfect circularity. According to (53), it can be calculated that the noncircularity of the two benchmark problems are 0.8936 and 0.4253 respectively, which indicates that the noncircular chaotic Ikeda map signal is of great noncircularity, whereas the other one is of less noncircularity. The CFNNSLs used in simulations are single output networks, and input weights and hidden biases are randomly generated by drawing the real parts and imaginary parts from a uniform distribution U (−0.1, 0.1).…”
Section: Simulation Resultsmentioning
confidence: 99%
“…where λ > 0 is the regularization parameter to balance the tradeoff between the training error and the output weight norm. According to Wirtinger calculus theory [41], [51]- [53], the solution of this optimization problem can be obtained by solving…”
Section: A Regularized Os-celmmentioning
confidence: 99%
“…Whereas BFGS algorithm also requires a number of storage units for each update encountering large-scale optimization problems, which weakens the memory efficiency of BFGS. Limited-memory BFGS (LBFGS) was further introduced, and also extended to the complex-valued domain in [29]. Compared with BFGS, LBFGS only needs to store the information of the latest m iterations to approximate the inverse of Hessian matrix, releasing the pressure of storage and showing outstanding convergence performance, both in real and complex domain [30].…”
Section: The Features Of Cvdnnmentioning
confidence: 99%