2003
DOI: 10.1016/s0893-6080(03)00085-6
|View full text |Cite
|
Sign up to set email alerts
|

On structure-exploiting trust-region regularized nonlinear least squares algorithms for neural-network learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
20
0

Year Published

2008
2008
2024
2024

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 24 publications
(21 citation statements)
references
References 17 publications
1
20
0
Order By: Relevance
“…In consequence, the use of the global Hessian would reduce the number of epochs required to achieve a target error level, although it would increase the time per epoch. (This is confirmed with the Gauss-Newton Hessian case in our previous work [18], [11] using the trust-region method [10], [12]). The posed tradeoff between the global and local Hessian matrices leads to a delicate balancing act to be pursued further.…”
Section: Discussionsupporting
confidence: 83%
See 1 more Smart Citation
“…In consequence, the use of the global Hessian would reduce the number of epochs required to achieve a target error level, although it would increase the time per epoch. (This is confirmed with the Gauss-Newton Hessian case in our previous work [18], [11] using the trust-region method [10], [12]). The posed tradeoff between the global and local Hessian matrices leads to a delicate balancing act to be pursued further.…”
Section: Discussionsupporting
confidence: 83%
“…(6) in order to exploit sparsity; for more details, see [19], [8], [12], [18]. This is just a block-diagonal approximation to the global Hessian matrix H L .…”
Section: A the Local Hessian Matricesmentioning
confidence: 99%
“…In recent years, as many important practical problems involve in great number of variables(say, at the magnitude of millions of variables), large scale optimization problems, including large scale nonlinear equations and large scale nonlinear least squares have been attracting more and more attention from researchers, for example see (Toint 1987;Gould et al 2005;Gould and Toint 2007). Most large scale problems have either sparse property or special structure, therefore special approaches, such as partial separability and structure-exploiting, should be and can be applied to such problems (Mizutani and Demmel 2003;Bouaricha and Moré 1997). We consider another approach, using subspace techniques, to large scale problems.…”
Section: Introductionmentioning
confidence: 99%
“…In all of the analyzed cases, the elements causing the most demanding calculations are the positive feedback matrices e i d j , here the complexity is of the order of O(n 2 ). As shown by formulas (15), (20) and in Table 12, the computational complexity O(n) are linked with the other elements of the matrix like c i , c 0 and d i . Interesting results were obtained by juxtaposing formulas describing the computational complexity of the block matrix inversion and the computational complexity resulting from the Gauss method.…”
Section: Discussionmentioning
confidence: 99%
“…64(4) 2016 T. Trawiński, A. Kochan, P. Kielan, and D. Kurzyk Square matrices with entries equal zero except for their main diagonal, one row and one column have many applications e.g. in wireless communication systems [19], neural-network models [20] as well as issues related with chemistry [21] or phisics [22]. In this paper we propose a method of inversion of symetric matrices containing non-zero blocks in their main diagonal, one column, one row and zeros in remaining entries.…”
Section: Introductionmentioning
confidence: 99%