2022
DOI: 10.1007/s10957-022-02003-4
|View full text |Cite
|
Sign up to set email alerts
|

An Augmented Lagrangian Method Exploiting an Active-Set Strategy and Second-Order Information

Abstract: In this paper, we consider nonlinear optimization problems with nonlinear equality constraints and bound constraints on the variables. For the solution of such problems, many augmented Lagrangian methods have been defined in the literature. Here, we propose to modify one of these algorithms, namely ALGENCAN by Andreani et al., in such a way to incorporate second-order information into the augmented Lagrangian framework, using an active-set strategy. We show that the overall algorithm has the same convergence p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(6 citation statements)
references
References 19 publications
0
6
0
Order By: Relevance
“…In the field of constrained optimization, several active-set techniques were proposed to identify the active (or binding) constraints, see, e.g., [4,6,10,11,13,16,17,22,23,34,36]. Active-set strategies were successfully used also to identify the zero variables in ℓ 1 -regularized problems [14,26,38,43,44] and in ℓ 1 -constrained problems [12].…”
Section: The Algorithmmentioning
confidence: 99%
See 4 more Smart Citations
“…In the field of constrained optimization, several active-set techniques were proposed to identify the active (or binding) constraints, see, e.g., [4,6,10,11,13,16,17,22,23,34,36]. Active-set strategies were successfully used also to identify the zero variables in ℓ 1 -regularized problems [14,26,38,43,44] and in ℓ 1 -constrained problems [12].…”
Section: The Algorithmmentioning
confidence: 99%
“…Given a feasible point x k = 0 produced by the algorithm, in the sequel we define two possible strategies to compute π k in (13) and to update the variables. The first strategy uses the whole gradient ∇ϕ and is more accurate in the active-set estimate, while the second strategy only uses partial derivatives and is computationally more efficient.…”
Section: The Algorithmmentioning
confidence: 99%
See 3 more Smart Citations