2019
DOI: 10.1016/j.jcp.2019.01.035
|View full text |Cite
|
Sign up to set email alerts
|

Some greedy algorithms for sparse polynomial chaos expansions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 27 publications
(6 citation statements)
references
References 38 publications
0
6
0
Order By: Relevance
“…At present, the commonly used methods for selecting the basis function include the convex relaxation algorithm (CR) [ 26 ] and orthogonal matching pursuit (OMP), etc. [ 12 , 30 ]. Among them, OMP determines the position of the nonzero item by verifying the orthogonality between the residual and the expansion item, and then the approximate solution of the L0 norm minimization problem is directly obtained by the least square method.…”
Section: Methodsmentioning
confidence: 99%
“…At present, the commonly used methods for selecting the basis function include the convex relaxation algorithm (CR) [ 26 ] and orthogonal matching pursuit (OMP), etc. [ 12 , 30 ]. Among them, OMP determines the position of the nonzero item by verifying the orthogonality between the residual and the expansion item, and then the approximate solution of the L0 norm minimization problem is directly obtained by the least square method.…”
Section: Methodsmentioning
confidence: 99%
“…Orthogonal matching pursuit (OMP) (Tropp and Gilbert, 2007;Doostan and Owhadi, 2011;Marelli and Sudret, 2019) is a classical forward selection algorithm in which orthonormalized regressors are added to the model one-by-one according to their correlation with the residual, and the coefficients are computed by least-squares. Baptista et al (2019) suggest extensions to OMP such as parallelization, randomization and a modified regressor selection procedure. Subspace pursuit (SP) (Dai and Milenkovic, 2009;Diaz et al, 2018;Diaz, 2018) is an iterative algorithm that repeatedly uses least squares on a subset of regressors.…”
Section: Solution Of the Minimization Problemmentioning
confidence: 99%
“…In this paper, an enhanced version of the algorithm proposed in [17] is adopted. In [17], a greedy algorithm [45] and a (dense) grid obtained as the Cartesian product of Gaussian Points of Legendre polynomials were adopted for the construction of the reduced order model, incurring in the curse of dimensionality problem. In this work, a random-based method for the construction of the reduced order model is used instead, thus avoiding the problem of the curse of dimensionality and significantly accelerating the algorithm without loosing accuracy, as shown in section IV.…”
Section: A Parametric Model Order Reductionmentioning
confidence: 99%