2021
DOI: 10.1093/mnras/staa3540
|View full text |Cite
|
Sign up to set email alerts
|

Machines learn to infer stellar parameters just by looking at a large number of spectra

Abstract: Machine learning has been widely applied to clearly defined problems of astronomy and astrophysics. However, deep learning and its conceptual differences to classical machine learning have been largely overlooked in these fields. The broad hypothesis behind our work is that letting the abundant real astrophysical data speak for itself, with minimal supervision and no labels, can reveal interesting patterns that may facilitate discovery of novel physical relationships. Here, as the first step, we seek to interp… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(9 citation statements)
references
References 39 publications
0
9
0
Order By: Relevance
“…The variational autoencoder (VAE; Kingma & Welling 2014) is a variant of AE, and is widely used in astronomy (e.g. Portillo et al 2020;Sedaghat et al 2021). Although the architecture of VAE is similar to that of AE, the concept behind is very different, and VAE is based on the variational Bayesian inference (Kingma & Welling 2014).…”
Section: Variational Autoencodermentioning
confidence: 99%
“…The variational autoencoder (VAE; Kingma & Welling 2014) is a variant of AE, and is widely used in astronomy (e.g. Portillo et al 2020;Sedaghat et al 2021). Although the architecture of VAE is similar to that of AE, the concept behind is very different, and VAE is based on the variational Bayesian inference (Kingma & Welling 2014).…”
Section: Variational Autoencodermentioning
confidence: 99%
“…for backpropagation when training DL models), such neural-network-based estimators do not necessarily return an accurate estimate of equation (1), are heavily dependent on the training hyperparameters, and have been shown to suffer from a poor variance-bias tradeoff [75]. The use of MI estimates for interpreting deep representation learning has recently been investigated as well [19,32,79,80]; however, exploiting MI to interpret deep representation learning requires a robust density estimate of the joint probability distribution between latent variables and relevant physical parameters, and the uncertainties on the MI estimate to be quantified, ensuring that any trends in MI are statistically significant.…”
Section: Introductionmentioning
confidence: 99%
“…Most of these methods require prior knowledge of the system of interest, for example a priori knowledge of the relevant variables or the underlying dimensionality. Recently, Sedaghat et al [33] adopted a similar architecture to SciNet in an unsupervised setting, where a neural network is trained to find a lowdimensional representation of stellar spectra. Similar to our work, they used mutual information for interpretability; however, their use of mutual information was limited to identifying potential correlations between the latent representation and previously known parameters.…”
Section: Introductionmentioning
confidence: 99%