2018 IEEE International Parallel and Distributed Processing Symposium (IPDPS) 2018
DOI: 10.1109/ipdps.2018.00015
|View full text |Cite
|
Sign up to set email alerts
|

Cataloging the Visible Universe Through Bayesian Inference at Petascale

Abstract: Astronomical catalogs derived from wide-field imaging surveys are an important tool for understanding the Universe. We construct an astronomical catalog from 55 TB of imaging data using Celeste, a Bayesian variational inference code written entirely in the high-productivity programming language Julia. Using over 1.3 million threads on 650,000 Intel Xeon Phi cores of the Cori Phase II supercomputer, Celeste achieves a peak rate of 1.54 DP PFLOP/s. Celeste is able to jointly optimize parameters for 188M stars an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
2
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 14 publications
0
8
0
Order By: Relevance
“…In this case, Julia codes can be as performant as other codes written using multiple libraries and languages: the typical case uses Python for most of the code and some optimized library (Numba, Fortran codes wrapped using f2py) for the most performance-critical routines. As an application of this use case we mention the Celeste project, which was able to load and process 178 TB of data from the SDSS catalogue in 14.6 minutes across 8192 nodes (Regier et al 2018). • Existing codes are monolithic and difficult to use interactively, and the expense of rewriting code in Julia can be rewarded by the possibility to run the code interactively, either in Julia's command line or in Jupyter notebooks.…”
Section: Discussionmentioning
confidence: 99%
“…In this case, Julia codes can be as performant as other codes written using multiple libraries and languages: the typical case uses Python for most of the code and some optimized library (Numba, Fortran codes wrapped using f2py) for the most performance-critical routines. As an application of this use case we mention the Celeste project, which was able to load and process 178 TB of data from the SDSS catalogue in 14.6 minutes across 8192 nodes (Regier et al 2018). • Existing codes are monolithic and difficult to use interactively, and the expense of rewriting code in Julia can be rewarded by the possibility to run the code interactively, either in Julia's command line or in Jupyter notebooks.…”
Section: Discussionmentioning
confidence: 99%
“…The runtime supports distributed parallel as well as multi-threaded execution. It has been demonstrated to perform at peta-scale on a high-performance computing platform [17], and it has strong support for scientific machine learning [18,19,20,21]. The language implementation is open source, available under the MIT license.…”
Section: The Julia Languagementioning
confidence: 99%
“…L OOP parallelization is the de-facto standard method for performing shared-memory data-parallel computation. Parallel computing frameworks such as OpenMP [1] have enabled the acceleration of advances in many scientific and engineering fields such as astronomical physics [2], climate analytics [3], and machine learning [4]. A major challenge in enabling efficient loop parallelization is to deal with the inherent imbalance in workloads [5].…”
Section: Introductionmentioning
confidence: 99%