Exploring the Cosmos while Preserving Spaceship Earth

Jun 16, 2022

Chun-Hao To

By Chun-Hao To

On-going and upcoming cosmological surveys, such as the Dark Energy Survey (DES), Rubin Observatory's Legacy Survey of Space and Time (LSST), and observations from the Nancy Grace Roman Space Telescope, as well as the Euclid Space Telescope, will produce tremendous volumes of astrophysical data in the near future. Extracting cosmological information from this data requires sophisticated theory modeling that combines physical models of our Universe, astrophysical models of galaxies, and telescope and survey-specific models of measurement uncertainties. The theory models involved are now quite sophisticated and typically have thirty to fifty parameters, making it hard to find plausible sets of parameters that describe the observational data (see our post from March 2022 for a discussion of state-of-the art analyses at DES). Therefore, the analyses of modern cosmological datasets are highly time-consuming, even with the most advanced algorithms and software.

Specifically, portions of DES's latest analysis took as long as twenty-one days on state-of-the-art supercomputing clusters. These runs will get significantly longer in the future, as datasets get correspondingly larger and upcoming analysis models become more sophisticated. The long runtime will increase energy usage and cost, as well as taking more time to get the cosmological constraints once the data are in hand. Further, substantial computational costs place a significant burden on our environment. In a typical cosmological analysis, one performs the analysis thousands of times to test different scenarios, resulting in considerable CO2 emissions from the electricity consumption. 

Outlines of LINNA structure. Given a set of standard models, LINNA uses the information to generate a new, approximate model and infers cosmological parameters from the data.  Here everything in green boxes is initial data and theory inputs, and the red oval at the bottom labeled “LINNA chain”  is the ultimate output, giving an estimate of the cosmological parameters we are attempting to determine for this particular analysis. (Credit: Adapted from To, et al., 2022.)
Fig 1: Outlines of LINNA structure. Given a set of standard models, LINNA uses the information to generate a new, approximate model and infers cosmological parameters from the data. Here, everything in green boxes is initial data and theory inputs, and the red oval at the bottom labeled “LINNA chain”  is the ultimate output, giving an estimate of the cosmological parameters we are attempting to determine for this particular analysis. (Credit: Adapted from To, et al., 2022.)

 

In this paper, we build a Likelihood Inference Neural Network Accelerator (LINNA, Fig. 1) to speed up the process of extracting cosmological information from the data. We achieve this using a specially designed deep neural network, a powerful machine learning method that can approximate complicated models accurately and is fast to evaluate. The deep neural network is then combined with standard Monte Carlo Markov Chain (MCMC) samplers, which, boiled down to basics, allow us to extract the cosmological information from the data using probability distributions. We find that the combination of neural networks with standard sampling techniques can accurately reproduce expensive standard analyses in a few iterations (Fig. 2), with total computational costs a factor of 8-50 smaller.

Posteriors (i.e., probability distributions of cosmological parameters extracted from the analysisinferred from LINNA in each iteration. (Credit: Adapted from To, et al., 2022.) 
Fig. 2  Posteriors (i.e., probability distributions of cosmological parameters extracted from the analysis—in this particular case what is shown is the logarithm of A_s, a parameter related to how concentrated galaxies and matter are, versus ⍵_m, the amount of total matter in the Universe) inferred from LINNA in each iteration. LINNA (brown contour) produces almost identical results compared to other standard (and expensive) methods, as shown by the black contours, but does so much more quickly. The steady shrinking of the contour indicates rapid convergence to a preferred stable parameter output, which is exactly that obtained from the much more time- and energy-intensive method. (Credit: Adapted from To, et al., 2022.)

 

This new tool will be particularly useful for ongoing and future surveys. If we use LINNA on the upcoming LSST year one (LSST-Y1) cosmological analyses, the time consumption on posterior inferences of cosmological parameters will be greatly reduced. This not only speeds up scientific developments but also has non-trivial environmental and economic impacts. Under very conservative assumptions, we find that applying LINNA on LSST first-year analyses can save $300k US on energy costs and reduce 2400 tons of CO2 emissions, an amount equivalent to the annual carbon footprint of approximately 65 astronomers.

With the use of LINNA, we can learn much more in efficient ways about the fundamental cosmological evolution of the Universe, and how we got to from the start all the way to living on this lovely blue and green marble floating through space, while at the same time minimizing the damage here to Spaceship Earth—the only home in this vast Universe that we humans have ever known, thus far.

 

-------------------- Extra Reading

Astronomers for Planet Earth -- The Astronomical Perspective