Astronomers carry out largest ever cosmological computer simulation to investigate universe’s evolution

The project, dubbed Flamingo, calculated the evolution of all components of the cosmos.

By Royal Astronomical Society

The background image shows the present-day distribution of matter in a slice through the largest FLAMINGO simulation

The background image shows the present-day distribution of matter in a slice through the largest FLAMINGO simulation, which is a cubic volume of 2.8 Gpc (9.1 billion light years) on a side. The luminosity of the background image gives the present-day distribution of dark matter, while the colour encodes the distribution of neutrinos. The insets show three consecutive zooms centred on the most massive cluster of galaxies; in order, these show the gas temperature, the dark matter density, and a virtual X-ray observation (from Figure 1 from Schaye et al. 2023). Credit: Josh Borrow, the FLAMINGO team and the Virgo Consortium. Licensed CC-BY-4.0

An international team of astronomers has carried out what is believed to be the largest ever cosmological computer simulation, tracking not only dark but also ordinary matter (such as planets, stars and galaxies), giving us a glimpse into how our Universe may have evolved.

The FLAMINGO simulations calculate the evolution of all components of the universe - ordinary matter, dark matter, and dark energy - according to the laws of physics. As the simulation progresses, virtual galaxies and clusters of galaxies emerge. Three papers have been published in Monthly Notices of the Royal Astronomical Society: one describing the methods, another presenting the simulations and the third examining how well the simulations reproduce the large-scale structure of the Universe.

Facilities such as the Euclid Space Telescope recently launched by the European Space Agency (ESA) and NASA’s JWST collect impressive amounts of data on galaxies, quasars, and stars. Simulations such as FLAMINGO play a key role in the scientific interpretation of the data by connecting predictions from theories of our universe to the observed data.

The cold dark matter model, at risk.

According to the theory, the properties of our entire universe are set by a few numbers called cosmological parameters (six of them in the simplest version of the theory). The values of these parameters can be measured very precisely in various ways.

One of these methods relies on the properties of the cosmic microwave background (CMB), a faint background glow left over from the early Universe. However, these values do not match those measured by other techniques that rely on the way in which the gravitational force of galaxies bends light (lensing). These tensions could signal the demise of the standard model of cosmology – the cold dark matter model.

Dark matter surface density, zooming in on a massive cluster in the centre. Credit: Flamingo

The computer simulations may be able to reveal the cause of these tensions because they can inform scientists about possible biases (systematic errors) in the measurements. If none of these prove sufficient to explain away the tensions, the theory will be in real trouble.

So far, the computer simulations used to compare to the observations only track cold dark matter. “Although the dark matter dominates gravity, the contribution of ordinary matter can no longer be neglected,” says research leader Joop Schaye (Leiden University), “since that contribution could be similar to the deviations between the models and the observations.”

The first results show that both neutrinos and ordinary matter are essential for making accurate predictions, but do not eliminate the tensions between the different cosmological observations.

Simulations that also track ordinary, baryonic matter (also known as baryonic matter) are much more challenging and require much more computing power. This is because ordinary matter - which makes up only sixteen per cent of all matter in the universe - feels not only gravity but also gas pressure, which can cause matter to be blown out of galaxies by active black holes and supernovae far into intergalactic space.

Dark matter, ordinary matter, and neutrinos.

The strength of these intergalactic winds depends on explosions in the interstellar medium and is very difficult to predict. On top of this, the contribution of neutrinos, subatomic particles of very small but not precisely known mass, is also important but their motion has not been simulated so far.

The astronomers have completed a series of computer simulations tracking structure formation in dark matter, ordinary matter, and neutrinos. PhD student Roi Kugel (Leiden University) explains: “The effect of galactic winds was calibrated using machine learning, by comparing the predictions of lots of different simulations of relatively small volumes with the observed masses of galaxies and the distribution of gas in clusters of galaxies.”

The researchers simulated the model that best describes the calibration observations with a supercomputer in different cosmic volumes and at different resolutions. In addition, they varied the parameters of the model, including the strength of galactic winds, the mass of neutrinos, and the cosmological parameters in simulations of slightly smaller but still large volumes.

Stellar surface density, zooming in on a massive cluster in the centre. Credit: Flamingo

The largest simulation uses 300 billion resolution elements (particles with the mass of a small galaxy) in a cubic volume with edges of ten billion light years. This is believed to be the largest cosmological computer simulation with ordinary matter ever completed. Matthieu Schaller (Leiden University): “To make this simulation possible, we developed a new code, SWIFT, which efficiently distributes the computational work over 30 thousand CPUs.”

The FLAMINGO simulations open a new virtual window on the universe that will help make the most of cosmological observations. In addition, the large amount of (virtual) data creates opportunities to make new theoretical discoveries and to test new data analysis techniques, including machine learning. Using machine learning, astronomers can then make predictions for random virtual universes. By comparing these with large-scale structure observations, they can measure the values of cosmological parameters. Moreover, they can measure the corresponding uncertainties by comparing with observations that constrain the effect of galactic winds.

Anterior
Anterior

Uncovering the mysteries of milk allergy

Siguiente
Siguiente

Black holes could come in ‘perfect pairs’ in an ever expanding Universe