TY - JOUR
T1 - Deep learning inference with the Event Horizon Telescope II. The ZINGULARITY framework for Bayesian artificial neural networks
AU - Janssen, M.
AU - Chan, C. K.
AU - Davelaar, J.
AU - Wielgus, M.
N1 - Publisher Copyright:
© The Authors 2025.
PY - 2025/5/1
Y1 - 2025/5/1
N2 - Context. In this second paper in our publication series, we present the open-source ZINGULARITY framework for parameter inference with deep Bayesian artificial neural networks. We carried out supervised learning with synthetic millimeter very long baseline interferometry observations of the Event Horizon Telescope (EHT). Our ground-truth models are based on general relativistic magnetohydrodynamic simulations of Sgr A∗ and M87∗ on horizon scales. The models predict the synchrotron emission produced by these accreting supermassive black hole systems. Aims. We investigated how well ZINGULARITY neural networks are able to infer key model parameters from EHT observations, such as the black hole spin and the magnetic state of the accretion disk, when uncertainties in the data are accurately taken into account. Methods. ZINGULARITY makes use of the TENSORFLOW PROBABILITY library and is able to handle large amounts of data with a combination of the efficient TFRecord data format plus the HOROVOD framework for distributed deep learning. Our approach is the first analysis of EHT data with Bayesian neural networks, where an unprecedented training data size, under consideration of a closely modeled EHT signal path, and the full information content of the observational data are used. ZINGULARITY infers parameters based on salient features in the data and is containerized for scientific reproducibility. Results. Through parameter surveys and dedicated validation tests, we identified neural network architectures, that are robust against internal stochastic processes and unaffected by noise in the observational and model data. We give examples of how different data properties affect the network training. We show how the Bayesian nature of our networks gives trustworthy uncertainties and uncovers failure modes for uncharacterizable data. Conclusions. It is easy to achieve low validation errors during training on synthetic data with neural networks, particularly when the forward modeling is too simplified. Through careful studies, we demonstrate that our trained networks can generalize well so that reliable results can be obtained from observational data.
AB - Context. In this second paper in our publication series, we present the open-source ZINGULARITY framework for parameter inference with deep Bayesian artificial neural networks. We carried out supervised learning with synthetic millimeter very long baseline interferometry observations of the Event Horizon Telescope (EHT). Our ground-truth models are based on general relativistic magnetohydrodynamic simulations of Sgr A∗ and M87∗ on horizon scales. The models predict the synchrotron emission produced by these accreting supermassive black hole systems. Aims. We investigated how well ZINGULARITY neural networks are able to infer key model parameters from EHT observations, such as the black hole spin and the magnetic state of the accretion disk, when uncertainties in the data are accurately taken into account. Methods. ZINGULARITY makes use of the TENSORFLOW PROBABILITY library and is able to handle large amounts of data with a combination of the efficient TFRecord data format plus the HOROVOD framework for distributed deep learning. Our approach is the first analysis of EHT data with Bayesian neural networks, where an unprecedented training data size, under consideration of a closely modeled EHT signal path, and the full information content of the observational data are used. ZINGULARITY infers parameters based on salient features in the data and is containerized for scientific reproducibility. Results. Through parameter surveys and dedicated validation tests, we identified neural network architectures, that are robust against internal stochastic processes and unaffected by noise in the observational and model data. We give examples of how different data properties affect the network training. We show how the Bayesian nature of our networks gives trustworthy uncertainties and uncovers failure modes for uncharacterizable data. Conclusions. It is easy to achieve low validation errors during training on synthetic data with neural networks, particularly when the forward modeling is too simplified. Through careful studies, we demonstrate that our trained networks can generalize well so that reliable results can be obtained from observational data.
KW - methods: data analysis
KW - techniques: high angular resolution
KW - techniques: interferometric
UR - https://www.scopus.com/pages/publications/105007524863
UR - https://www.scopus.com/pages/publications/105007524863#tab=citedBy
U2 - 10.1051/0004-6361/202553785
DO - 10.1051/0004-6361/202553785
M3 - Article
AN - SCOPUS:105007524863
SN - 0004-6361
VL - 698
JO - Astronomy and astrophysics
JF - Astronomy and astrophysics
M1 - A61
ER -