The open-access journal Living Reviews in Relativity has published a new review article on 27 February 2025:
Cuoco, E., Cavaglia’, M., Heng, I.S. et al.
Applications of machine learning in gravitational-wave research with current interferometric detectors.
Living Rev Relativ 28, 2 (2025).
https://doi.org/10.1007/s41114-024-00055-8
Abstract:
This article provides an overview of the current state of machine learning in gravitational-wave research with interferometric detectors. Such applications are often still in their early days, but have reached sufficient popularity to warrant an assessment of their impact across various domains, including detector studies, noise and signal simulations, and the detection and interpretation of astrophysical signals. In detector studies, machine learning could be useful to optimize instruments like LIGO, Virgo, KAGRA, and future detectors. Algorithms could predict and help in mitigating environmental disturbances in real time, ensuring detectors operate at peak performance. Furthermore, machine-learning tools for characterizing and cleaning data after it is taken have already become crucial tools for achieving the best sensitivity of the LIGO-Virgo-KAGRA network. In data analysis, machine learning has already been applied as an alternative to traditional methods for signal detection, source localization, noise reduction, and parameter estimation. For some signal types, it can already yield improved efficiency and robustness, though in many other areas traditional methods remain dominant. As the field evolves, the role of machine learning in advancing gravitational-wave research is expected to become increasingly prominent. This report highlights recent advancements, challenges, and perspectives for the current detector generation, with a brief outlook to the next generation of gravitational-wave detectors.