Loading…

Comparison of neural network architectures for feature extraction from binary black hole merger waveforms

We evaluate several neural-network architectures, both convolutional and recurrent, for gravitational-wave time-series feature extraction by performing point parameter estimation on noisy waveforms from binary-black-hole mergers. We build datasets of 100 000 elements for each of four different wavef...

Full description

Saved in:
Bibliographic Details
Published in:Machine learning: science and technology 2024-03, Vol.5 (1), p.15036
Main Authors: Gramaxo Freitas, Osvaldo, Calderón Bustillo, Juan, Font, José A, Nunes, Solange, Onofre, Antonio, Torres-Forné, Alejandro
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We evaluate several neural-network architectures, both convolutional and recurrent, for gravitational-wave time-series feature extraction by performing point parameter estimation on noisy waveforms from binary-black-hole mergers. We build datasets of 100 000 elements for each of four different waveform models (or approximants) in order to test how approximant choice affects feature extraction. Our choices include SEOBNRv4P and IMRPhenomPv3 , which contain only the dominant quadrupole emission mode, alongside IMRPhenomPv3HM and NRHybSur3dq8 , which also account for high-order modes. Each dataset element is injected into detector noise corresponding to the third observing run of the LIGO-Virgo-KAGRA (LVK) collaboration. We identify the temporal convolutional network architecture as the overall best performer in terms of training and validation losses and absence of overfitting to data. Comparison of results between datasets shows that the choice of waveform approximant for the creation of a dataset conditions the feature extraction ability of a trained network. Hence, care should be taken when building a dataset for the training of neural networks, as certain approximants may result in better network convergence of evaluation metrics. However, this performance does not necessarily translate to data which is more faithful to numerical relativity simulations. We also apply this network on actual signals from LVK runs, finding that its feature-extracting performance can be effective on real data.
ISSN:2632-2153
2632-2153
DOI:10.1088/2632-2153/ad2972