# Assignment 7: Gravitational Waves

**Ouvert le :**jeudi 14 mars 2024, 00:00

**À remettre :**jeudi 21 mars 2024, 23:59

For this assignment, you are to mimic one of the methods used by the Laser Interferometer Gravitational Wave Observatory (LIGO) for detecting gravitational waves (GWs) by analyzing signals from the detector, which led to the 2017 Nobel prize in physics for the detection of GWs.

We will use data generated by real numerical relativity simulations kindly made publicly available by the Center for Computational Relativity and Gravitation, see Campanelli, Lousto, Nakano, and Zlochower, Phys. Rev. D 79:084010 (2009).

An oversimplified representation of LIGOs "Matched Filtered Algorithm" goes as follows:

- Consider a bank of predicted GW waveforms generated by numerical relativity simulations of binary black hole mergers. Each wave form is given as a complex-valued time series.
- Collect a set of measured signals from the detector (also complex-valued time series).
- Perform a power spectrum analysis, correlating the GWs signals with the signals in the simulated waveform bank. This correlation analysis is performed in the Fourier domain to make the results more robust, as the signals are usually embedded in noise.
- Select candidate events with a high correlation coefficients with one of the predicted waveforms.

**Assignment**

In this assignment, you will implement something similar. You only need to consider one predicted GW signal from simulations of the merger of two black holes in the netcdf file GWprediction.nc, and 32 detector signals in the netcdf files detection01.nc, ..., detection32.nc. This data can be found on the teach cluster in the directory /scinet/course/phy1610/gwdata.

The file structure is as follows. Each file contains a variable called f which stores the values of the wave signal at regularly spaced time points. The file also contains a variable with the actual times at which the samples were taken but this information is not really needed in this assignment. The wave signal is a set of double-precision complex numbers, so each point has a double-precision real and imaginary part. Because NetCDF does not have native complex number support, the f variable is a nt x 2 matrix, where the time dimension nt is stored in the file. This is in fact also how complex numbers are laid out in memory in C++, so it's in fact fairly easy to read in these files, e.g., like this:

#include <rarray>

#include <netcdf>

#include <complex>

using namespace netCDF;

...

NcFile ncfile("detection01.nc", NcFile::read);

rvector<std::complex<double>> frvector(ncfile.getDim("nt").getSize());

ncfile.getVar("f").getVar(frvector.data());

...

You will have to compute the correlations of the power spectrum of the prediction with each mock detection. The power spectrum F of a signal f is related to the fourier transform of that signal: for each wave number k (in this case they are really frequency numbers), the power spectrum is the square norm of the fourier component with that wave number, i.e.,

$$F_k=|\hat f_k|^2=\hat f_k \hat f^*_k$$

Note that the power spectrum is real-valued.

The correlation of two power spectra F and G is given by

$$C_{F,G} = \frac{<F,G>}{\sqrt{<F,F><G,G>}}$$

where <,> denotes an inner product, which here is simply:

$$<A,B> = \sum_{k=0}^{n-1} A_k B_k$$

The algorithm that you have to implement is structured as follows:

- Read the predicted GW signal from GWprediction.nc.
- Read one of the GW signal from observations detection01.nc . . . detection32.nc.
- Compute the FFT of the two complex quantities, using FFTW.
- Compute the power spectrum of both signals.
- Compute the correlation coefficient between the power spectra as defined above.
- Store the correlation coefficient.
- Repeat steps 2-to-6 for each of the signals in the observation set.
- Finally, have the code determine the 5 most significant candidates (those with the 5 largest values of the correlation coefficient) from the observations set.

For point 3, you will need an fft implementation. Usage the fftw module on the teach cluster for that.

As a reference, the following graph plots the values of the correlation coefficients for the 32 candidates sorted from smallest to largest. Do note that the precise values may vary depending on your implementation so you should not focus on the exact values but instead the approximate ranges and distribution overall.

Use git version control for the assignment. To your git repository, add a plain text file with the results of your analysis, i.e. which are the 5 most significant candidates (those with the 5 largest values of the correlation coefficient) from the observations set.

We continue to expect you to use best practices like commenting your code and several meaningful git commit messages. While there is not a lot of opportunity for modularity here, structuring your code into functions is a good idea.

Submit your git repo for the code (without the data!) by midnight March 21st, 2024.