The project was originally forked from the lfigw project. It is licensed under the MIT License. Our work was inspired and developed upon the research in "Complete parameter inference for GW150914 using deep learning" by Stephen R. Green et al. Please visit https://github.com/stephengreen/lfi-gw or the paper by Stephen R. Green and Jonathan Gair for further details and contributions.
He Wang 1, Zhoujian Cao2, Yue Zhou3, Zong-Kuan Guo1, Zhixiang Ren3*
1. Institute of Theoretical Physics, Chinese Academy of Sciences, Beijing, China
2. Department of Astronomy, Beijing Normal University, Beijing, China
3. Research Center for AI, Peng Cheng Laboratory, Shenzhen, China
Extracting knowledge from high-dimensional data has been notoriously difficult, mainly due to the so-called "curse of dimensionality" and the complex joint distributions of the dimensions. This is a particularly profound issue for 24,096-dimensional gravitational-wave data analysis, where one needs to conduct Bayesian inference and estimate the joint posterior distributions. In this work, we incorporate physical prior knowledge by sampling from desired interim distributions and construct the training dataset. In practice, we draw the training samples from the interim distribution with a certain probability alpha and also from an evenly distributed set with a probability 1-\alpha. Accordingly, the more relevant regions in the high-dimensional feature space are covered by data points so that the model can learn the subtle but critical details. We also adapt the normalizing flow method to be more expressive and trainable so that the information can be extracted and represented effectively by the transformation between the prior and target distributions. The evaluation of our approach verifies the efficacy and efficiency for gravitational-wave data inference and points to a promising direction for similar research.
The paper is available on Big Data Mining and Analytics. This release contains scripts and data files of the research.
Create and activate a conda environment with the correct software dependencies:
conda env create -f environment.yml
conda activate env
-
Download data (optional).
I have already download and pre-processing the strain data for GWTC-1 as in
./data
. And one can download the samples data of domain knowledge from Bilby GWTC-1 indownsampled_posterior_samples_v1.0.0
. -
Generate the reduced-order representation.
Use
prepreparing4RBasis.py
to generate reduced-order representation for each BBH events in GWTC-1. Usetest4RBasis.py
to test the reduced performance. -
Train the network.
./run.sh # alpha=0 or 1 ./run_alpha.sh # 0 < alpha < 1
Network hyperparameters and training options can be specified by modifying this script.
To perform inference, execute the code in evaluate_PCL_test_sh.py
or evaluate_PCL_test_whentrain.py
This loads the strain data for each BBH events in GWTC-1 and conditions the trained normalizing flow on them.
It draws samples from the base space, and applies the flow to generate samples from the conditional density estimator.
To quantitatively measure the prior sampling effect on
The resulting posterior distributions for GW150914 can be compared, and we see that they are indistinguishable, both qualitatively and quantitatively, with the ground truth:
Marginalized one- (1-sigma) and two- dimensional posterior distributions for the benchmarking event (GW150914) over the complete 15 physical parameters, with our approach (orange) and the ground truth (blue). The JS divergences are showed on the top and the contours represent 1-, 2- and 3-sigma significance level.
Copyright (c) 2020 Stephen Green
Copyright (c) 2021 Peng Cheng Laboratory.
Licensed under the MIT license. For copyright information, consult the LICENSE file.