LASER INTERFEROMETER GRAVITATIONAL WAVE OBSERVATORY - LIGO CALIFORNIA INSTITUTE OF TECHNOLOGY MASSACHUSETTS INSTITUTE OF TECHNOLOGY Document Type 2003/05/02 PSU proposal for contributing to S2 burst analysis plan Michael Ashley, Lee Samuel Finn, John McNabb, Eric Rotthoff, Kevin Schlaufman, Amber Stuver, Tiffany Summerscales, Patrick Sutton, Matthew Tibbits, Kristina Zaleski Distribution of this draft: LIGO I Collaboration California Institute of Technology LIGO Project - MS 51-33 Pasadena CA 91125 Phone (626) 395-2129 Fax (626) 304-9834 E-mail: [email protected] Massachusetts Institute of Technology LIGO Project - MS 20B-145 Cambridge, MA 01239 Phone (617) 253-4824 Fax (617) 253-7014 E-mail: [email protected] WWW: http://www.ligo.caltech.edu/ Processed with LATEX on 2003/05/02 Abstract Contents 1 Introduction 2 2 Pipeline Overview 2.1 Veto generation and application . . . . . . . . . . . . . . . . . . . . 2.2 Data preconditioning . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Event Trigger Generation . . . . . . . . . . . . . . . . . . . . . . . 2.4 Characterizing single IFO events from Block-Normal change-points 2.5 Coincidence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 3 4 5 6 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Astrophysical Interpretation of Candidate Events 7 4 Issue for discussion: L1, H1, and H2 coincidence 8 5 Simulations 9 6 Personnel 9 7 References 9 1 Introduction This document describes the work that the Penn State group proposes to contribute to the Burst Group analysis of S2 data. The object of the work proposed here is to provide answers to the following questions: • What is the bound in rate for gravitational wave bursts detected by the LIGO detectors? • Is the energy distribution of events seen at zero-delay consistent or inconsistent with the distribution expected from background at confidence level p (TBD)? • What is the bound in (energy, rate) space for model gravitational wave bursts incident on the LIGO detectors? • What is the energy spectrum (i.e., energy per frequency band) of all observed bursts events expected less frequently than (say) once per year, based on background estimation? • How do our sensitivity and event rates compare with those produced by the resonant acoustic detector community (including, especially but not exclusively, the “Rome events”)? The remainder of this document describes how we will achieve these goals and the personnel assigned principal responsibility for leading the different components of the effort. page 2 of 9 Simulation Single Interferometer Event Trigger Generator Data Conditioning AS_Q Diagnostics Veto Generator Event Triggers Veto Triggers and Epoch Vetoes Single Interferometer Event Characterizaion Single IFO Events Coincidence GW Candidates Interpretation Figure 1: Schematic of the Pipeline 2 Pipeline Overview Figure 1 shows, in schematic form, the analysis pipeline we will use to produce the events that will be used to address the questions set forth in the introduction above. At the most abstract level the pipeline does not differ from that used in the S1 analysis. The details of veto generation, data conditioning, event trigger generation, and especially coincidence will be different from that previous analysis, however. The analysis described here will be carried out using only science mode, triple-coincidence data which has been “blessed” by the Detector and Detector Characterization groups. The Penn State group will concentrate its effort on the use of the Block-Normal Event Trigger Generator in addressing the questions raised in the introduction. Where possible we will re-use or adapt existing software and analysis infrastructure; however, an important ancillary goal of this analysis is the development of an analysis infrastructure that scales, or can be scaled, to encompass more than the LIGO detectors and analyses beyond S2. Consequently, we anticipate the need to develop new software and analysis infrastructure in some key areas. 2.1 Veto generation and application The Syracuse group is proposing to undertake the study and development of vetoes for the S2 data set and analyses. We intend to rely heavily on their leadership in this regard and will contribute one FTE to that effort (see section 6). The rest of this section describes briefly the different types of vetoes we envision using in this analysis and draws attention to where particular attention to veto validation is necessary. Prior to bringing the data from the several interferometers together we identify in each interferometer’s data stream a set of events that that we have reason to believe may contain gravitational wave signals. Only some fraction of these events will be of gravitational wave origin. The goal of veto generation and application is to select-out from a preliminary set of events those that have a strong likelihood of being due to non-gravitational wave instrumental or environmental effects. page 3 of 9 We distinguish two different veto classes: Diagnostic vetoes (represented by veto triggers) mark transient behavior (e.g., “glitches”) in channels other than AS Q which are measured to correlate with event triggers in AS Q. These triggers are of durations comparable to those of the gravitational-wave signals being looked for. They are subdivided by their origin into two sets: IFO vetoes are derived from interferometer length-sensing and control and optics channels. Since these channels may have a coupling to AS Q, IFO vetoes need to be carefully considered to insure we are not selecting against gravitational wave events. PEM vetoes are derived from physical environment monitors such as the seismometers, magnetometers and temperature sensors around the interferometer. Vetoes derived from these channels are effectively insensitive to gravitational waves or AS Q events. Epoch vetoes are based on the AS Q channel. They mark relatively long periods of time (minutes or more) when the interferometer is performing anomalously in some way that makes events generated during the period of anomalous behavior suspect. 2.2 Data preconditioning Data preconditioning in this analysis serves two purposes: • it places the data into the form expected by the event trigger generator; and • it separates the data by frequency band, enabling the identification of event energy spectra and more sophisticated downstream coincidence analysis. The analysis described here will be based on the Block-Normal Event Trigger Generator. Block-Normal operates best on data that is white: i.e., has constant power spectral density. The data preconditioning we will undertake separates data into sub-bands and whitens the data in each sub-band. The process of data conditioning can be broken down into three basic steps: prefiltering, basebanding and bandlimiting, and final whitening. These three steps are described in detail below along with the ordering needed to simplify successive steps. 2.2.1 Prefiltering The object of prefiltering is to remove the most outstanding features of the data. Prefiltering is done before the data is broken into different frequency bands. This includes line removal and a coarse flattening of the power spectrum. We will use Kalman filtering to remove violin modes [1]. We will investigate Kalman filtering and regression for power lines. 2.2.2 Base-banding and bandlimiting The base-banding [2] and bandlimiting operation is performed for each identified frequency band. Focus attention on the frequency band (fc − ∆f /2, fc + ∆f /2), where 0 < ∆f ≤ 2fc . By the following steps we identify a timeseries yk with sample rate 2/∆f from a higher bandwidth page 4 of 9 timeseries xk with sample rate fs > 2(fc + ∆f /2), that captures just the power in the band (fc − ∆f /2, fc + ∆f /2) of xk : 1. Mix the timeseries x down by frequency fc : i.e., construct zk equal to xk e2πikfc /fs . The band of interest is centered at zero frequency in the timeseries zk . 2. Resample (with anti-aliasing) z to bandwidth 2∆f : i.e., frequencies (−∆f, ∆f ). 3. Lowpass filter the (resampled) timeseries z with a halfband filter: i.e., with a lowpass filter whose band-edge is at half the Nyquist frequency. Note that the only power that remains in z is the power in the band of interest. 4. Mix the timeseries z up by frequency ∆f /2: i.e., construct y k equal to zk e−2πik∆f /2fs . The contribution to x in the band (fc − ∆f /2, fc + ∆f /2) is equal to the contribution to y in the band (0, ∆f ) and there is no power in y outside this band. 5. Finally, take the real part of y to create a real time series. 2.2.3 Final Whitening Final whitening is performed separately on each bandlimited timeseries produced in the basebanding step described above. We will whiten the data in each base-banded timeseries using autoregressive (AR) parametric filter, trained on the nearest science mode playground segment, to remove any remaining spectral features. If xk is the time series representation of a particular frequency band, then the AR model filter satisfies xj + Na ak xj−k = ej (1) k=1 where ej is a white noise process: i.e., the result of filtering the timeseries x according to the transfer function Na A(z) = 1 + ak z −k (2) k=1 is a white process. 2.3 Event Trigger Generation The Block-Normal event trigger generator identifies places in time where the mean or variance of the data changes. Since gravitational waves are uncorrelated with detector noise, an incident gravitational wave burst must change one or both of these quantities during the period of the burst. Block-Normal operates in the time domain. The errors in its statistical determination of where or when the statistics change are minimized when it operates on data that is white. Block-Normal reports the location, by either sample or GPS time, of each change-point in the statistics of the input timeseries, probability measures (odds) associated with the identification of a change point, and the mean and variance of the data to either side of the putative change point. page 5 of 9 Identifying change points is not quite the same as identifying gravitational wave events. The events we want to identify are epochs during which the signal mean or variance are away from their nominal values. Correspondingly, we must construct events from change-points. This step is described in the next subsection. 2.4 Characterizing single IFO events from Block-Normal change-points Block-Normal identifies change-points in the statistics of its input. We will identify events as the largest set of consecutive intervals in any single band, bounded by change-points, where the mean or variance or both are different than their nominal values in that band. Each single interferometer event will have the following attributes: • IFO • Frequency Band • Start Time • End Time or duration • Calibrated Energy In addition there may be uncertainties associated with the start time, end time, and calibrated energy. This conversion from an event trigger to an “event” will require input not only from the trigger, but also from calibrations and simulations. 2.5 Coincidence The coincidence step of the data processing pipeline takes as input single interferometer events from each detector and produces as output a list of candidate gravitational wave bursts. The coincidence step also requires knowledge of the uncertainties in the determination of start time, end time or duration, and calibrated energy. These will be determined by simulation. The coincidence methodology creates the output from the input through the following steps: 1. For each frequency band in turn create clusters of events with start/end times consistent with the incidence of a plane gravitational wave on the detectors. Discard single IFO events that do not have corresponding events in the same frequency band in the other interferometers. 2. Focus on the clusters identified in the previous step. Still within each frequency band, select clusters with calibrated energy consistent with the incidence of a plane gravitational wave on the detectors. Discard all other clusters. 3. Form a global estimate of the calibrated energy in the band for each cluster. 4. Focus on clusters selected in previous step. Cluster all clusters, regardless of frequency band, that are within a time window of short (e.g., 1s) duration. page 6 of 9 5. Perform time-series cross-correlation, using the Cadonati r-statistic. This test produces for each detector pair a cross correlation statistic and the time difference that maximizes that statistic. Select only those clusters whose cross-correlation statistic is sufficiently large and for which the returned time differences between detectors are consistent with real gravitational waves, discarding all others. 6. Form, for each of the clusters identified above, a map of the energy spectrum over time for the cluster. The result is a set of candidate gravitational wave events characterized by the following attributes: • Start time; • End time or duration; • Total calibrated energy; • Energy spectrum as function of time; • Outputs of the cross correlation routine for each pair of detectors. 3 Astrophysical Interpretation of Candidate Events Here we describe, question by question, how we will address the science goals set forth in the introduction: • What is the bound in rate for gravitational wave bursts detected by the LIGO detectors? We will address this question by estimating, using time-delays, the expected background event number and comparing, using standard classical confidence interval techniques, with the zero-delay event number. • Is the energy distribution of events seen at zero-delay consistent or inconsistent with the distribution expected from background at confidence level p (TBD)? It is our intent to set a relatively low threshold so that we have a significant number of events passing our coincidence cut. Using time-delay methods we will establish a background event population in total event energy. We will compare this background population with the zerodelay population using the Mann-Whitney test, which is non-parametric and distribution independent. • What is the bound in (energy, rate) space for model gravitational wave bursts incident on the LIGO detectors? Using simulations we will determine the efficiency for the detection of prototypical gravitational wave bursts taken from a galactic population. Several models will be considered, drawn from a set including but not limited to – Black hole ringdowns; page 7 of 9 – Gravitational wave bursts from broadband axisymmetric sources; – Gravitational wave bursts from broadband non-axisymmetric sources. Each model implies an energy spectrum. When calculating the number of events in background and at zero delay an additional cut, on the energy spectrum, will be applied to the events that pass coincidence. An overall bound in (rate, energy) space will also be investigated. Such a bound is more subtle than in the S1 analysis since our events will also be characterized by a total calibrated strain energy deposited in the detector. • What is the energy spectrum (i.e., energy per frequency band) of all observed bursts events expected less frequently than (say) once per year, based on background estimation? Background estimation, based on a large number of timeshifts, can be used to estimate the expected event rate for rates much lower than once per livetime. For any zero-delay event whose occurrence in the data is found, through background estimation, to be very unlikely (e.g., less likely than once per year livetime) we will examine the event energy spectrum and attempt to understand the event origin. • How do our sensitivity and event rates compare with those produced by the resonant acoustic detector community (including, especially but not exclusively, the “Rome events”)? We will select a particular narrow band, from approximately 900 to 930 Hz, for analysis. Events identified in this band, regardless of their character in other bands, are similar to those that would be identified in the ALLEGRO, AURIGA, EXPLORER and NAUTILUS resonant acoustic detectors. Answers to the first three questions described above for just these events, and focused on just this band, can be unambiguously compared to the same questions asked of these other detectors. 4 Issue for discussion: L1, H1, and H2 coincidence As described above we operate only on triple coincidence science mode data “blessed” by the detector and detector characterization groups and require that all events be identified independently in all three interferometers. Correspondingly our sensitivity is controlled by the detector with the worst combination of sensitivity, data quality and livetime. There are several variations we can consider that would mitigate this limitation on our sensitivity; two are described below: • Require L1, H1 coincidence, and a minimum H1, H2 and L1, H2 r statistic. This option would reduce our dependence on H2, our least sensitive detector. It would still be the case that we require evidence of a signal in H2 (through the r statistic), but we would not require that the event be identified independently there. • Require L1, H1 coincidence only when consistent with thresholds and L1 event energy, but always require a minimum L1/H1, L1/H2 and H1/H2 r statistic. This option would maximize our “reach”, based on L1’s great sensitivity. As above, however, we still require confirmation of an event through correlation between the timeseries of all three detectors. page 8 of 9 These or other variations would increase our sensitivity at some cost in confidence that an event is present. Our current favor is to require triple coincidence as described in the main body of the proposal; however, we are cognizant that this represents one extreme of the balance between safety and risk, which should be settled through broader discussion within the burst group and the collaboration. 5 Simulations Simulations will be required to determine uncertainties in various parameters, including but not limited to, • uncertainties in event start and end times, and energies in each band; • efficiencies for detection of different model events and r-statistic false rates as a function of thresholds; • background estimation. We will contribute to the work of the Caltech group on simulations. 6 Personnel Penn State personnel and primary areas of responsibility are described below. Leaders of the Penn State component of the effort in these areas are designated by italics. • Vetoes: Ashley, Zaleski, Schlaufman • Data preconditioning: Summerscales, Rotthoff • Block-Normal ETG and single IFO event generation: Finn, Stuver, Tibbits • Coincidence: McNabb, TBD • Statistics and interpretation: Finn, Sutton • Simulations: McNabb, TBD 7 References [1] Lee Samuel Finn and Soma Mukherjee. Data conditioning for gravitational wave detectors: A Kalman filter for regressing suspension violin modes. Phys. Rev. D, 63:062004, 2001. grqc/0009012. [2] Lee Samuel Finn. Basebanding vs. high-pass filtering for burst searches. Technical Report T030027, Laser Interferometer Gravitational Wave Observatory and The VIRGO Experiment, LIGO Document Control Center, California Institute of Technology, 2003. page 9 of 9
© Copyright 2026 Paperzz