1
Grid Development and Deployment of LIGO Scientific Collaboration Data Analysis Software Onasys LSC Users Pipeline Generation Glue LALApps LAL LDR LSC Data Grid / Open Science Grid The ONline Analysis SYStem (onasys) software provides tools for automating the real-time analysis of gravitational wave data. Start Get segment list Compute unused segments Pause segment database used segments Segments remain? No Yes Generate DAG for segment Submit DAG to compute cluster Executes data analysis pipelines created by LSC scientists in real-time to process data from the three LIGO gravitational wave interferometers as it is taken. Responsible for the execution and monitoring of data analysis pipelines which generate astrophysical results. Results are returned to the user or can be published for replication around the LSC data grid. Built on top of the Grid LSC User Environment (Glue) and uses the scientific data analysis pipeline software deployed by the LSC. Online monitor- ing of analysis is available via a web interface which queries job information metadata data- bases. Onasys periodically, and on a short time- scale, performs the fol- lowing sequence: Identify the data to analyze using a query to a GSI-authenticating instrument status and data quality server (LSCsegFind). Locate the data using a query to a GSI-authenticating data location server (LDRdataFind). Configure a data analysis pipeline using a user- supplied pipeline construction tool. Execute the data analysis pipeline on the grid. Users search for gravitational waves in LIGO data by running data analysis pipelines on the grid. Simple text files contain the scientific parameters of the desired search. Pipelines are run on the grid to produce scientific results Complicated workflows can be constructed to performs the all steps necessary to search data from the four LSC detectors: the 3 LIGO interferometers and the GEO600 interferometer. A subset of a binary inspiral workflow. Workflows used in analysis pipelines typically have over 10,000 Interferometer(s) Backup Frame Builder (fb0) Primary Frame Builder (fb1) /fb0_frames (~ 2 days) /frames (~ 1 week) /archive /cluster Publish /frames and state vector mySQL + RLS LDR LDRdataFindServer LSCsegFindServer LDRmetadata, RLS, BOSS, LSCSegFind Databases Disk to Disk Copy Publish R, RDS from /archive Secure web server for online job monitoring gateway ldas-grid onasysd (inspiral) onasysd (power) Condor schedd Condor negotiator Condor collector Observatory Condor pool GSI sshd (user login) Cluster home directories Secure GSI Authenticated TCP Internal Network TCP UNIX Socket Fiberchannel or NFS Condor startd The Grid LSC User Environment (Glue) provides a high level infrastructure for developing LSC grid applications and metadata services. User applications which perform specific data analysis tasks as part of a pipeline. Replicates LIGO and GEO data around the LSC grid. LSC Algorithm Library (LAL) contains data analysis routines. Provides underlying hardware and operating system software for executing LSC data analysis code. LHO LLO CIT MIT UWM PSU AEI Maintains and distributes metadata about interferometer data. Replicates data from LIGO observato- ries and between LSC compute sites. Provides services to users and applica- tions for mapping metadata to physical data locations. Built on top of GridFTP, MySQL, Globus RLS and pyGlobus. Computing clusters at Caltech, MIT, LIGO Liv- ingston Observatory, LIGO Hanford Observa- tory, UWM, Penn State, AEI, Cardiff, University of Birmingham comprise the LSC Data Grid. Online analysis is performed at the observa- tory sites (LHO and LLO). Offline analysis and follow-up of online results runs at other sites. The LSC Data Grid provides a standard set of grid tools for users to execute pipelines. Access to the LSC Data Grid is via ssh login or Globus job managers. LSC Data Grid usage can be monitored using Ganglia. DATAFIND TMPLTBANK INSPIRAL SINCA THINCA TRIGBANK INSPVETO THINCA2 thinca2_H1L1_1 thinca2_H2L1_1 thinca2_slide_H1L1_1 thinca2_slide_H2L1_1 trigbank_L1_1 insp_L1_V_1 trigbank_L1_2 insp_L1_V_2 trigbank_L1_3 insp_L1_V_3 trigbank_L1_4 insp_L1_V_4 trigbank_L1_5 insp_L1_V_5 trigbank_L1_6 insp_L1_V_6 trigbank_L1_7 insp_L1_V_7 insp_H2_1 thinca_H2L1_1 thinca_slide_H2L1_1 insp_H2_2 thinca_H1H2L1_1 thinca_slide_H1H2L1_1 insp_H2_3 insp_H2_4 trigbank_H2_1 insp_H2_V_1 trigbank_H2_2 insp_H2_V_2 trigbank_H2_3 insp_H2_V_3 trigbank_H2_4 insp_H2_V_4 trigbank_H2_5 insp_H2_V_5 thinca2_H1H2L1_1 thinca2_slide_H1H2L1_1 datafind_L_1 insp_L1_1 insp_L1_2 insp_L1_3 insp_L1_4 insp_L1_5 insp_L1_6 tmplt_L1_1 tmplt_L1_2 tmplt_L1_3 tmplt_L1_4 tmplt_L1_5 tmplt_L1_6 trigbank_H1_1 insp_H1_V_1 trigbank_H1_2 insp_H1_V_2 trigbank_H1_3 insp_H1_V_3 trigbank_H1_4 insp_H1_V_4 trigbank_H1_5 insp_H1_V_5 trigbank_H1_6 insp_H1_V_6 insp_H1_1 insp_H1_2 thinca_H1L1_1 thinca_slide_H1L1_1 insp_H1_3 insp_H1_4 sinca_H1_1 sinca_L1_1 tmplt_H1_1 tmplt_H1_2 tmplt_H1_3 tmplt_H1_4 tmplt_H2_1 tmplt_H2_2 tmplt_H2_3 tmplt_H2_4 datafind_H_1 datafind_H_2 Pipeline generation is built on top of Glue and the LALApps analysis codes. Different analysis strategies can be pursued by constructing various work- flows. Different pipelines are typically con- structed for online analysis, offline follow-up of online triggers, and offline analysis used in large scale Monte Carlo simulation and parameter tuning. Pipelines are created as Condor DAGs or VDS abstract workflows (DAX) Workflows constructed by Pegasus from the LSC workflows can be executed on the Open Science Grid. OSG will be used for compute inten- sive workflows which are too large for the LSC grid. Condor and VDT Provides grid middleware used to execute searches. Condor provides the underlying infrastruc- ture for execution and management of workflows on the grid, either directly or through submission to Globus job manag- ers. The LSC Data Grid client and server pack- ages are built on top of the VDT. The client and server provide the subset of the VDT used by the LSC and add LSC specific grid tools. Pacman makes this bundling and distribution simple. Pegasus Takes abstract workflows created by data analy- sis pipeline generators. Queries metadata catalogs to discover locations of data needed by each node in pipeline. Adds additional data transfer and publication nodes needed to run pipeline on the grid. Resulting concrete workflows can be ex- ecuted on a wide vari- ety of grids using Condor and Globus. Coincidence Filter Filter Coincidence Filter Transfer Data Publish Data Transfer Data Filter Abstract Workflow Concrete Workflow Allows analysis to run on grids which do not have LIGO data. Time Algorithms are written in ANSI C89 for wide portability. Routines are contributed to LAL by LSC members. 6 FIG. 3: The top panel shows the sensitivity in MW halos NH of the search to the target population as a function of the loudest SNRρmax. The largest SNR observed in this analysis was ρ 2 max = 67.4 mean- ing that the search was sensitive to a fraction NH =0.95 MWH of the halo. The middle panel shows NH as a function of total mass M = m1+m2 of the injected signal. The error bars show the statis- tical error due to the finite number of injections in the Monte Carlo simulation. The lower panel shows NH as a function of the symmet- ric mass ratio η = m1m2/M 2 . We can see that the efficiency is a weak function of the total mass, as the amplitude of the inspiral sig- nal is a function of the total mass. The efficiency of the search does not depend strongly upon η. NH =0.95 MWH. The various contributions to the error in the measured value of detection efficiency are described in detail in [2]. In summary, the systematic errors are due to uncertainties in the instrumental response, errors in the wave- form due to differences between the true inspiral signal, and the finite number of injections in the Monte Carlo simulation. In this analysis, we neglect errors due to the spatial distribu- tion of the PBH binaries as studies show that the upper limit is relatively insensitive to the shape of the Milky Way halo. This is because the maximum range of all three detectors is greater than 50 kpc for PBH masses 0.2 M. The systematic errors will affect the rate through the measured SNR of the loudest event. We can see that the efficiency of the search depends very weakly on the SNR of the loudest event, again due to the range of the search compared to the halo radius. The statistical errors in the Monte Carlo analysis dominate the errors in NH. The combined error due to waveform mismatch and the cali- bration uncertainty is found to be O(10 4 ) MWH. The effects of spin were ignored both in the population and in the wave- forms used to detect inspiral signals. Estimates based on the work of Apostolatos[24] suggest that the mismatch between the signal from spinning PBHs and our templates will not sig- nificantly affect the upper limit. To be conservative, however, we place an upper limit only on non-spinning PBHs; we will address this issue quantatively in future analysis. Combining the errors in quadrature and assuming the downward excur- sion of NH to be conservative, we obtain an observational upper limit on the rate of PBH binary coalescence with com- FIG. 4: The shaded region shows rates excluded at 90% confidence by the observational upper limit on PBH binary coalescence pre- sented in this paper as a function of total mass M = m1 + m2 of the binary. The three points show the rates estimated using Eq. (5) for halo models S (M =1.58 M), F (M =0.44 M) and B (M =1.84 M) of [5]. ponent masses 0.21.0 Min the Milky Way halo to be R90% = 63 yr 1 MWH 1 . (4) By considering numerical simulations of three body PBH interactions in the early universe Ioka et. al. [15] obtain a probability distribution for the formation rate and coalescence time of PBH binaries. This depends on the PBH mass m, which we assume to be the MACHO mass. From this distri- bution, we may obtain an estimate of the rate of PBH coales- cence at the present time, given by R =1 × 10 13 M M m M32 37 yr 1 MWH 1 (5) where m is the MACHO mass and M is the mass of the halo in MACHOs, which is obtained from microlensing ob- servations. These measured values depend on the halo model used in the analysis of the microlensing results [25, 26]. The halo model in Eq. (1) corresponds to model S of the MACHO Collaboration [25]. The microlensing observations and PBH formation models assume a δ-function mass distribution, as does the rate estimate in Eq. (5). We can see from Fig 3 that our detection efficiency is not strongly dependent on the ra- tio of the binary masses η, and so we can marginalize over this parameter to obtain the rate as a function of total PBH mass M, which can be compared with the predicted rates from microlensing for different halo models. The analysis of 5.7 yrs of photometry of 11.9 million stars in the LMC sug- gests a MACHO mass of m =0.79 +0.32 0.24 and a halo MACHO mass M = 10 +4 3 × 10 10 Mfor halo model S [5]. Assum- ing all the MACHOs are PBHs, we obtain the rate estimate Glue provides an infrastructure to simplify the construction of workflows by treating data analysis applications as modules that can be chained together. Glue’s use of metadata (e.g. data quality information) allows complicated workflows to be easily constructed. Glue also contains certain LSC specific metadata clients and servers, such as data discovery tools. The figure shows the online analysis infrastructure which uses components of Glue. LAL algorithms include: Data conditioning, gravitational wave simulation, correlation of data with binary inspi- ral signals, excess power filters, etc. * = LALApps is a suite of stand-alone scientific data analysis applications built on top of LAL. This suite contains programs for analyzing interferometer data as well as manipulation of data products. It also contains code to abstract the scientific applications to allow pipeline construction in conjunction with Glue. 19 19.2 19.4 19.6 19.8 20 20.2 20.4 20.6 20.8 21 0 100 200 300 400 500 600 700 Seconds Frequency Excess Power Triggers (Start: 791599005 GPS Seconds) 100 200 300 400 500 600 700 A simulated inspiral analyzed with the LALApps excess power search. Duncan A. Brown for the LIGO Scientific Collaboration LIGO-G050334-00-Z

Grid Development and Deployment of LIGO Scientific ...OSG will be used for compute inten-sive workflows which are too large for the LSC grid. Condor and VDT Provides grid middleware

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Grid Development and Deployment of LIGO Scientific ...OSG will be used for compute inten-sive workflows which are too large for the LSC grid. Condor and VDT Provides grid middleware

Grid Development and Deployment of LIGO Scientific Collaboration Data Analysis Software

Onasys LSC Users

Pipeline Generation

Glue LALApps

LALLDR

LSC Data Grid / Open Science Grid

The ONline Analysis SYStem (onasys) software provides tools for automating the real-time analysis of gravitational wave data.

Start

Get segment list

Compute unused segments

Pause

segment database

used segments

Segmentsremain?

No Yes Generate DAGfor segment

Submit DAG tocompute cluster

Executes data analysis pipelines created by LSC scientists in real-time to process data from the three LIGO gravitational wave interferometers as it is taken.

Responsible for the execution and monitoring of data analysis pipelines which generate astrophysical results. Results are returned to the user or can be published for replication around the LSC data grid.

Built on top of the Grid LSC User Environment (Glue) and uses the scientific data analysis pipeline software deployed by the LSC.

Online monitor-ing of analysis is available via a web interface which queries job information metadata data-bases.

Onasys periodically, and on a short time-scale, performs the fol-lowing sequence:

Identify the data to analyze using a query to a GSI-authenticating instrument status and data quality server(LSCsegFind).

Locate the data using a query to a GSI-authenticating data location server (LDRdataFind).

Configure a data analysis pipeline using a user-supplied pipeline construction tool.

Execute the data analysis pipeline on the grid.

Users search for gravitational waves in LIGO data by running data analysis pipelines on the grid.

Simple text files contain the scientific parameters of the desired search.

Pipelines are run on the grid to produce scientific results

Complicated workflows can be constructed to performs the all steps necessary to search data from the four LSC detectors: the 3 LIGO interferometers and the GEO600 interferometer.

A subset of a binary inspiral workflow. Workflows used in analysis pipelines typically have over 10,000

Interferometer(s)

Backup FrameBuilder (fb0)

Primary FrameBuilder (fb1)

/fb0_frames(~ 2 days)

/frames(~ 1 week)

/archive/cluster

Publish /framesand state vector

mySQL+ RLS

LDR LDRdataFindServer

LSCsegFindServer

LDRmetadata,RLS, BOSS, LSCSegFindDatabases

Disk to Disk Copy

Publish R, RDSfrom /archive

Secure web server for online job

monitoring

gateway

ldas-gridonasysd (inspiral)

onasysd (power)

Condor scheddCondor negotiatorCondor collector

ObservatoryCondor pool

GSI sshd(user login)

Clusterhome

directories

Secure GSI Authenticated TCP

Internal Network TCP

UNIX Socket

Fiberchannel or NFS

Condor startd

The Grid LSC User Environment (Glue) provides a high level infrastructure for developing LSC grid applications and metadata services.

User applications which perform specific data analysis tasks as part of a pipeline.

Replicates LIGO and GEO data around the LSC grid.

LSC Algorithm Library (LAL) contains data analysis routines.

Provides underlying hardware and operating system software for executing LSC data analysis code.

LHO LLO

CIT

MIT UWM

PSU AEI

Maintains and distributes metadata about interferometer data.

Replicates data from LIGO observato-ries and between LSC compute sites.

Provides services to users and applica-tions for mapping metadata to physical data locations.

Built on top of GridFTP, MySQL, Globus RLS and pyGlobus.

Computing clusters at Caltech, MIT, LIGO Liv-ingston Observatory, LIGO Hanford Observa-tory, UWM, Penn State, AEI, Cardiff, University of Birmingham comprise the LSC Data Grid.

Online analysis is performed at the observa-tory sites (LHO and LLO). Offline analysis and follow-up of online results runs at other sites.

The LSC Data Grid provides a standard set of grid tools for users to execute pipelines.

Access to the LSC Data Grid is via ssh login or Globus job managers.

LSC Data Grid usage can be monitored using Ganglia.

DATAFIND

TMPLTBANK

INSPIRAL

SINCA THINCA

TRIGBANK

INSPVETO

THINCA2 thinca2_H1L1_1 thinca2_H2L1_1thinca2_slide_H1L1_1 thinca2_slide_H2L1_1

trigbank_L1_1

insp_L1_V_1

trigbank_L1_2

insp_L1_V_2

trigbank_L1_3

insp_L1_V_3

trigbank_L1_4

insp_L1_V_4

trigbank_L1_5

insp_L1_V_5

trigbank_L1_6

insp_L1_V_6

trigbank_L1_7

insp_L1_V_7

insp_H2_1

thinca_H2L1_1thinca_slide_H2L1_1

insp_H2_2

thinca_H1H2L1_1 thinca_slide_H1H2L1_1

insp_H2_3 insp_H2_4

trigbank_H2_1

insp_H2_V_1

trigbank_H2_2

insp_H2_V_2

trigbank_H2_3

insp_H2_V_3

trigbank_H2_4

insp_H2_V_4

trigbank_H2_5

insp_H2_V_5

thinca2_H1H2L1_1 thinca2_slide_H1H2L1_1

datafind_L_1

insp_L1_1insp_L1_2insp_L1_3insp_L1_4insp_L1_5insp_L1_6

tmplt_L1_1tmplt_L1_2tmplt_L1_3tmplt_L1_4tmplt_L1_5tmplt_L1_6

trigbank_H1_1

insp_H1_V_1

trigbank_H1_2

insp_H1_V_2

trigbank_H1_3

insp_H1_V_3

trigbank_H1_4

insp_H1_V_4

trigbank_H1_5

insp_H1_V_5

trigbank_H1_6

insp_H1_V_6

insp_H1_1insp_H1_2

thinca_H1L1_1thinca_slide_H1L1_1

insp_H1_3insp_H1_4

sinca_H1_1 sinca_L1_1

tmplt_H1_1tmplt_H1_2 tmplt_H1_3tmplt_H1_4 tmplt_H2_1tmplt_H2_2tmplt_H2_3 tmplt_H2_4

datafind_H_1 datafind_H_2Pipeline generation is built on top of Glue and the LALApps analysis codes.

Different analysis strategies can be pursued by constructing various work-flows.

Different pipelines are typically con-structed for online analysis, offline follow-up of online triggers, and offline analysis used in large scale Monte Carlo simulation and parameter tuning.

Pipelines are created as Condor DAGs or VDS abstract workflows (DAX)

Workflows constructed by Pegasus from the LSC workflows can be executed on the Open Science Grid.

OSG will be used for compute inten-sive workflows which are too large for the LSC grid.

Condor and VDTProvides grid middleware used to execute searches.

Condor provides the underlying infrastruc-ture for execution and management of workflows on the grid, either directly or through submission to Globus job manag-ers.

The LSC Data Grid client and server pack-ages are built on top of the VDT. The client and server provide the subset of the VDT used by the LSC and add LSC specific grid tools. Pacman makes this bundling and distribution simple.

Pegasus

Takes abstract workflows created by data analy-sis pipeline generators.

Queries metadata catalogs to discover locations of data needed by each node in pipeline.

Adds additional data transfer and publication nodes needed to run pipeline on the grid.

Resulting concrete workflows can be ex-ecuted on a wide vari-ety of grids using Condor and Globus. Coincidence

Filter Filter

Coincidence

Filter

Transfer Data

Publish Data

Transfer Data

Filter

Abstract Workflow

Concrete Workflow

Allows analysis to run on grids which do not have LIGO data.

Time

Algorithms are written in ANSI C89 for wide portability.

Routines are contributed to LAL by LSC members.

6

FIG. 3: The top panel shows the sensitivity in MW halosNH of thesearch to the target population as a function of the loudest SNRρmax.The largest SNR observed in this analysis wasρ2

max = 67.4 mean-ing that the search was sensitive to a fractionNH = 0.95 MWH ofthe halo. The middle panel showsNH as a function of total massM = m1 +m2 of the injected signal. The error bars show the statis-tical error due to the finite number of injections in the MonteCarlosimulation. The lower panel showsNH as a function of the symmet-ric mass ratioη = m1m2/M

2. We can see that the efficiency is aweak function of the total mass, as the amplitude of the inspiral sig-nal is a function of the total mass. The efficiency of the search doesnot depend strongly uponη.

NH = 0.95 MWH. The various contributions to the errorin the measured value of detection efficiency are described indetail in [2]. In summary, the systematic errors are due touncertainties in the instrumental response, errors in the wave-form due to differences between the true inspiral signal, andthe finite number of injections in the Monte Carlo simulation.In this analysis, we neglect errors due to the spatial distribu-tion of the PBH binaries as studies show that the upper limit isrelatively insensitive to the shape of the Milky Way halo. Thisis because the maximum range of all three detectors is greaterthan50 kpc for PBH masses≥ 0.2 M⊙. The systematic errorswill affect the rate through the measured SNR of the loudestevent. We can see that the efficiency of the search dependsvery weakly on the SNR of the loudest event, again due to therange of the search compared to the halo radius. The statisticalerrors in the Monte Carlo analysis dominate the errors inNH .The combined error due to waveform mismatch and the cali-bration uncertainty is found to beO(10−4) MWH. The effectsof spin were ignored both in the population and in the wave-forms used to detect inspiral signals. Estimates based on thework of Apostolatos[24] suggest that the mismatch betweenthe signal from spinning PBHs and our templates will not sig-nificantly affect the upper limit. To be conservative, however,we place an upper limit only on non-spinning PBHs; we willaddress this issue quantatively in future analysis. Combiningthe errors in quadrature and assuming the downward excur-sion of NH to be conservative, we obtain an observationalupper limit on the rate of PBH binary coalescence with com-

FIG. 4: The shaded region shows rates excluded at90% confidenceby the observational upper limit on PBH binary coalescence pre-sented in this paper as a function of total massM = m1 + m2

of the binary. The three points show the rates estimated using Eq. (5)for halo models S (M = 1.58 M⊙), F (M = 0.44 M⊙) and B(M = 1.84 M⊙) of [5].

ponent masses0.2–1.0 M⊙ in the Milky Way halo to be

R90% = 63 yr−1 MWH−1. (4)

By considering numerical simulations of three body PBHinteractions in the early universe Iokaet. al. [15] obtain aprobability distribution for the formation rate and coalescencetime of PBH binaries. This depends on the PBH massm,which we assume to be the MACHO mass. From this distri-bution, we may obtain an estimate of the rate of PBH coales-cence at the present time, given by

R = 1 × 10−13

(

M

M⊙

) (

m

M⊙

)−32

37

yr−1 MWH−1 (5)

wherem is the MACHO mass andM is the mass of thehalo in MACHOs, which is obtained from microlensing ob-servations. These measured values depend on the halo modelused in the analysis of the microlensing results [25, 26]. Thehalo model in Eq. (1) corresponds to model S of the MACHOCollaboration [25]. The microlensing observations and PBHformation models assume aδ-function mass distribution, asdoes the rate estimate in Eq. (5). We can see from Fig 3 thatour detection efficiency is not strongly dependent on the ra-tio of the binary massesη, and so we can marginalize overthis parameter to obtain the rate as a function of total PBHmassM , which can be compared with the predicted ratesfrom microlensing for different halo models. The analysis of5.7 yrs of photometry of11.9 million stars in the LMC sug-gests a MACHO mass ofm = 0.79+0.32

−0.24 and a halo MACHOmassM = 10+4

−3 × 1010 M⊙ for halo model S [5]. Assum-ing all the MACHOs are PBHs, we obtain the rate estimate

Glue provides an infrastructure to simplify the construction of workflows by treating data analysis applications as modules that can be chained together.

Glue’s use of metadata (e.g. data quality information) allows complicated workflows to be easily constructed.

Glue also contains certain LSC specific metadata clients and servers, such as data discovery tools.

The figure shows the online analysis infrastructure which uses components of Glue.

LAL algorithmsinclude:

Data conditioning, gravitational wave simulation, correlation of data with binary inspi-ral signals,excess power filters, etc.

=

LALApps is a suite of stand-alone scientific data analysis applications built on top of LAL.

This suite contains programs for analyzing interferometer data as well as manipulation of data products.

It also contains code to abstract the scientific applications to allow pipeline construction inconjunction with Glue.

19 19.2 19.4 19.6 19.8 20 20.2 20.4 20.6 20.8 210

100

200

300

400

500

600

700

Seconds

Freq

uenc

y

Excess Power Triggers (Start: 791599005 GPS Seconds)

100

200

300

400

500

600

700

A simulated inspiral analyzed with the LALApps excess power search.

Duncan A. Brown for the LIGO Scientific Collaboration

LIGO-G050334-00-Z