20
Event reconstruction in the LHCb Online Farm Markus Frank (CERN) & Albert Puig (UB)

Event reconstruction in the LHCb Online Farm

  • Upload
    nascha

  • View
    37

  • Download
    5

Embed Size (px)

DESCRIPTION

Markus Frank (CERN) & Albert Puig (UB). Event reconstruction in the LHCb Online Farm. An opportunity (Motivation) Adopted approach Implementation specifics Status Conclusions. Outline. The LHCb Online Farm: schematics. Readout Network. Online cluster (event selection). Data logging - PowerPoint PPT Presentation

Citation preview

Page 1: Event reconstruction in the LHCb Online Farm

Event reconstruction in the LHCb Online FarmMarkus Frank (CERN) & Albert Puig (UB)

Page 2: Event reconstruction in the LHCb Online Farm

Outline An opportunity (Motivation) Adopted approach Implementation specifics Status Conclusions

2

Page 3: Event reconstruction in the LHCb Online Farm

The LHCb Online Farm: schematics

Data loggingfacility

Readout Network

Switch

Switch

Switch

CPU

CPU

CPU

CPU

CPU

CPU

CPU

CPU

CPU

Online cluster(event selection)

Storage

3

Page 4: Event reconstruction in the LHCb Online Farm

The LHCb Online Farm: numbers

4

~16000 CPU cores foreseen (~1000 boxes) Environmental constraints:

2000 1U boxes space limit 50 x 11 kW cooling/power limit

Computing power equivalent to that provided by all Tier 1’s to LHCb.

Storage system: 40 TB installed 400-500 MB/s

Data loggingfacility

Readout Network

Switch

Switch

Switch

CPU

CPU

CPU

CPU

CPU

CPU

CPU

CPU

CPU

Online cluster(event selection)

Storage

Page 5: Event reconstruction in the LHCb Online Farm

An opportunity Significant idle time of the farm

During LHC winter shutdown (~ months) During beam period, experiment and

machine downtime (~ hours)

5

Could we use it for reconstruction? Farm is fully LHCb controlled

Good internal network connectivity- Slow disk access (only fast for a very few nodes via

Fiber Channel interface)

Page 6: Event reconstruction in the LHCb Online Farm

Design considerations Background information:

+ 1 file (2GB) contains 60.000 events.+ It takes 1-2s to reconstruct an event.

Cannot reprocess à la Tier-1 (1 file per core) Cannot perform reconstruction in short idle periods:

Each file takes 1-2 s/evt * 60k evt ~ 1 day. Insufficient storage or CPUs not used efficiently:

Input: 32 TB (16000 files * 2 GB/file) Output: ~44 TB (16000 * 60k evt * 50 kB/evt)

A different approach is needed Distributed reconstruction architecture.

6

Page 7: Event reconstruction in the LHCb Online Farm

Adopted approach: parallel reconstruction Files are split in events and distributed to

many cores, which perform reconstruction: First idea: full parallelization (1 file/16k cores)

Reconstruction time: 4-8 s Full speed not reachable (only one file open!)

Practical approach: split the farm in slices of subfarms (1 file/n subfarms).

Example: 4 concurrent open files yield a reconstruction time of 30s/file.

7

Page 8: Event reconstruction in the LHCb Online Farm

Implementation

8

ECSControl and allocation of resources

Reco Manager

Database

Job steering

DIRACConnection to LHCb production system

Storage

Switch

Storage Nodes

Subfarms

… …

… …

Farm

See A.Tsaregorodtsev talk, DIRAC3 - the new generation of the LHCb grid software

Page 9: Event reconstruction in the LHCb Online Farm

Resource management

9

ECSControl and allocation of resources

Farm

Storage

Switch

Subfarms

… …

… … Reco Manager

Database

Job steering

DIRACConnexion to production system

Storage Nodes

Page 10: Event reconstruction in the LHCb Online Farm

Farm and process control

10

Control using standard LHCb ECS software: Reuse of existing components for storage and subfarms. New components for reconstruction tree management. See Clara Gaspar’s talk (LHCb Run Control System).

Allocate, configure, start/stop resources (storage and subfarms).

Task initialization slow, so tasks don’t restart on file change. Idea: tasks sleep during data-taking, and are only

restarted on configuration change.

Page 11: Event reconstruction in the LHCb Online Farm

subfarm

Farm Control Tree

11

PVSS Control

x 50 subfarms1 control PC each4 PC each

x 8 cores/PC1 Reco task/coreData management tasks

Reco

RecoRecoReco

Reco Reco Reco

Reco Output

Input

Event (data) processing

Event (data) management

Page 12: Event reconstruction in the LHCb Online Farm

Target Node

The building blocks

12

Consumer

Producer

Buffer manager

Processing Node Data processing block Producers put events in a buffer

manager (MBM) Consumers receive events from the

MBM

Buffer

Sender

Receiver

Data transfer block Senders access events from MBM Receivers get data and declare it

to MBM

Source Node

Buffer

Page 13: Event reconstruction in the LHCb Online Farm

Putting everything together…

13

Storage

Storage Reader

Receiver

BrunelBrunelBrunelReco

Storage Writer

Sender

Storage nodes

Worker node

1 per coreOutput

Events Events

Sender Receiver

Input

Page 14: Event reconstruction in the LHCb Online Farm

Reconstruction management

14

ECSControl and allocation of resources

Reco Manager

Database

Job steering

DIRACConnection to LHCb production system

Storage

Switch

Subfarms

… …

… …

Farm

Storage Nodes

Page 15: Event reconstruction in the LHCb Online Farm

Task management

15

Granularity to file level. Individual event flow handled

automatically by allocated resource slices.

Reconstruction: specific actions in specific order. Each file is treated as a Finite

State Machine (FSM) Reconstruction information

stored in a database. System status Protection against crashes

PREPARED

PROCESSING

PREPARING

TODO

DONE ERROR

Page 16: Event reconstruction in the LHCb Online Farm

Job steering: the Reco Manager

16

Job steering done by a Reco Manager: Holds the each FSM instance and moves it

through all the states based on the feedback from the static resources.

Sends commands to the readers and writers: files to read and filenames to write to.

Interacts with the database.

Page 17: Event reconstruction in the LHCb Online Farm

Connection to the LHCb Production System

17

The Online Farm will be treated like a CE connected to the LHCb Production system.

Reconstruction is formulated as DIRAC jobs, and managed by DIRAC WMS Agents.

DIRAC interacts with the Reco Manager through a thin client, not directly with the DB.

Data transfer in and out of the Online farm managed by DIRAC DMS Agents.

Page 18: Event reconstruction in the LHCb Online Farm

Status and tests

18

Current performance constrained by hardware. Reading from disk: ~130 MB/s.

FC saturated with 3 readers. Reader saturates CPU at 45MB/s.

Test with dummy reconstruction Just copy data from input to output

Stable throughput 105 MB/s Constrained by Gigabit network

Upgrade to 10 Gigabit planned

Storage

Storage Reader

Receiver

Reco

Storage Writer

Sender

Sender Receiver

Resource handling and Reco Manager implemented.

Integration in LHCb Production system recently decided, not implemented yet.

Page 19: Event reconstruction in the LHCb Online Farm

Future plans

19

Software Pending implementation of thin client for

interfacing with DIRAC. Hardware

Network upgrade to 10 Gbit in storage nodes before summer.

More subfarms and PCs to be installed. From current ~4800 cores to the planned

16000.

Page 20: Event reconstruction in the LHCb Online Farm

Conclusions

20

The LHCb Online cluster needs huge resources for data event selection of LHC collisions.

These resources have much idle time (50% of the time).

They can be used on idle periods by applying a parallelized architecture to data reprocessing.

A working system is already in place, pending integration in the LHCb Production system.

Planned hardware upgrades to meet DAQ requirements should overcome current bandwidth constraints.