COMPASS offline computing n the COMPASS experiment n

  • Slides: 21
Download presentation
COMPASS off-line computing n the COMPASS experiment n the analysis model n the off-line

COMPASS off-line computing n the COMPASS experiment n the analysis model n the off-line system hardware n software n CHEP 2000 *

The COMPASS Experiment (Common Muon and Proton Apparatus for Structure and Spectroscopy) 1. Fixed

The COMPASS Experiment (Common Muon and Proton Apparatus for Structure and Spectroscopy) 1. Fixed target experiment at the CERN SPS 2. approved in February 1997 3. commissioning from May 2000 4. data taking for at least 5 years 5. collaboration about 200 physicists from Europe and Japan q diversified physics programme n muon beam gluon contribution to nucleon spin, quark spin distribution functions n hadron beam glueballs, charmed baryons, Primakoff reactions all measurements require high statistics CHEP 2000

Experimental Apparatus Two stage spectrometer (LAS, SAS) Several new detectors GEMs, micro. Mega, straw

Experimental Apparatus Two stage spectrometer (LAS, SAS) Several new detectors GEMs, micro. Mega, straw trackers, scintillating fibers, RICH, and silicon detectors, Calorimeters, Drift and MWP Chambers (440 K electronic channels) Not an easy geometry highly inhomogeneous magnetic field (SM 1, PTM) CHEP 2000

Expected Rates 1. beam intensity: 108 muons/s with duty cycle of 2. 4 s/14

Expected Rates 1. beam intensity: 108 muons/s with duty cycle of 2. 4 s/14 s 2. RAW event size: ~ 20 - 30 k. B 3. trigger rate: 104 events/spill DAQ designed for 105 events/spill (hadron programme) on-line filtering 4. continuous data acquisition flux: 35 MB/s 5. data taking period ~100 days/year: ~ 1010 events/year, ~ 300 TB/year of RAW data CHEP 2000

COMPASS analysis model The RAW data l l will be stored at CERN (no

COMPASS analysis model The RAW data l l will be stored at CERN (no copy foreseen) and have to be accessible during all the experiment lifetime will be processed at CERN, in parallel to and at the same speed of data acquisition assuming no pre-processing for calibrations ~ 1 reprocessing of the full data set processing time 2 SPECint 95 -sec/event n calibrations “on-line”, powerful on- off-line monitoring, small data subset reprocessing if fast raw data access the needed CPU power is 2000 Spec. Int 95 (~ 20 000 CU) Physics analysis will be performed at the home institutes, as well as specific studies and MC production n CHEP 2000 the relevant sets of data must have a much smaller size remote and concurrent access to raw data important (“PANDA” model…)

General choices In 1997 COMPASS decided to l build a completely new software system

General choices In 1997 COMPASS decided to l build a completely new software system l use OO techniques m m C++ as programming language ODB to store the data. Given the short time scale, the ‘small’ collaboration, the novelty, and the well known difficulty of the tasks, it was mandatory to l l l collaborate with the IT division foresee the use LHC++ and commercial products (Hep. ODBMS, Objectivity/DB) look at other developments (ROOT) CHEP 2000

Off-line system Hardware n n central data recording COMPASS Computing Farm (CCF) (see M.

Off-line system Hardware n n central data recording COMPASS Computing Farm (CCF) (see M. Lamanna presentation, Feb. 7, session E) Software n n CHEP 2000 Data structures and access CORAL (Compass Reconstruction and Ana. Lysis) program

Central data recording (CDR) updated version of the CERN Central Data Recording (CDR) scheme

Central data recording (CDR) updated version of the CERN Central Data Recording (CDR) scheme the on-line system o performs the event building (and filtering) - ALICE DATE system o writes RAW data on local disks files in byte-stream format (10 -20 parallel streams), keeping a "run" structure (typical sizes of 50 GB) the Central Data Recording system o transfers the files to the COMPASS Computing Farm, at the computer center (rate of 35 MB/s) the COMPASS Computing Farm CCF o formats the data into a federated database (Objectivity/DB) converts the RAW events in simple persistent objects o o performs fast event tagging or clusterisation (if necessary) sends the DB to HMS for storage CHEP 2000

COMPASS Computing Farm (CCF) Beginning 1998: n IT/PDP Task Force: computing farms for high-rate

COMPASS Computing Farm (CCF) Beginning 1998: n IT/PDP Task Force: computing farms for high-rate experiments (NA 48, NA 45, and COMPASS). Proposed model for the CCF: hybrid farm with n about 10 Proprietary Unix servers (“data servers”) n about 200 PCs (”CPU clients”), 2000 SPECint 95 (0. 2 s/ev) n 3 to 10 TB of disk space Present model: farm with n PCs as “data servers” and ”CPU clients” order of 100 dual PIII machines standard PCs running CERN certified Linux (now: RH 5. 1 with kernel 2. 2. 10/12) CHEP 2000

CCF CHEP 2000

CCF CHEP 2000

COMPASS Computing Farm (cont. ) The data servers will m m handle the network

COMPASS Computing Farm (cont. ) The data servers will m m handle the network traffic from the CDR, format the RAW events into a federated DB , and send them to the HSM and m m m receive the data to be processed from the HSM, if needed, distribute the RAW events to the PCs for reconstruction receive back the output (persistent objects), and send it to the HSM. The CPU clients will m process the RAW events (reconstruction of different runs/files has to run in parallel) a real challenge: 1000 ev/sec to be stored and processed by 100 dual PCs tests with prototypes are going on since two years; good results CHEP 2000

Off-line system Software l Data structures m m l CHEP 2000 Event DB Experimental

Off-line system Software l Data structures m m l CHEP 2000 Event DB Experimental conditions DB Reconstruction quality control monitoring DB MC data CORAL: Compass Reconstruction and Ana. Lysis program

Data structures Event DB n event headers containers: small dimensions (on disk), basic information

Data structures Event DB n event headers containers: small dimensions (on disk), basic information like tag, time, . . . n n q RAW event containers: just one object with event (DATE) buffer reconstructed data containers: objects for physics analysis direct access to objects run: file structure not visible q association to avoid duplications q q direct: raw - reco. data via “time”: raw - mon. ev CHEP 2000

Data structures (cont. ) Experimental conditions DB n includes all information for processing and

Data structures (cont. ) Experimental conditions DB n includes all information for processing and physics analysis (on-line calibrations, geometrical description of the apparatus. . . ) n based on CERN porting of BABAR Condition Database package (included in Hep. ODBMS) n versioning of objects n access to valid information using event time Reconstruction quality control monitoring data n n includes all quantities needed for monitoring the stability of the reconstruction and of the apparatus performances stored in Objectivity/DB Monte Carlo data n n we are using Geant 3 (Geant 4: under investigation, not in the short term) ntuples, Zebra banks CHEP 2000

status Event DB version 1 ready Experimental conditions DB in progress: implementation started Reconstruction

status Event DB version 1 ready Experimental conditions DB in progress: implementation started Reconstruction quality control monitoring data starting Monte Carlo data ready CHEP 2000

CORAL COmpass Reconstruction and Ana. Lysis program 1. fully written in C++, OO technique

CORAL COmpass Reconstruction and Ana. Lysis program 1. fully written in C++, OO technique 2. 3. modular architecture with 4. a framework providing all basic functionalities m m well defined interfaces for all components needed for event reconstruction insulation layers for all "external" packages m access to the experimental conditions and event DB (reading and writing persistent objects) - Hep. ODBMS to assure flexibility in changing both reconstruction components and external packages m l components for event reconstruction developed in parallel detector decoding, pattern rec. in geom. regions, track fit, RICH and Calorimeter rec. , … CHEP 2000

CORAL CHEP 2000

CORAL CHEP 2000

CORAL status 1. development and tests on Linux 2. we try to keep portability

CORAL status 1. development and tests on Linux 2. we try to keep portability on other platforms (Solaris) 3. framework: almost ready work going on to interface new reconstruction components and access to experimental conditions DB 4. reconstruction components: 1. integrated inside CORAL and tested MC event reading and decoding, track pattern recognition, track fit, … 2. integration foreseen soon RICH pattern recognition, Calorimeter reconstruction, vertex fit, . . . 3. under development detector (DATE buffer) decoding, in parallel with on-line, . . . Goal: version 1 ready at the end of April 2000 all basic functionalities, even if not optimised as for all other off-line system components CHEP 2000

General comments most of the problems we had are related to the fact that

General comments most of the problems we had are related to the fact that we are still in a transition period: q no stable environment both for available software (LHC++) and OS (Linux) q lack of standard “HEP made” tools and packages commercial products seem not to be always a solution q too few examples of HEP software systems using new techniques q expertise and resources having a large number of physicists knowing the new programming language (and techniques) requires time all the work has been done by a very small enthusiastic team (3 to 10 fte in 2 years) Still, we think we made the right choice CHEP 2000

from the minutes of the 16 th meeting of FOCUS held on December 2,

from the minutes of the 16 th meeting of FOCUS held on December 2, 1999: “FOCUS …. recognises the role that the experiment has as a "test-bed" for the LHC experiments. ” CHEP 2000

CHEP 2000

CHEP 2000