1 / 24

LHC Computing and Analysis Workshop

LHC Computing and Analysis Workshop. Physics Cases - ATLAS 25-26 August 2005 CSCS Manno. HansPeter Beck LHEP – Uni Bern. ATLAS Cavern status. 28 July 2005. Swiss Contribution to ATLAS. Semiconductor Tracker Barrel Support Structure Forward Modules Geneva. Calorimeter

pinnes
Download Presentation

LHC Computing and Analysis Workshop

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. LHC Computing and Analysis Workshop Physics Cases - ATLAS 25-26 August 2005 CSCS Manno HansPeter Beck LHEP – Uni Bern

  2. ATLAS Cavern status 28 July 2005

  3. Swiss Contribution to ATLAS Semiconductor Tracker Barrel Support Structure Forward Modules Geneva Calorimeter Readout Electronics Geneva Trigger & DAQ Dataflow High Level Triggers Bern Barrel Toroid Superconductor Coil Casings Bern and Geneva

  4. n n   e e µ µ p p n n n - e e q µ + - c 1 µ - ~ q q Z ~ p g H p p p ~ q Z +  µ + - q ~  c 0 µ 2 - ~ c 0 1 p-p Collisions @ LHC Nominal LHC Parameters: 7 TeV Proton Energy 1034cm-2s-1 Luminosity 2808 Bunches per Beam 1011 Protons per Bunch 0.6 A 25ns 7.5m Bunch Crossings 4x107 Hz Proton-Proton Collisions 109 Hz Quark/Gluon Collisions Heavy particle production 10+3…-7 Hz(W, Z, t, Higgs, SUSY,…)

  5. OFF-line ON-line Level-1 Trigger 40 MHz Hardware (ASIC, FPGA) Massive parallel Architecture Pipelines LHC is a Standard Model Factory bb ~2 MHz W, Z ~1 kHz tt ~2 Hz Rate (Hz) QED - 8 10 Level-2 Trigger 100 kHz s/w PC farm Local Reconstruction - 6 10 2 µs 4 10 1 sec W,Z 2 10 Top 10 ms Reconstruction & Analyses Z * TIER0/1/2/3/4 0 10 Centres Higgs -2 10 Level-3 Trigger 3 kHz s/w PC farm Full Reconstruction -4 10 New Physics – the real raison d’être of the LHC HSM , HSM  4,… SUSY ~mHz-Hz Leptoquarks sec 25ns 3µs ms year hour 10 10 10 -9 -6 -3 10 10 10 sec 0 3 6 Physics Rates Event rate  Precisions Measurements of W Mass to 15 MeV i.e. 0.2‰ Top Mass to 2 GeV i.e. 1.2% Level-1  Level-2  Mass Storage  B-Physics Mass amounts of B0sJ/yf, J/yh. Potential to do research on rare decays B0dK0*m+m- ,B0dr0m+m- ,B0sf0m+m- B0s Mixing Offline Analyses

  6. Physics Selection Strategy • ATLAS has an inclusive trigger strategy • LVL1 Triggers on individual signatures / objects • EM / Had Cluster • Total Energy • Missing Energy • Muon track • LVL2 confirms & refines LVL1 signature • seeding of LVL2 with LVL1 result – i.e. Region of Interest [RoI] • EventFilter confirms & refines LVL2 signature • seedig of EventFilter with LVL2 result • tags accepted events according to physics selection • Offline Analysis is based on trigger samples • an individual analysis will always run over a (tag) of events • need to understand trigger object selection efficiencies

  7. LVL1 - Muons & Calorimetry Toroid Muon Trigger looking for coincidences in muon trigger chambers 2 out of 3 (low-pT; >6 GeV) and 3 out of 3 (high-pT; > 20 GeV) Triger efficiency 99% (low-pT) and 98% (high-pT) Calorimetry Trigger looking for e/g/t + jets • Various combinations of cluster sums and isolation criteria • SETem,had , ETmiss

  8. LVL1 Trigger Rates LVL1 rate is dominated by electromagnetic clusters: 78% of physics triggers

  9. Inclusive Higher Level Trigger Event Selection HLT rate reduces e/g a lot: 45% of physics triggers

  10. DAQ/HLT/TIER0-1-2-3 are all based on PC Farms

  11. Inner Detector Channels Fragment size - kB Muon Spectrometer Channels Fragment size - kB Pixels 1.4x108 60 MDT 3.7x105 154 SCT 6.2x106 110 CSC 6.7x104 256 TRT 3.7x105 307 RPC 3.5x105 12 TGC 4.4x105 6 Calorimetry Channels Fragment size - kB Trigger Channels Fragment size - kB LAr 1.8x105 576 LVL1 28 Tile 104 48 ATLAS Event Size Affordable mass storage b/w: 300 MB/sec  200 Hz ☛ 3 PB/yearof carefully triggered data for offline analysis Atlas event size: 1.5 MBytes140 million channelsorganized into ~1600 Readout Links

  12. Raw data (“byte stream”, what the DAQ writes) Raw data (“byte stream”, what the DAQ writes) (1.5 MB/event) (1.5 MB/event) reconstruction reconstruction ESD (Event Summary Data) ESD (Event Summary Data) (0.5 MB/event) (0.5 MB/event) AOD production AOD production back navigation back navigation back navigation back navigation AOD (Analysis Object Data) AOD (Analysis Object Data) (10 kB/event) (10 kB/event) analysis in ATHENA analysis in ATHENA analysis in ROOT n-tuple final histograms final histograms How to Analyze 3 PB/year while still getting the Physics out? Today: Joe Physicist analyzing ROOT n-tuples TIER 0/1 TIER 0/1 TIER 2/3 TIER 2/3/4 Tomorrow: Joe Physicist analyzing AODs directly (using ATHENA, Python, ROOT,…) today still too slow to be practical TIER 2/3/4

  13. Analysis of ATLAS data (cont) • What Joe Physicist (aka Heiri/Henri Füsikus, Vreni/Giselle Füsika, ) will do • Joe is organized in a physics working group • e.g. Higgs • Joe is looking at a specific/inclusive channel • e.g. Higgs  four electrons • Joe needs to understand how electrons are • triggered • LVL1, HLT • reconstructed • efficiencies • fakes • receive there measured properties • 4vector; i.e. energy and 3 momenta • error matrix (correlations) • Joe runs over many different data samples • real data on tight trigger selection (full statistics) • Background samples (i.e. complementarory trigger samples; e.g. jet-samples) • real data on loosened trigger samples (full statistics not needed – need to keep the systematical errors under control) • Monte Carlo data (needs to be generated first, simulated, reconstructed) • MC produced ATLAS wide via Higgs group • Special dedicated MC samples produced by Joe • amount defined by precision needed. balance statistical and systematica error • With increasing understanding of his task, Joe will do this on • ntuples full statistics • AODs full statistics • ESD maybe full statistics –but more likely only a moderate subset of ESD needed • RAW only a small subset a RAW needed

  14. Raw data (“byte stream”, what the DAQ writes) (1.5 MB/event) reconstruction ESD (Event Summary Data) (0.5 MB/event) AOD production back navigation back navigation AOD (Analysis Object Data) (10 kB/event) analysis in ATHENA final histograms Analysis of ATLAS data (cont) TIER 0/1 • Final Analysis of Joe Physicist focused on TIER 2/3/4 centres • TIER4 (Laptop): data presentation, job preparation, coding of user specific code • TIER3/4: running over small samples • skimmed data providing full statistic (specialized AODs or ntuples) • test samples (AOD, ESD, RAW) small fraction for debugging and testing • producing some dedicated (Joe’s) Monte Carlo data (gen, sim, rec, ESD, AOD) • TIER2: running over big samples • full statistics of Joe’s pre-selection (AOD and sometimes even ESD) needs to be available • real data and Monte Carlo data • fast data movement from TIER1, enough storage at TIER2 • Joe producing a lot of dedicated Monte Carlo data (gen, sim, rec, ESD, AOD) • TIER 1/2 • ATLAS physics groups decide on global production needs • i.e. production of a very big Higgs sample can be proposed by Joe – but priorities will be decided ATLAS wide • Joe Physicist is only competitive if these TIERs give him high priorities for his jobs. I.e. do not try to do any of above on lxplus at Cern • There is not only Joe there, also Vreni and Heiri have their own and independent analysis… TIER 0/1 TIER 2/3/4

  15. Swiss ATLAS Groups will do…. • Exploiting ATLAS Physics potential • based on tagged AOD (and ESD) data • specific trigger samples, but full statistics • every physicist has his own preferred trigger samples • Higgs, SUSY, Beyond Standard Model, Standard Model • Heavy Ion ?? • Detector studies • based on AOD, ESD and RAW data • calibration, alignment, efficiencies • events from calibration stream • often, full statistic sample not needed (a sub-sample is often enough= • SCT & LAr • more systems will be needed in the course of analyzing data • RAW data will just be accessed on a small sub-sample of the events available – but needed for debugging/improving of reconstruction code • Trigger studies • based on AOD and ESD data • efficiencies • specific trigger samples and control samples • needed for LVL1, LVL2 and EF • Based on Real Data and Monte Carlo Data • Real Data needs to be processed and re-processed (calibration, conditions and alignment) • Monte Carlo Data needs to be produced and processed and re-processed (calibration, conditions, alignment) TIER2/3 TIER2/3 TIER2/3 TIER 1/2/3

  16. Swiss ATLAS Analysis Actual Examples and Experiences… • Monte Carlo based studies of • Offline and Trigger analyses • HZZ*  4e • H  ZZ*  2e2 • H  WW*  2e2 via VBF • SUSY • Participation in DataChallenges for MC production • sparticle masses (edges) at various mSUGRA points. Including stop and stau coannihilation; bulk region; etc. • Real Data studied from Combined Test Beam • LAr energy calibration • electron/pion separation • track reconstruction

  17. … and Problems seen on Phoenix, Ubelix,… • Releases installation • Many flavors of ATLAS s/w on the market, caused confusion • libraries (system + ATLAS specific) were missing • differences seen between OS versions • Reco (ESD) 9.0.4 went to infinite loop with SL3 build, Not with RH73 build (under SL3!) (Phoenix) • pythom boost lib libboost_python-gcc.so changed by admin triggered crashes on 32 bit machines • /usr/bin/time directory missing in some Ubelix nodes • Grid • NFS problems with nordugrid/Ubelix front end. (NFS failed also on some Phoenix nodes) • Problems in handling too many jobs submission, even if below the (500) limit(lheppc10). • #time limit of 500’ from some hidden default settings in SGE scheduler, fixed (Ubelix). • gridftp server problems (various, wrong permission ) (Ubelix) • Oracle DB at cern get stuck (too many requests) • Atlas Software • Memory leak in reco, memory exceeds requirements (1GB). Problem in 9.0.4. Still not solved in • 10.0.1. Workaround: require 1.5Gb to reco 50 evt. • Tauola bug (prodution restarted from scratch) • Hardware • PCs that did not supported the load, over-heated, etc.

  18. What Joe Physicsist Expectsapplies also to Heiri/Henri Phüsikus and Vreni/Giselle Phüsika • Analysis of ATLAS data is a complex procedure • need to go not only over ntuples or AOD data • but some access needed to ESD and RAW data • easy use of middleware • i.e. Joe Physicist doesn’t even want to know about • available ATLAS s/w releases • more than one in parallel • Joe may need version x while Vreni wants y and Henri z • local Database access • condition, calibration, alignment data • infrastructure robust against s/w still under development • memory leaks, infinite loops, crashes • flexible environment • transparent use of TIER 4, TIER 3 and TIER 2 usage • Joe does not want to learn every time a new environment • availability of data • at various TIERs • Independence from Cern lxplus/lxshare • provided there are enough resources • CPU, storage, bandwidth

  19. Conclusions • ATLAS is well on track • First data can be expected by mid 2007 • However • Combined Testbeam data already now • Cosmics data starting up now (individual detector components) • Monte Carlo data existing now • Can start implementing and applying policies alread now • A flexible and robust TIER 2/3/4 system needed for individual analysis • otherwise no impact of Swiss physicists to LHC exploitation possible • usable resources (CPU, storage, bandwidth) for Swiss physicists concentrated in TIER2 (i.e. Manno) fits well with analysis procedure presented here • need for TIER3 (i.e. institutes cluster) • for fast turn around in development and debug cycles • final analysis over skimmed data sets • Need to establish policies on how to share and use resources like the Manno cluster • this workshop

  20. BackUp Slides

  21. From Bunch Crossings to Physics Analyses OFF-line ON-line Level-1 Trigger 40 MHz Hardware (ASIC, FPGA) Massive parallel Architecture Pipelines Rate (Hz) QED Event rate  8 10 Level-2 Trigger 100 kHz s/w PC farm Local Reconstruction 6 10 Level-1  2 µs 4 10 Level-2  1 sec W,Z Mass Storage  2 10 Top 10 ms Reconstruction & Analyses Z * TIER0/1/2/3/4 0 10 Centres Higgs -2 10 Level-3 Trigger 3 kHz s/w PC farm Full Reconstruction -4 10 Offline Analyses sec 25ns 3µs ms year hour 10 10 10 -9 -6 -3 10 10 10 sec 0 3 6

  22. LVL1 decision made with calorimeter data with coarse granularity and muon trigger chambers data. Buffering on detector LVL2 uses Region of Interest data (ca. 2%) with full granularity and combines information from all detectors; performs fast rejection. Buffering in ROBs EventFilter refines the selection, can perform event reconstruction at full granularity using latest alignment and calibration data. Buffering in EB & EF ATLAS Three Level Trigger Architecture 2.5 ms ~10 ms ~ sec.

  23. 40 MHz 1 PB/s FE Pipelines 75 kHz ROD ROD ROD RoI Read-Out Drivers 120 GB/s 120 GB/s RoI data = 1-2% Read-Out Links ROS ROB ROB ROB RoI requests ~ 10 ms LVL2 Read-Out Buffers ROIB L2SV RoI Builder L2 Supervisor Read-Out Sub-systems L2N L2P L2 N/work L2 Proc Unit ~2 kHz ~2+4 GB/s Lvl2 acc = ~2 kHz Dataflow Manager EB EBN Event Building N/work ~ sec Event Filter SFI Sub-Farm Input DFM ~4 GB/s Event Builder EFN Event Filter Processors Event Filter N/work EFP EFP EFP EFP EFacc = ~0.2 kHz SFO Sub-Farm Output ~ 200 Hz ~ 300 MB/s ATLAS TDAQ Architecture Trigger DAQ Calo MuTrCh Other detectors 40 MHz specialized h/w ASICsFPGA LV L1 D E T R/O 2.5 ms Lvl1 acc = 75 kHz H L T D A T A F L O W

  24. Resources needed for Simulating 100k SUSY events • Generation • 100k events: 20 files of 5k evgen events • SIM+DIGI+REC: File size: 50 events/files (limited by memory leak; i.e. s/w bugs)20 x 100 jobs, about 10h each • CPU • Time to process 1 event:650 sec on a 2.8 GHz processor • GEN 0.5 kSI2-sec • SIM+DIGI 100 kSI2-sec • REC 15 kSI2-sec • Analyze 0.5 kSI2-sec • 100k hours GHz –or- 4.5k day GHz – or- 15 days on a 100x3GHz cluster • DISK SPACE • Event size • Generated : 0.07 MB/evt • Simulated and digitalized (raw data): 2.1 MB/evt • Reconstructed (ESD) 0.9MB/evt AOD’s 0.025 MB/evt • Total 3MB/event • Need about 300GB to store 100k events

More Related