230 likes | 312 Views
The LHC Computing Grid A World-Wide Computer Centre. Pour St éphanie Benz Journaliste – l’Expansion. Fr édéric Hemmer Deputy IT Department Head April 13, 2007. The LHC Accelerator.
E N D
The LHC Computing GridA World-Wide Computer Centre Pour Stéphanie BenzJournaliste – l’Expansion Frédéric HemmerDeputy IT Department HeadApril 13, 2007 Frédéric Hemmer, CERN, IT Department
The LHC Accelerator The accelerator generates 40 million particle collisions (events) every second at the centre of each of the four experiments’ detectors Frédéric Hemmer, CERN, IT Department
Which are recorded on disk and magnetic tapeat 100-1,000 MegaBytes/sec ~15 PetaBytes per year for all four experiments LHC DATA This is reduced by online computers that filter out a few hundred “good” events per sec. Frédéric Hemmer, CERN, IT Department
simulation CERN Data Handling and Computation for Physics Analysis reconstruction event filter (selection & reconstruction) detector analysis processed data event summary data raw data batch physics analysis event reprocessing analysis objects (extracted by physics topic) event simulation interactive physics analysis les.robertson@cern.ch Frédéric Hemmer, CERN, IT Department
The LHC Data Challenge • The accelerator will be completed in 2007 and run for 10-15 years • Experiments will produce about15 Million Gigabytes of data each year (about 20 million CDs!) • LHC data analysis requires a computing power equivalent to ~100,000 of today's fastest PC processors • Requires many cooperating computer centres, as CERN can only provide ~20% of the capacity Frédéric Hemmer, CERN, IT Department
Solution: the Grid • Use the Grid to unite computing resources of particle physics institutes around the world • The World Wide Web provides seamless access to information that is stored in many millions of different geographical locations • The Grid is an infrastructure that provides seamless access to computing power and data storage capacity distributed over the globe Frédéric Hemmer, CERN, IT Department
Tier-0 – the accelerator centre • Data acquisition & initial processing • Long-term data curation • Data Distribution to Tier-1 centres Tier-1 – “online” to the data acquisition process high availability • Managed Mass Storage – grid-enabled data service • All re-processing passes • Data-heavy analysis • National, regional support Canada – Triumf (Vancouver) France – IN2P3 (Lyon) Germany –Karlsruhe Italy – CNAF (Bologna) Netherlands – NIKHEF/SARA (Amsterdam) Nordic countries – distributed Tier-1 Spain – PIC (Barcelona) Taiwan – Academia SInica (Taipei) UK – CLRC (Oxford) US – FermiLab (Illinois) – Brookhaven (NY) LCG Service Hierarchy Tier-2 – ~100 centres in ~40 countries • Simulation • End-user analysis – batch and interactive • Services, including Data Archive and Delivery, from Tier-1s Frédéric Hemmer, CERN, IT Department
Distribution of Computing Services CPU Disk Tape Frédéric Hemmer, CERN, IT Department
LHC Computing Grid Project - a Collaboration • The physicists and computing specialists from the LHC experiments • The national and regional projects in Europe and the US that have been developing Grid middleware • The regional and national computing centres that provide resources for LHC • The research networks Building and operating the LHC Grid – a global collaboration between Researchers Computer Scientists &Software Engineers Service Providers Frédéric Hemmer, CERN, IT Department
The Grid used by CERN and its partners • The EGEEand OSG projects are the basis of the Worldwide LHC Computing Grid Project Frédéric Hemmer, CERN, IT Department
The new European Network Backbone • LCG working group with Tier-1s and national/ regional research network organisations • New GÉANT 2 – research network backbone Strong correlation with major European LHC centres • Swiss PoP at CERN Frédéric Hemmer, CERN, IT Department
Impact of the LHC Computing Gridin Europe • LCG has been the driving force for the European multi-science Grid EGEE (Enabling Grids for E-sciencE) • EGEE is now a global effort, and the largest Grid infrastructure worldwide • Co-funded by the European Commission (~130 M€ over 4 years) • EGEE already used for >20 applications, including… Bio-informatics Education, Training Medical Imaging Frédéric Hemmer, CERN, IT Department
The EGEE Project • Infrastructure operation • Currently includes >200 sites across 40 countries • Continuous monitoring of grid services & automated site configuration/management http://gridportal.hep.ph.ic.ac.uk/rtm/ • Middleware • Production quality middleware distributed under business friendly open source licence • User Support - Managed process from first contact through to production usage • Training • Documentation • Expertise in grid-enabling applications • Online helpdesk • Networking events (User Forum, Conferences etc.) • Interoperability • Expanding interoperability with related infrastructures Frédéric Hemmer, CERN, IT Department
Status • ~17.5 million jobs run (6450 cpu-years) in 2006; • Workloads of the “other VOs” start to be significant – approaching 8-10K jobs per day; and 1000 cpu-months/month • one year ago this was the overall scale of work for all VOs Frédéric Hemmer, CERN, IT Department
Use of the EGEE Infrastructure >20k jobs running simultaneously Frédéric Hemmer, CERN, IT Department
Use for massive data transfer Large LHC experiments now transferring ~ 1PB/month each Frédéric Hemmer, CERN, IT Department
Site availability – SAM Frédéric Hemmer, CERN, IT Department
Applications on EGEE • More than 20 applications from 7 domains • Astrophysics • MAGIC, Planck • Computational Chemistry • Earth Sciences • Earth Observation, Solid Earth Physics, Hydrology, Climate • Financial Simulation • E-GRID • Fusion • Geophysics • EGEODE • High Energy Physics • 4 LHC experiments (ALICE, ATLAS, CMS, LHCb) • BaBar, CDF, DØ, ZEUS • Life Sciences • Bioinformatics (Drug Discovery, GPS@, Xmipp_MLrefine, etc.) • Medical imaging (GATE, CDSS, gPTM3D, SiMRI 3D, etc.) • Multimedia • Material Sciences • … Frédéric Hemmer, CERN, IT Department
Example: EGEE Attacks Avian Flu • EGEE used to analyse 300,000 possible potential drug compounds against bird flu virus, H5N1. • 2000 computers at 60 computer centres in Europe, Russia, Taiwan, Israel ran during four weeks in April - the equivalent of 100 years on a single computer. • Potential drug compounds now being identified and ranked Neuraminidase, one of the two major surface proteins of influenza viruses, facilitating the release of virions from infected cells. Image Courtesy Ying-Ta Wu, AcademiaSinica. Frédéric Hemmer, CERN, IT Department
Example: Geocluster industrial application • The first industrial application successfully running on EGEE • Developed by the Compagnie Générale de Géophysique (CGG) in France, doing geophysical simulations for oil, gas, mining and environmental industries. • EGEE technology helps CGG to federate its computing resources around the globe. Frédéric Hemmer, CERN, IT Department
EDG EGEE EGEE-II EGEE-III Routine Usage Testbeds Utility Service Evolution European e-InfrastructureCoordination Frédéric Hemmer, CERN, IT Department