210 likes | 238 Views
CERN. The Grid Project. Fabrizio Gagliardi. CERN. Information Technology Division. May, 2000. F.Gagliardi@cern.ch. TERENA Lisbon. Summary. CERN. High Energy Physics and CERN computing problem. . An excellent computing model: the GRID. . The Data Grid Initiative. .
E N D
CERN The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000 F.Gagliardi@cern.ch TERENA Lisbon
Summary CERN High Energy Physics and CERN computing problem An excellent computing model: the GRID The Data Grid Initiative (http://www.cern.ch/grid/) TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 2
CERN organization CERN European International Center for Particle Physics Research Largest Particle Physics lab in the world Budget: 1020 M CHF 2700 staff 7000 physicist users TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 3
The LHC Detectors CMS CERN ATLAS 3.5 PetaBytes / year ~108events/year LHCb TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 4
The HEP Problem - Part I CERN The scale... TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 5
CERN Estimated CPU Capacity at CERN 2,500 2,000 1,500 K SI95 Non-LHC 1,000 ~10K SI95 1200 processors LHC 500 0 1998 1999 2000 2001 2002 2003 2004 2005 2006 year technology-price curve (40% annual price improvement) Capacity that can purchased for the value of the equipment present in 2000 TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 6
CERN Estimated DISK Capacity ay CERN 1800 1600 1400 1200 TeraBytes 1000 Non-LHC 800 LHC 600 400 200 0 1998 1999 2000 2001 2002 2003 2004 2005 2006 year technology-price curve (40% annual price improvement) TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 7
Long Term Tape Storage Estimates CERN 14'000 12'000 10'000 TeraBytes 8'000 LHC 6'000 4'000 Current Experiments COMPASS 2'000 0 1996 2006 2000 2003 2005 1995 1997 1998 1999 2001 2002 2004 Year TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 8
HPC or HTC CERN High Throughput Computing mass of modest, independent problems computing in parallel – not parallel computing throughput rather than single-program performance resilience rather than total system reliability Have learned to exploit inexpensive mass market components But we need to marry these with inexpensive highly scalable management tools Much in common with other sciences (see EU-US Annapolis Workshop at www.cacr.caltech.edu/euus): Astronomy, Earth Observation, Bioinformatics, and commercial/industrial: data mining, Internet computing, e-commerce facilities, …… TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 9
Generic component model of a computing farm network servers application servers tape servers disk servers
The HEP Problem - Part II CERN Geography, Sociology, Funding and Politics... TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 11
World Wide Collaboration distributed computing & storage capacity CERN CMS: 1800 physicists 150 institutes 32 countries TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 12
Regional Centres - a Multi-Tier Model CERN CERN – Tier 0 IN2P3 RAL Tier 1 FNAL Uni n Lab a Tier2 Uni b Lab c Department Desktop MONARC report: http://home.cern.ch/~barone/monarc/RCArchitecture.html TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 13
Are Grids a solution? CERN Change of orientation of US Meta-computing activity From inter-connected super-computers … .. towards a more general concept of a computational Grid (The Grid – Ian Foster, Carl Kesselman) Has initiated a flurry of activity in HEP US – Particle Physics Data Grid (PPDG) GriPhyN – data grid proposal submitted to NSF Grid technology evaluation project in INFN UK proposal for funding for a prototype grid NASA Information Processing Grid TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 14
The Grid CERN “Dependable, consistent, pervasive access to [high-end] resources” • Dependable: • provides performance and functionality guarantees • Consistent: • uniform interfaces to a wide variety of resources • Pervasive: • ability to “plug in” from anywhere TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 15
R&D required CERN Local fabric Management of giant computing fabrics auto-installation, configuration management, resilience, self- healing Mass storage management multi-PetaByte data storage, “real-time” data recording requirement, active tape layer – 1,000s of users Wide-area - building on an existing framework & RN (e.g.Globus, Geant) workload management no central status local access policies data management caching, replication, synchronisation object database model application monitoring TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 16
HEP Data Grid Initiative CERN European level coordination of national initiatives & projects Principal goals: Middleware for fabric & Grid management Large scale testbed - major fraction of one LHC experiment Production quality HEP demonstrations “mock data”, simulation analysis, current experiments Other science demonstrations Three year phased developments & demos Complementary to other GRID projects EuroGrid: Uniform access to parallel supercomputing resources Synergy to be developed (GRID Forum, Industry and Research Forum) TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 17
Participants CERN Main partners: CERN, INFN(I), CNRS(F), PPARC(UK), NIKEF(NL), ESA-Earth Observation Other sciences: KNMI(NL), Biology, Medicine Industrial participation: CS SI/F, DataMat/I, IBM/UK Associated partners: Czech Republic, Finland, Germany, Hungary, Spain, Sweden (mostly computer scientists) Formal collaboration with USA Industry and Research Project Forum with representatives from: Denmark, Greece, Israel, Japan, Norway, Poland, Portugal, Russia TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 18
Status CERN Prototype work already started at CERN and in most of collaborating institutes Proposal to RN2 submitted Network requirements discussed with Dante/Geant TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 19
WAN Requirements CERN High bandwidth from CERN to Tier 1 centres (5-6) VPN, Quality of Service Guaranteed performance during limited test periods and at the end of the project for production quality services Target requirements (2003) 2.5 Gb/s + 622 Mb/s + 155 Mb/s Could saturate for limited amount of test time 2.5 Gb/s (100 MB/s out from a 100 PC farm, we plan for 1000’s PC farm) Reliability is an important factor: from WEB client-server model to GRID peer distributed computing model TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 20
Conclusions CERN This project, motivated by HEP and other high data and computing demanding sciences, will contribute to develop and implement a new world-wide distributed computing model: The GRID An ideal computing model for the next generation Internet An excellent test case for the next generation of high- performance research networks TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 21