200 likes | 374 Views
Российский GRID для операций с большими массивами научных данных. Ильин В.А. (НИИЯФ МГУ) проекты LCG ( LHC Computing GRID) и EGEE ( Enabling Grids for E_science in Europe). Таруса, 6 февраля 2004 г. CERN. 40 MHz (40 TB/sec). level 1 - special hardware. 75 KHz (75 GB/sec).
E N D
Российский GRID для операций с большими массивами научных данных Ильин В.А. (НИИЯФ МГУ) проекты LCG (LHC Computing GRID) и EGEE (Enabling Grids for E_science in Europe) Таруса, 6 февраля 2004 г.
40 MHz (40 TB/sec) level 1 - special hardware 75 KHz (75 GB/sec) level 2 - embedded processors 5 KHz (5 GB/sec) level 3 - PCs 100 Hz (100 MB/sec) Data recording & offline analysis • Online system • Multi-level trigger • Filter out background • Reduce data volume • Online reduction 107 • Trigger menus • Select interesting events • Filter out less interesting
1-100 ГБ/сек Отбор событий и первичная реконструкция детектор 200 MБ/сек Архивное хранение 1-6 ПБ/год Суммарные данные по событию ESD 0.5-1 ПБ/год сырые данные 200 TБ/год Подготовка данных для анализа Реконструкция события ~100 MБ/сек Данные для анализа (выделенные по физ. каналам) Моделиро- вание событий Интерактивный физический анализ тысячи ученых Большой адронный коллайдер (БАК) потоки данных, этапы обработки и анализа Tier1 0.1-1 ГБ/сек Tier2 РИВК БАК 5-10%
LHC Challenges: Scale (in 2006) • Data written to tape ~40 Petabytes/Year and UP (1 PB = 10**9 MBytes) • Processing capacity 100 - TIPS and UP • (1 TIPS = 10**6 MIPS) • Typical networks Few Gbps Per Link • Lifetime of experiment 2-3 Decades (start in 2007) • Users ~ 5000 physicists • Software developers ~ 300 (Four Experiments)
Russian Tier2-ClusterRussian regional center for LHC computing Cluster of institutional computing centers with Tier2 functionality and summary resources at 50-70% level of the canonical Tier1 center for each experiment (ALICE, ATLAS, CMS, LHCb): analysis; simulations; users data support. Participating institutes: Moscow ITEP, KI, MSU, LPI, MEPhI… Moscow region JINR, IHEP, INR RAS St.Petersburg PNPI RAS, … Novosibirsk BINP SB RAS Coherent use of distributed resources by means of LCG (EDG, VDT, …) technologies. Active participation in the LCG Phase1 Prototyping and Data Challenges (at 5% level).
Russia in LCGhttp://www.cern.ch/lcg • We have started activity in LCG (LHC Computing GRID project) in autumn 2002. • Russia is joining to the LCG-1 infrastructure now. First SINP MSU, then JINR, ITEP and IHEP. • Goal – to have in Russia in Q4 operational segment of world-wide LCG infrastructure and be ready to DataChallenges in 2004. • Manpower contribution to LCG (started in May 2003): • the Agreement is under signing by CERN and Russia and JINR officials, • 3 tasks for our responsibility: • 1) testing new GRID mw to be used in LCG • 2) evaluation of new-on-the-market GRID mw (first task – evaluation of GT3 and WebSphere) • 3) common solutions for event generators (event data bases).
Программы генерации модельных (Монте Карло) данных для будущих экспериментов на LHC. Типичная нагрузка на каналы связи сейчас – передача 100 Гбайт данных из Москвы в ЦЕРН за рабочий день 50 Мбит/сек ! LHC Data Challenges Но это не средняя нагрузка – «пиковая»:
LCG - Goals • The goal of the LCG project is to prototype and deploy the computing environment for the LHC experiments • Two phases: • Phase 1: 2002 – 2005 • Build a service prototype, based on existing grid middleware • Gain experience in running a production grid service • Produce the TDR for the final system • Phase 2: 2006 – 2008 • Build and commission the initial LHC computing environment LCG is not a development project – • it relies on other grid projects for grid middleware development and support
Grid The virtualLHC Computing organizations Building a Grid Collaborating Computer Centres Alice VO CMS VO
lhc20.sinp.msu.ru НИИЯФ МГУ grid011.pd.infn.it lhc01.sinp.msu.ru lhc02.sinp.msu.ru lhc03.sinp.msu.ru Падуя SINP MSU RB+ InformationIndex lhc04.sinp.msu.ru CE lxshare0220.cern.ch WN SE Пользователь ЦЕРН Пример использования EDG (LCG) middleware (CMS VO)
Distribution of GRID Service Activities over Europe: • Operations Management at CERN; • Core Infrastructure Centres (CICs) in the UK, France, Italy, Russia and at CERN, responsible for managing the overall Gridinfrastructure; • Regional Operations Centres (ROCs), responsible for coordinating regional resources, regional deployment and support of services. Россия: CIC – МГУ, РНЦ КИ, ОИЯИ, ИТЭФ, ИПМ РАН ROC – ИФВЭ, ОИЯИ, ПИЯФРАН, ИМПБ РАН
Pan-European Multi-Gigabit Backbone (33 Countries)January 2004 Planning Underway for “GEANT2” (GN2) Multi-Lambda Backbone, to Start In 2005
International Committee for Future Accelerators (ICFA)Standing Committee on Inter-Regional Connectivity (SCIC) http://icfa-scic.web.cern.ch/ICFA-SCIC/ ICFA SCIC Reports Networking for High Energy and Nuclear Physics - Feb/2004 ( Doc - A4 / Pdf - A4 ) Report on the Digital Divide in Russia - Feb/2004 ( Doc - A4 / Pdf - A4 ) Network Monitoring Report - Feb/2004 ( Doc - A4 / Pdf - A4 ) Advanced Technologies Interim Report - Feb/2003 ( Doc ) Digital Divide Executive Report - Feb/2003 ( Doc )
“Little Gloriad” (OC3) Launched January 12; to OC192 in 2005 GLORIAD: Global Optical Ring (US-Ru-Cn) Also Important for Intra-Russia Connectivity
SCIC Monitoring WG – Throughput Improvements 1995-2004 Bandwidth of TCP < MSS/(RTT*Sqrt(Loss)) (1) 60% annual improvement Factor ~100/10 yr Some Regions ~5-10 Years Behind SE Europe and Parts of Asia May be Catching Up (Slowly) Progress: but Digital Divide is Mostly Maintained (1) Matthis et al., Computer Communication Review 27(3), July 1997
REGIONAL and INTERNATIONAL CONNECTIVITY for RUSSIA High Energy Physics Moscow 1 Gbps IHEP 8 Mbps (m/w), under construction 100 Mbps fiber-optic (Q1-Q2 2004?) JINR 45Mbps, 100-155 Mbps (Q1-Q2 2004), Gbps (2004-2005) INR RAS 2 Mbps+2x4Mbps(m/w) BINP 1 Mbps, 45 Mbps (2004 ?), … PNPI 512 Kbps (commodity), and 34 Mbps f/o but budget is only for 2 Mbps (!) USA NaukaNET 155 Mbps GEANT 155 Mbps basic link, plus 155 Mbps additional link for GRID projects Japan through USA by FastNET, 512 Kbps Novosibirsk(BINP) - KEK GLORIAD 10 Gbps