230 likes | 247 Views
The CrossGrid Project aims to develop new programming tools, enhance grid interoperability, and implement standards for distributed computing. It focuses on biomedical applications, flooding crisis support, and data analysis in high-energy physics, while also collaborating with other grid projects.
E N D
Forschungszentrum Karlsruhein der Helmholtz-Gemeinschaft The European CrossGrid Project Marcel Kunze Abteilung Grid-Computing und e-Science Forschungszentrum Karlsruhe GmbH On behalf of the CrossGrid Collaboration Special thanks to Marian Bubak and Jesus Marco www.eu-crossgrid.org
Outline • The CrossGrid Project • CrossGrid Aplications • Testbed Status • Deliverables and Plans
-Links with European National efforts - Links with US projects (GriPhyN, PPDG, iVDGL,…) Applications EGSO CROSSGRID GRIA EUROGRID DATAGRID GRIP GRIDLAB DAMIEN Middleware & Tools DATATAG Underlying Infrastructures Industry / business Science EU Funded Grid Project Space(Kyriakos Baxevanidis)
CrossGrid Collaboration 21 institutes 11 countries Ireland: TCD Dublin Poland: Cyfronet & INP Cracow PSNC Poznan ICM & IPJ Warsaw Germany: FZK Karlsruhe TUM Munich USTU Stuttgart Netherlands: UvA Amsterdam Slovakia: II SAS Bratislava Austria: U.Linz Spain: CSIC Santander Valencia & RedIris UAB Barcelona USC Santiago & CESGA Greece: Algosystems Demo Athens AuTh Thessaloniki Portugal: LIP Lisbon Italy: DATAMAT Cyprus: UCY Nikosia
Main Objectives • EU Vth Framework Programme (IST) Project, started March 2002 • 21 partners from 11 countries • New category of Grid enabled applications • Computing and data intensive • Distributed • Interactive, near real time response (a person in a loop) • Layered • New programming tools • Grid more user friendly, secure and efficient • Interoperability with other Grids • Implementation of standards
Collaboration with other Grid Projects • Exchange of • Information • Software components • Partners • DATAGRID • DATATAG • GRIDLAB • EUROGRID and GRIP • GRIDSTART • Participation in GGF
Workpackages • WP1 – CrossGrid Application Development • Biomedical simulation and visualization • Flooding crisis support • Interactive distributed data analysis in HEP • Weather forecast and air pollution modeling • WP2 - Grid Application Programming Environments • Tools for parallel programming and debugging on the Grid (MPI) • WP3 – New Grid Services and Tools • Portals and roaming access • Resource management • Monitoring • Optimisation of data access • WP4 - International Testbed Organisation • Integration team, certification authority, support and test procedures • Based on EDG 1.2 (see talk of Marcus Hardt in WP6 parallel session) • WP5 - Project Management • Coordination, architecture, dissemination
Biomedical Application CT / MRI scan Segmentation Visualization LB flow simulation Medical Medical HDB VE DB DB WD PC PDA 10 simulations/day 60 GB/simulation > 20 MB/s Interaction
Cascade of Flood Simulations Data sources Meteorological simulations Hydrological simulations Users Hydraulic simulations Output visualization
Distributed Data Analysis in HEP • Objectives • Distributed data access • Distributed data mining techniques with neural networks • Issues • Typical interactive requests will run on o(TB) distributed data • Transfer/replication times for the whole data about one hour • Data transfers once and in advance of the interactive session • Allocation, installation and set-up of corresponding database servers before the interactive session • Integration of user-friendly interactive access (based on PROOF)
Selection Parameters TagDB CPU Procedure PROOF DB1 RDB CPU Proc.C DB2 Proc.C DB3 CPU Proc.C DB4 CPU Proc.C DB5 CPU Proc.C DB6 CPU Parallel ROOT Facility: PROOF Local Remote http://root.cern.ch
Weather Forecast and Air Pollution Modeling • Distributed/parallel codes on the Grid • Coupled Ocean/Atmosphere Mesoscale Prediction System • STEM-II Air Pollution Code • Integration of distributed databases • Data mining applied to downscaling weather forecast
Key Features of CrossGrid Applications • Data • Data sources and data bases geographically distributed • To be selected on demand • Processing • Large processing capacity required; both HPC & HTC • Interactive • Presentation • Complex data requires versatile 3D visualisation • Support for interaction and feedback to other components
Overview of the CrossGrid Architecture Applications 1.1 BioMed 1.2 Flooding 1.3 Interactive Distributed Data Access 1.3 Data Mining on Grid (NN) 1.4 Meteo Pollution 2.2 MPI Verification 2.3 Metrics and Benchmarks 2.4 Performance Analysis 3.1 Portal & Migrating Desktop Supporting Tools Applications Development Support MPICH-G 1.1, 1.2 HLA and others App. Spec Services 1.1 User Interaction Services 1.1 Grid Visualisation Kernel 3.2 Scheduling Agents 3.4 Optimization of Grid Data Access 3.3 Grid Monitoring 3.1 Roaming Access DataGrid Replica Manager Globus Replica Manager Generic Services DataGrid Job Submission Service GRAM GridFTP GIS / MDS GSI Globus-IO Replica Catalog Replica Catalog Fabric Resource Manager (CE) Resource Manager (SE) Resource Manager Resource Manager 3.4 Optimization of Local Data Access CPU Secondary Storage Instruments ( Satelites, Radars) Tertiary Storage
Statusafter M6++ • Software Requirements Specifications together with use cases • CrossGrid Architecture defined • Detailed Design documents for tools and the new Grid services (OO approach, UML) • Analysis of security issues and the first proposal of solutions • Detailed description of the test and integration procedures • Testbed first experience • Sites: LIP, FZK, CSIC+USC, PSNC, AuTH+Demo • Basic: EDG release 1.2 • Applications: • EDG HEP simulations (Atlas,CMS) • first distributed prototypes using MPI: • NN distributed training • Evolutionary Algorithms
CrossGrid Testbed Map Géant TCD Dublin PSNC Poznan UvA Amsterdam ICM & IPJ Warsaw FZK Karlsruhe CYFRONET Cracow CSIC-UC IFCA Santander II SAS Bratislava USC Santiago LIP Lisbon Auth Thessaloniki UAB Barcelona CSIC RedIris Madrid CSIC IFIC Valencia DEMO Athens UCY Nikosia
Sample Testbed Sites Valencia GoG farm, Santander (GridWall), FZK
Plans for the Future • Participation in production testbed with DataGrid • All sites will be ready to join by end of September • Common DEMO at IST 2002, Copenhagen, November 4th-6th • Collaboration with DataGrid in specific points (e.g. user support and helpdesk software) • Conference together with RI Forum and the “Across Grids” • Santiago de Compostella, Spain, February 9th-14th,2003 • With Proceedings (reviewed papers) • CrossGrid workshop, Linz (w/ EuroPVM/MPI 2002), September 28th-29th
Linz CrossGrid Workshop Sep.28th-29th • Evaluate the current status of all tasks • Contact partners who are developing software we are going to use: discuss interfaces and functionality • Understand what we may expect as first prototypes • Coordinate the operation of testbeds • Agree about common rules for software development:Standard Operational Procedures (SOP) document written • Start to organize the first CrossGrid EU review • Meet with EU DataGrid representatives • Discuss the technology for the future (OGSA) Details at http://www.gup.uni-linz.ac.at/crossgrid/workshop/
CrossGrid Schedule and Deliverables • M6 (now) • WPs 1-3 start to develop prototypes of the applications, tool environment, and the new grid services • WP4 are working on launching the first testbed. • M9 (Dec.2002) • WP4 internal status report • WP5 CrossGrid website • M10 (Jan. 2003) • WP4 First testbed prototype release • M11 (Feb. 2003) • WP5 Project leaflet/brochure • WP5 Dissemination and exploitation report • CrossGrid conference in Santiago di Compostella, February 2002 • M12 (March 2003) • WP1 First software release • WP2 First prototypes • WP3 First prototypes • WP5 Report on requirements on integration and interoperability with DataGrid • WP5 Yearly report
Grid-enabled Applications Prototype Grid Infrastructures Gèant: World Class Networking 1980s: Internet 1990s: Web 2000s: Grid • Where do we need to get to ? • Applications to support an “e-society” (“Cyber-Infrastructure”) • A Grid infrastructure which hides the complexities from the users (“Invisible Computing”) • A powerful and flexible network infrastructure • Where do we need to invest ? • Applications targeted at realistic problems in “e-science” • Prototypes of Grid infrastructures • Maintain and improve the GEANT network • Expression of Interest for EU FP6 program: • “Enabling Grids and e-Science in Europe (EGEE)”