180 likes | 354 Views
SpaceGRID EO GRID Workshop ESRIN 6-7 May 2002 Pier Giorgio Marchetti Pier.giorgio.Marchetti@esa.int see also http://esagrid.esa.int/spacegrid. SpaceGRID Goals. Assess how GRID technology can serve requirements across a large variety of space disciplines and applications .
E N D
SpaceGRID EO GRID Workshop ESRIN 6-7 May 2002 Pier Giorgio Marchetti Pier.giorgio.Marchetti@esa.int see also http://esagrid.esa.int/spacegrid
SpaceGRID Goals • Assess how GRID technology can serve requirements across a large variety of space disciplines and applications. • Foster collaboration and enable shared efforts across space applications • Sketch the design of an ESA-wide (horizontal and common) GRID infrastructure • Proof of concept through prototyping • Involve both industry and research centres • Keep Europe up with GRID efforts !!
SpaceGRID Study • Financed under Agency’s General Study Programme 800KEuro • Datamat, main contractor, subcontractors: Alactel Space Industries, CS Systémes d’Information , Qinetic, Rutherford Appleton Laboratories, Science Systems (Space) Ltd • 1 Technical Officer + 4 Domain experts from Agency: • Eamonn Daly, Hugh Evans (Space Weather) • Hans-Peter De Koning (Spacecraft Engineering) • Gerhard Kreiner, Salim Ansari (Solar System Research) • Patrick Grimont (Future EO Ground Segments) • September 2001-July2003 • Co-operation with EC financed projects DataGRID, EGSO
Radiation TransportGeant4 Application • Monte-Carlo simulation of high energy particle interactions inside a spacecraft system, components or detector. • Large numbers (>107) of source particles are required to ensure good statistics. • Easily adaptable to a parallel processing system (independent particle histories).
Spacecraft/Plasma Simulation • Monte-Carlo 3-d time dependent electrostatic kinetic Particle-in-Cell Method. • Large Number of particles (>109) to be tracked per iteration. • Strong coupling between cells required to solve the field. • Challenging to develop sufficiently independent subtasks for parallel processing.
Space Weather Simulation • Simulation of a Coronal Mass Ejection and interplanetary shock from the Sun to the Earth and subsequent effects on the Earth’s magnetosphere • Magneto Hydrodynamic Simulation (MHD). • Particle simulation of shock effects on trapped particle belt population (adiabatic responses). • Trade-off GRID vs LCMPP and “@home” solutions
SpaceGRID - Objectivies for ESA Science • Evaluation of GRID technology in support of Solar System Research activities: • access to multiple distributed data archives • online data manipulation, analysis and visualisation of complex data sets • interoperability between legacy systems and heterogeneous databases • collaborative environment (e.g. sharing of algorithms for analysis and visualisation)
ENVISAT • launched 28 February 2002 • 10 instruments on board • 200 Mbps data rate to ground • 400 Tbytes data archived/year • ~100 “standard” products • 10+ dedicated facilities in Europe • ~700 approved science user projects
Ka-Band USERS Space Communication - ARTEMIS X-Band Ka-Band KIRUNA ESOC FOCC ESRIN PDS
Mechnical Engineering Usage ScenarioGRID-based analysis and test support Test team at test facility (ESTEC, Intespace, IABG) Test results data • Perform test scenario • Make test data available Video/audio teleconferencing Updated analysis predicitions Engineering analysis teamat (sub)contractor site (anywhere in Europe) • Remote monitoring of test • Near-real-time correlation analysis model with test results • Near-real-time investigation of test anomalies
Why Service Chaining SC permits exploitation of multiple Services 1,2,3 Service 5: Meteo info Service 1: NDVI Service 4: Land use • new improved Service: Risk Index corrected with meteo • new Service: precision farming (NDVI + land use + meteo) • … Service 2: LST Service 3=1+2 Risk Index
Simple Service Chaining Service C Request Receive Request A Receive Request B Receive Request Process Service C=A+B Service C Status Process Service A Process Service B Deliver Service C=A+B Deliver Service A Deliver Service B Service C Delivery Info Portal Service C Service A Service B
High Level GRID Requirements • Distributed access to • computing and storage resources • real-time data and archives • Seamless integration with W3C Web Services • Easy access to metadata repositories • Seamless integration with flow control and management • Domain based and geographic based Portals and Service directories • Network availability • Group to Group Collaborative Environment