150 likes | 275 Views
Status and requirements of PLANCK. NA4/SA1 meeting. Brief intro on application. Planck: ESA satellite mission (2007); LevelS: mission simulation s/w foreach instrument { foreach frequency { foreach cosmology { …. Some Montecarlo jobs; Links with VObs. The LevelS. Pipeline:
E N D
Status and requirements of PLANCK NA4/SA1 meeting
Brief intro on application • Planck: ESA satellite mission (2007); • LevelS: mission simulation s/w • foreach instrument { • foreach frequency { • foreach cosmology { • …. • Some Montecarlo jobs; • Links with VObs.
The LevelS • Pipeline: • Chained but not parallel; • Stages are C/C++/Fortran…. • Shell/perl scripts; Cmb maps foregrounds Scanning F90! Noise Cmb maps analysis Plank simulation is a set of 70 instances of the Pipeline.
Some benchmarks + HFI (50 channels)
Questions: • Is it parallel? NO it runs concurrently. • Do you need MPI/parallel? Yes. In later phase 16/32 CPUs in the site. • What is the BW? > Gigabit! • How long does it run? From 6h up to 24h
Status of the Application • Application Setup: • Basic Gridification; • First tests; • IT; • People (MPA!!); • Refined gridification; • Data&metadata; • Tests: • Runs; • Data; • VO Setup: • Management; • Technical Management; • VO manager; • Site managers; • RLS; • Planck users cert; • Planck sites setup; • EGEE site support.
Slow startup… Technical setup; Two sites: OATS+IFC; Two members. Problem of European users to join the VO… Knowledge: Heterogeneous! Contacts with EGEE sites; MPA looks for EGEE in Munich; Training: User tutorial; Site manager tutorial; Data and replica!!! DBMS and Metadata! VO status & needs
VO evolution • User join the VO 15-30 members; • UI in each site; • Quantum-grid in each site;
Application status • Basic gridification: • Customized scripts; • WN env; • Data handling. • Basic tests: • IT • LFI (22 channels) > 12 time faster!!! but ~5% failures
10-15 terabytes ˜20.000 CD-ROM 1 Eiffel Tower unit Lesson learned • Massive data production on WN ( > 40 GB): • Big disks; • Complex site topology (parallel/distributed FS); • Compressing/RM-CR/removing file program; • FITSIO with fgal/gsiftp support; • Data handling: • Complex data structure; • Complex data structure; • 1 GB RAM.
Application needs • Massive storage of ~ > 5 TB • Data storing/replica (automatic!) • Tier or not Tier? • User common data front-end: • web portal or data browser; • DSE support (metadata) for Grid/non-Grid data: • G-DSE; • External DB; • More then 200 CPUS.
Application deployment:status & strategy • Software deployment: • Dynamic; • Licences: • SW • Compilers. • MPI support intrasite: 16/32 CPUs; • Specific io libs; • Grid-UI • Submission tools • TEST (summer 2005) • Grid-UI • Data browsing • Network & storage tests (end 2005).
Grid added values (…not just CPUS) • Data sharing! • distributed data for distributed users; • Replica and security; • Common interface to SW and data; • Collaborative work for simulations and reduction: • less time, less space, less frustration….
What we have… what we need • VO and RLS • RB • Basic Grid-FS browsing tools: grid-ls, grid-cp etc. • Beowulf/parallel sys as one WN. • DB connection + WS. • Easiest WN env setup (we are Astrophysics…) • Documentation!!!! We are young and we need time to grow… Discuss later our needs for EGEE-2 ?