1 / 32

Experiment Applications: applying the power of the grid to real science

Experiment Applications: applying the power of the grid to real science. Rick Cavanaugh University of Florida. GriPhyN/iVDGL External Advisory Committee 13 January, 2002. GriPhyN/iVDGL and ATLAS. Argonne, Boston, Brookhaven, Chicago, Indiana, Berkeley, Texas. ATLAS at SC2002.

jin
Download Presentation

Experiment Applications: applying the power of the grid to real science

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Experiment Applications:applying the power of the grid to real science Rick Cavanaugh University of Florida GriPhyN/iVDGL External Advisory Committee 13 January, 2002

  2. GriPhyN/iVDGLand ATLAS Argonne, Boston, Brookhaven, Chicago, Indiana, Berkeley, Texas

  3. ATLAS at SC2002 • GrappaManages the overall grid experience • MagdaDistributed data management and replication • PacmanDefines and produces software environments • Dc1 production with gratData challenge simulations for Atlas • Instrumented athenaGrid monitoring of Atlas analysis applications • vo-gridmapVirtual organization management • GridviewMonitoring U.S. Atlas resources • WorldgridWorld-wide US/EU grid infrastructure EAC Review

  4. Pacman at SC2002 • How did we install our software for this demo? • % pacman –get iVDGL:WorldGrid ScienceGrid • Pacman lets you define how a mixed tarball/rpm/gpt/native software environment is • Fetched • Installed • Setup • Updated • This can be figured out once and exported to the rest of the world via caches • % pacman –get atlas_testbed EAC Review

  5. The caches you have decided to trust Dependencies are automatically resolved Installed software, pointer to local documentation Pacman at SC2002 • How did we install our software for this demo? • % pacman –get iVDGL:WorldGrid ScienceGrid • Pacman lets you define how a mixed tarball/rpm/gpt/native software environment is • Fetched • Installed • Setup • Updated • This can be figured out once and exported to the rest of the world via caches • % pacman –get atlas_testbed EAC Review

  6. Grappa at SC2002 • Web-based interface for Athena job submission to Grid resources • Based on XCAT Science Portal technology developed at Indiana • EDG JDL backend to Grappa • Common submission to US gatekeepers and EDG resource broker (through EDG “user interface” machine) EAC Review

  7. Grappa Communications Flow Script-Based Submisson interactive or cron-job http: JavaScript Cactus framework https - JavaScript Input files Web Browsing Machine (JavaScript) Netscape/Mozilla/Int.Expl/PalmScape CoG : Submission, Monitoring http:// browse catalogue Grappa Portal Machine: XCAT tomcat server MAGDA: registers file/location registers file metadata Magda . . . CoG : Data Storage: - Data Disk - HPSS (spider) Resource A Resource Z Data Copy EAC Review Compute Resources

  8. Instrumented Athena at SC2002 • Part of SuperComputing 2002 ATLAS demo • Prophesy (http://prophesy.mcs.anl.gov/) • An Infrastructure for Analyzing & Modeling the Performance of Parallel & Distributed Applications • Normally a Parse & auto-instrument approach (C & FORTRAN). • NetLogger (http://www-didc.lbl.gov/NetLogger/) • End-to-End Monitoring & Analysis of Distributed Systems • C, C++, Java, Python, Perl, Tcl APIs • Web Service Activation EAC Review

  9. GriPhyN/iVDGL and CMS Caltech, Fermilab, Florida, San Diego, Wisconsin

  10. Bandwidth Gluttony at SC2002 • "Grid-Enabled" particle physics analysis application • issued remote database selection queries; prepared data object collections, • moved collections across the WAN using specially enhanced TCP/IP stacks • rendered the results in real time on the analysis client workstation in Baltimore. EAC Review

  11. MonaLisa at SC2002 • MonaLisa (Caltech) • Deployed on the US-CMS Test-bed • Dynamic information/resource discovery mechanism using agents • Implemented in • Java / Jini with interfaces to SNMP, MDS, and Ganglia • WDSL / SOAP with UDDI • Proved critical during live CMS production runs EAC Review Pictures taken from Iosif Legrand

  12. MOP and Clarens at SC2002 VDT Server 1 Condor • Simple, robust grid planner integrated with CMS production software • 1.5 million simulated CMS events produced over 2 months (~30 CPU years) VDT Client MCRunJob DAGMan/ Condor-G Linker ScriptGen GridFTP Clarens Server Config Master mop-submitter Req. Self Des GridFTP Clarens Client VDT Server N Condor GridFTP Clarens Server EAC Review

  13. Chimera Production at SC2002 params exec. data • Used VDL to describe virtual data products and their dependencies • Used the Chimera Planners to map abstract workflows onto concrete grid resources • Implemented a WorkRunner to continously schedule jobs across all grid sites Generator Production Simulator Formator Reconstructor Ntuple Example CMS concrete DAG Analysis Stage File In Execute Job Stage File Out Register File EAC Review

  14. Data Provenance at SC2002 A virtual space of simulated data is created for future use by scientists... mass = 200 mass = 200 decay = ZZ mass = 200 decay = WW stability = 3 mass = 200 decay = WW mass = 200 decay = WW stability = 1 mass = 200 event = 8 mass = 200 decay = WW stability = 1 event = 8 mass = 200 plot = 1 mass = 200 decay = WW event = 8 mass = 200 decay = WW stability = 1 plot = 1 mass = 200 decay = WW plot = 1 EAC Review

  15. Data Provenance at SC2002 Search for WW decays of the Higgs Boson and where only stable, final state particles are recorded: mass = 200; decay = WW; stability = 1 mass = 200 mass = 200 decay = ZZ mass = 200 decay = WW stability = 3 mass = 200 decay = WW mass = 200 decay = WW stability = 1 mass = 200 event = 8 mass = 200 decay = WW stability = 1 event = 5 mass = 200 plot = 1 mass = 200 decay = WW event = 8 mass = 200 decay = WW stability = 1 plot = 1 mass = 200 decay = WW plot = 1 EAC Review

  16. Data Provenance at SC2002 mass = 200 mass = 200 decay = ZZ mass = 200 decay = WW stability = 3 mass = 200 decay = WW stability = 1 LowPt = 20 HighPt = 10000 mass = 200 decay = WW ...The scientist adds a new derived data branch... mass = 200 decay = WW stability = 1 and continues to investigate ! mass = 200 event = 8 mass = 200 decay = WW stability = 1 event = 8 mass = 200 plot = 1 mass = 200 decay = WW event = 8 mass = 200 decay = WW stability = 1 plot = 1 mass = 200 decay = WW plot = 1 EAC Review

  17. GriPhyN and LIGO (Laser Interferometer Gravitational-wave Observatory) ISI, Caltech, Milwaukee

  18. archive Interferometer Extract channel Short Fourier Transform transpose Long time frames 30 minutes Short time frames Single Frame Time-frequency Image Extract frequency range Hz Find Candidate event DB Construct image Store Time raw channels LIGO’s Pulsar Search EAC Review

  19. Pegasus: Planning for Execution in Grids • Developed at ISI as part of the GriPhyN project • Configurable system that can map and execute complex workflows on the Grid • Integrated with the GriPhyN Chimera system • It Receives an abstract workflow (AW) description from Chimera, produces a concrete workflow (CW) • Submits the CW to DAGMan for execution. • Optimizations of CW are done from the point of view of Virtual Data. • Can perform AW planning based on application-level metadata attributes. • Given attributes such as time interval, frequency of interest, location in the sky, etc., Pegasus is currently able to produce any virtual data products present in the LIGO pulsar search EAC Review

  20. Metadata Driven Configuration EAC Review

  21. LIGO’s pulsar search at SC2002 • The pulsar search conducted at SC 2002 • Used LIGO’s data collected during the first scientific run of the instrument • Targeted a set of 1000 locations of known pulsar as well as random locations in the sky • Results of the analysis were published via LDAS (LIGO Data Analysis System) to the LIGO Scientific Collaboration • performed using LDAS and compute and storage resources at Caltech, University of Southern California, University of Wisconsin Milwaukee. EAC Review

  22. SC 2002 demo Over 58 pulsar searches Total of 330 tasks 469 data transfers 330 output files The total runtime was 11:24:35 To date 185 pulsar searches Total of 975 tasks 1365 data transfers 975 output files Total runtime 96:49:47 Results EAC Review

  23. Virtual Galaxy Cluster System:An Application of the GriPhyN Virtual Data Toolkit to Sloan Digital Sky Survey Data Chicago, Argonne, Fermilab

  24. The Brightest Cluster Galaxy Pipeline maxBcg is a series of transformations • 1: extracts galaxies from the full tsObj data set. • 2:filter the field for Bright Red Galaxies. • 3: calculate the weighted BCG likelihood for each galaxy, most expensive. • 4: is this galaxy the most likely galaxy in the neighborhood? • 5: remove extraneous data, and store in a compact format. Interesting intermediate data reuse made possible by Chimera: Cluster finding works well with 1 Mpc radius apertures. If one instead was looking for the sites of gravitational lensing, one would rather use a 1/4 Mpc radius. This would start at transformation 3. EAC Review

  25. The DAG Catalog Cluster Core BRG EAC Review

  26. A DAG for 50 Fields • 744 files, 387 nodes, 40 minutes 50 60 168 108 EAC Review

  27. Example:Sloan Galaxy Cluster Analysis Sloan Data DAG Galaxy cluster size distribution With Jim Annis & Steve Kent, FNAL EAC Review

  28. Conclusion • Built a virtual cluster system based on Chimera and SDSS cluster finding. • Described the five stages and data dependencies in VDL. • Tested the system on a virtual data grid. • Conducting performance analysis. • Helped improve Chimera. EAC Review

  29. Some CMS Issues/Challenges • How to generate more buy-in from the experiments? Sociological trust problem, not technical. • More exploition of (virtual) collections of objects and further use of web services (work already well underway). • What is required to store the complete provenance of data generated in a grid environment? • Creation of collaborative peer-to-peer environments. • Data Challenge 2003-4: generate and analyze 5% of the expected data at startup (~1/2 year of continuous production). • What is the relationship between WorldGRID and the LCG? • Robust, portable applications! • Virtual Organization Management and Policy Enforcement. EAC Review

  30. Some ATLAS Issues/Challenges • How to generate more buy-in from the experiments? Sociological trust problem, not technical. • Fleshing out the notion of Pacman "Projects" and prototyping them • What is the best integration path for chimera infrastructure with international atlas catalog systems? Need standardized Virtual Data API? • Packaging and distribution of ATLAS SW releases for each step in the production/analysis chain: gen, sim, reco, analysis. • LCG SW application development env. is now SCRAM: ATLAS evaluating possible migration from CMT to SCRAM EAC Review

  31. SDSS Challenges • Cluster Finding • Distribution of clusters in the universe • Evolution of the mass function • Balanced I/O and compute • Power Spectrum • Distribution of galaxies in the universe • Direct constraints on cosmological parameters • Compute intensive, prefer MPI systems • Premium on discovering similar results • Analyses based on pixel data • Weak lensing analysis of the SDSS coadded southern survey data • Near Earth asteroid searches • Galaxy morphological properties: NVO Galaxy Morphology Demo • All involve moving around terabytes of data • Or choosing not to EAC Review

  32. LIGO Challenges EAC Review

More Related