1 / 26

Cybersar

Explore CYBERSAR, a cooperative effort providing computational support to RTD activities. Learn about its dual purpose, members, funding, and network and computing cluster anatomy.

davidpalmer
Download Presentation

Cybersar

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Gianluigi Zanetti (gianluigi.zanetti@crsr4.it) Cybersar RTD on computational infrastructure

  2. CYBERSAR: a cooperative effort for a shared HPC facility • Dual purpose • Computational Support to RTD • RTD on computational infrastructure • Members of the co-op • CRS4, INAF, INFN, UNICA, UNISS, Nice, Tiscali • Partially Funded by MIUR • March 2006 – December 2008 • Supported by RAS (Sardinian Regional Authority) • access to dark-fibers & regional network INFN AOB

  3. Cybersar flyby

  4. Cybersar: a cooperative effort for a shared HPC facility • Dual purpose • Computational Support to RTD • RTD on computational infrastructure • Members of the co-op • CRS4, INAF, INFN, UNICA, UNISS, Nice, Tiscali • Partially Funded by MUR • March 2006 – December 2008 • Supported by RAS (Sardinian Regional Authority) • access to dark-fibers & regional network INFN AOB

  5. INFN AOB Cybersar Cluster anatomy • Standard node configuration • 48 x (2x AMD F2218, 8GB RAM, 500 GB HD, 2 1GbE)‏ • 24 x (2x AMD F2218, 16GB RAM, 500 GB HD, IB)‏ Storage 40TB

  6. INFN AOB Cybersar Network anatomy

  7. Configuration computing

  8. ‘Computable’ deep computing stack reconfiguration application middleware OS OS OS OS HW HW HW HW Network application application middleware middleware OS OS OS OS HW HW HW HW Network Network

  9. Configuration computing: atomic operations and programs • ‘Bare hardware’ services • IPMI control (boot/shutdown)‏ • Remote install/repair • VLAN setup/configuration • Virtualization layer services • VM deployment • VM cluster deployment • Orchestration services • WS-BPEL • Jolie • Specialized scripts • Modified GE/AR for resource allocation and scheduling

  10. A research problem • Work on massive collections of ‘data with structure’ • Biology data • New sequencing machines (e.g., 454) produce 1TB/3days each • Medical data • Modalities at > 1GB per acquisition are now common • High definitions 3d models collections • Fast, very high resolution, 3D laser scanners • Terrains data • ~10cm resolution, country size • Have configurable computing & network Adapt computing facility to problem

  11. Massively parallel sequencing (1/5) http://www.454.com 400,000 wells * 300 bases / 7hr 1Terabases /3 days

  12. Massively parallel sequencing (2/5) • Large data accumulations in productions sites • E.g., Sanger Center can produce 300TB/month • Deep, highly non-linear, information content • Complex data analysis based on whole dataset • Analysis method itself object of research • Multiple computational cultures & approaches

  13. Massively parallel sequencing (3/5) storage Data production center

  14. storage Data production center Massively parallel sequencing (4/5) Computing site A NETWORK 3 Tbytes / (10Gbit/sec ) ~ 1 hr Computing site B

  15. Computing site B (virtual) NETWORK Massively parallel sequencing (5/5) Lab A storage xml + images ~ 4Gbytes 5 secs Computing site A (virtual) Data production center

  16. State of the art CRS4@cybersar traditional cluster 4’ from scratch to running 72 phys. nodes 8’ from scratch to running 288 vn on 72 phys. n. control plane computer Computing site (virtual) non traditional computing e.g., map-reduce facility for biocomputing virtual computing site hosting

  17. map-reduce cluster 288 virtual nodes P2P storage sim. 988 virtual nodes gLite cluster 32 virtual nodes bio-cluster EET 25 virtual nodes production experimental Virtual bubble clusters CRS4@cybersar • vs grid ‘5000 • vs Science Clouds

  18. Configuration computing

  19. Configurable computing facilities • Computing pipeline • Not necessarily a single large machine • One of the clusters is ‘attached’ to the data • Clusters can be dynamically configured • User cluster configuration in sand-box, safe for hosting facility • Network is abundant • Dynamic allocation of lambda (a la OptIPuter)

  20. Reference application: “holodeck” 50Mpixel 0.8° resolution50-70° FOV

  21. Geographical computational pipelines

  22. Light field display, the movie

  23. Light field display Front-end PC Back-end Rendering Cluster RAM 3 5 9 1 GPU1 LRU Cache (Ids of current working set chunks) (Geometry/color for new chunks) Cache 2 Partitioning and simplification Mmap I/O Adaptive multires. renderer Holovizio wrapper front-end Holovizio wrapper back-end Network 50Mpixel 0.8° resolution50-70° FOV GPUm n Off-line Cache 4 7 2 Multiresolution DAG Cut Multiresolution structure (data+dependency) Agocs, Balogh, Forgács, Bettio, Gobbetti, and Zanetti. A Large Scale Interactive Holographic Display. In Proc. IEEE VR 2006 Workshop on Emerging Display Technologies, 2006. CD ROM Proceedings.

  24. Conclusions & future • CYBERSAR Good platform to support research on • Computing control plane • Virtual clusters • OptIPuter architecture • Future work • Virtual cluster appliances for genome analysis • Virtualization of coprocessors (GPU, FPGA,…) • Use GARR-X for National & European scale test and developments INFN AOB

  25. `HPC for the masses’ Massive amounts of data are looming… questa non è detto che sia una nuvola… Digital camera 39Megapixel (6726x5040 px) Lightfield display ‘real 3d’

  26. Questions?

More Related