210 likes | 366 Views
“UCF”. Computing Capabilities at UNM HPC. Timothy L. Thomas UNM Dept of Physics and Astronomy. I have a 200K SU ( 150K LL CPU hour ) grant from the NRAC of the NSF/NCSA, with which UNM HPC (“AHPCC”) is affiliated. (Stolen from Andrew…). Peripheral Data Vs Simulation.
E N D
“UCF” Computing Capabilities at UNM HPC Timothy L. Thomas UNM Dept of Physics and Astronomy Timothy L. Thomas
I have a 200K SU (150K LL CPU hour) grant from the NRAC of the NSF/NCSA, with which UNM HPC (“AHPCC”) is affiliated.
(Stolen from Andrew…) Peripheral Data Vs Simulation Simulation: Muons From Central Hijing (QM02 Project07) Data: Centrality by Perp > 60
(Stolen from Andrew…) Simulated Decay Muons • QM’02 Project07 PISA files (Central HIJING) • Closest cuts possible from PISA file to match data(PT parent >1 GeV/c, Theta Porig Parent 155-161) Investigating possibility of keeping only muon and parent hits for reconstruction. • 17100 total events distributed over Z=±10, ±20, ±38 More events available but only a factor for smallest error bar Zeff ~75 cm Not in fit "(IDPART==5 || IDPART==6) && IDPARENT >6 &&IDPARENT < 13 && PTHE_PRI >155 && PTHE_PRI < 161 && IPLANE == 1 && IARM == 2 && LASTGAP > 2002 && PTOT_PRI*sin(PTHE_PRI*acos(0)/90.) > 1."
Now at UNM HPC: • PBS • Globus 2.2.x • Condor-G / Condor • (GDMP) • …all supported by HPC staff. • In Progress: • A new 1.2 TB RAID 5 disk server, to host: • AFS cache PHENIX software • ARGO file catalog (PostgreSQL) • Local Objectivity mirror • Globus 2.2.x (GridFTP and more…)
Pre-QM2002 experience • with globus-url-copy… • Easily saturated UNM • bandwidth limitations • (as they were at that time) • PKI infrastructure and • sophisticated error-handling • are a real bonus over bbftp. • (One bug, known at the time • is being / has been addressed.) • (at left: 10 streams used) KB/sec
LLDIMU.HPC.UNM.EDU Timothy L. Thomas
Resources Assume 90KByte/event and 0.1GByte/hour/CPU Filtered event can be analyzed, but not ALL PRDF event Many trigger has overlap.
Rough calculation of real-data processing (I/O-intensive) capabilities: 10 M events, PRDF-to-{DST+x}, both mut & mutoo; assume 3 sec/event (*1.3 for LL), 200200 KB/event. One pass: 7 days on 50 CPUs (25 boxes), using 56% of LL local network capacity. My 200K “SU” (~150K LL CPU hours) allocation allows for 18 of these passes (4.2 months) 3 MB/sec Internet2 connection = 1.6 TB / 12 nights (MUIDN_1D1S&NTCN) (Presently) LL is most effective for CPU-intensive tasks: simulations can easily fill the 512 CPUs; e.g, QM02 Project 07. Caveats: “LLDIMU” is a front-end machine; LL worker node environment is different from CAS/RCS node ( P.Power…)