1 / 21

SKA Data Flow and Processing – a key SKA design driver

SKA Data Flow and Processing – a key SKA design driver. Paul Alexander and Jaap Bregman. Overview. We “know” that the SKA will produce vast amounts of data Following the flow of data and information is crucial to the overall system design of the SKA

kateb
Download Presentation

SKA Data Flow and Processing – a key SKA design driver

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. SKA Data Flow and Processing – a key SKA design driver Paul Alexander and Jaap Bregman

  2. Overview We “know” that the SKA will produce vast amounts of data Following the flow of data and information is crucial to the overall system design of the SKA Can we process the data the SKA will produce How does the data processing tension the SKA design? This analysis based only on imaging requirements

  3. SKA Overall System Mass Storage 0.3-1.2 GHz Wide FoV Tile & Station Processing Central Processing Facility - CPF To 250 AA Stations Dense AA ... Correlator – AA & Dish 16 Tb/s ... Data .. 70-450 MHz Wide FoV .. Time Post Processor Control Sparse AA ... 0.8-10 GHz Single Pixel or Focal plane array DSP Control Processors & User interface 12-15m Dishes 80 Gb/s DSP Time Standard ... To 2400 Dishes User interface via Internet

  4. Some numbers to remember Fastest HPC today – Roadrunner ~ 1.2 PFlop Total internet data rate ~ 20 Tb/s BUT on your desktop you can have ~1 TFlop

  5. Data rates to the correlator Data rate from each collectorG1 = 2 NpDfNbitNb = 4 DfNbitNb AA, Number of elements Ne ~ 65000; Nb << Ne limited by data rate250 sq-deg across band Nb ~ 1200 (Memo 100)G1 ~ 16 Tb/s Dense AA Correlator Visibility processors Image formation Science analysis & archive switch AA: 250 x 16 Tb/s Sparse AA Dish: 3000 x 64 Gb/s 12-15m Dishes Dishes G1 ~ 64 Gbs PAFs FoV is constant across the bandNb ~ 7 to give 20 sq-deg across the band G1 ~ 60 Gb/s (Memo 100)

  6. Data rates from the correlator Standard results for integration/dump time and channel width Naive data rate then given by Can reduce this using baseline-dependent integration times and channel widths

  7. Reduction in data rate using baseline-dependence (> 0.06) 1/3 B = B1B >> B1 Data rate out of correlator exceeds input data rate for 15-m dishes for baselines exceeding ~ 130km (36km if single integration time) At best for dishes output data rate ~ input; AA’s reduction by ~104 Data rates from the correlator B / km a b 1 Dishes 0.08 0.11 2.8 Aperture Array 0.33 0.13 2.2

  8. Data rates from the correlator

  9. Processing model

  10. Post correlator Processing on the raw UV data is highly parallel UV data needs to be buffered based on current algorithms UV processing includes: Gridding and de-gridding from current sky model Peeling and subtraction of current sky model Image size: a2Nch (B/D)2Nb Ratio UV to “image” data Major reduction in data rate occurs between UV data and image data

  11. The SKA Processing Challenge Correlator Visibility processors Image formation Science analysis, user interface & archive switch AA: 250 x 16 Tb/s Dish: 3000 x 60 Gb/s ~ 200 Pflop to 2.5 Eflop ~10 PFlop ~ ? PFlop ~1 – 500 Tb/s Software complexity

  12. Model for UV processor Highly parallel – consider something achievable – NVIDIA promises 20 TFlop in 2 years – assume 50 TFlop Approximate analysis of ops/sample: 20000/calibration loop Processor: €1000, 5 calibration loops, 50% efficiency, each processor processes ~ 1 GB/s of data Requirement: 100 PFlop (AA) 2 EFlop (dishes) Buffer 1 hr of data therefore we need to buffer 7.2 TB in a fast store Assume €200 per TB. UV-processing cost: Cost = €2.5m (TB/s) AA < €10m Dishes ~ €125m

  13. The SKA Processing Challenge Correlator Visibility processors Image formation Science analysis, user interface & archive switch AA: 250 x 16 Tb/s Dish: 3000 x 60 Gb/s ~ 200 Pflop to 2.5 Eflop ~10 PFlop ~ ? PFlop ~1 – 500 Tb/s Software complexity

  14. Data Products • ~0.5 – 10 PB/day of image data • Source count ~106 sources per square degree • ~1010 sources in the accessible SKA sky, 104 numbers/record • ~1 PB for the catalogued data 100 Pbytes – 3 EBytes / year of fully processed data

  15. Summary • Processing challenge depends critically on overall system and experiment • For dishes, the data rate drops significantly only after the UV processor • For many experiments the data rate can increase through the correlator. • Design of processing system needs a complete analysis of the data flow • Separate UV-processor and imaging processors • Processing UV data is a highly parallel excellent scaleability • Local buffer for UV processor • Maximum baseline and number of correlated elements drive costs • Processing is expensive, but achievable • BUT must consider whole system – the processor MUST match the requirements of the system – PrepSKA

  16. Non PDRA resources Data flow and overview Alexander, Diamond, Garrington, ASTRON (de Vos, de Bruyn?) Pipeline processing: ASTRON LOFAR team CASU – Cambridge Astronomical Survey Unit – Irwin TDP – to be determined Data products and VO UK VO team – Walton, Richards; Astrowise Groningen Hardware architecture and implementation Centre for scientific computing (Cambridge), Oxford e-Science centre TDP - Kembal

  17. Data from the correlator Data rate depends on the experiment f = 1000 MHz. Minimum data rate for dishes 0.8 square degrees, 2500 15-m dishes out to 200 km baseline Df = 500MHz High data rate for Aperture Array 75 square degrees, 250 56-m stations out to 200 km baseline Df = 500MHz >24 PB temporary buffer required

  18. Processing For each sample Subtract Ncs confusing bright sources from global sky model DB Grid sample allowing for sky curvature Repeat Nloop Effective ~20000 ops/sample using current algorithms Form image FFT + deconvolution Update global sky model Calibrate: solve for telescope errors Store ~50TB/day of Observations

  19. SKA Overall Structure Beam Data Mass Storage 0.3-1.2 GHz Wide FoV Tile & Station Processing Central Processing Facility - CPF To 250 AA Stations Dense AA ... Correlator – AA & Dish 16 Tb/s ... Data .. 70-450 MHz Wide FoV .. Time Post Processor Control Sparse AA ... 0.8-10 GHz Single Pixel or Focal plane array DSP Control Processors & User interface 12-15m Dishes 80 Gb/s DSP Time Standard ... To 2400 Dishes User interface via Internet

More Related