130 likes | 322 Views
Common I/O needs. A set of processesRead input data from a single fileWrite output data to a single fileCheck point during computationsPerform out of core computations. Need for parallel I/O. Sequential I/O becomes a bottleneckTedious pre- and post-processing is necessary to split and merge glo
E N D
1. Parallel I/O using MPI I/O Giri Chukkapalli
SDSC
2. Common I/O needs A set of processes
Read input data from a single file
Write output data to a single file
Check point during computations
Perform out of core computations
3. Need for parallel I/O Sequential I/O becomes a bottleneck
Tedious pre- and post-processing is necessary to split and merge global files if parallel I/O is not used
Standard unix I/O is not portable
True parallel I/O requires multi-process access from a parallel file system
4. Advantages of MPI I/O Performance
portability
Convenience
Public domain MPI I/O API ROMIO from ANL
5. Features of MPI I/O Parallel read/write
non-contiguous data read/write
non-blocking read/write
collective read write
portable data representation across platforms
HPF style distributed array syntax
6. Derived data types New type is a combination of existing types and spaces
Derived data types are used to communicate noncontiguous data
Sender’s data type is mapped onto receiver’s data type
writer’s data type is mapped onto fileview and vice versa in MPI I/O
7. Derived data types MPI_Type
_contiguous
_vector/_hvector
_indexed/_hindexed
_struct
_subarray/_darray
8. Fileview Displacement, etype and filetype creates a fileview
fileview allows simultaneous writing/reading of noncontiguous interleaved data by multiple processes
MPI_File_set_view call
each process has a different fileview of a single file
9. MPI I/O standard All calls start with MPI_File_
open, read, write, seek, close
Asynchronous modifier “i”: iread etc.
Absolute position modifier “_at”: read_at
Collective modifier “_all”: read_all etc
split collective modifier “_begin” “_end”
shared file pointer modifier: “_shared”
MPI_Type to create derived data types
10. Examples of MPI I/O Usage Simple read/write of contiguous data
read/write of noncontiguous data from a file to contiguous memory location
Non-blocking read/write of data
collective I/O to a single file
read/write of distributed arrays
look in /usr/local/apps/examples/UsingMPI2 on BlueHorizon
11. Blue Horizon Specific Always compile with thread safe compilers
mpcc_r, mpxlf90_r etc.
Use GPFS file system /gpfs/userid
5TB of disk and 12 I/O servers and can achieve ~ 1 GB/s bandwidth
Not fully MPI I/O compliant yet.
Look in /usr/lpp/ppe.poe/include/mpi.h for function call support
Look in /usr/local/apps/examples/UsingMPI2 for example codes
12. MPI I/O performance issues To achieve good performance:
Write as large chunks as possible
use derived data types to read/write non-contiguous data
Use collective I/O calls
use non-blocking I/O calls
provide hints through “info” parameter
provide complete picture of the total I/O operation on the whole file by all the processes
13. References Http://www.mcs.anl.gov/software : MPI-2 reference manual
http://www.classes.cs.uchicago.edu/classes/archive/2000/fall/CS103-01/Lectures/mpi-io : slides from a very good talk
http://hpcf.nersc.gov/software/libs/io/mpiio.html : Intro to MPI I/O at NERSC
http://www.llnl.gov/asci/sc99fliers/mpi_io_pg1.html : MPI I/O efforts at Livermore
http://www-unix.mcs.anl.gov/romio/papers.html : good papers on MPI I/O
http://www.cs.dartmouth.edu/pario/ : A comprehensive parallel I/O archive