TY - GEN
T1 - Towards an understanding of the performance of MPI-IO in Lustre file systems
AU - Logan, Jeremy
AU - Dickens, Phillip
PY - 2008
Y1 - 2008
N2 - Lustre is becoming an increasingly important file system for large-scale computing clusters. The problem, however, is that many data-intensive applications use MPI-IO for their I/O requirements, and MPI-IO performs poorly in a Lustre file system environment. While this poor performance has been well documented, the reasons for such performance are currently not well understood. Our research suggests that the primary performance issues have to do with the assumptions underpinning most of the parallel I/O optimizations implemented in MPI-IO, which do not appear to hold in a Lustre environment. Perhaps the most important assumption is that optimal performance is obtained by performing large, contiguous I/O operations. However, the research results presented in this poster show that this is often the worst approach to take in a Lustre file system. In fact, we found that the best performance is often achieved when each process performs a series of smaller, non-contiguous I/O requests. In this poster, we provide experimental results supporting these non-intuitive ideas, and provide alternative approaches that significantly enhance the performance of MPI-IO in a Lustre file system.
AB - Lustre is becoming an increasingly important file system for large-scale computing clusters. The problem, however, is that many data-intensive applications use MPI-IO for their I/O requirements, and MPI-IO performs poorly in a Lustre file system environment. While this poor performance has been well documented, the reasons for such performance are currently not well understood. Our research suggests that the primary performance issues have to do with the assumptions underpinning most of the parallel I/O optimizations implemented in MPI-IO, which do not appear to hold in a Lustre environment. Perhaps the most important assumption is that optimal performance is obtained by performing large, contiguous I/O operations. However, the research results presented in this poster show that this is often the worst approach to take in a Lustre file system. In fact, we found that the best performance is often achieved when each process performs a series of smaller, non-contiguous I/O requests. In this poster, we provide experimental results supporting these non-intuitive ideas, and provide alternative approaches that significantly enhance the performance of MPI-IO in a Lustre file system.
UR - http://www.scopus.com/inward/record.url?scp=57949084306&partnerID=8YFLogxK
U2 - 10.1109/CLUSTR.2008.4663791
DO - 10.1109/CLUSTR.2008.4663791
M3 - Conference contribution
AN - SCOPUS:57949084306
SN - 9781424426409
T3 - Proceedings - IEEE International Conference on Cluster Computing, ICCC
SP - 330
EP - 335
BT - Proceedings of the 2008 IEEE International Conference on Cluster Computing, CCGRID 2008
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2008 IEEE International Conference on Cluster Computing, ICCC 2008
Y2 - 29 September 2008 through 1 October 2008
ER -