TY - GEN
T1 - Decision trees and MPI collective algorithm selection problem
AU - Pješivac-Grbović, Jelena
AU - Bosilca, George
AU - Fagg, Graham E.
AU - Angskun, Thara
AU - Dongarra, Jack J.
PY - 2007
Y1 - 2007
N2 - Selecting the close-to-optimal collective algorithm based on the parameters of the collective call at run time is an important step for achieving good performance of MPI applications. In this paper, we explore the applicability of C4.5 decision trees to the MPI collective algorithm selection problem. We construct C4.5 decision trees from the measured algorithm performance data and analyze both the decision tree properties and the expected run time performance penalty. In cases we considered, results show that the C4.5 decision trees can be used to generate a reasonably small and very accurate decision function. For example, the broadcast decision tree with only 21 leaves was able to achieve a mean performance penalty of 2.08%. Similarly, combining experimental data for reduce and broadcast and generating a decision function from the combined decision trees resulted in less than 2.5% relative performance penalty. The results indicate that C4.5 decision trees are applicable to this problem and should be more widely used in this domain.
AB - Selecting the close-to-optimal collective algorithm based on the parameters of the collective call at run time is an important step for achieving good performance of MPI applications. In this paper, we explore the applicability of C4.5 decision trees to the MPI collective algorithm selection problem. We construct C4.5 decision trees from the measured algorithm performance data and analyze both the decision tree properties and the expected run time performance penalty. In cases we considered, results show that the C4.5 decision trees can be used to generate a reasonably small and very accurate decision function. For example, the broadcast decision tree with only 21 leaves was able to achieve a mean performance penalty of 2.08%. Similarly, combining experimental data for reduce and broadcast and generating a decision function from the combined decision trees resulted in less than 2.5% relative performance penalty. The results indicate that C4.5 decision trees are applicable to this problem and should be more widely used in this domain.
UR - http://www.scopus.com/inward/record.url?scp=38049134284&partnerID=8YFLogxK
U2 - 10.1007/978-3-540-74466-5_13
DO - 10.1007/978-3-540-74466-5_13
M3 - Conference contribution
AN - SCOPUS:38049134284
SN - 9783540744658
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 107
EP - 117
BT - Euro-Par 2007 Parallel Processing - 13th International Euro-Par Conference, Proceedings
PB - Springer Verlag
T2 - 13th International Euro-Par Conference on Parallel Processing, Euro-Par 2007
Y2 - 28 August 2007 through 31 August 2007
ER -