TY - GEN
T1 - Is In-Context Learning Feasible for HPC Performance Autotuning?
AU - Randall, Thomas
AU - Bondapalli, Akhilesh
AU - Ge, Rong
AU - Balaprakash, Prasanna
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - We examine whether in-context learning with Large Language Models (LLMs) can effectively address the challenges of High-Performance Computing (HPC) autotuning. LLMs have demonstrated remarkable natural language processing and artificial intelligence (AI) capabilities, sparking interest in their application across various domains, including HPC. Performance autotuning - the process of automatically optimizing system configurations to maximize efficiency through empirical evaluation - offers significant promise for enhancing application performance on larger systems and emerging architectures. However, this process remains computationally expensive due to the combinatorial explosion of configuration parameters and the complex, nonlinear relationships between configurations and performance outcomes.We pose a critical question: Can LLMs, without task-specific fine-tuning, accurately infer performance-configuration patterns by combining in-context examples with latent knowledge? To explore this, we leverage empirical performance data from real-world HPC systems, designing structured prompts and queries to evaluate LLMs' capabilities. Our experiments reveal inherent limitations in applying in-context learning to performance autotuning, particularly for tasks requiring precise mathematical reasoning and analysis of complex multivariate dependencies. We provide empirical evidence of these shortcomings and discuss potential research directions to overcome these challenges.
AB - We examine whether in-context learning with Large Language Models (LLMs) can effectively address the challenges of High-Performance Computing (HPC) autotuning. LLMs have demonstrated remarkable natural language processing and artificial intelligence (AI) capabilities, sparking interest in their application across various domains, including HPC. Performance autotuning - the process of automatically optimizing system configurations to maximize efficiency through empirical evaluation - offers significant promise for enhancing application performance on larger systems and emerging architectures. However, this process remains computationally expensive due to the combinatorial explosion of configuration parameters and the complex, nonlinear relationships between configurations and performance outcomes.We pose a critical question: Can LLMs, without task-specific fine-tuning, accurately infer performance-configuration patterns by combining in-context examples with latent knowledge? To explore this, we leverage empirical performance data from real-world HPC systems, designing structured prompts and queries to evaluate LLMs' capabilities. Our experiments reveal inherent limitations in applying in-context learning to performance autotuning, particularly for tasks requiring precise mathematical reasoning and analysis of complex multivariate dependencies. We provide empirical evidence of these shortcomings and discuss potential research directions to overcome these challenges.
KW - HPC
KW - In Context Learning
KW - LLM
KW - Performance Autotuning
UR - https://www.scopus.com/pages/publications/105015468245
U2 - 10.1109/IPDPSW66978.2025.00152
DO - 10.1109/IPDPSW66978.2025.00152
M3 - Conference contribution
AN - SCOPUS:105015468245
T3 - Proceedings - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
SP - 978
EP - 985
BT - Proceedings - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2025 IEEE International Parallel and Distributed Processing Symposium Workshops, IPDPSW 2025
Y2 - 3 June 2025 through 7 June 2025
ER -