Cost-Effective Methodology for Complex Tuning Searches in HPC: Navigating Interdependencies and Dimensionality (2403.08131v1)
Abstract: Tuning searches are pivotal in High-Performance Computing (HPC), addressing complex optimization challenges in computational applications. The complexity arises not only from finely tuning parameters within routines but also potential interdependencies among them, rendering traditional optimization methods inefficient. Instead of scrutinizing interdependencies among parameters and routines, practitioners often face the dilemma of conducting independent tuning searches for each routine, thereby overlooking interdependence, or pursuing a more resource-intensive joint search for all routines. This decision is driven by the consideration that some interdependence analysis and high-dimensional decomposition techniques in literature may be prohibitively expensive in HPC tuning searches. Our methodology adapts and refines these methods to ensure computational feasibility while maximizing performance gains in real-world scenarios. Our methodology leverages a cost-effective interdependence analysis to decide whether to merge several tuning searches into a joint search or conduct orthogonal searches. Tested on synthetic functions with varying levels of parameter interdependence, our methodology efficiently explores the search space. In comparison to Bayesian-optimization-based full independent or fully joint searches, our methodology suggested an optimized breakdown of independent and merged searches that led to final configurations up to 8% more accurate, reducing the search time by up to 95%. When applied to GPU-offloaded Real-Time Time-Dependent Density Functional Theory (RT-TDDFT), an application in computational materials science that challenges modern HPC autotuners, our methodology achieved an effective tuning search. Its adaptability and efficiency extend beyond RT-TDDFT, making it valuable for related applications in HPC.
- L. Wang, “A survey of codes and algorithms used in nersc material science allocations,” 2006. [Online]. Available: https://escholarship.org/uc/item/9kh2q0cd
- E. Brochu, V. Cora, and N. Freitas, “A tutorial on bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning,” CoRR, 2010.
- R. Moriconi, M. P. Deisenroth, and K. S. S. Kumar, “High-dimensional bayesian optimization using low-dimensional feature spaces,” Machine Learning (Springer), vol. 109, pp. 1925–1943, 2020.
- K. Kandasamy, J. Schneider, and B. Poczos, “High dimensional bayesian optimisation and bandits via additive models,” in Proceedings of the 32nd International Conference on Machine Learning, vol. 37, 2015, pp. 295–304.
- M. B. Oviedo and B. M. Wong, “Real-time quantum dynamics reveals complex, many-body interactions in solvated nanodroplets,” Journal of Chemical Theory and Computation, vol. 12, no. 4, pp. 1862–1871, 2016.
- P. Liashchynskyi and P. Liashchynskyi, “Grid search, random search, genetic algorithm: A big comparison for nas,” 2019.
- M. Malu, G. Dasarathy, and A. Spanias, “Bayesian optimization in high-dimensional spaces: A brief survey,” in 2021 12th International Conference on Information, Intelligence, Systems & Applications (IISA), 2021, pp. 1–8.
- Z. Wang, M. Zoghi, F. Hutter, D. Matheson, and N. De Freitas, “Bayesian optimization in high dimensions via random embeddings,” in Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, ser. IJCAI ’13, 2013, p. 1778–1784.
- B. Letham, R. Calandra, A. Rai, and E. Bakshy, “Re-examining linear embeddings for high-dimensional bayesian optimization,” in Proceedings of the 34th International Conference on Neural Information Processing Systems, ser. NIPS’20, 2020.
- C. Li, S. Gupta, S. Rana, V. Nguyen, S. Venkatesh, and A. Shilton, “High dimensional bayesian optimization using dropout,” ser. IJCAI’17, 2017, p. 2096–2102.
- X. Wu, P. Balaprakash, M. Kruse, J. Koo, B. Videau, P. Hovland, and et al., “ytopt: Autotuning scientific applications for energy efficiency at large scales,” 2023.
- M. Dorier, R. Egele, and P. B. et al., “HPC storage service autotuning using variational- autoencoder -guided asynchronous bayesian optimization,” in 2022 IEEE International Conference on Cluster Computing (CLUSTER), sep 2022.
- Y. Liu, W. M. Sid-Lakhdar, O. Marques, X. Zhu, C. Meng, J. W. Demmel, and X. S. Li, “Gptune: Multitask learning for autotuning exascale applications,” in Proceedings of the 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, ser. PPoPP ’21, 2021, p. 234–246.
- A. P. Dieguez, M. Choi, X. Zhu, B. M. Wong, and K. Z. Ibrahim, “Ml-based performance portability for time-dependent density functional theory in hpc environments,” in 2022 IEEE Workshop on Performance Modeling, Benchmarking and Simulation of High Performance Computer Systems (PMBS), 2022, pp. 1–12.
- W. Kohn and L. J. Sham, “Self-consistent equations including exchange and correlation effects,” Phys. Rev., vol. 140, pp. A1133–A1138, Nov 1965. [Online]. Available: https://link.aps.org/doi/10.1103/PhysRev.140.A1133
- F. Bloch, “Über die quantenmechanik der elektronen in kristallgittern,” Z. Physik, vol. 52, pp. 555–600, 1929. [Online]. Available: https://link.springer.com/article/10.1007/BF01339455#citeas
- Y. Cho, J. W. Demmel, J. King, X. S. Li, Y. Liu, and H. Luo, “Harnessing the crowd for autotuning high-performance computing applications,” in 2023 IEEE International Parallel and Distributed Processing Symposium (IPDPS), 2023, pp. 635–645.
- R.-C. Chen, C. Dewi, S.-W. Huang, and R. E. Caraka, “Selecting critical features for data classification based on machine learning methods,” Journal of Big Data, vol. 7, no. 1, p. 52, 2020.
- F. Gygi, R. K. Yates, J. Lorenz, and et al., “Large-scale first-principles molecular dynamics simulations on the bluegene/l platform using the qbox code.” in SC, 2005, p. 24.