EC-NAS: Energy Consumption Aware Tabular Benchmarks for Neural Architecture Search (2210.06015v4)
Abstract: Energy consumption from the selection, training, and deployment of deep learning models has seen a significant uptick recently. This work aims to facilitate the design of energy-efficient deep learning models that require less computational resources and prioritize environmental sustainability by focusing on the energy consumption. Neural architecture search (NAS) benefits from tabular benchmarks, which evaluate NAS strategies cost-effectively through precomputed performance statistics. We advocate for including energy efficiency as an additional performance criterion in NAS. To this end, we introduce an enhanced tabular benchmark encompassing data on energy consumption for varied architectures. The benchmark, designated as EC-NAS, has been made available in an open-source format to advance research in energy-conscious NAS. EC-NAS incorporates a surrogate model to predict energy consumption, aiding in diminishing the energy expenditure of the dataset creation. Our findings emphasize the potential of EC-NAS by leveraging multi-objective optimization algorithms, revealing a balance between energy usage and accuracy. This suggests the feasibility of identifying energy-lean architectures with little or no compromise in performance.
- “A comprehensive survey of neural architecture search: Challenges and solutions,” ACM Computing Surveys, vol. 54, no. 4, pp. 1–34, 2021.
- “Zen-NAS: A Zero-Shot NAS for High-Performance Deep Image Recognition,” in International Conference on Computer Vision (ICCV), 2021.
- “Accelerating neural architecture search using performance prediction,” in International Conference on Learning Representations (ICLR) - Workshop Track, 2017.
- “Efficientnet: Rethinking model scaling for convolutional neural networks,” in International Conference on Machine Learning (ICML), 2019.
- “Green AI,” Communications of the ACM, vol. 63, no. 12, pp. 54–63, 2020.
- “Carbontracker: Tracking and Predicting the Carbon Footprint of Training Deep Learning Models,” ICML Workshop on Challenges in Deploying and monitoring Machine Learning Systems, 2020.
- “Compute trends across three eras of machine learning,” in International Joint Conference on Neural Networks (IJCNN), 2022.
- “Tabular benchmarks for joint architecture and hyperparameter optimization,” Arxiv, 2019.
- “Surrogate NAS benchmarks: Going beyond the limited search spaces of tabular NAS benchmarks,” in International Conference on Learning Representations (ICLR), 2022.
- “NAS-Bench-101: Towards reproducible neural architecture search,” in International Conference on Machine Learning (ICML), 2019.
- “NAS-Bench-1Shot1: Benchmarking and dissecting one-shot neural architecture search,” in International Conference on Learning Representations (ICLR), 2020.
- “EA-HAS-bench: Energy-aware hyperparameter and architecture search benchmark,” in International Conference on Learning Representations (ICLR), 2023.
- “Neural architecture search for speech emotion recognition,” Arxiv, 2022.
- “Search for efficient deep visual-inertial odometry through neural architecture search,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023.
- Alex Krizhevsky, “Learning multiple layers of features from tiny images,” Tech. Rep., Univeristy of Toronto, 2009.
- “Measuring the carbon intensity of ai in cloud instances,” in Conference on Fairness, Accountability, and Transparency (FAccT), 2022.
- “Multi-objective optimization with unbounded solution sets,” in NeurIPS Workshop on Bayesian Optimization (BayesOpt 2016), 2016.
- “Bag of baselines for multi-objective joint neural architecture search and hyperparameter optimization,” in ICML Workshop on Automated Machine Learning (AutoML), 2021.
- “Nas-bench-x11 and the power of learning curves,” Advances in Neural Information Processing Systems (NeurIPS), 2021.
- “KNAS: green neural architecture search,” in International Conference on Machine Learning (ICML), 2021.
- “Carbon footprint of selecting and training deep learning models for medical image analysis,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2022.
- “Few-shot neural architecture search,” in International Conference on Machine Learning (ICML), 2021.
- “Uptime Institute global data center survey 2020,” Tech. Rep., Uptime Institute, 07 2020.
- “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems (NeurIPS)). 2019.
- “Adam: A method for stochastic optimization,” in International Conference on Learning Representations, 2015.
- “Towards the systematic reporting of the energy and carbon footprints of machine learning,” Journal of Machine Learning Research, vol. 21, no. 248, pp. 1–43, 2020.
- “MobileNets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.
- “Constructing fast network through deconstruction of convolution,” Advances in Neural Information Processing Systems, 2018.
- “Hyperparameter power impact in transformer language model training,” in Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, Virtual, Nov. 2021, pp. 96–118, Association for Computational Linguistics.
- “Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach,” IEEE Transactions on Evolutionary Computation, vol. 3, no. 4, pp. 257–271, 1999.
- “Performance assessment of multiobjective optimizers: An analysis and review,” IEEE Transactions on Evolutionary Computation, vol. 7, no. 2, pp. 117–132, 2003.
- “Speeding up many-objective optimization by Monte Carlo approximations,” Artificial Intelligence, vol. 204, pp. 22–29, 2013.
- “Approximation quality of the hypervolume indicator,” Artificial Intelligence, vol. 195, pp. 265–290, 2013.
- “SMS-EMOA: Multiobjective selection based on dominated hypervolume,” European Journal of Operational Research, vol. 181, no. 3, pp. 1653–1669, 2007.
- “Covariance matrix adaptation for multi-objective optimization,” Evolutionary Computation, vol. 15, no. 1, pp. 1–28, 2007.
- “HypE: An algorithm for fast hypervolume-based many-objective optimization,” Evolutionary computation, vol. 19, no. 1, pp. 45–76, 2011.
- James Edward Baker, “Adaptive selection methods for genetic algorithms,” in International Conference on Genetic Algorithms and their Applications, 1985, vol. 1.
- “How genetic algorithms work: A critical look at implicit parallelism,” in International Conference on Genetic Algorithms, 1989, pp. 20–27.