Uncovering Energy-Efficient Practices in Deep Learning Training

Preliminary Steps Towards Green AI

Conference Paper (2023)
Author(s)

Tim Yarally (Student TU Delft)

Luis Cruz (TU Delft - Software Engineering)

Daniel Feitosa (Rijksuniversiteit Groningen)

June Sallou (TU Delft - Software Engineering)

Arie Deursen (TU Delft - Software Technology)

Research Group
Software Engineering
Copyright
© 2023 Tim Yarally, Luis Cruz, Daniel Feitosa, J. Sallou, A. van Deursen
DOI related publication
https://doi.org/10.1109/CAIN58948.2023.00012
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Tim Yarally, Luis Cruz, Daniel Feitosa, J. Sallou, A. van Deursen
Research Group
Software Engineering
Pages (from-to)
25-36
ISBN (electronic)
979-8-3503-0113-7
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Modern AI practices all strive towards the same goal: better results. In the context of deep learning, the term "results"often refers to the achieved accuracy on a competitive problem set. In this paper, we adopt an idea from the emerging field of Green AI to consider energy consumption as a metric of equal importance to accuracy and to reduce any irrelevant tasks or energy usage. We examine the training stage of the deep learning pipeline from a sustainability perspective, through the study of hyperparameter tuning strategies and the model complexity, two factors vastly impacting the overall pipeline's energy consumption. First, we investigate the effectiveness of grid search, random search and Bayesian optimisation during hyperparameter tuning, and we find that Bayesian optimisation significantly dominates the other strategies. Furthermore, we analyse the architecture of convolutional neural networks with the energy consumption of three prominent layer types: convolutional, linear and ReLU layers. The results show that convolutional layers are the most computationally expensive by a strong margin. Additionally, we observe diminishing returns in accuracy for more energy-hungry models. The overall energy consumption of training can be halved by reducing the network complexity. In conclusion, we highlight innovative and promising energy-efficient practices for training deep learning models. To expand the application of Green AI, we advocate for a shift in the design of deep learning models, by considering the trade-off between energy efficiency and accuracy.

Files

Uncovering_Energy_Efficient_Pr... (pdf)
(pdf | 1.27 Mb)
- Embargo expired in 04-01-2024
License info not available