Reducing energy consumption of Neural Architecture Search: An inference latency prediction framework. (April 2021)
- Record Type:
- Journal Article
- Title:
- Reducing energy consumption of Neural Architecture Search: An inference latency prediction framework. (April 2021)
- Main Title:
- Reducing energy consumption of Neural Architecture Search: An inference latency prediction framework
- Authors:
- Lu, Longfei
Lyu, Bo - Abstract:
- Abstract: Benefit from the success of NAS (Neural Architecture Search) in deep learning, humans are hopefully been released from the tremendous labor of manual tuning of structure and hyper-parameters. However, the success of NAS comes at the cost of much more computational resource consumption, thousands of times more computational power than ordinary training of manual-designed models, especially for the resource-aware multi-objective NAS, which must be serialized as a sequential loop of sampling, training, deployment, and inference. Recent research has shown that deep learning leads to huge energy consumption and CO2 emission (training of the namely Transformer can emit CO2 as much as five cars in their lifetimes Strubell et al. (2019)). Aiming to alleviate this issue, we propose the end-to-end inference latency prediction framework to empower the NAS process with a direct resource-aware efficiency indicator. Namely, we first propose the end-to-end latency prediction framework, which can predict latency quickly and accurately based on the dataset collected by ourselves. Eventually, we experimentally show that with the encoding scheme we designed, our proposed best model, LSTM-GBDT Latency Predictor(LGLP) achieves an excellent result of 0.9349 MSE, 0.5249 MAE, 0.9842 R 2, and 0.9925 corrcoef . In other words, our limited dataset and encoding scheme already provide the precise knowledge representation of this large search space. By equipping NAS with the proposed framework,Abstract: Benefit from the success of NAS (Neural Architecture Search) in deep learning, humans are hopefully been released from the tremendous labor of manual tuning of structure and hyper-parameters. However, the success of NAS comes at the cost of much more computational resource consumption, thousands of times more computational power than ordinary training of manual-designed models, especially for the resource-aware multi-objective NAS, which must be serialized as a sequential loop of sampling, training, deployment, and inference. Recent research has shown that deep learning leads to huge energy consumption and CO2 emission (training of the namely Transformer can emit CO2 as much as five cars in their lifetimes Strubell et al. (2019)). Aiming to alleviate this issue, we propose the end-to-end inference latency prediction framework to empower the NAS process with a direct resource-aware efficiency indicator. Namely, we first propose the end-to-end latency prediction framework, which can predict latency quickly and accurately based on the dataset collected by ourselves. Eventually, we experimentally show that with the encoding scheme we designed, our proposed best model, LSTM-GBDT Latency Predictor(LGLP) achieves an excellent result of 0.9349 MSE, 0.5249 MAE, 0.9842 R 2, and 0.9925 corrcoef . In other words, our limited dataset and encoding scheme already provide the precise knowledge representation of this large search space. By equipping NAS with the proposed framework, taking NEMO for example, it will save 1588 kWh ⋅ PUE energy, 1515 pounds CO2 emissions, and $3176 cloud compute cost of AWS . For NAS is now widely exploited in research or industry applications, this will bring incalculable benefits to society and the environment. Highlights: Neural Architecture Search (NAS) suffers from abundant energy consumption. We benchmark a dataset and present LGLP model to predict the latency of CNN. Latency predictor empowers NAS with offline resource-aware efficiency indicator. Latency prediction-based NAS dramatically reduce the energy consumption. … (more)
- Is Part Of:
- Sustainable cities and society. Volume 67(2021)
- Journal:
- Sustainable cities and society
- Issue:
- Volume 67(2021)
- Issue Display:
- Volume 67, Issue 2021 (2021)
- Year:
- 2021
- Volume:
- 67
- Issue:
- 2021
- Issue Sort Value:
- 2021-0067-2021-0000
- Page Start:
- Page End:
- Publication Date:
- 2021-04
- Subjects:
- Deep learning -- Latency prediction -- Neural architecture search -- Energy saving -- Society and environment
Sustainable urban development -- Periodicals
Sustainable buildings -- Periodicals
Urban ecology (Sociology) -- Periodicals
307.76 - Journal URLs:
- http://www.sciencedirect.com/science/journal/22106707/ ↗
http://www.sciencedirect.com/ ↗
http://www.journals.elsevier.com/sustainable-cities-and-society ↗ - DOI:
- 10.1016/j.scs.2021.102747 ↗
- Languages:
- English
- ISSNs:
- 2210-6707
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - BLDSS-3PM
British Library HMNTS - ELD Digital store - Ingest File:
- 16022.xml