Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

E2-Train: Training State-of-the-art CNNs with Over 80% Energy Savings (1910.13349v4)

Published 29 Oct 2019 in cs.LG and stat.ML

Abstract: Convolutional neural networks (CNNs) have been increasingly deployed to edge devices. Hence, many efforts have been made towards efficient CNN inference in resource-constrained platforms. This paper attempts to explore an orthogonal direction: how to conduct more energy-efficient training of CNNs, so as to enable on-device training. We strive to reduce the energy cost during training, by dropping unnecessary computations from three complementary levels: stochastic mini-batch dropping on the data level; selective layer update on the model level; and sign prediction for low-cost, low-precision back-propagation, on the algorithm level. Extensive simulations and ablation studies, with real energy measurements from an FPGA board, confirm the superiority of our proposed strategies and demonstrate remarkable energy savings for training. For example, when training ResNet-74 on CIFAR-10, we achieve aggressive energy savings of >90% and >60%, while incurring a top-1 accuracy loss of only about 2% and 1.2%, respectively. When training ResNet-110 on CIFAR-100, an over 84% training energy saving is achieved without degrading inference accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yue Wang (676 papers)
  2. Ziyu Jiang (16 papers)
  3. Xiaohan Chen (30 papers)
  4. Pengfei Xu (57 papers)
  5. Yang Zhao (382 papers)
  6. Yingyan Lin (67 papers)
  7. Zhangyang Wang (375 papers)
Citations (79)

Summary

We haven't generated a summary for this paper yet.