On the Impact of Device-Level Techniques on Energy-Efficiency of Neural Network Accelerators

Energy-efficiency is a key concern for neural network applications. To alleviate this issue, hardware acceleration using FPGAs or GPUs can provide better energy-efficiency than general-purpose processors. However, further improvement of the energy-efficiency of such accelerators will be extremely beneficial specially to deploy neural network in power-constrained edge computing environments. In this paper, we experimentally explore the potential of device-level energy-efficiency techniques (e.g.,supply voltage underscaling, frequency scaling, and data quantization) for representative off-the-shelf FPGAs compared to GPUs. Frequency scaling in both platforms can improve the power and energy consumption but with performance overhead, e.g.,in GPUs it improves the power consumption and GOPs/J by up to 34 However, leveraging reduced-precision instructions improves power (up to 13 energy (up to 20 reduction in accuracy of neural network accuracy.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset